target/loongarch: Implement vmul/vmuh/vmulw{ev/od}

This patch includes:
- VMUL.{B/H/W/D};
- VMUH.{B/H/W/D}[U];
- VMULW{EV/OD}.{H.B/W.H/D.W/Q.D}[U];
- VMULW{EV/OD}.{H.BU.B/W.HU.H/D.WU.W/Q.DU.D}.

Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
Signed-off-by: Song Gao <gaosong@loongson.cn>
Message-Id: <20230504122810.4094787-15-gaosong@loongson.cn>
This commit is contained in:
Song Gao 2023-05-04 20:27:40 +08:00
parent 9ab29520f7
commit cd1c49ada0
No known key found for this signature in database
GPG Key ID: 40A2FFF239263EDF
5 changed files with 732 additions and 0 deletions

View File

@ -972,3 +972,41 @@ INSN_LSX(vmini_bu, vv_i)
INSN_LSX(vmini_hu, vv_i)
INSN_LSX(vmini_wu, vv_i)
INSN_LSX(vmini_du, vv_i)
INSN_LSX(vmul_b, vvv)
INSN_LSX(vmul_h, vvv)
INSN_LSX(vmul_w, vvv)
INSN_LSX(vmul_d, vvv)
INSN_LSX(vmuh_b, vvv)
INSN_LSX(vmuh_h, vvv)
INSN_LSX(vmuh_w, vvv)
INSN_LSX(vmuh_d, vvv)
INSN_LSX(vmuh_bu, vvv)
INSN_LSX(vmuh_hu, vvv)
INSN_LSX(vmuh_wu, vvv)
INSN_LSX(vmuh_du, vvv)
INSN_LSX(vmulwev_h_b, vvv)
INSN_LSX(vmulwev_w_h, vvv)
INSN_LSX(vmulwev_d_w, vvv)
INSN_LSX(vmulwev_q_d, vvv)
INSN_LSX(vmulwod_h_b, vvv)
INSN_LSX(vmulwod_w_h, vvv)
INSN_LSX(vmulwod_d_w, vvv)
INSN_LSX(vmulwod_q_d, vvv)
INSN_LSX(vmulwev_h_bu, vvv)
INSN_LSX(vmulwev_w_hu, vvv)
INSN_LSX(vmulwev_d_wu, vvv)
INSN_LSX(vmulwev_q_du, vvv)
INSN_LSX(vmulwod_h_bu, vvv)
INSN_LSX(vmulwod_w_hu, vvv)
INSN_LSX(vmulwod_d_wu, vvv)
INSN_LSX(vmulwod_q_du, vvv)
INSN_LSX(vmulwev_h_bu_b, vvv)
INSN_LSX(vmulwev_w_hu_h, vvv)
INSN_LSX(vmulwev_d_wu_w, vvv)
INSN_LSX(vmulwev_q_du_d, vvv)
INSN_LSX(vmulwod_h_bu_b, vvv)
INSN_LSX(vmulwod_w_hu_h, vvv)
INSN_LSX(vmulwod_d_wu_w, vvv)
INSN_LSX(vmulwod_q_du_d, vvv)

View File

@ -243,3 +243,33 @@ DEF_HELPER_FLAGS_4(vmaxi_bu, TCG_CALL_NO_RWG, void, ptr, ptr, i64, i32)
DEF_HELPER_FLAGS_4(vmaxi_hu, TCG_CALL_NO_RWG, void, ptr, ptr, i64, i32)
DEF_HELPER_FLAGS_4(vmaxi_wu, TCG_CALL_NO_RWG, void, ptr, ptr, i64, i32)
DEF_HELPER_FLAGS_4(vmaxi_du, TCG_CALL_NO_RWG, void, ptr, ptr, i64, i32)
DEF_HELPER_FLAGS_4(vmuh_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmuh_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmuh_w, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmuh_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmuh_bu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmuh_hu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmuh_wu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmuh_du, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwev_h_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwev_w_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwev_d_w, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwod_h_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwod_w_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwod_d_w, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwev_h_bu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwev_w_hu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwev_d_wu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwod_h_bu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwod_w_hu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwod_d_wu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwev_h_bu_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwev_w_hu_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwev_d_wu_w, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwod_h_bu_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwod_w_hu_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
DEF_HELPER_FLAGS_4(vmulwod_d_wu_w, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)

View File

@ -1514,3 +1514,553 @@ TRANS(vmaxi_bu, gvec_vv_i, MO_8, do_vmaxi_u)
TRANS(vmaxi_hu, gvec_vv_i, MO_16, do_vmaxi_u)
TRANS(vmaxi_wu, gvec_vv_i, MO_32, do_vmaxi_u)
TRANS(vmaxi_du, gvec_vv_i, MO_64, do_vmaxi_u)
TRANS(vmul_b, gvec_vvv, MO_8, tcg_gen_gvec_mul)
TRANS(vmul_h, gvec_vvv, MO_16, tcg_gen_gvec_mul)
TRANS(vmul_w, gvec_vvv, MO_32, tcg_gen_gvec_mul)
TRANS(vmul_d, gvec_vvv, MO_64, tcg_gen_gvec_mul)
static void gen_vmuh_w(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
{
TCGv_i32 discard = tcg_temp_new_i32();
tcg_gen_muls2_i32(discard, t, a, b);
}
static void gen_vmuh_d(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
{
TCGv_i64 discard = tcg_temp_new_i64();
tcg_gen_muls2_i64(discard, t, a, b);
}
static void do_vmuh_s(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
{
static const GVecGen3 op[4] = {
{
.fno = gen_helper_vmuh_b,
.vece = MO_8
},
{
.fno = gen_helper_vmuh_h,
.vece = MO_16
},
{
.fni4 = gen_vmuh_w,
.fno = gen_helper_vmuh_w,
.vece = MO_32
},
{
.fni8 = gen_vmuh_d,
.fno = gen_helper_vmuh_d,
.vece = MO_64
},
};
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
}
TRANS(vmuh_b, gvec_vvv, MO_8, do_vmuh_s)
TRANS(vmuh_h, gvec_vvv, MO_16, do_vmuh_s)
TRANS(vmuh_w, gvec_vvv, MO_32, do_vmuh_s)
TRANS(vmuh_d, gvec_vvv, MO_64, do_vmuh_s)
static void gen_vmuh_wu(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
{
TCGv_i32 discard = tcg_temp_new_i32();
tcg_gen_mulu2_i32(discard, t, a, b);
}
static void gen_vmuh_du(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
{
TCGv_i64 discard = tcg_temp_new_i64();
tcg_gen_mulu2_i64(discard, t, a, b);
}
static void do_vmuh_u(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
{
static const GVecGen3 op[4] = {
{
.fno = gen_helper_vmuh_bu,
.vece = MO_8
},
{
.fno = gen_helper_vmuh_hu,
.vece = MO_16
},
{
.fni4 = gen_vmuh_wu,
.fno = gen_helper_vmuh_wu,
.vece = MO_32
},
{
.fni8 = gen_vmuh_du,
.fno = gen_helper_vmuh_du,
.vece = MO_64
},
};
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
}
TRANS(vmuh_bu, gvec_vvv, MO_8, do_vmuh_u)
TRANS(vmuh_hu, gvec_vvv, MO_16, do_vmuh_u)
TRANS(vmuh_wu, gvec_vvv, MO_32, do_vmuh_u)
TRANS(vmuh_du, gvec_vvv, MO_64, do_vmuh_u)
static void gen_vmulwev_s(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
{
TCGv_vec t1, t2;
int halfbits = 4 << vece;
t1 = tcg_temp_new_vec_matching(a);
t2 = tcg_temp_new_vec_matching(b);
tcg_gen_shli_vec(vece, t1, a, halfbits);
tcg_gen_sari_vec(vece, t1, t1, halfbits);
tcg_gen_shli_vec(vece, t2, b, halfbits);
tcg_gen_sari_vec(vece, t2, t2, halfbits);
tcg_gen_mul_vec(vece, t, t1, t2);
}
static void gen_vmulwev_w_h(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
{
TCGv_i32 t1, t2;
t1 = tcg_temp_new_i32();
t2 = tcg_temp_new_i32();
tcg_gen_ext16s_i32(t1, a);
tcg_gen_ext16s_i32(t2, b);
tcg_gen_mul_i32(t, t1, t2);
}
static void gen_vmulwev_d_w(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
{
TCGv_i64 t1, t2;
t1 = tcg_temp_new_i64();
t2 = tcg_temp_new_i64();
tcg_gen_ext32s_i64(t1, a);
tcg_gen_ext32s_i64(t2, b);
tcg_gen_mul_i64(t, t1, t2);
}
static void do_vmulwev_s(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
{
static const TCGOpcode vecop_list[] = {
INDEX_op_shli_vec, INDEX_op_sari_vec, INDEX_op_mul_vec, 0
};
static const GVecGen3 op[3] = {
{
.fniv = gen_vmulwev_s,
.fno = gen_helper_vmulwev_h_b,
.opt_opc = vecop_list,
.vece = MO_16
},
{
.fni4 = gen_vmulwev_w_h,
.fniv = gen_vmulwev_s,
.fno = gen_helper_vmulwev_w_h,
.opt_opc = vecop_list,
.vece = MO_32
},
{
.fni8 = gen_vmulwev_d_w,
.fniv = gen_vmulwev_s,
.fno = gen_helper_vmulwev_d_w,
.opt_opc = vecop_list,
.vece = MO_64
},
};
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
}
TRANS(vmulwev_h_b, gvec_vvv, MO_8, do_vmulwev_s)
TRANS(vmulwev_w_h, gvec_vvv, MO_16, do_vmulwev_s)
TRANS(vmulwev_d_w, gvec_vvv, MO_32, do_vmulwev_s)
static void tcg_gen_mulus2_i64(TCGv_i64 rl, TCGv_i64 rh,
TCGv_i64 arg1, TCGv_i64 arg2)
{
tcg_gen_mulsu2_i64(rl, rh, arg2, arg1);
}
#define VMUL_Q(NAME, FN, idx1, idx2) \
static bool trans_## NAME (DisasContext *ctx, arg_vvv *a) \
{ \
TCGv_i64 rh, rl, arg1, arg2; \
\
rh = tcg_temp_new_i64(); \
rl = tcg_temp_new_i64(); \
arg1 = tcg_temp_new_i64(); \
arg2 = tcg_temp_new_i64(); \
\
get_vreg64(arg1, a->vj, idx1); \
get_vreg64(arg2, a->vk, idx2); \
\
tcg_gen_## FN ##_i64(rl, rh, arg1, arg2); \
\
set_vreg64(rh, a->vd, 1); \
set_vreg64(rl, a->vd, 0); \
\
return true; \
}
VMUL_Q(vmulwev_q_d, muls2, 0, 0)
VMUL_Q(vmulwod_q_d, muls2, 1, 1)
VMUL_Q(vmulwev_q_du, mulu2, 0, 0)
VMUL_Q(vmulwod_q_du, mulu2, 1, 1)
VMUL_Q(vmulwev_q_du_d, mulus2, 0, 0)
VMUL_Q(vmulwod_q_du_d, mulus2, 1, 1)
static void gen_vmulwod_s(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
{
TCGv_vec t1, t2;
int halfbits = 4 << vece;
t1 = tcg_temp_new_vec_matching(a);
t2 = tcg_temp_new_vec_matching(b);
tcg_gen_sari_vec(vece, t1, a, halfbits);
tcg_gen_sari_vec(vece, t2, b, halfbits);
tcg_gen_mul_vec(vece, t, t1, t2);
}
static void gen_vmulwod_w_h(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
{
TCGv_i32 t1, t2;
t1 = tcg_temp_new_i32();
t2 = tcg_temp_new_i32();
tcg_gen_sari_i32(t1, a, 16);
tcg_gen_sari_i32(t2, b, 16);
tcg_gen_mul_i32(t, t1, t2);
}
static void gen_vmulwod_d_w(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
{
TCGv_i64 t1, t2;
t1 = tcg_temp_new_i64();
t2 = tcg_temp_new_i64();
tcg_gen_sari_i64(t1, a, 32);
tcg_gen_sari_i64(t2, b, 32);
tcg_gen_mul_i64(t, t1, t2);
}
static void do_vmulwod_s(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
{
static const TCGOpcode vecop_list[] = {
INDEX_op_sari_vec, INDEX_op_mul_vec, 0
};
static const GVecGen3 op[3] = {
{
.fniv = gen_vmulwod_s,
.fno = gen_helper_vmulwod_h_b,
.opt_opc = vecop_list,
.vece = MO_16
},
{
.fni4 = gen_vmulwod_w_h,
.fniv = gen_vmulwod_s,
.fno = gen_helper_vmulwod_w_h,
.opt_opc = vecop_list,
.vece = MO_32
},
{
.fni8 = gen_vmulwod_d_w,
.fniv = gen_vmulwod_s,
.fno = gen_helper_vmulwod_d_w,
.opt_opc = vecop_list,
.vece = MO_64
},
};
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
}
TRANS(vmulwod_h_b, gvec_vvv, MO_8, do_vmulwod_s)
TRANS(vmulwod_w_h, gvec_vvv, MO_16, do_vmulwod_s)
TRANS(vmulwod_d_w, gvec_vvv, MO_32, do_vmulwod_s)
static void gen_vmulwev_u(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
{
TCGv_vec t1, t2, mask;
t1 = tcg_temp_new_vec_matching(a);
t2 = tcg_temp_new_vec_matching(b);
mask = tcg_constant_vec_matching(t, vece, MAKE_64BIT_MASK(0, 4 << vece));
tcg_gen_and_vec(vece, t1, a, mask);
tcg_gen_and_vec(vece, t2, b, mask);
tcg_gen_mul_vec(vece, t, t1, t2);
}
static void gen_vmulwev_w_hu(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
{
TCGv_i32 t1, t2;
t1 = tcg_temp_new_i32();
t2 = tcg_temp_new_i32();
tcg_gen_ext16u_i32(t1, a);
tcg_gen_ext16u_i32(t2, b);
tcg_gen_mul_i32(t, t1, t2);
}
static void gen_vmulwev_d_wu(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
{
TCGv_i64 t1, t2;
t1 = tcg_temp_new_i64();
t2 = tcg_temp_new_i64();
tcg_gen_ext32u_i64(t1, a);
tcg_gen_ext32u_i64(t2, b);
tcg_gen_mul_i64(t, t1, t2);
}
static void do_vmulwev_u(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
{
static const TCGOpcode vecop_list[] = {
INDEX_op_mul_vec, 0
};
static const GVecGen3 op[3] = {
{
.fniv = gen_vmulwev_u,
.fno = gen_helper_vmulwev_h_bu,
.opt_opc = vecop_list,
.vece = MO_16
},
{
.fni4 = gen_vmulwev_w_hu,
.fniv = gen_vmulwev_u,
.fno = gen_helper_vmulwev_w_hu,
.opt_opc = vecop_list,
.vece = MO_32
},
{
.fni8 = gen_vmulwev_d_wu,
.fniv = gen_vmulwev_u,
.fno = gen_helper_vmulwev_d_wu,
.opt_opc = vecop_list,
.vece = MO_64
},
};
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
}
TRANS(vmulwev_h_bu, gvec_vvv, MO_8, do_vmulwev_u)
TRANS(vmulwev_w_hu, gvec_vvv, MO_16, do_vmulwev_u)
TRANS(vmulwev_d_wu, gvec_vvv, MO_32, do_vmulwev_u)
static void gen_vmulwod_u(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
{
TCGv_vec t1, t2;
int halfbits = 4 << vece;
t1 = tcg_temp_new_vec_matching(a);
t2 = tcg_temp_new_vec_matching(b);
tcg_gen_shri_vec(vece, t1, a, halfbits);
tcg_gen_shri_vec(vece, t2, b, halfbits);
tcg_gen_mul_vec(vece, t, t1, t2);
}
static void gen_vmulwod_w_hu(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
{
TCGv_i32 t1, t2;
t1 = tcg_temp_new_i32();
t2 = tcg_temp_new_i32();
tcg_gen_shri_i32(t1, a, 16);
tcg_gen_shri_i32(t2, b, 16);
tcg_gen_mul_i32(t, t1, t2);
}
static void gen_vmulwod_d_wu(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
{
TCGv_i64 t1, t2;
t1 = tcg_temp_new_i64();
t2 = tcg_temp_new_i64();
tcg_gen_shri_i64(t1, a, 32);
tcg_gen_shri_i64(t2, b, 32);
tcg_gen_mul_i64(t, t1, t2);
}
static void do_vmulwod_u(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
{
static const TCGOpcode vecop_list[] = {
INDEX_op_shri_vec, INDEX_op_mul_vec, 0
};
static const GVecGen3 op[3] = {
{
.fniv = gen_vmulwod_u,
.fno = gen_helper_vmulwod_h_bu,
.opt_opc = vecop_list,
.vece = MO_16
},
{
.fni4 = gen_vmulwod_w_hu,
.fniv = gen_vmulwod_u,
.fno = gen_helper_vmulwod_w_hu,
.opt_opc = vecop_list,
.vece = MO_32
},
{
.fni8 = gen_vmulwod_d_wu,
.fniv = gen_vmulwod_u,
.fno = gen_helper_vmulwod_d_wu,
.opt_opc = vecop_list,
.vece = MO_64
},
};
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
}
TRANS(vmulwod_h_bu, gvec_vvv, MO_8, do_vmulwod_u)
TRANS(vmulwod_w_hu, gvec_vvv, MO_16, do_vmulwod_u)
TRANS(vmulwod_d_wu, gvec_vvv, MO_32, do_vmulwod_u)
static void gen_vmulwev_u_s(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
{
TCGv_vec t1, t2, mask;
int halfbits = 4 << vece;
t1 = tcg_temp_new_vec_matching(a);
t2 = tcg_temp_new_vec_matching(b);
mask = tcg_constant_vec_matching(t, vece, MAKE_64BIT_MASK(0, 4 << vece));
tcg_gen_and_vec(vece, t1, a, mask);
tcg_gen_shli_vec(vece, t2, b, halfbits);
tcg_gen_sari_vec(vece, t2, t2, halfbits);
tcg_gen_mul_vec(vece, t, t1, t2);
}
static void gen_vmulwev_w_hu_h(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
{
TCGv_i32 t1, t2;
t1 = tcg_temp_new_i32();
t2 = tcg_temp_new_i32();
tcg_gen_ext16u_i32(t1, a);
tcg_gen_ext16s_i32(t2, b);
tcg_gen_mul_i32(t, t1, t2);
}
static void gen_vmulwev_d_wu_w(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
{
TCGv_i64 t1, t2;
t1 = tcg_temp_new_i64();
t2 = tcg_temp_new_i64();
tcg_gen_ext32u_i64(t1, a);
tcg_gen_ext32s_i64(t2, b);
tcg_gen_mul_i64(t, t1, t2);
}
static void do_vmulwev_u_s(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
{
static const TCGOpcode vecop_list[] = {
INDEX_op_shli_vec, INDEX_op_sari_vec, INDEX_op_mul_vec, 0
};
static const GVecGen3 op[3] = {
{
.fniv = gen_vmulwev_u_s,
.fno = gen_helper_vmulwev_h_bu_b,
.opt_opc = vecop_list,
.vece = MO_16
},
{
.fni4 = gen_vmulwev_w_hu_h,
.fniv = gen_vmulwev_u_s,
.fno = gen_helper_vmulwev_w_hu_h,
.opt_opc = vecop_list,
.vece = MO_32
},
{
.fni8 = gen_vmulwev_d_wu_w,
.fniv = gen_vmulwev_u_s,
.fno = gen_helper_vmulwev_d_wu_w,
.opt_opc = vecop_list,
.vece = MO_64
},
};
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
}
TRANS(vmulwev_h_bu_b, gvec_vvv, MO_8, do_vmulwev_u_s)
TRANS(vmulwev_w_hu_h, gvec_vvv, MO_16, do_vmulwev_u_s)
TRANS(vmulwev_d_wu_w, gvec_vvv, MO_32, do_vmulwev_u_s)
static void gen_vmulwod_u_s(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
{
TCGv_vec t1, t2;
int halfbits = 4 << vece;
t1 = tcg_temp_new_vec_matching(a);
t2 = tcg_temp_new_vec_matching(b);
tcg_gen_shri_vec(vece, t1, a, halfbits);
tcg_gen_sari_vec(vece, t2, b, halfbits);
tcg_gen_mul_vec(vece, t, t1, t2);
}
static void gen_vmulwod_w_hu_h(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
{
TCGv_i32 t1, t2;
t1 = tcg_temp_new_i32();
t2 = tcg_temp_new_i32();
tcg_gen_shri_i32(t1, a, 16);
tcg_gen_sari_i32(t2, b, 16);
tcg_gen_mul_i32(t, t1, t2);
}
static void gen_vmulwod_d_wu_w(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
{
TCGv_i64 t1, t2;
t1 = tcg_temp_new_i64();
t2 = tcg_temp_new_i64();
tcg_gen_shri_i64(t1, a, 32);
tcg_gen_sari_i64(t2, b, 32);
tcg_gen_mul_i64(t, t1, t2);
}
static void do_vmulwod_u_s(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
{
static const TCGOpcode vecop_list[] = {
INDEX_op_shri_vec, INDEX_op_sari_vec, INDEX_op_mul_vec, 0
};
static const GVecGen3 op[3] = {
{
.fniv = gen_vmulwod_u_s,
.fno = gen_helper_vmulwod_h_bu_b,
.opt_opc = vecop_list,
.vece = MO_16
},
{
.fni4 = gen_vmulwod_w_hu_h,
.fniv = gen_vmulwod_u_s,
.fno = gen_helper_vmulwod_w_hu_h,
.opt_opc = vecop_list,
.vece = MO_32
},
{
.fni8 = gen_vmulwod_d_wu_w,
.fniv = gen_vmulwod_u_s,
.fno = gen_helper_vmulwod_d_wu_w,
.opt_opc = vecop_list,
.vece = MO_64
},
};
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
}
TRANS(vmulwod_h_bu_b, gvec_vvv, MO_8, do_vmulwod_u_s)
TRANS(vmulwod_w_hu_h, gvec_vvv, MO_16, do_vmulwod_u_s)
TRANS(vmulwod_d_wu_w, gvec_vvv, MO_32, do_vmulwod_u_s)

View File

@ -668,3 +668,41 @@ vmini_bu 0111 00101001 01100 ..... ..... ..... @vv_ui5
vmini_hu 0111 00101001 01101 ..... ..... ..... @vv_ui5
vmini_wu 0111 00101001 01110 ..... ..... ..... @vv_ui5
vmini_du 0111 00101001 01111 ..... ..... ..... @vv_ui5
vmul_b 0111 00001000 01000 ..... ..... ..... @vvv
vmul_h 0111 00001000 01001 ..... ..... ..... @vvv
vmul_w 0111 00001000 01010 ..... ..... ..... @vvv
vmul_d 0111 00001000 01011 ..... ..... ..... @vvv
vmuh_b 0111 00001000 01100 ..... ..... ..... @vvv
vmuh_h 0111 00001000 01101 ..... ..... ..... @vvv
vmuh_w 0111 00001000 01110 ..... ..... ..... @vvv
vmuh_d 0111 00001000 01111 ..... ..... ..... @vvv
vmuh_bu 0111 00001000 10000 ..... ..... ..... @vvv
vmuh_hu 0111 00001000 10001 ..... ..... ..... @vvv
vmuh_wu 0111 00001000 10010 ..... ..... ..... @vvv
vmuh_du 0111 00001000 10011 ..... ..... ..... @vvv
vmulwev_h_b 0111 00001001 00000 ..... ..... ..... @vvv
vmulwev_w_h 0111 00001001 00001 ..... ..... ..... @vvv
vmulwev_d_w 0111 00001001 00010 ..... ..... ..... @vvv
vmulwev_q_d 0111 00001001 00011 ..... ..... ..... @vvv
vmulwod_h_b 0111 00001001 00100 ..... ..... ..... @vvv
vmulwod_w_h 0111 00001001 00101 ..... ..... ..... @vvv
vmulwod_d_w 0111 00001001 00110 ..... ..... ..... @vvv
vmulwod_q_d 0111 00001001 00111 ..... ..... ..... @vvv
vmulwev_h_bu 0111 00001001 10000 ..... ..... ..... @vvv
vmulwev_w_hu 0111 00001001 10001 ..... ..... ..... @vvv
vmulwev_d_wu 0111 00001001 10010 ..... ..... ..... @vvv
vmulwev_q_du 0111 00001001 10011 ..... ..... ..... @vvv
vmulwod_h_bu 0111 00001001 10100 ..... ..... ..... @vvv
vmulwod_w_hu 0111 00001001 10101 ..... ..... ..... @vvv
vmulwod_d_wu 0111 00001001 10110 ..... ..... ..... @vvv
vmulwod_q_du 0111 00001001 10111 ..... ..... ..... @vvv
vmulwev_h_bu_b 0111 00001010 00000 ..... ..... ..... @vvv
vmulwev_w_hu_h 0111 00001010 00001 ..... ..... ..... @vvv
vmulwev_d_wu_w 0111 00001010 00010 ..... ..... ..... @vvv
vmulwev_q_du_d 0111 00001010 00011 ..... ..... ..... @vvv
vmulwod_h_bu_b 0111 00001010 00100 ..... ..... ..... @vvv
vmulwod_w_hu_h 0111 00001010 00101 ..... ..... ..... @vvv
vmulwod_d_wu_w 0111 00001010 00110 ..... ..... ..... @vvv
vmulwod_q_du_d 0111 00001010 00111 ..... ..... ..... @vvv

View File

@ -370,3 +370,79 @@ VMINMAXI(vmaxi_bu, 8, UB, DO_MAX)
VMINMAXI(vmaxi_hu, 16, UH, DO_MAX)
VMINMAXI(vmaxi_wu, 32, UW, DO_MAX)
VMINMAXI(vmaxi_du, 64, UD, DO_MAX)
#define DO_VMUH(NAME, BIT, E1, E2, DO_OP) \
void HELPER(NAME)(void *vd, void *vj, void *vk, uint32_t v) \
{ \
int i; \
VReg *Vd = (VReg *)vd; \
VReg *Vj = (VReg *)vj; \
VReg *Vk = (VReg *)vk; \
typedef __typeof(Vd->E1(0)) T; \
\
for (i = 0; i < LSX_LEN/BIT; i++) { \
Vd->E2(i) = ((T)Vj->E2(i)) * ((T)Vk->E2(i)) >> BIT; \
} \
}
void HELPER(vmuh_d)(void *vd, void *vj, void *vk, uint32_t v)
{
uint64_t l, h1, h2;
VReg *Vd = (VReg *)vd;
VReg *Vj = (VReg *)vj;
VReg *Vk = (VReg *)vk;
muls64(&l, &h1, Vj->D(0), Vk->D(0));
muls64(&l, &h2, Vj->D(1), Vk->D(1));
Vd->D(0) = h1;
Vd->D(1) = h2;
}
DO_VMUH(vmuh_b, 8, H, B, DO_MUH)
DO_VMUH(vmuh_h, 16, W, H, DO_MUH)
DO_VMUH(vmuh_w, 32, D, W, DO_MUH)
void HELPER(vmuh_du)(void *vd, void *vj, void *vk, uint32_t v)
{
uint64_t l, h1, h2;
VReg *Vd = (VReg *)vd;
VReg *Vj = (VReg *)vj;
VReg *Vk = (VReg *)vk;
mulu64(&l, &h1, Vj->D(0), Vk->D(0));
mulu64(&l, &h2, Vj->D(1), Vk->D(1));
Vd->D(0) = h1;
Vd->D(1) = h2;
}
DO_VMUH(vmuh_bu, 8, UH, UB, DO_MUH)
DO_VMUH(vmuh_hu, 16, UW, UH, DO_MUH)
DO_VMUH(vmuh_wu, 32, UD, UW, DO_MUH)
#define DO_MUL(a, b) (a * b)
DO_EVEN(vmulwev_h_b, 16, H, B, DO_MUL)
DO_EVEN(vmulwev_w_h, 32, W, H, DO_MUL)
DO_EVEN(vmulwev_d_w, 64, D, W, DO_MUL)
DO_ODD(vmulwod_h_b, 16, H, B, DO_MUL)
DO_ODD(vmulwod_w_h, 32, W, H, DO_MUL)
DO_ODD(vmulwod_d_w, 64, D, W, DO_MUL)
DO_EVEN(vmulwev_h_bu, 16, UH, UB, DO_MUL)
DO_EVEN(vmulwev_w_hu, 32, UW, UH, DO_MUL)
DO_EVEN(vmulwev_d_wu, 64, UD, UW, DO_MUL)
DO_ODD(vmulwod_h_bu, 16, UH, UB, DO_MUL)
DO_ODD(vmulwod_w_hu, 32, UW, UH, DO_MUL)
DO_ODD(vmulwod_d_wu, 64, UD, UW, DO_MUL)
DO_EVEN_U_S(vmulwev_h_bu_b, 16, H, UH, B, UB, DO_MUL)
DO_EVEN_U_S(vmulwev_w_hu_h, 32, W, UW, H, UH, DO_MUL)
DO_EVEN_U_S(vmulwev_d_wu_w, 64, D, UD, W, UW, DO_MUL)
DO_ODD_U_S(vmulwod_h_bu_b, 16, H, UH, B, UB, DO_MUL)
DO_ODD_U_S(vmulwod_w_hu_h, 32, W, UW, H, UH, DO_MUL)
DO_ODD_U_S(vmulwod_d_wu_w, 64, D, UD, W, UW, DO_MUL)