target/loongarch: Implement vaddw/vsubw
This patch includes: - VADDW{EV/OD}.{H.B/W.H/D.W/Q.D}[U]; - VSUBW{EV/OD}.{H.B/W.H/D.W/Q.D}[U]; - VADDW{EV/OD}.{H.BU.B/W.HU.H/D.WU.W/Q.DU.D}. Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Signed-off-by: Song Gao <gaosong@loongson.cn> Message-Id: <20230504122810.4094787-10-gaosong@loongson.cn>
This commit is contained in:
parent
c037fbc97d
commit
2d5f950c05
@ -865,3 +865,46 @@ INSN_LSX(vhsubw_hu_bu, vvv)
|
||||
INSN_LSX(vhsubw_wu_hu, vvv)
|
||||
INSN_LSX(vhsubw_du_wu, vvv)
|
||||
INSN_LSX(vhsubw_qu_du, vvv)
|
||||
|
||||
INSN_LSX(vaddwev_h_b, vvv)
|
||||
INSN_LSX(vaddwev_w_h, vvv)
|
||||
INSN_LSX(vaddwev_d_w, vvv)
|
||||
INSN_LSX(vaddwev_q_d, vvv)
|
||||
INSN_LSX(vaddwod_h_b, vvv)
|
||||
INSN_LSX(vaddwod_w_h, vvv)
|
||||
INSN_LSX(vaddwod_d_w, vvv)
|
||||
INSN_LSX(vaddwod_q_d, vvv)
|
||||
INSN_LSX(vsubwev_h_b, vvv)
|
||||
INSN_LSX(vsubwev_w_h, vvv)
|
||||
INSN_LSX(vsubwev_d_w, vvv)
|
||||
INSN_LSX(vsubwev_q_d, vvv)
|
||||
INSN_LSX(vsubwod_h_b, vvv)
|
||||
INSN_LSX(vsubwod_w_h, vvv)
|
||||
INSN_LSX(vsubwod_d_w, vvv)
|
||||
INSN_LSX(vsubwod_q_d, vvv)
|
||||
|
||||
INSN_LSX(vaddwev_h_bu, vvv)
|
||||
INSN_LSX(vaddwev_w_hu, vvv)
|
||||
INSN_LSX(vaddwev_d_wu, vvv)
|
||||
INSN_LSX(vaddwev_q_du, vvv)
|
||||
INSN_LSX(vaddwod_h_bu, vvv)
|
||||
INSN_LSX(vaddwod_w_hu, vvv)
|
||||
INSN_LSX(vaddwod_d_wu, vvv)
|
||||
INSN_LSX(vaddwod_q_du, vvv)
|
||||
INSN_LSX(vsubwev_h_bu, vvv)
|
||||
INSN_LSX(vsubwev_w_hu, vvv)
|
||||
INSN_LSX(vsubwev_d_wu, vvv)
|
||||
INSN_LSX(vsubwev_q_du, vvv)
|
||||
INSN_LSX(vsubwod_h_bu, vvv)
|
||||
INSN_LSX(vsubwod_w_hu, vvv)
|
||||
INSN_LSX(vsubwod_d_wu, vvv)
|
||||
INSN_LSX(vsubwod_q_du, vvv)
|
||||
|
||||
INSN_LSX(vaddwev_h_bu_b, vvv)
|
||||
INSN_LSX(vaddwev_w_hu_h, vvv)
|
||||
INSN_LSX(vaddwev_d_wu_w, vvv)
|
||||
INSN_LSX(vaddwev_q_du_d, vvv)
|
||||
INSN_LSX(vaddwod_h_bu_b, vvv)
|
||||
INSN_LSX(vaddwod_w_hu_h, vvv)
|
||||
INSN_LSX(vaddwod_d_wu_w, vvv)
|
||||
INSN_LSX(vaddwod_q_du_d, vvv)
|
||||
|
@ -148,3 +148,48 @@ DEF_HELPER_4(vhsubw_hu_bu, void, env, i32, i32, i32)
|
||||
DEF_HELPER_4(vhsubw_wu_hu, void, env, i32, i32, i32)
|
||||
DEF_HELPER_4(vhsubw_du_wu, void, env, i32, i32, i32)
|
||||
DEF_HELPER_4(vhsubw_qu_du, void, env, i32, i32, i32)
|
||||
|
||||
DEF_HELPER_FLAGS_4(vaddwev_h_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwev_w_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwev_d_w, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwev_q_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwod_h_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwod_w_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwod_d_w, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwod_q_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
|
||||
DEF_HELPER_FLAGS_4(vsubwev_h_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vsubwev_w_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vsubwev_d_w, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vsubwev_q_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vsubwod_h_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vsubwod_w_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vsubwod_d_w, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vsubwod_q_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
|
||||
DEF_HELPER_FLAGS_4(vaddwev_h_bu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwev_w_hu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwev_d_wu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwev_q_du, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwod_h_bu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwod_w_hu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwod_d_wu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwod_q_du, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
|
||||
DEF_HELPER_FLAGS_4(vsubwev_h_bu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vsubwev_w_hu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vsubwev_d_wu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vsubwev_q_du, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vsubwod_h_bu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vsubwod_w_hu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vsubwod_d_wu, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vsubwod_q_du, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
|
||||
DEF_HELPER_FLAGS_4(vaddwev_h_bu_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwev_w_hu_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwev_d_wu_w, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwev_q_du_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwod_h_bu_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwod_w_hu_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwod_d_wu_w, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
DEF_HELPER_FLAGS_4(vaddwod_q_du_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
|
||||
|
@ -174,3 +174,798 @@ TRANS(vhsubw_hu_bu, gen_vvv, gen_helper_vhsubw_hu_bu)
|
||||
TRANS(vhsubw_wu_hu, gen_vvv, gen_helper_vhsubw_wu_hu)
|
||||
TRANS(vhsubw_du_wu, gen_vvv, gen_helper_vhsubw_du_wu)
|
||||
TRANS(vhsubw_qu_du, gen_vvv, gen_helper_vhsubw_qu_du)
|
||||
|
||||
static void gen_vaddwev_s(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
|
||||
{
|
||||
TCGv_vec t1, t2;
|
||||
|
||||
int halfbits = 4 << vece;
|
||||
|
||||
t1 = tcg_temp_new_vec_matching(a);
|
||||
t2 = tcg_temp_new_vec_matching(b);
|
||||
|
||||
/* Sign-extend the even elements from a */
|
||||
tcg_gen_shli_vec(vece, t1, a, halfbits);
|
||||
tcg_gen_sari_vec(vece, t1, t1, halfbits);
|
||||
|
||||
/* Sign-extend the even elements from b */
|
||||
tcg_gen_shli_vec(vece, t2, b, halfbits);
|
||||
tcg_gen_sari_vec(vece, t2, t2, halfbits);
|
||||
|
||||
tcg_gen_add_vec(vece, t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vaddwev_w_h(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
|
||||
{
|
||||
TCGv_i32 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i32();
|
||||
t2 = tcg_temp_new_i32();
|
||||
tcg_gen_ext16s_i32(t1, a);
|
||||
tcg_gen_ext16s_i32(t2, b);
|
||||
tcg_gen_add_i32(t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vaddwev_d_w(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
|
||||
{
|
||||
TCGv_i64 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i64();
|
||||
t2 = tcg_temp_new_i64();
|
||||
tcg_gen_ext32s_i64(t1, a);
|
||||
tcg_gen_ext32s_i64(t2, b);
|
||||
tcg_gen_add_i64(t, t1, t2);
|
||||
}
|
||||
|
||||
static void do_vaddwev_s(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
|
||||
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
|
||||
{
|
||||
static const TCGOpcode vecop_list[] = {
|
||||
INDEX_op_shli_vec, INDEX_op_sari_vec, INDEX_op_add_vec, 0
|
||||
};
|
||||
static const GVecGen3 op[4] = {
|
||||
{
|
||||
.fniv = gen_vaddwev_s,
|
||||
.fno = gen_helper_vaddwev_h_b,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_16
|
||||
},
|
||||
{
|
||||
.fni4 = gen_vaddwev_w_h,
|
||||
.fniv = gen_vaddwev_s,
|
||||
.fno = gen_helper_vaddwev_w_h,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_32
|
||||
},
|
||||
{
|
||||
.fni8 = gen_vaddwev_d_w,
|
||||
.fniv = gen_vaddwev_s,
|
||||
.fno = gen_helper_vaddwev_d_w,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_64
|
||||
},
|
||||
{
|
||||
.fno = gen_helper_vaddwev_q_d,
|
||||
.vece = MO_128
|
||||
},
|
||||
};
|
||||
|
||||
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
|
||||
}
|
||||
|
||||
TRANS(vaddwev_h_b, gvec_vvv, MO_8, do_vaddwev_s)
|
||||
TRANS(vaddwev_w_h, gvec_vvv, MO_16, do_vaddwev_s)
|
||||
TRANS(vaddwev_d_w, gvec_vvv, MO_32, do_vaddwev_s)
|
||||
TRANS(vaddwev_q_d, gvec_vvv, MO_64, do_vaddwev_s)
|
||||
|
||||
static void gen_vaddwod_w_h(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
|
||||
{
|
||||
TCGv_i32 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i32();
|
||||
t2 = tcg_temp_new_i32();
|
||||
tcg_gen_sari_i32(t1, a, 16);
|
||||
tcg_gen_sari_i32(t2, b, 16);
|
||||
tcg_gen_add_i32(t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vaddwod_d_w(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
|
||||
{
|
||||
TCGv_i64 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i64();
|
||||
t2 = tcg_temp_new_i64();
|
||||
tcg_gen_sari_i64(t1, a, 32);
|
||||
tcg_gen_sari_i64(t2, b, 32);
|
||||
tcg_gen_add_i64(t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vaddwod_s(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
|
||||
{
|
||||
TCGv_vec t1, t2;
|
||||
|
||||
int halfbits = 4 << vece;
|
||||
|
||||
t1 = tcg_temp_new_vec_matching(a);
|
||||
t2 = tcg_temp_new_vec_matching(b);
|
||||
|
||||
/* Sign-extend the odd elements for vector */
|
||||
tcg_gen_sari_vec(vece, t1, a, halfbits);
|
||||
tcg_gen_sari_vec(vece, t2, b, halfbits);
|
||||
|
||||
tcg_gen_add_vec(vece, t, t1, t2);
|
||||
}
|
||||
|
||||
static void do_vaddwod_s(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
|
||||
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
|
||||
{
|
||||
static const TCGOpcode vecop_list[] = {
|
||||
INDEX_op_sari_vec, INDEX_op_add_vec, 0
|
||||
};
|
||||
static const GVecGen3 op[4] = {
|
||||
{
|
||||
.fniv = gen_vaddwod_s,
|
||||
.fno = gen_helper_vaddwod_h_b,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_16
|
||||
},
|
||||
{
|
||||
.fni4 = gen_vaddwod_w_h,
|
||||
.fniv = gen_vaddwod_s,
|
||||
.fno = gen_helper_vaddwod_w_h,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_32
|
||||
},
|
||||
{
|
||||
.fni8 = gen_vaddwod_d_w,
|
||||
.fniv = gen_vaddwod_s,
|
||||
.fno = gen_helper_vaddwod_d_w,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_64
|
||||
},
|
||||
{
|
||||
.fno = gen_helper_vaddwod_q_d,
|
||||
.vece = MO_128
|
||||
},
|
||||
};
|
||||
|
||||
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
|
||||
}
|
||||
|
||||
TRANS(vaddwod_h_b, gvec_vvv, MO_8, do_vaddwod_s)
|
||||
TRANS(vaddwod_w_h, gvec_vvv, MO_16, do_vaddwod_s)
|
||||
TRANS(vaddwod_d_w, gvec_vvv, MO_32, do_vaddwod_s)
|
||||
TRANS(vaddwod_q_d, gvec_vvv, MO_64, do_vaddwod_s)
|
||||
|
||||
static void gen_vsubwev_s(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
|
||||
{
|
||||
TCGv_vec t1, t2;
|
||||
|
||||
int halfbits = 4 << vece;
|
||||
|
||||
t1 = tcg_temp_new_vec_matching(a);
|
||||
t2 = tcg_temp_new_vec_matching(b);
|
||||
|
||||
/* Sign-extend the even elements from a */
|
||||
tcg_gen_shli_vec(vece, t1, a, halfbits);
|
||||
tcg_gen_sari_vec(vece, t1, t1, halfbits);
|
||||
|
||||
/* Sign-extend the even elements from b */
|
||||
tcg_gen_shli_vec(vece, t2, b, halfbits);
|
||||
tcg_gen_sari_vec(vece, t2, t2, halfbits);
|
||||
|
||||
tcg_gen_sub_vec(vece, t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vsubwev_w_h(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
|
||||
{
|
||||
TCGv_i32 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i32();
|
||||
t2 = tcg_temp_new_i32();
|
||||
tcg_gen_ext16s_i32(t1, a);
|
||||
tcg_gen_ext16s_i32(t2, b);
|
||||
tcg_gen_sub_i32(t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vsubwev_d_w(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
|
||||
{
|
||||
TCGv_i64 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i64();
|
||||
t2 = tcg_temp_new_i64();
|
||||
tcg_gen_ext32s_i64(t1, a);
|
||||
tcg_gen_ext32s_i64(t2, b);
|
||||
tcg_gen_sub_i64(t, t1, t2);
|
||||
}
|
||||
|
||||
static void do_vsubwev_s(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
|
||||
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
|
||||
{
|
||||
static const TCGOpcode vecop_list[] = {
|
||||
INDEX_op_shli_vec, INDEX_op_sari_vec, INDEX_op_sub_vec, 0
|
||||
};
|
||||
static const GVecGen3 op[4] = {
|
||||
{
|
||||
.fniv = gen_vsubwev_s,
|
||||
.fno = gen_helper_vsubwev_h_b,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_16
|
||||
},
|
||||
{
|
||||
.fni4 = gen_vsubwev_w_h,
|
||||
.fniv = gen_vsubwev_s,
|
||||
.fno = gen_helper_vsubwev_w_h,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_32
|
||||
},
|
||||
{
|
||||
.fni8 = gen_vsubwev_d_w,
|
||||
.fniv = gen_vsubwev_s,
|
||||
.fno = gen_helper_vsubwev_d_w,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_64
|
||||
},
|
||||
{
|
||||
.fno = gen_helper_vsubwev_q_d,
|
||||
.vece = MO_128
|
||||
},
|
||||
};
|
||||
|
||||
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
|
||||
}
|
||||
|
||||
TRANS(vsubwev_h_b, gvec_vvv, MO_8, do_vsubwev_s)
|
||||
TRANS(vsubwev_w_h, gvec_vvv, MO_16, do_vsubwev_s)
|
||||
TRANS(vsubwev_d_w, gvec_vvv, MO_32, do_vsubwev_s)
|
||||
TRANS(vsubwev_q_d, gvec_vvv, MO_64, do_vsubwev_s)
|
||||
|
||||
static void gen_vsubwod_s(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
|
||||
{
|
||||
TCGv_vec t1, t2;
|
||||
|
||||
int halfbits = 4 << vece;
|
||||
|
||||
t1 = tcg_temp_new_vec_matching(a);
|
||||
t2 = tcg_temp_new_vec_matching(b);
|
||||
|
||||
/* Sign-extend the odd elements for vector */
|
||||
tcg_gen_sari_vec(vece, t1, a, halfbits);
|
||||
tcg_gen_sari_vec(vece, t2, b, halfbits);
|
||||
|
||||
tcg_gen_sub_vec(vece, t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vsubwod_w_h(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
|
||||
{
|
||||
TCGv_i32 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i32();
|
||||
t2 = tcg_temp_new_i32();
|
||||
tcg_gen_sari_i32(t1, a, 16);
|
||||
tcg_gen_sari_i32(t2, b, 16);
|
||||
tcg_gen_sub_i32(t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vsubwod_d_w(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
|
||||
{
|
||||
TCGv_i64 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i64();
|
||||
t2 = tcg_temp_new_i64();
|
||||
tcg_gen_sari_i64(t1, a, 32);
|
||||
tcg_gen_sari_i64(t2, b, 32);
|
||||
tcg_gen_sub_i64(t, t1, t2);
|
||||
}
|
||||
|
||||
static void do_vsubwod_s(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
|
||||
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
|
||||
{
|
||||
static const TCGOpcode vecop_list[] = {
|
||||
INDEX_op_sari_vec, INDEX_op_sub_vec, 0
|
||||
};
|
||||
static const GVecGen3 op[4] = {
|
||||
{
|
||||
.fniv = gen_vsubwod_s,
|
||||
.fno = gen_helper_vsubwod_h_b,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_16
|
||||
},
|
||||
{
|
||||
.fni4 = gen_vsubwod_w_h,
|
||||
.fniv = gen_vsubwod_s,
|
||||
.fno = gen_helper_vsubwod_w_h,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_32
|
||||
},
|
||||
{
|
||||
.fni8 = gen_vsubwod_d_w,
|
||||
.fniv = gen_vsubwod_s,
|
||||
.fno = gen_helper_vsubwod_d_w,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_64
|
||||
},
|
||||
{
|
||||
.fno = gen_helper_vsubwod_q_d,
|
||||
.vece = MO_128
|
||||
},
|
||||
};
|
||||
|
||||
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
|
||||
}
|
||||
|
||||
TRANS(vsubwod_h_b, gvec_vvv, MO_8, do_vsubwod_s)
|
||||
TRANS(vsubwod_w_h, gvec_vvv, MO_16, do_vsubwod_s)
|
||||
TRANS(vsubwod_d_w, gvec_vvv, MO_32, do_vsubwod_s)
|
||||
TRANS(vsubwod_q_d, gvec_vvv, MO_64, do_vsubwod_s)
|
||||
|
||||
static void gen_vaddwev_u(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
|
||||
{
|
||||
TCGv_vec t1, t2, t3;
|
||||
|
||||
t1 = tcg_temp_new_vec_matching(a);
|
||||
t2 = tcg_temp_new_vec_matching(b);
|
||||
t3 = tcg_constant_vec_matching(t, vece, MAKE_64BIT_MASK(0, 4 << vece));
|
||||
tcg_gen_and_vec(vece, t1, a, t3);
|
||||
tcg_gen_and_vec(vece, t2, b, t3);
|
||||
tcg_gen_add_vec(vece, t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vaddwev_w_hu(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
|
||||
{
|
||||
TCGv_i32 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i32();
|
||||
t2 = tcg_temp_new_i32();
|
||||
tcg_gen_ext16u_i32(t1, a);
|
||||
tcg_gen_ext16u_i32(t2, b);
|
||||
tcg_gen_add_i32(t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vaddwev_d_wu(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
|
||||
{
|
||||
TCGv_i64 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i64();
|
||||
t2 = tcg_temp_new_i64();
|
||||
tcg_gen_ext32u_i64(t1, a);
|
||||
tcg_gen_ext32u_i64(t2, b);
|
||||
tcg_gen_add_i64(t, t1, t2);
|
||||
}
|
||||
|
||||
static void do_vaddwev_u(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
|
||||
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
|
||||
{
|
||||
static const TCGOpcode vecop_list[] = {
|
||||
INDEX_op_add_vec, 0
|
||||
};
|
||||
static const GVecGen3 op[4] = {
|
||||
{
|
||||
.fniv = gen_vaddwev_u,
|
||||
.fno = gen_helper_vaddwev_h_bu,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_16
|
||||
},
|
||||
{
|
||||
.fni4 = gen_vaddwev_w_hu,
|
||||
.fniv = gen_vaddwev_u,
|
||||
.fno = gen_helper_vaddwev_w_hu,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_32
|
||||
},
|
||||
{
|
||||
.fni8 = gen_vaddwev_d_wu,
|
||||
.fniv = gen_vaddwev_u,
|
||||
.fno = gen_helper_vaddwev_d_wu,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_64
|
||||
},
|
||||
{
|
||||
.fno = gen_helper_vaddwev_q_du,
|
||||
.vece = MO_128
|
||||
},
|
||||
};
|
||||
|
||||
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
|
||||
}
|
||||
|
||||
TRANS(vaddwev_h_bu, gvec_vvv, MO_8, do_vaddwev_u)
|
||||
TRANS(vaddwev_w_hu, gvec_vvv, MO_16, do_vaddwev_u)
|
||||
TRANS(vaddwev_d_wu, gvec_vvv, MO_32, do_vaddwev_u)
|
||||
TRANS(vaddwev_q_du, gvec_vvv, MO_64, do_vaddwev_u)
|
||||
|
||||
static void gen_vaddwod_u(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
|
||||
{
|
||||
TCGv_vec t1, t2;
|
||||
|
||||
int halfbits = 4 << vece;
|
||||
|
||||
t1 = tcg_temp_new_vec_matching(a);
|
||||
t2 = tcg_temp_new_vec_matching(b);
|
||||
|
||||
/* Zero-extend the odd elements for vector */
|
||||
tcg_gen_shri_vec(vece, t1, a, halfbits);
|
||||
tcg_gen_shri_vec(vece, t2, b, halfbits);
|
||||
|
||||
tcg_gen_add_vec(vece, t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vaddwod_w_hu(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
|
||||
{
|
||||
TCGv_i32 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i32();
|
||||
t2 = tcg_temp_new_i32();
|
||||
tcg_gen_shri_i32(t1, a, 16);
|
||||
tcg_gen_shri_i32(t2, b, 16);
|
||||
tcg_gen_add_i32(t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vaddwod_d_wu(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
|
||||
{
|
||||
TCGv_i64 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i64();
|
||||
t2 = tcg_temp_new_i64();
|
||||
tcg_gen_shri_i64(t1, a, 32);
|
||||
tcg_gen_shri_i64(t2, b, 32);
|
||||
tcg_gen_add_i64(t, t1, t2);
|
||||
}
|
||||
|
||||
static void do_vaddwod_u(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
|
||||
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
|
||||
{
|
||||
static const TCGOpcode vecop_list[] = {
|
||||
INDEX_op_shri_vec, INDEX_op_add_vec, 0
|
||||
};
|
||||
static const GVecGen3 op[4] = {
|
||||
{
|
||||
.fniv = gen_vaddwod_u,
|
||||
.fno = gen_helper_vaddwod_h_bu,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_16
|
||||
},
|
||||
{
|
||||
.fni4 = gen_vaddwod_w_hu,
|
||||
.fniv = gen_vaddwod_u,
|
||||
.fno = gen_helper_vaddwod_w_hu,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_32
|
||||
},
|
||||
{
|
||||
.fni8 = gen_vaddwod_d_wu,
|
||||
.fniv = gen_vaddwod_u,
|
||||
.fno = gen_helper_vaddwod_d_wu,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_64
|
||||
},
|
||||
{
|
||||
.fno = gen_helper_vaddwod_q_du,
|
||||
.vece = MO_128
|
||||
},
|
||||
};
|
||||
|
||||
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
|
||||
}
|
||||
|
||||
TRANS(vaddwod_h_bu, gvec_vvv, MO_8, do_vaddwod_u)
|
||||
TRANS(vaddwod_w_hu, gvec_vvv, MO_16, do_vaddwod_u)
|
||||
TRANS(vaddwod_d_wu, gvec_vvv, MO_32, do_vaddwod_u)
|
||||
TRANS(vaddwod_q_du, gvec_vvv, MO_64, do_vaddwod_u)
|
||||
|
||||
static void gen_vsubwev_u(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
|
||||
{
|
||||
TCGv_vec t1, t2, t3;
|
||||
|
||||
t1 = tcg_temp_new_vec_matching(a);
|
||||
t2 = tcg_temp_new_vec_matching(b);
|
||||
t3 = tcg_constant_vec_matching(t, vece, MAKE_64BIT_MASK(0, 4 << vece));
|
||||
tcg_gen_and_vec(vece, t1, a, t3);
|
||||
tcg_gen_and_vec(vece, t2, b, t3);
|
||||
tcg_gen_sub_vec(vece, t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vsubwev_w_hu(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
|
||||
{
|
||||
TCGv_i32 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i32();
|
||||
t2 = tcg_temp_new_i32();
|
||||
tcg_gen_ext16u_i32(t1, a);
|
||||
tcg_gen_ext16u_i32(t2, b);
|
||||
tcg_gen_sub_i32(t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vsubwev_d_wu(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
|
||||
{
|
||||
TCGv_i64 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i64();
|
||||
t2 = tcg_temp_new_i64();
|
||||
tcg_gen_ext32u_i64(t1, a);
|
||||
tcg_gen_ext32u_i64(t2, b);
|
||||
tcg_gen_sub_i64(t, t1, t2);
|
||||
}
|
||||
|
||||
static void do_vsubwev_u(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
|
||||
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
|
||||
{
|
||||
static const TCGOpcode vecop_list[] = {
|
||||
INDEX_op_sub_vec, 0
|
||||
};
|
||||
static const GVecGen3 op[4] = {
|
||||
{
|
||||
.fniv = gen_vsubwev_u,
|
||||
.fno = gen_helper_vsubwev_h_bu,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_16
|
||||
},
|
||||
{
|
||||
.fni4 = gen_vsubwev_w_hu,
|
||||
.fniv = gen_vsubwev_u,
|
||||
.fno = gen_helper_vsubwev_w_hu,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_32
|
||||
},
|
||||
{
|
||||
.fni8 = gen_vsubwev_d_wu,
|
||||
.fniv = gen_vsubwev_u,
|
||||
.fno = gen_helper_vsubwev_d_wu,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_64
|
||||
},
|
||||
{
|
||||
.fno = gen_helper_vsubwev_q_du,
|
||||
.vece = MO_128
|
||||
},
|
||||
};
|
||||
|
||||
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
|
||||
}
|
||||
|
||||
TRANS(vsubwev_h_bu, gvec_vvv, MO_8, do_vsubwev_u)
|
||||
TRANS(vsubwev_w_hu, gvec_vvv, MO_16, do_vsubwev_u)
|
||||
TRANS(vsubwev_d_wu, gvec_vvv, MO_32, do_vsubwev_u)
|
||||
TRANS(vsubwev_q_du, gvec_vvv, MO_64, do_vsubwev_u)
|
||||
|
||||
static void gen_vsubwod_u(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
|
||||
{
|
||||
TCGv_vec t1, t2;
|
||||
|
||||
int halfbits = 4 << vece;
|
||||
|
||||
t1 = tcg_temp_new_vec_matching(a);
|
||||
t2 = tcg_temp_new_vec_matching(b);
|
||||
|
||||
/* Zero-extend the odd elements for vector */
|
||||
tcg_gen_shri_vec(vece, t1, a, halfbits);
|
||||
tcg_gen_shri_vec(vece, t2, b, halfbits);
|
||||
|
||||
tcg_gen_sub_vec(vece, t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vsubwod_w_hu(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
|
||||
{
|
||||
TCGv_i32 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i32();
|
||||
t2 = tcg_temp_new_i32();
|
||||
tcg_gen_shri_i32(t1, a, 16);
|
||||
tcg_gen_shri_i32(t2, b, 16);
|
||||
tcg_gen_sub_i32(t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vsubwod_d_wu(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
|
||||
{
|
||||
TCGv_i64 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i64();
|
||||
t2 = tcg_temp_new_i64();
|
||||
tcg_gen_shri_i64(t1, a, 32);
|
||||
tcg_gen_shri_i64(t2, b, 32);
|
||||
tcg_gen_sub_i64(t, t1, t2);
|
||||
}
|
||||
|
||||
static void do_vsubwod_u(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
|
||||
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
|
||||
{
|
||||
static const TCGOpcode vecop_list[] = {
|
||||
INDEX_op_shri_vec, INDEX_op_sub_vec, 0
|
||||
};
|
||||
static const GVecGen3 op[4] = {
|
||||
{
|
||||
.fniv = gen_vsubwod_u,
|
||||
.fno = gen_helper_vsubwod_h_bu,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_16
|
||||
},
|
||||
{
|
||||
.fni4 = gen_vsubwod_w_hu,
|
||||
.fniv = gen_vsubwod_u,
|
||||
.fno = gen_helper_vsubwod_w_hu,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_32
|
||||
},
|
||||
{
|
||||
.fni8 = gen_vsubwod_d_wu,
|
||||
.fniv = gen_vsubwod_u,
|
||||
.fno = gen_helper_vsubwod_d_wu,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_64
|
||||
},
|
||||
{
|
||||
.fno = gen_helper_vsubwod_q_du,
|
||||
.vece = MO_128
|
||||
},
|
||||
};
|
||||
|
||||
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
|
||||
}
|
||||
|
||||
TRANS(vsubwod_h_bu, gvec_vvv, MO_8, do_vsubwod_u)
|
||||
TRANS(vsubwod_w_hu, gvec_vvv, MO_16, do_vsubwod_u)
|
||||
TRANS(vsubwod_d_wu, gvec_vvv, MO_32, do_vsubwod_u)
|
||||
TRANS(vsubwod_q_du, gvec_vvv, MO_64, do_vsubwod_u)
|
||||
|
||||
static void gen_vaddwev_u_s(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
|
||||
{
|
||||
TCGv_vec t1, t2, t3;
|
||||
|
||||
int halfbits = 4 << vece;
|
||||
|
||||
t1 = tcg_temp_new_vec_matching(a);
|
||||
t2 = tcg_temp_new_vec_matching(b);
|
||||
t3 = tcg_constant_vec_matching(t, vece, MAKE_64BIT_MASK(0, halfbits));
|
||||
|
||||
/* Zero-extend the even elements from a */
|
||||
tcg_gen_and_vec(vece, t1, a, t3);
|
||||
|
||||
/* Sign-extend the even elements from b */
|
||||
tcg_gen_shli_vec(vece, t2, b, halfbits);
|
||||
tcg_gen_sari_vec(vece, t2, t2, halfbits);
|
||||
|
||||
tcg_gen_add_vec(vece, t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vaddwev_w_hu_h(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
|
||||
{
|
||||
TCGv_i32 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i32();
|
||||
t2 = tcg_temp_new_i32();
|
||||
tcg_gen_ext16u_i32(t1, a);
|
||||
tcg_gen_ext16s_i32(t2, b);
|
||||
tcg_gen_add_i32(t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vaddwev_d_wu_w(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
|
||||
{
|
||||
TCGv_i64 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i64();
|
||||
t2 = tcg_temp_new_i64();
|
||||
tcg_gen_ext32u_i64(t1, a);
|
||||
tcg_gen_ext32s_i64(t2, b);
|
||||
tcg_gen_add_i64(t, t1, t2);
|
||||
}
|
||||
|
||||
static void do_vaddwev_u_s(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
|
||||
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
|
||||
{
|
||||
static const TCGOpcode vecop_list[] = {
|
||||
INDEX_op_shli_vec, INDEX_op_sari_vec, INDEX_op_add_vec, 0
|
||||
};
|
||||
static const GVecGen3 op[4] = {
|
||||
{
|
||||
.fniv = gen_vaddwev_u_s,
|
||||
.fno = gen_helper_vaddwev_h_bu_b,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_16
|
||||
},
|
||||
{
|
||||
.fni4 = gen_vaddwev_w_hu_h,
|
||||
.fniv = gen_vaddwev_u_s,
|
||||
.fno = gen_helper_vaddwev_w_hu_h,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_32
|
||||
},
|
||||
{
|
||||
.fni8 = gen_vaddwev_d_wu_w,
|
||||
.fniv = gen_vaddwev_u_s,
|
||||
.fno = gen_helper_vaddwev_d_wu_w,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_64
|
||||
},
|
||||
{
|
||||
.fno = gen_helper_vaddwev_q_du_d,
|
||||
.vece = MO_128
|
||||
},
|
||||
};
|
||||
|
||||
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
|
||||
}
|
||||
|
||||
TRANS(vaddwev_h_bu_b, gvec_vvv, MO_8, do_vaddwev_u_s)
|
||||
TRANS(vaddwev_w_hu_h, gvec_vvv, MO_16, do_vaddwev_u_s)
|
||||
TRANS(vaddwev_d_wu_w, gvec_vvv, MO_32, do_vaddwev_u_s)
|
||||
TRANS(vaddwev_q_du_d, gvec_vvv, MO_64, do_vaddwev_u_s)
|
||||
|
||||
static void gen_vaddwod_u_s(unsigned vece, TCGv_vec t, TCGv_vec a, TCGv_vec b)
|
||||
{
|
||||
TCGv_vec t1, t2;
|
||||
|
||||
int halfbits = 4 << vece;
|
||||
|
||||
t1 = tcg_temp_new_vec_matching(a);
|
||||
t2 = tcg_temp_new_vec_matching(b);
|
||||
|
||||
/* Zero-extend the odd elements from a */
|
||||
tcg_gen_shri_vec(vece, t1, a, halfbits);
|
||||
/* Sign-extend the odd elements from b */
|
||||
tcg_gen_sari_vec(vece, t2, b, halfbits);
|
||||
|
||||
tcg_gen_add_vec(vece, t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vaddwod_w_hu_h(TCGv_i32 t, TCGv_i32 a, TCGv_i32 b)
|
||||
{
|
||||
TCGv_i32 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i32();
|
||||
t2 = tcg_temp_new_i32();
|
||||
tcg_gen_shri_i32(t1, a, 16);
|
||||
tcg_gen_sari_i32(t2, b, 16);
|
||||
tcg_gen_add_i32(t, t1, t2);
|
||||
}
|
||||
|
||||
static void gen_vaddwod_d_wu_w(TCGv_i64 t, TCGv_i64 a, TCGv_i64 b)
|
||||
{
|
||||
TCGv_i64 t1, t2;
|
||||
|
||||
t1 = tcg_temp_new_i64();
|
||||
t2 = tcg_temp_new_i64();
|
||||
tcg_gen_shri_i64(t1, a, 32);
|
||||
tcg_gen_sari_i64(t2, b, 32);
|
||||
tcg_gen_add_i64(t, t1, t2);
|
||||
}
|
||||
|
||||
static void do_vaddwod_u_s(unsigned vece, uint32_t vd_ofs, uint32_t vj_ofs,
|
||||
uint32_t vk_ofs, uint32_t oprsz, uint32_t maxsz)
|
||||
{
|
||||
static const TCGOpcode vecop_list[] = {
|
||||
INDEX_op_shri_vec, INDEX_op_sari_vec, INDEX_op_add_vec, 0
|
||||
};
|
||||
static const GVecGen3 op[4] = {
|
||||
{
|
||||
.fniv = gen_vaddwod_u_s,
|
||||
.fno = gen_helper_vaddwod_h_bu_b,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_16
|
||||
},
|
||||
{
|
||||
.fni4 = gen_vaddwod_w_hu_h,
|
||||
.fniv = gen_vaddwod_u_s,
|
||||
.fno = gen_helper_vaddwod_w_hu_h,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_32
|
||||
},
|
||||
{
|
||||
.fni8 = gen_vaddwod_d_wu_w,
|
||||
.fniv = gen_vaddwod_u_s,
|
||||
.fno = gen_helper_vaddwod_d_wu_w,
|
||||
.opt_opc = vecop_list,
|
||||
.vece = MO_64
|
||||
},
|
||||
{
|
||||
.fno = gen_helper_vaddwod_q_du_d,
|
||||
.vece = MO_128
|
||||
},
|
||||
};
|
||||
|
||||
tcg_gen_gvec_3(vd_ofs, vj_ofs, vk_ofs, oprsz, maxsz, &op[vece]);
|
||||
}
|
||||
|
||||
TRANS(vaddwod_h_bu_b, gvec_vvv, MO_8, do_vaddwod_u_s)
|
||||
TRANS(vaddwod_w_hu_h, gvec_vvv, MO_16, do_vaddwod_u_s)
|
||||
TRANS(vaddwod_d_wu_w, gvec_vvv, MO_32, do_vaddwod_u_s)
|
||||
TRANS(vaddwod_q_du_d, gvec_vvv, MO_64, do_vaddwod_u_s)
|
||||
|
@ -559,3 +559,46 @@ vhsubw_hu_bu 0111 00000101 10100 ..... ..... ..... @vvv
|
||||
vhsubw_wu_hu 0111 00000101 10101 ..... ..... ..... @vvv
|
||||
vhsubw_du_wu 0111 00000101 10110 ..... ..... ..... @vvv
|
||||
vhsubw_qu_du 0111 00000101 10111 ..... ..... ..... @vvv
|
||||
|
||||
vaddwev_h_b 0111 00000001 11100 ..... ..... ..... @vvv
|
||||
vaddwev_w_h 0111 00000001 11101 ..... ..... ..... @vvv
|
||||
vaddwev_d_w 0111 00000001 11110 ..... ..... ..... @vvv
|
||||
vaddwev_q_d 0111 00000001 11111 ..... ..... ..... @vvv
|
||||
vaddwod_h_b 0111 00000010 00100 ..... ..... ..... @vvv
|
||||
vaddwod_w_h 0111 00000010 00101 ..... ..... ..... @vvv
|
||||
vaddwod_d_w 0111 00000010 00110 ..... ..... ..... @vvv
|
||||
vaddwod_q_d 0111 00000010 00111 ..... ..... ..... @vvv
|
||||
vsubwev_h_b 0111 00000010 00000 ..... ..... ..... @vvv
|
||||
vsubwev_w_h 0111 00000010 00001 ..... ..... ..... @vvv
|
||||
vsubwev_d_w 0111 00000010 00010 ..... ..... ..... @vvv
|
||||
vsubwev_q_d 0111 00000010 00011 ..... ..... ..... @vvv
|
||||
vsubwod_h_b 0111 00000010 01000 ..... ..... ..... @vvv
|
||||
vsubwod_w_h 0111 00000010 01001 ..... ..... ..... @vvv
|
||||
vsubwod_d_w 0111 00000010 01010 ..... ..... ..... @vvv
|
||||
vsubwod_q_d 0111 00000010 01011 ..... ..... ..... @vvv
|
||||
|
||||
vaddwev_h_bu 0111 00000010 11100 ..... ..... ..... @vvv
|
||||
vaddwev_w_hu 0111 00000010 11101 ..... ..... ..... @vvv
|
||||
vaddwev_d_wu 0111 00000010 11110 ..... ..... ..... @vvv
|
||||
vaddwev_q_du 0111 00000010 11111 ..... ..... ..... @vvv
|
||||
vaddwod_h_bu 0111 00000011 00100 ..... ..... ..... @vvv
|
||||
vaddwod_w_hu 0111 00000011 00101 ..... ..... ..... @vvv
|
||||
vaddwod_d_wu 0111 00000011 00110 ..... ..... ..... @vvv
|
||||
vaddwod_q_du 0111 00000011 00111 ..... ..... ..... @vvv
|
||||
vsubwev_h_bu 0111 00000011 00000 ..... ..... ..... @vvv
|
||||
vsubwev_w_hu 0111 00000011 00001 ..... ..... ..... @vvv
|
||||
vsubwev_d_wu 0111 00000011 00010 ..... ..... ..... @vvv
|
||||
vsubwev_q_du 0111 00000011 00011 ..... ..... ..... @vvv
|
||||
vsubwod_h_bu 0111 00000011 01000 ..... ..... ..... @vvv
|
||||
vsubwod_w_hu 0111 00000011 01001 ..... ..... ..... @vvv
|
||||
vsubwod_d_wu 0111 00000011 01010 ..... ..... ..... @vvv
|
||||
vsubwod_q_du 0111 00000011 01011 ..... ..... ..... @vvv
|
||||
|
||||
vaddwev_h_bu_b 0111 00000011 11100 ..... ..... ..... @vvv
|
||||
vaddwev_w_hu_h 0111 00000011 11101 ..... ..... ..... @vvv
|
||||
vaddwev_d_wu_w 0111 00000011 11110 ..... ..... ..... @vvv
|
||||
vaddwev_q_du_d 0111 00000011 11111 ..... ..... ..... @vvv
|
||||
vaddwod_h_bu_b 0111 00000100 00000 ..... ..... ..... @vvv
|
||||
vaddwod_w_hu_h 0111 00000100 00001 ..... ..... ..... @vvv
|
||||
vaddwod_d_wu_w 0111 00000100 00010 ..... ..... ..... @vvv
|
||||
vaddwod_q_du_d 0111 00000100 00011 ..... ..... ..... @vvv
|
||||
|
@ -85,3 +85,193 @@ void HELPER(vhsubw_qu_du)(CPULoongArchState *env,
|
||||
Vd->Q(0) = int128_sub(int128_make64((uint64_t)Vj->D(1)),
|
||||
int128_make64((uint64_t)Vk->D(0)));
|
||||
}
|
||||
|
||||
#define DO_EVEN(NAME, BIT, E1, E2, DO_OP) \
|
||||
void HELPER(NAME)(void *vd, void *vj, void *vk, uint32_t v) \
|
||||
{ \
|
||||
int i; \
|
||||
VReg *Vd = (VReg *)vd; \
|
||||
VReg *Vj = (VReg *)vj; \
|
||||
VReg *Vk = (VReg *)vk; \
|
||||
typedef __typeof(Vd->E1(0)) TD; \
|
||||
for (i = 0; i < LSX_LEN/BIT; i++) { \
|
||||
Vd->E1(i) = DO_OP((TD)Vj->E2(2 * i) ,(TD)Vk->E2(2 * i)); \
|
||||
} \
|
||||
}
|
||||
|
||||
#define DO_ODD(NAME, BIT, E1, E2, DO_OP) \
|
||||
void HELPER(NAME)(void *vd, void *vj, void *vk, uint32_t v) \
|
||||
{ \
|
||||
int i; \
|
||||
VReg *Vd = (VReg *)vd; \
|
||||
VReg *Vj = (VReg *)vj; \
|
||||
VReg *Vk = (VReg *)vk; \
|
||||
typedef __typeof(Vd->E1(0)) TD; \
|
||||
for (i = 0; i < LSX_LEN/BIT; i++) { \
|
||||
Vd->E1(i) = DO_OP((TD)Vj->E2(2 * i + 1), (TD)Vk->E2(2 * i + 1)); \
|
||||
} \
|
||||
}
|
||||
|
||||
void HELPER(vaddwev_q_d)(void *vd, void *vj, void *vk, uint32_t v)
|
||||
{
|
||||
VReg *Vd = (VReg *)vd;
|
||||
VReg *Vj = (VReg *)vj;
|
||||
VReg *Vk = (VReg *)vk;
|
||||
|
||||
Vd->Q(0) = int128_add(int128_makes64(Vj->D(0)), int128_makes64(Vk->D(0)));
|
||||
}
|
||||
|
||||
DO_EVEN(vaddwev_h_b, 16, H, B, DO_ADD)
|
||||
DO_EVEN(vaddwev_w_h, 32, W, H, DO_ADD)
|
||||
DO_EVEN(vaddwev_d_w, 64, D, W, DO_ADD)
|
||||
|
||||
void HELPER(vaddwod_q_d)(void *vd, void *vj, void *vk, uint32_t v)
|
||||
{
|
||||
VReg *Vd = (VReg *)vd;
|
||||
VReg *Vj = (VReg *)vj;
|
||||
VReg *Vk = (VReg *)vk;
|
||||
|
||||
Vd->Q(0) = int128_add(int128_makes64(Vj->D(1)), int128_makes64(Vk->D(1)));
|
||||
}
|
||||
|
||||
DO_ODD(vaddwod_h_b, 16, H, B, DO_ADD)
|
||||
DO_ODD(vaddwod_w_h, 32, W, H, DO_ADD)
|
||||
DO_ODD(vaddwod_d_w, 64, D, W, DO_ADD)
|
||||
|
||||
void HELPER(vsubwev_q_d)(void *vd, void *vj, void *vk, uint32_t v)
|
||||
{
|
||||
VReg *Vd = (VReg *)vd;
|
||||
VReg *Vj = (VReg *)vj;
|
||||
VReg *Vk = (VReg *)vk;
|
||||
|
||||
Vd->Q(0) = int128_sub(int128_makes64(Vj->D(0)), int128_makes64(Vk->D(0)));
|
||||
}
|
||||
|
||||
DO_EVEN(vsubwev_h_b, 16, H, B, DO_SUB)
|
||||
DO_EVEN(vsubwev_w_h, 32, W, H, DO_SUB)
|
||||
DO_EVEN(vsubwev_d_w, 64, D, W, DO_SUB)
|
||||
|
||||
void HELPER(vsubwod_q_d)(void *vd, void *vj, void *vk, uint32_t v)
|
||||
{
|
||||
VReg *Vd = (VReg *)vd;
|
||||
VReg *Vj = (VReg *)vj;
|
||||
VReg *Vk = (VReg *)vk;
|
||||
|
||||
Vd->Q(0) = int128_sub(int128_makes64(Vj->D(1)), int128_makes64(Vk->D(1)));
|
||||
}
|
||||
|
||||
DO_ODD(vsubwod_h_b, 16, H, B, DO_SUB)
|
||||
DO_ODD(vsubwod_w_h, 32, W, H, DO_SUB)
|
||||
DO_ODD(vsubwod_d_w, 64, D, W, DO_SUB)
|
||||
|
||||
void HELPER(vaddwev_q_du)(void *vd, void *vj, void *vk, uint32_t v)
|
||||
{
|
||||
VReg *Vd = (VReg *)vd;
|
||||
VReg *Vj = (VReg *)vj;
|
||||
VReg *Vk = (VReg *)vk;
|
||||
|
||||
Vd->Q(0) = int128_add(int128_make64((uint64_t)Vj->D(0)),
|
||||
int128_make64((uint64_t)Vk->D(0)));
|
||||
}
|
||||
|
||||
DO_EVEN(vaddwev_h_bu, 16, UH, UB, DO_ADD)
|
||||
DO_EVEN(vaddwev_w_hu, 32, UW, UH, DO_ADD)
|
||||
DO_EVEN(vaddwev_d_wu, 64, UD, UW, DO_ADD)
|
||||
|
||||
void HELPER(vaddwod_q_du)(void *vd, void *vj, void *vk, uint32_t v)
|
||||
{
|
||||
VReg *Vd = (VReg *)vd;
|
||||
VReg *Vj = (VReg *)vj;
|
||||
VReg *Vk = (VReg *)vk;
|
||||
|
||||
Vd->Q(0) = int128_add(int128_make64((uint64_t)Vj->D(1)),
|
||||
int128_make64((uint64_t)Vk->D(1)));
|
||||
}
|
||||
|
||||
DO_ODD(vaddwod_h_bu, 16, UH, UB, DO_ADD)
|
||||
DO_ODD(vaddwod_w_hu, 32, UW, UH, DO_ADD)
|
||||
DO_ODD(vaddwod_d_wu, 64, UD, UW, DO_ADD)
|
||||
|
||||
void HELPER(vsubwev_q_du)(void *vd, void *vj, void *vk, uint32_t v)
|
||||
{
|
||||
VReg *Vd = (VReg *)vd;
|
||||
VReg *Vj = (VReg *)vj;
|
||||
VReg *Vk = (VReg *)vk;
|
||||
|
||||
Vd->Q(0) = int128_sub(int128_make64((uint64_t)Vj->D(0)),
|
||||
int128_make64((uint64_t)Vk->D(0)));
|
||||
}
|
||||
|
||||
DO_EVEN(vsubwev_h_bu, 16, UH, UB, DO_SUB)
|
||||
DO_EVEN(vsubwev_w_hu, 32, UW, UH, DO_SUB)
|
||||
DO_EVEN(vsubwev_d_wu, 64, UD, UW, DO_SUB)
|
||||
|
||||
void HELPER(vsubwod_q_du)(void *vd, void *vj, void *vk, uint32_t v)
|
||||
{
|
||||
VReg *Vd = (VReg *)vd;
|
||||
VReg *Vj = (VReg *)vj;
|
||||
VReg *Vk = (VReg *)vk;
|
||||
|
||||
Vd->Q(0) = int128_sub(int128_make64((uint64_t)Vj->D(1)),
|
||||
int128_make64((uint64_t)Vk->D(1)));
|
||||
}
|
||||
|
||||
DO_ODD(vsubwod_h_bu, 16, UH, UB, DO_SUB)
|
||||
DO_ODD(vsubwod_w_hu, 32, UW, UH, DO_SUB)
|
||||
DO_ODD(vsubwod_d_wu, 64, UD, UW, DO_SUB)
|
||||
|
||||
#define DO_EVEN_U_S(NAME, BIT, ES1, EU1, ES2, EU2, DO_OP) \
|
||||
void HELPER(NAME)(void *vd, void *vj, void *vk, uint32_t v) \
|
||||
{ \
|
||||
int i; \
|
||||
VReg *Vd = (VReg *)vd; \
|
||||
VReg *Vj = (VReg *)vj; \
|
||||
VReg *Vk = (VReg *)vk; \
|
||||
typedef __typeof(Vd->ES1(0)) TDS; \
|
||||
typedef __typeof(Vd->EU1(0)) TDU; \
|
||||
for (i = 0; i < LSX_LEN/BIT; i++) { \
|
||||
Vd->ES1(i) = DO_OP((TDU)Vj->EU2(2 * i) ,(TDS)Vk->ES2(2 * i)); \
|
||||
} \
|
||||
}
|
||||
|
||||
#define DO_ODD_U_S(NAME, BIT, ES1, EU1, ES2, EU2, DO_OP) \
|
||||
void HELPER(NAME)(void *vd, void *vj, void *vk, uint32_t v) \
|
||||
{ \
|
||||
int i; \
|
||||
VReg *Vd = (VReg *)vd; \
|
||||
VReg *Vj = (VReg *)vj; \
|
||||
VReg *Vk = (VReg *)vk; \
|
||||
typedef __typeof(Vd->ES1(0)) TDS; \
|
||||
typedef __typeof(Vd->EU1(0)) TDU; \
|
||||
for (i = 0; i < LSX_LEN/BIT; i++) { \
|
||||
Vd->ES1(i) = DO_OP((TDU)Vj->EU2(2 * i + 1), (TDS)Vk->ES2(2 * i + 1)); \
|
||||
} \
|
||||
}
|
||||
|
||||
void HELPER(vaddwev_q_du_d)(void *vd, void *vj, void *vk, uint32_t v)
|
||||
{
|
||||
VReg *Vd = (VReg *)vd;
|
||||
VReg *Vj = (VReg *)vj;
|
||||
VReg *Vk = (VReg *)vk;
|
||||
|
||||
Vd->Q(0) = int128_add(int128_make64((uint64_t)Vj->D(0)),
|
||||
int128_makes64(Vk->D(0)));
|
||||
}
|
||||
|
||||
DO_EVEN_U_S(vaddwev_h_bu_b, 16, H, UH, B, UB, DO_ADD)
|
||||
DO_EVEN_U_S(vaddwev_w_hu_h, 32, W, UW, H, UH, DO_ADD)
|
||||
DO_EVEN_U_S(vaddwev_d_wu_w, 64, D, UD, W, UW, DO_ADD)
|
||||
|
||||
void HELPER(vaddwod_q_du_d)(void *vd, void *vj, void *vk, uint32_t v)
|
||||
{
|
||||
VReg *Vd = (VReg *)vd;
|
||||
VReg *Vj = (VReg *)vj;
|
||||
VReg *Vk = (VReg *)vk;
|
||||
|
||||
Vd->Q(0) = int128_add(int128_make64((uint64_t)Vj->D(1)),
|
||||
int128_makes64(Vk->D(1)));
|
||||
}
|
||||
|
||||
DO_ODD_U_S(vaddwod_h_bu_b, 16, H, UH, B, UB, DO_ADD)
|
||||
DO_ODD_U_S(vaddwod_w_hu_h, 32, W, UW, H, UH, DO_ADD)
|
||||
DO_ODD_U_S(vaddwod_d_wu_w, 64, D, UD, W, UW, DO_ADD)
|
||||
|
Loading…
Reference in New Issue
Block a user