alpha: convert a few more instructions to TCG

Signed-off-by: Aurelien Jarno <aurelien@aurel32.net>

git-svn-id: svn://svn.savannah.nongnu.org/qemu/trunk@5152 c046a42c-6fe2-441c-8c8c-71466251a162
This commit is contained in:
aurel32 2008-09-04 04:36:20 +00:00
parent 3761035f2c
commit 4f821e1757
2 changed files with 28 additions and 90 deletions

View File

@ -148,13 +148,6 @@ void OPPROTO op_no_op (void)
#include "op_mem.h" #include "op_mem.h"
#endif #endif
/* Special operation for load and store */
void OPPROTO op_n7 (void)
{
T0 &= ~(uint64_t)0x7;
RETURN();
}
/* Misc */ /* Misc */
void OPPROTO op_excp (void) void OPPROTO op_excp (void)
{ {
@ -259,18 +252,6 @@ void OPPROTO op_sublv (void)
RETURN(); RETURN();
} }
void OPPROTO op_s4 (void)
{
T0 <<= 2;
RETURN();
}
void OPPROTO op_s8 (void)
{
T0 <<= 3;
RETURN();
}
void OPPROTO op_mull (void) void OPPROTO op_mull (void)
{ {
T0 = (int64_t)((int32_t)T0 * (int32_t)T1); T0 = (int64_t)((int32_t)T0 * (int32_t)T1);
@ -646,19 +627,6 @@ void OPPROTO op_cmplbc (void)
RETURN(); RETURN();
} }
/* Branches */
void OPPROTO op_branch (void)
{
env->pc = T0 & ~3;
RETURN();
}
void OPPROTO op_addq1 (void)
{
T1 += T0;
RETURN();
}
#if 0 // Qemu does not know how to do this... #if 0 // Qemu does not know how to do this...
void OPPROTO op_bcond (void) void OPPROTO op_bcond (void)
{ {

View File

@ -273,15 +273,11 @@ static always_inline void gen_load_mem (DisasContext *ctx,
gen_op_nop(); gen_op_nop();
} else { } else {
if (rb != 31) if (rb != 31)
tcg_gen_mov_i64(cpu_T[0], cpu_ir[rb]); tcg_gen_addi_i64(cpu_T[0], cpu_ir[rb], disp16);
else else
tcg_gen_movi_i64(cpu_T[0], 0); tcg_gen_movi_i64(cpu_T[0], disp16);
if (disp16 != 0) {
tcg_gen_movi_i64(cpu_T[1], disp16);
gen_op_addq();
}
if (clear) if (clear)
gen_op_n7(); tcg_gen_andi_i64(cpu_T[0], cpu_T[0], ~0x7);
(*gen_load_op)(ctx); (*gen_load_op)(ctx);
if (ra != 31) if (ra != 31)
tcg_gen_mov_i64(cpu_ir[ra], cpu_T[1]); tcg_gen_mov_i64(cpu_ir[ra], cpu_T[1]);
@ -294,15 +290,11 @@ static always_inline void gen_store_mem (DisasContext *ctx,
int clear) int clear)
{ {
if (rb != 31) if (rb != 31)
tcg_gen_mov_i64(cpu_T[0], cpu_ir[rb]); tcg_gen_addi_i64(cpu_T[0], cpu_ir[rb], disp16);
else else
tcg_gen_movi_i64(cpu_T[0], 0); tcg_gen_movi_i64(cpu_T[0], disp16);
if (disp16 != 0) {
tcg_gen_movi_i64(cpu_T[1], disp16);
gen_op_addq();
}
if (clear) if (clear)
gen_op_n7(); tcg_gen_andi_i64(cpu_T[0], cpu_T[0], ~0x7);
if (ra != 31) if (ra != 31)
tcg_gen_mov_i64(cpu_T[1], cpu_ir[ra]); tcg_gen_mov_i64(cpu_T[1], cpu_ir[ra]);
else else
@ -315,13 +307,9 @@ static always_inline void gen_load_fmem (DisasContext *ctx,
int ra, int rb, int32_t disp16) int ra, int rb, int32_t disp16)
{ {
if (rb != 31) if (rb != 31)
tcg_gen_mov_i64(cpu_T[0], cpu_ir[rb]); tcg_gen_addi_i64(cpu_T[0], cpu_ir[rb], disp16);
else else
tcg_gen_movi_i64(cpu_T[0], 0); tcg_gen_movi_i64(cpu_T[0], disp16);
if (disp16 != 0) {
tcg_gen_movi_i64(cpu_T[1], disp16);
gen_op_addq();
}
(*gen_load_fop)(ctx); (*gen_load_fop)(ctx);
gen_store_fir(ctx, ra, 1); gen_store_fir(ctx, ra, 1);
} }
@ -331,13 +319,9 @@ static always_inline void gen_store_fmem (DisasContext *ctx,
int ra, int rb, int32_t disp16) int ra, int rb, int32_t disp16)
{ {
if (rb != 31) if (rb != 31)
tcg_gen_mov_i64(cpu_T[0], cpu_ir[rb]); tcg_gen_addi_i64(cpu_T[0], cpu_ir[rb], disp16);
else else
tcg_gen_movi_i64(cpu_T[0], 0); tcg_gen_movi_i64(cpu_T[0], disp16);
if (disp16 != 0) {
tcg_gen_movi_i64(cpu_T[1], disp16);
gen_op_addq();
}
gen_load_fir(ctx, ra, 1); gen_load_fir(ctx, ra, 1);
(*gen_store_fop)(ctx); (*gen_store_fop)(ctx);
} }
@ -346,13 +330,7 @@ static always_inline void gen_bcond (DisasContext *ctx,
void (*gen_test_op)(void), void (*gen_test_op)(void),
int ra, int32_t disp16) int ra, int32_t disp16)
{ {
if (disp16 != 0) { tcg_gen_movi_i64(cpu_T[1], ctx->pc + (int64_t)(disp16 << 2));
tcg_gen_movi_i64(cpu_T[0], ctx->pc);
tcg_gen_movi_i64(cpu_T[1], disp16 << 2);
gen_op_addq1();
} else {
tcg_gen_movi_i64(cpu_T[1], ctx->pc);
}
if (ra != 31) if (ra != 31)
tcg_gen_mov_i64(cpu_T[0], cpu_ir[ra]); tcg_gen_mov_i64(cpu_T[0], cpu_ir[ra]);
else else
@ -365,13 +343,7 @@ static always_inline void gen_fbcond (DisasContext *ctx,
void (*gen_test_op)(void), void (*gen_test_op)(void),
int ra, int32_t disp16) int ra, int32_t disp16)
{ {
if (disp16 != 0) { tcg_gen_movi_i64(cpu_T[1], ctx->pc + (int64_t)(disp16 << 2));
tcg_gen_movi_i64(cpu_T[0], ctx->pc);
tcg_gen_movi_i64(cpu_T[1], disp16 << 2);
gen_op_addq1();
} else {
tcg_gen_movi_i64(cpu_T[1], ctx->pc);
}
gen_load_fir(ctx, ra, 0); gen_load_fir(ctx, ra, 0);
(*gen_test_op)(); (*gen_test_op)();
_gen_op_bcond(ctx); _gen_op_bcond(ctx);
@ -484,50 +456,50 @@ static always_inline void gen_itf (DisasContext *ctx,
static always_inline void gen_s4addl (void) static always_inline void gen_s4addl (void)
{ {
gen_op_s4(); tcg_gen_shli_i64(cpu_T[0], cpu_T[0], 2);
gen_op_addl(); gen_op_addl();
} }
static always_inline void gen_s4subl (void) static always_inline void gen_s4subl (void)
{ {
gen_op_s4(); tcg_gen_shli_i64(cpu_T[0], cpu_T[0], 2);
gen_op_subl(); gen_op_subl();
} }
static always_inline void gen_s8addl (void) static always_inline void gen_s8addl (void)
{ {
gen_op_s8(); tcg_gen_shli_i64(cpu_T[0], cpu_T[0], 3);
gen_op_addl(); gen_op_addl();
} }
static always_inline void gen_s8subl (void) static always_inline void gen_s8subl (void)
{ {
gen_op_s8(); tcg_gen_shli_i64(cpu_T[0], cpu_T[0], 3);
gen_op_subl(); gen_op_subl();
} }
static always_inline void gen_s4addq (void) static always_inline void gen_s4addq (void)
{ {
gen_op_s4(); tcg_gen_shli_i64(cpu_T[0], cpu_T[0], 2);
gen_op_addq(); tcg_gen_add_i64(cpu_T[0], cpu_T[0], cpu_T[1]);
} }
static always_inline void gen_s4subq (void) static always_inline void gen_s4subq (void)
{ {
gen_op_s4(); tcg_gen_shli_i64(cpu_T[0], cpu_T[0], 2);
gen_op_subq(); tcg_gen_sub_i64(cpu_T[0], cpu_T[0], cpu_T[1]);
} }
static always_inline void gen_s8addq (void) static always_inline void gen_s8addq (void)
{ {
gen_op_s8(); tcg_gen_shli_i64(cpu_T[0], cpu_T[0], 3);
gen_op_addq(); tcg_gen_add_i64(cpu_T[0], cpu_T[0], cpu_T[1]);
} }
static always_inline void gen_s8subq (void) static always_inline void gen_s8subq (void)
{ {
gen_op_s8(); tcg_gen_shli_i64(cpu_T[0], cpu_T[0], 3);
gen_op_subq(); tcg_gen_sub_i64(cpu_T[0], cpu_T[0], cpu_T[1]);
} }
static always_inline void gen_amask (void) static always_inline void gen_amask (void)
@ -1383,7 +1355,7 @@ static always_inline int translate_one (DisasContext *ctx, uint32_t insn)
else else
tcg_gen_movi_i64(cpu_T[0], 0); tcg_gen_movi_i64(cpu_T[0], 0);
tcg_gen_movi_i64(cpu_T[1], disp12); tcg_gen_movi_i64(cpu_T[1], disp12);
gen_op_addq(); tcg_gen_add_i64(cpu_T[0], cpu_T[0], cpu_T[1]);
switch ((insn >> 12) & 0xF) { switch ((insn >> 12) & 0xF) {
case 0x0: case 0x0:
/* Longword physical access */ /* Longword physical access */
@ -1638,7 +1610,7 @@ static always_inline int translate_one (DisasContext *ctx, uint32_t insn)
else else
tcg_gen_movi_i64(cpu_T[0], 0); tcg_gen_movi_i64(cpu_T[0], 0);
tcg_gen_movi_i64(cpu_T[1], (((int64_t)insn << 51) >> 51)); tcg_gen_movi_i64(cpu_T[1], (((int64_t)insn << 51) >> 51));
gen_op_addq(); tcg_gen_add_i64(cpu_T[0], cpu_T[0], cpu_T[1]);
gen_op_hw_ret(); gen_op_hw_ret();
} }
ret = 2; ret = 2;
@ -1652,11 +1624,9 @@ static always_inline int translate_one (DisasContext *ctx, uint32_t insn)
if (!ctx->pal_mode) if (!ctx->pal_mode)
goto invalid_opc; goto invalid_opc;
if (ra != 31) if (ra != 31)
tcg_gen_mov_i64(cpu_T[0], cpu_ir[rb]); tcg_gen_addi_i64(cpu_T[0], cpu_ir[rb], disp12);
else else
tcg_gen_movi_i64(cpu_T[0], 0); tcg_gen_movi_i64(cpu_T[0], disp12);
tcg_gen_movi_i64(cpu_T[1], disp12);
gen_op_addq();
if (ra != 31) if (ra != 31)
tcg_gen_mov_i64(cpu_T[1], cpu_ir[ra]); tcg_gen_mov_i64(cpu_T[1], cpu_ir[ra]);
else else