提交 ab20bdc1 编写于 作者: R Richard Henderson

tcg/sparc: Introduce TCG_REG_TB

Signed-off-by: NRichard Henderson <rth@twiddle.net>
上级 55129955
...@@ -85,6 +85,9 @@ static const char * const tcg_target_reg_names[TCG_TARGET_NB_REGS] = { ...@@ -85,6 +85,9 @@ static const char * const tcg_target_reg_names[TCG_TARGET_NB_REGS] = {
# define TCG_GUEST_BASE_REG TCG_REG_I5 # define TCG_GUEST_BASE_REG TCG_REG_I5
#endif #endif
#define TCG_REG_TB TCG_REG_I1
#define USE_REG_TB (sizeof(void *) > 4)
static const int tcg_target_reg_alloc_order[] = { static const int tcg_target_reg_alloc_order[] = {
TCG_REG_L0, TCG_REG_L0,
TCG_REG_L1, TCG_REG_L1,
...@@ -249,6 +252,8 @@ static const int tcg_target_call_oarg_regs[] = { ...@@ -249,6 +252,8 @@ static const int tcg_target_call_oarg_regs[] = {
#define MEMBAR (INSN_OP(2) | INSN_OP3(0x28) | INSN_RS1(15) | (1 << 13)) #define MEMBAR (INSN_OP(2) | INSN_OP3(0x28) | INSN_RS1(15) | (1 << 13))
#define NOP (SETHI | INSN_RD(TCG_REG_G0) | 0)
#ifndef ASI_PRIMARY_LITTLE #ifndef ASI_PRIMARY_LITTLE
#define ASI_PRIMARY_LITTLE 0x88 #define ASI_PRIMARY_LITTLE 0x88
#endif #endif
...@@ -423,10 +428,11 @@ static inline void tcg_out_movi_imm13(TCGContext *s, TCGReg ret, int32_t arg) ...@@ -423,10 +428,11 @@ static inline void tcg_out_movi_imm13(TCGContext *s, TCGReg ret, int32_t arg)
tcg_out_arithi(s, ret, TCG_REG_G0, arg, ARITH_OR); tcg_out_arithi(s, ret, TCG_REG_G0, arg, ARITH_OR);
} }
static void tcg_out_movi(TCGContext *s, TCGType type, static void tcg_out_movi_int(TCGContext *s, TCGType type, TCGReg ret,
TCGReg ret, tcg_target_long arg) tcg_target_long arg, bool in_prologue)
{ {
tcg_target_long hi, lo = (int32_t)arg; tcg_target_long hi, lo = (int32_t)arg;
tcg_target_long test, lsb;
/* Make sure we test 32-bit constants for imm13 properly. */ /* Make sure we test 32-bit constants for imm13 properly. */
if (type == TCG_TYPE_I32) { if (type == TCG_TYPE_I32) {
...@@ -455,6 +461,27 @@ static void tcg_out_movi(TCGContext *s, TCGType type, ...@@ -455,6 +461,27 @@ static void tcg_out_movi(TCGContext *s, TCGType type,
return; return;
} }
/* A 21-bit constant, shifted. */
lsb = ctz64(arg);
test = (tcg_target_long)arg >> lsb;
if (check_fit_tl(test, 13)) {
tcg_out_movi_imm13(s, ret, test);
tcg_out_arithi(s, ret, ret, lsb, SHIFT_SLLX);
return;
} else if (lsb > 10 && test == extract64(test, 0, 21)) {
tcg_out_sethi(s, ret, test << 10);
tcg_out_arithi(s, ret, ret, lsb - 10, SHIFT_SLLX);
return;
}
if (USE_REG_TB && !in_prologue) {
intptr_t diff = arg - (uintptr_t)s->code_gen_ptr;
if (check_fit_ptr(diff, 13)) {
tcg_out_arithi(s, ret, TCG_REG_TB, diff, ARITH_ADD);
return;
}
}
/* A 64-bit constant decomposed into 2 32-bit pieces. */ /* A 64-bit constant decomposed into 2 32-bit pieces. */
if (check_fit_i32(lo, 13)) { if (check_fit_i32(lo, 13)) {
hi = (arg - lo) >> 32; hi = (arg - lo) >> 32;
...@@ -470,6 +497,12 @@ static void tcg_out_movi(TCGContext *s, TCGType type, ...@@ -470,6 +497,12 @@ static void tcg_out_movi(TCGContext *s, TCGType type,
} }
} }
static inline void tcg_out_movi(TCGContext *s, TCGType type,
TCGReg ret, tcg_target_long arg)
{
tcg_out_movi_int(s, type, ret, arg, false);
}
static inline void tcg_out_ldst_rr(TCGContext *s, TCGReg data, TCGReg a1, static inline void tcg_out_ldst_rr(TCGContext *s, TCGReg data, TCGReg a1,
TCGReg a2, int op) TCGReg a2, int op)
{ {
...@@ -512,6 +545,11 @@ static inline bool tcg_out_sti(TCGContext *s, TCGType type, TCGArg val, ...@@ -512,6 +545,11 @@ static inline bool tcg_out_sti(TCGContext *s, TCGType type, TCGArg val,
static void tcg_out_ld_ptr(TCGContext *s, TCGReg ret, uintptr_t arg) static void tcg_out_ld_ptr(TCGContext *s, TCGReg ret, uintptr_t arg)
{ {
intptr_t diff = arg - (uintptr_t)s->code_gen_ptr;
if (USE_REG_TB && check_fit_ptr(diff, 13)) {
tcg_out_ld(s, TCG_TYPE_PTR, ret, TCG_REG_TB, diff);
return;
}
tcg_out_movi(s, TCG_TYPE_PTR, ret, arg & ~0x3ff); tcg_out_movi(s, TCG_TYPE_PTR, ret, arg & ~0x3ff);
tcg_out_ld(s, TCG_TYPE_PTR, ret, ret, arg & 0x3ff); tcg_out_ld(s, TCG_TYPE_PTR, ret, ret, arg & 0x3ff);
} }
...@@ -543,7 +581,7 @@ static void tcg_out_div32(TCGContext *s, TCGReg rd, TCGReg rs1, ...@@ -543,7 +581,7 @@ static void tcg_out_div32(TCGContext *s, TCGReg rd, TCGReg rs1,
static inline void tcg_out_nop(TCGContext *s) static inline void tcg_out_nop(TCGContext *s)
{ {
tcg_out_sethi(s, TCG_REG_G0, 0); tcg_out32(s, NOP);
} }
static const uint8_t tcg_cond_to_bcond[] = { static const uint8_t tcg_cond_to_bcond[] = {
...@@ -812,7 +850,8 @@ static void tcg_out_addsub2_i64(TCGContext *s, TCGReg rl, TCGReg rh, ...@@ -812,7 +850,8 @@ static void tcg_out_addsub2_i64(TCGContext *s, TCGReg rl, TCGReg rh,
tcg_out_mov(s, TCG_TYPE_I64, rl, tmp); tcg_out_mov(s, TCG_TYPE_I64, rl, tmp);
} }
static void tcg_out_call_nodelay(TCGContext *s, tcg_insn_unit *dest) static void tcg_out_call_nodelay(TCGContext *s, tcg_insn_unit *dest,
bool in_prologue)
{ {
ptrdiff_t disp = tcg_pcrel_diff(s, dest); ptrdiff_t disp = tcg_pcrel_diff(s, dest);
...@@ -820,14 +859,15 @@ static void tcg_out_call_nodelay(TCGContext *s, tcg_insn_unit *dest) ...@@ -820,14 +859,15 @@ static void tcg_out_call_nodelay(TCGContext *s, tcg_insn_unit *dest)
tcg_out32(s, CALL | (uint32_t)disp >> 2); tcg_out32(s, CALL | (uint32_t)disp >> 2);
} else { } else {
uintptr_t desti = (uintptr_t)dest; uintptr_t desti = (uintptr_t)dest;
tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_T1, desti & ~0xfff); tcg_out_movi_int(s, TCG_TYPE_PTR, TCG_REG_T1,
desti & ~0xfff, in_prologue);
tcg_out_arithi(s, TCG_REG_O7, TCG_REG_T1, desti & 0xfff, JMPL); tcg_out_arithi(s, TCG_REG_O7, TCG_REG_T1, desti & 0xfff, JMPL);
} }
} }
static void tcg_out_call(TCGContext *s, tcg_insn_unit *dest) static void tcg_out_call(TCGContext *s, tcg_insn_unit *dest)
{ {
tcg_out_call_nodelay(s, dest); tcg_out_call_nodelay(s, dest, false);
tcg_out_nop(s); tcg_out_nop(s);
} }
...@@ -915,7 +955,7 @@ static void build_trampolines(TCGContext *s) ...@@ -915,7 +955,7 @@ static void build_trampolines(TCGContext *s)
/* Set the env operand. */ /* Set the env operand. */
tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_O0, TCG_AREG0); tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_O0, TCG_AREG0);
/* Tail call. */ /* Tail call. */
tcg_out_call_nodelay(s, qemu_ld_helpers[i]); tcg_out_call_nodelay(s, qemu_ld_helpers[i], true);
tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_O7, ra); tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_O7, ra);
} }
...@@ -964,7 +1004,7 @@ static void build_trampolines(TCGContext *s) ...@@ -964,7 +1004,7 @@ static void build_trampolines(TCGContext *s)
/* Set the env operand. */ /* Set the env operand. */
tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_O0, TCG_AREG0); tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_O0, TCG_AREG0);
/* Tail call. */ /* Tail call. */
tcg_out_call_nodelay(s, qemu_st_helpers[i]); tcg_out_call_nodelay(s, qemu_st_helpers[i], true);
tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_O7, ra); tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_O7, ra);
} }
} }
...@@ -992,11 +1032,17 @@ static void tcg_target_qemu_prologue(TCGContext *s) ...@@ -992,11 +1032,17 @@ static void tcg_target_qemu_prologue(TCGContext *s)
#ifndef CONFIG_SOFTMMU #ifndef CONFIG_SOFTMMU
if (guest_base != 0) { if (guest_base != 0) {
tcg_out_movi(s, TCG_TYPE_PTR, TCG_GUEST_BASE_REG, guest_base); tcg_out_movi_int(s, TCG_TYPE_PTR, TCG_GUEST_BASE_REG, guest_base, true);
tcg_regset_set_reg(s->reserved_regs, TCG_GUEST_BASE_REG); tcg_regset_set_reg(s->reserved_regs, TCG_GUEST_BASE_REG);
} }
#endif #endif
/* We choose TCG_REG_TB such that no move is required. */
if (USE_REG_TB) {
QEMU_BUILD_BUG_ON(TCG_REG_TB != TCG_REG_I1);
tcg_regset_set_reg(s->reserved_regs, TCG_REG_TB);
}
tcg_out_arithi(s, TCG_REG_G0, TCG_REG_I1, 0, JMPL); tcg_out_arithi(s, TCG_REG_G0, TCG_REG_I1, 0, JMPL);
/* delay slot */ /* delay slot */
tcg_out_nop(s); tcg_out_nop(s);
...@@ -1156,7 +1202,7 @@ static void tcg_out_qemu_ld(TCGContext *s, TCGReg data, TCGReg addr, ...@@ -1156,7 +1202,7 @@ static void tcg_out_qemu_ld(TCGContext *s, TCGReg data, TCGReg addr,
func = qemu_ld_trampoline[memop & (MO_BSWAP | MO_SSIZE)]; func = qemu_ld_trampoline[memop & (MO_BSWAP | MO_SSIZE)];
} }
tcg_debug_assert(func != NULL); tcg_debug_assert(func != NULL);
tcg_out_call_nodelay(s, func); tcg_out_call_nodelay(s, func, false);
/* delay slot */ /* delay slot */
tcg_out_movi(s, TCG_TYPE_I32, param, oi); tcg_out_movi(s, TCG_TYPE_I32, param, oi);
...@@ -1235,7 +1281,7 @@ static void tcg_out_qemu_st(TCGContext *s, TCGReg data, TCGReg addr, ...@@ -1235,7 +1281,7 @@ static void tcg_out_qemu_st(TCGContext *s, TCGReg data, TCGReg addr,
func = qemu_st_trampoline[memop & (MO_BSWAP | MO_SIZE)]; func = qemu_st_trampoline[memop & (MO_BSWAP | MO_SIZE)];
tcg_debug_assert(func != NULL); tcg_debug_assert(func != NULL);
tcg_out_call_nodelay(s, func); tcg_out_call_nodelay(s, func, false);
/* delay slot */ /* delay slot */
tcg_out_movi(s, TCG_TYPE_I32, param, oi); tcg_out_movi(s, TCG_TYPE_I32, param, oi);
...@@ -1269,30 +1315,67 @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, ...@@ -1269,30 +1315,67 @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
if (check_fit_ptr(a0, 13)) { if (check_fit_ptr(a0, 13)) {
tcg_out_arithi(s, TCG_REG_G0, TCG_REG_I7, 8, RETURN); tcg_out_arithi(s, TCG_REG_G0, TCG_REG_I7, 8, RETURN);
tcg_out_movi_imm13(s, TCG_REG_O0, a0); tcg_out_movi_imm13(s, TCG_REG_O0, a0);
} else { break;
tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_I0, a0 & ~0x3ff); } else if (USE_REG_TB) {
tcg_out_arithi(s, TCG_REG_G0, TCG_REG_I7, 8, RETURN); intptr_t tb_diff = a0 - (uintptr_t)s->code_gen_ptr;
tcg_out_arithi(s, TCG_REG_O0, TCG_REG_O0, a0 & 0x3ff, ARITH_OR); if (check_fit_ptr(tb_diff, 13)) {
tcg_out_arithi(s, TCG_REG_G0, TCG_REG_I7, 8, RETURN);
/* Note that TCG_REG_TB has been unwound to O1. */
tcg_out_arithi(s, TCG_REG_O0, TCG_REG_O1, tb_diff, ARITH_ADD);
break;
}
} }
tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_I0, a0 & ~0x3ff);
tcg_out_arithi(s, TCG_REG_G0, TCG_REG_I7, 8, RETURN);
tcg_out_arithi(s, TCG_REG_O0, TCG_REG_O0, a0 & 0x3ff, ARITH_OR);
break; break;
case INDEX_op_goto_tb: case INDEX_op_goto_tb:
if (s->tb_jmp_insn_offset) { if (s->tb_jmp_insn_offset) {
/* direct jump method */ /* direct jump method */
s->tb_jmp_insn_offset[a0] = tcg_current_code_size(s); if (USE_REG_TB) {
/* Make sure to preserve links during retranslation. */ /* make sure the patch is 8-byte aligned. */
tcg_out32(s, CALL | (*s->code_ptr & ~INSN_OP(-1))); if ((intptr_t)s->code_ptr & 4) {
tcg_out_nop(s);
}
s->tb_jmp_insn_offset[a0] = tcg_current_code_size(s);
tcg_out_sethi(s, TCG_REG_T1, 0);
tcg_out_arithi(s, TCG_REG_T1, TCG_REG_T1, 0, ARITH_OR);
tcg_out_arith(s, TCG_REG_G0, TCG_REG_TB, TCG_REG_T1, JMPL);
tcg_out_arith(s, TCG_REG_TB, TCG_REG_TB, TCG_REG_T1, ARITH_ADD);
} else {
s->tb_jmp_insn_offset[a0] = tcg_current_code_size(s);
tcg_out32(s, CALL);
tcg_out_nop(s);
}
} else { } else {
/* indirect jump method */ /* indirect jump method */
tcg_out_ld_ptr(s, TCG_REG_T1, tcg_out_ld_ptr(s, TCG_REG_TB,
(uintptr_t)(s->tb_jmp_target_addr + a0)); (uintptr_t)(s->tb_jmp_target_addr + a0));
tcg_out_arithi(s, TCG_REG_G0, TCG_REG_T1, 0, JMPL); tcg_out_arithi(s, TCG_REG_G0, TCG_REG_TB, 0, JMPL);
tcg_out_nop(s);
}
s->tb_jmp_reset_offset[a0] = c = tcg_current_code_size(s);
/* For the unlinked path of goto_tb, we need to reset
TCG_REG_TB to the beginning of this TB. */
if (USE_REG_TB) {
c = -c;
if (check_fit_i32(c, 13)) {
tcg_out_arithi(s, TCG_REG_TB, TCG_REG_TB, c, ARITH_ADD);
} else {
tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_T1, c);
tcg_out_arith(s, TCG_REG_TB, TCG_REG_TB,
TCG_REG_T1, ARITH_ADD);
}
} }
tcg_out_nop(s);
s->tb_jmp_reset_offset[a0] = tcg_current_code_size(s);
break; break;
case INDEX_op_goto_ptr: case INDEX_op_goto_ptr:
tcg_out_arithi(s, TCG_REG_G0, a0, 0, JMPL); tcg_out_arithi(s, TCG_REG_G0, a0, 0, JMPL);
tcg_out_nop(s); if (USE_REG_TB) {
tcg_out_arith(s, TCG_REG_TB, a0, TCG_REG_G0, ARITH_OR);
} else {
tcg_out_nop(s);
}
break; break;
case INDEX_op_br: case INDEX_op_br:
tcg_out_bpcc(s, COND_A, BPCC_PT, arg_label(a0)); tcg_out_bpcc(s, COND_A, BPCC_PT, arg_label(a0));
...@@ -1709,13 +1792,40 @@ void tcg_register_jit(void *buf, size_t buf_size) ...@@ -1709,13 +1792,40 @@ void tcg_register_jit(void *buf, size_t buf_size)
void tb_target_set_jmp_target(uintptr_t tc_ptr, uintptr_t jmp_addr, void tb_target_set_jmp_target(uintptr_t tc_ptr, uintptr_t jmp_addr,
uintptr_t addr) uintptr_t addr)
{ {
uint32_t *ptr = (uint32_t *)jmp_addr; intptr_t tb_disp = addr - tc_ptr;
uintptr_t disp = addr - jmp_addr; intptr_t br_disp = addr - jmp_addr;
tcg_insn_unit i1, i2;
/* We can reach the entire address space for ILP32.
For LP64, the code_gen_buffer can't be larger than 2GB. */
tcg_debug_assert(tb_disp == (int32_t)tb_disp);
tcg_debug_assert(br_disp == (int32_t)br_disp);
if (!USE_REG_TB) {
atomic_set((uint32_t *)jmp_addr, deposit32(CALL, 0, 30, br_disp >> 2));
flush_icache_range(jmp_addr, jmp_addr + 4);
return;
}
/* We can reach the entire address space for 32-bit. For 64-bit /* This does not exercise the range of the branch, but we do
the code_gen_buffer can't be larger than 2GB. */ still need to be able to load the new value of TCG_REG_TB.
tcg_debug_assert(disp == (int32_t)disp); But this does still happen quite often. */
if (check_fit_ptr(tb_disp, 13)) {
/* ba,pt %icc, addr */
i1 = (INSN_OP(0) | INSN_OP2(1) | INSN_COND(COND_A)
| BPCC_ICC | BPCC_PT | INSN_OFF19(br_disp));
i2 = (ARITH_ADD | INSN_RD(TCG_REG_TB) | INSN_RS1(TCG_REG_TB)
| INSN_IMM13(tb_disp));
} else if (tb_disp >= 0) {
i1 = SETHI | INSN_RD(TCG_REG_T1) | ((tb_disp & 0xfffffc00) >> 10);
i2 = (ARITH_OR | INSN_RD(TCG_REG_T1) | INSN_RS1(TCG_REG_T1)
| INSN_IMM13(tb_disp & 0x3ff));
} else {
i1 = SETHI | INSN_RD(TCG_REG_T1) | ((~tb_disp & 0xfffffc00) >> 10);
i2 = (ARITH_XOR | INSN_RD(TCG_REG_T1) | INSN_RS1(TCG_REG_T1)
| INSN_IMM13((tb_disp & 0x3ff) | -0x400));
}
atomic_set(ptr, deposit32(CALL, 0, 30, disp >> 2)); atomic_set((uint64_t *)jmp_addr, deposit64(i2, 32, 32, i1));
flush_icache_range(jmp_addr, jmp_addr + 4); flush_icache_range(jmp_addr, jmp_addr + 8);
} }
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册