提交 61462af6 编写于 作者: P Peter Maydell

Merge remote-tracking branch 'remotes/aurel/tags/pull-target-sh4-20170530' into staging

Queued target/sh4 patches

# gpg: Signature made Tue 30 May 2017 20:12:10 BST
# gpg:                using RSA key 0xBA9C78061DDD8C9B
# gpg: Good signature from "Aurelien Jarno <aurelien@aurel32.net>"
# gpg:                 aka "Aurelien Jarno <aurelien@jarno.fr>"
# gpg:                 aka "Aurelien Jarno <aurel32@debian.org>"
# gpg: WARNING: This key is not certified with sufficiently trusted signatures!
# gpg:          It is not certain that the signature belongs to the owner.
# Primary key fingerprint: 7746 2642 A9EF 94FD 0F77  196D BA9C 7806 1DDD 8C9B

* remotes/aurel/tags/pull-target-sh4-20170530:
  target/sh4: fix RTE instruction delay slot
  target/sh4: ignore interrupts in a delay slot
  target/sh4: introduce DELAY_SLOT_MASK
  target/sh4: fix reset when using a kernel and an initrd
  target/sh4: log unauthorized accesses using qemu_log_mask
Signed-off-by: NPeter Maydell <peter.maydell@linaro.org>
...@@ -91,8 +91,10 @@ ...@@ -91,8 +91,10 @@
#define FPSCR_RM_NEAREST (0 << 0) #define FPSCR_RM_NEAREST (0 << 0)
#define FPSCR_RM_ZERO (1 << 0) #define FPSCR_RM_ZERO (1 << 0)
#define DELAY_SLOT_MASK 0x7
#define DELAY_SLOT (1 << 0) #define DELAY_SLOT (1 << 0)
#define DELAY_SLOT_CONDITIONAL (1 << 1) #define DELAY_SLOT_CONDITIONAL (1 << 1)
#define DELAY_SLOT_RTE (1 << 2)
typedef struct tlb_t { typedef struct tlb_t {
uint32_t vpn; /* virtual page number */ uint32_t vpn; /* virtual page number */
...@@ -263,7 +265,13 @@ void cpu_load_tlb(CPUSH4State * env); ...@@ -263,7 +265,13 @@ void cpu_load_tlb(CPUSH4State * env);
#define MMU_USER_IDX 1 #define MMU_USER_IDX 1
static inline int cpu_mmu_index (CPUSH4State *env, bool ifetch) static inline int cpu_mmu_index (CPUSH4State *env, bool ifetch)
{ {
return (env->sr & (1u << SR_MD)) == 0 ? 1 : 0; /* The instruction in a RTE delay slot is fetched in privileged
mode, but executed in user mode. */
if (ifetch && (env->flags & DELAY_SLOT_RTE)) {
return 0;
} else {
return (env->sr & (1u << SR_MD)) == 0 ? 1 : 0;
}
} }
#include "exec/cpu-all.h" #include "exec/cpu-all.h"
...@@ -380,7 +388,7 @@ static inline void cpu_get_tb_cpu_state(CPUSH4State *env, target_ulong *pc, ...@@ -380,7 +388,7 @@ static inline void cpu_get_tb_cpu_state(CPUSH4State *env, target_ulong *pc,
{ {
*pc = env->pc; *pc = env->pc;
*cs_base = 0; *cs_base = 0;
*flags = (env->flags & (DELAY_SLOT | DELAY_SLOT_CONDITIONAL)) /* Bits 0-1 */ *flags = (env->flags & DELAY_SLOT_MASK) /* Bits 0- 2 */
| (env->fpscr & (FPSCR_FR | FPSCR_SZ | FPSCR_PR)) /* Bits 19-21 */ | (env->fpscr & (FPSCR_FR | FPSCR_SZ | FPSCR_PR)) /* Bits 19-21 */
| (env->sr & ((1u << SR_MD) | (1u << SR_RB))) /* Bits 29-30 */ | (env->sr & ((1u << SR_MD) | (1u << SR_RB))) /* Bits 29-30 */
| (env->sr & (1u << SR_FD)) /* Bit 15 */ | (env->sr & (1u << SR_FD)) /* Bit 15 */
......
...@@ -21,6 +21,7 @@ ...@@ -21,6 +21,7 @@
#include "cpu.h" #include "cpu.h"
#include "exec/exec-all.h" #include "exec/exec-all.h"
#include "exec/log.h" #include "exec/log.h"
#include "sysemu/sysemu.h"
#if !defined(CONFIG_USER_ONLY) #if !defined(CONFIG_USER_ONLY)
#include "hw/sh4/sh_intc.h" #include "hw/sh4/sh_intc.h"
...@@ -92,7 +93,14 @@ void superh_cpu_do_interrupt(CPUState *cs) ...@@ -92,7 +93,14 @@ void superh_cpu_do_interrupt(CPUState *cs)
if (env->sr & (1u << SR_BL)) { if (env->sr & (1u << SR_BL)) {
if (do_exp && cs->exception_index != 0x1e0) { if (do_exp && cs->exception_index != 0x1e0) {
cs->exception_index = 0x000; /* masked exception -> reset */ /* In theory a masked exception generates a reset exception,
which in turn jumps to the reset vector. However this only
works when using a bootloader. When using a kernel and an
initrd, they need to be reloaded and the program counter
should be loaded with the kernel entry point.
qemu_system_reset_request takes care of that. */
qemu_system_reset_request(SHUTDOWN_CAUSE_GUEST_RESET);
return;
} }
if (do_irq && !env->in_sleep) { if (do_irq && !env->in_sleep) {
return; /* masked */ return; /* masked */
...@@ -164,11 +172,11 @@ void superh_cpu_do_interrupt(CPUState *cs) ...@@ -164,11 +172,11 @@ void superh_cpu_do_interrupt(CPUState *cs)
env->sgr = env->gregs[15]; env->sgr = env->gregs[15];
env->sr |= (1u << SR_BL) | (1u << SR_MD) | (1u << SR_RB); env->sr |= (1u << SR_BL) | (1u << SR_MD) | (1u << SR_RB);
if (env->flags & (DELAY_SLOT | DELAY_SLOT_CONDITIONAL)) { if (env->flags & DELAY_SLOT_MASK) {
/* Branch instruction should be executed again before delay slot. */ /* Branch instruction should be executed again before delay slot. */
env->spc -= 2; env->spc -= 2;
/* Clear flags for exception/interrupt routine. */ /* Clear flags for exception/interrupt routine. */
env->flags &= ~(DELAY_SLOT | DELAY_SLOT_CONDITIONAL); env->flags &= ~DELAY_SLOT_MASK;
} }
if (do_exp) { if (do_exp) {
...@@ -420,7 +428,7 @@ static int get_physical_address(CPUSH4State * env, target_ulong * physical, ...@@ -420,7 +428,7 @@ static int get_physical_address(CPUSH4State * env, target_ulong * physical,
if (!(env->sr & (1u << SR_MD)) if (!(env->sr & (1u << SR_MD))
&& (address < 0xe0000000 || address >= 0xe4000000)) { && (address < 0xe0000000 || address >= 0xe4000000)) {
/* Unauthorized access in user mode (only store queues are available) */ /* Unauthorized access in user mode (only store queues are available) */
fprintf(stderr, "Unauthorized access\n"); qemu_log_mask(LOG_GUEST_ERROR, "Unauthorized access\n");
if (rw == 0) if (rw == 0)
return MMU_DADDR_ERROR_READ; return MMU_DADDR_ERROR_READ;
else if (rw == 1) else if (rw == 1)
...@@ -863,8 +871,16 @@ int cpu_sh4_is_cached(CPUSH4State * env, target_ulong addr) ...@@ -863,8 +871,16 @@ int cpu_sh4_is_cached(CPUSH4State * env, target_ulong addr)
bool superh_cpu_exec_interrupt(CPUState *cs, int interrupt_request) bool superh_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
{ {
if (interrupt_request & CPU_INTERRUPT_HARD) { if (interrupt_request & CPU_INTERRUPT_HARD) {
superh_cpu_do_interrupt(cs); SuperHCPU *cpu = SUPERH_CPU(cs);
return true; CPUSH4State *env = &cpu->env;
/* Delay slots are indivisible, ignore interrupts */
if (env->flags & DELAY_SLOT_MASK) {
return false;
} else {
superh_cpu_do_interrupt(cs);
return true;
}
} }
return false; return false;
} }
...@@ -185,6 +185,9 @@ void superh_cpu_dump_state(CPUState *cs, FILE *f, ...@@ -185,6 +185,9 @@ void superh_cpu_dump_state(CPUState *cs, FILE *f,
} else if (env->flags & DELAY_SLOT_CONDITIONAL) { } else if (env->flags & DELAY_SLOT_CONDITIONAL) {
cpu_fprintf(f, "in conditional delay slot (delayed_pc=0x%08x)\n", cpu_fprintf(f, "in conditional delay slot (delayed_pc=0x%08x)\n",
env->delayed_pc); env->delayed_pc);
} else if (env->flags & DELAY_SLOT_RTE) {
cpu_fprintf(f, "in rte delay slot (delayed_pc=0x%08x)\n",
env->delayed_pc);
} }
} }
...@@ -217,8 +220,7 @@ static inline void gen_save_cpu_state(DisasContext *ctx, bool save_pc) ...@@ -217,8 +220,7 @@ static inline void gen_save_cpu_state(DisasContext *ctx, bool save_pc)
if (ctx->delayed_pc != (uint32_t) -1) { if (ctx->delayed_pc != (uint32_t) -1) {
tcg_gen_movi_i32(cpu_delayed_pc, ctx->delayed_pc); tcg_gen_movi_i32(cpu_delayed_pc, ctx->delayed_pc);
} }
if ((ctx->tbflags & (DELAY_SLOT | DELAY_SLOT_CONDITIONAL)) if ((ctx->tbflags & DELAY_SLOT_MASK) != ctx->envflags) {
!= ctx->envflags) {
tcg_gen_movi_i32(cpu_flags, ctx->envflags); tcg_gen_movi_i32(cpu_flags, ctx->envflags);
} }
} }
...@@ -329,7 +331,7 @@ static inline void gen_store_fpr64 (TCGv_i64 t, int reg) ...@@ -329,7 +331,7 @@ static inline void gen_store_fpr64 (TCGv_i64 t, int reg)
#define DREG(x) FREG(x) /* Assumes lsb of (x) is always 0 */ #define DREG(x) FREG(x) /* Assumes lsb of (x) is always 0 */
#define CHECK_NOT_DELAY_SLOT \ #define CHECK_NOT_DELAY_SLOT \
if (ctx->envflags & (DELAY_SLOT | DELAY_SLOT_CONDITIONAL)) { \ if (ctx->envflags & DELAY_SLOT_MASK) { \
gen_save_cpu_state(ctx, true); \ gen_save_cpu_state(ctx, true); \
gen_helper_raise_slot_illegal_instruction(cpu_env); \ gen_helper_raise_slot_illegal_instruction(cpu_env); \
ctx->bstate = BS_EXCP; \ ctx->bstate = BS_EXCP; \
...@@ -339,7 +341,7 @@ static inline void gen_store_fpr64 (TCGv_i64 t, int reg) ...@@ -339,7 +341,7 @@ static inline void gen_store_fpr64 (TCGv_i64 t, int reg)
#define CHECK_PRIVILEGED \ #define CHECK_PRIVILEGED \
if (IS_USER(ctx)) { \ if (IS_USER(ctx)) { \
gen_save_cpu_state(ctx, true); \ gen_save_cpu_state(ctx, true); \
if (ctx->envflags & (DELAY_SLOT | DELAY_SLOT_CONDITIONAL)) { \ if (ctx->envflags & DELAY_SLOT_MASK) { \
gen_helper_raise_slot_illegal_instruction(cpu_env); \ gen_helper_raise_slot_illegal_instruction(cpu_env); \
} else { \ } else { \
gen_helper_raise_illegal_instruction(cpu_env); \ gen_helper_raise_illegal_instruction(cpu_env); \
...@@ -351,7 +353,7 @@ static inline void gen_store_fpr64 (TCGv_i64 t, int reg) ...@@ -351,7 +353,7 @@ static inline void gen_store_fpr64 (TCGv_i64 t, int reg)
#define CHECK_FPU_ENABLED \ #define CHECK_FPU_ENABLED \
if (ctx->tbflags & (1u << SR_FD)) { \ if (ctx->tbflags & (1u << SR_FD)) { \
gen_save_cpu_state(ctx, true); \ gen_save_cpu_state(ctx, true); \
if (ctx->envflags & (DELAY_SLOT | DELAY_SLOT_CONDITIONAL)) { \ if (ctx->envflags & DELAY_SLOT_MASK) { \
gen_helper_raise_slot_fpu_disable(cpu_env); \ gen_helper_raise_slot_fpu_disable(cpu_env); \
} else { \ } else { \
gen_helper_raise_fpu_disable(cpu_env); \ gen_helper_raise_fpu_disable(cpu_env); \
...@@ -428,8 +430,9 @@ static void _decode_opc(DisasContext * ctx) ...@@ -428,8 +430,9 @@ static void _decode_opc(DisasContext * ctx)
CHECK_NOT_DELAY_SLOT CHECK_NOT_DELAY_SLOT
gen_write_sr(cpu_ssr); gen_write_sr(cpu_ssr);
tcg_gen_mov_i32(cpu_delayed_pc, cpu_spc); tcg_gen_mov_i32(cpu_delayed_pc, cpu_spc);
ctx->envflags |= DELAY_SLOT; ctx->envflags |= DELAY_SLOT_RTE;
ctx->delayed_pc = (uint32_t) - 1; ctx->delayed_pc = (uint32_t) - 1;
ctx->bstate = BS_STOP;
return; return;
case 0x0058: /* sets */ case 0x0058: /* sets */
tcg_gen_ori_i32(cpu_sr, cpu_sr, (1u << SR_S)); tcg_gen_ori_i32(cpu_sr, cpu_sr, (1u << SR_S));
...@@ -1784,7 +1787,7 @@ static void _decode_opc(DisasContext * ctx) ...@@ -1784,7 +1787,7 @@ static void _decode_opc(DisasContext * ctx)
fflush(stderr); fflush(stderr);
#endif #endif
gen_save_cpu_state(ctx, true); gen_save_cpu_state(ctx, true);
if (ctx->envflags & (DELAY_SLOT | DELAY_SLOT_CONDITIONAL)) { if (ctx->envflags & DELAY_SLOT_MASK) {
gen_helper_raise_slot_illegal_instruction(cpu_env); gen_helper_raise_slot_illegal_instruction(cpu_env);
} else { } else {
gen_helper_raise_illegal_instruction(cpu_env); gen_helper_raise_illegal_instruction(cpu_env);
...@@ -1798,14 +1801,14 @@ static void decode_opc(DisasContext * ctx) ...@@ -1798,14 +1801,14 @@ static void decode_opc(DisasContext * ctx)
_decode_opc(ctx); _decode_opc(ctx);
if (old_flags & (DELAY_SLOT | DELAY_SLOT_CONDITIONAL)) { if (old_flags & DELAY_SLOT_MASK) {
/* go out of the delay slot */ /* go out of the delay slot */
ctx->envflags &= ~(DELAY_SLOT | DELAY_SLOT_CONDITIONAL); ctx->envflags &= ~DELAY_SLOT_MASK;
tcg_gen_movi_i32(cpu_flags, ctx->envflags); tcg_gen_movi_i32(cpu_flags, ctx->envflags);
ctx->bstate = BS_BRANCH; ctx->bstate = BS_BRANCH;
if (old_flags & DELAY_SLOT_CONDITIONAL) { if (old_flags & DELAY_SLOT_CONDITIONAL) {
gen_delayed_conditional_jump(ctx); gen_delayed_conditional_jump(ctx);
} else if (old_flags & DELAY_SLOT) { } else {
gen_jump(ctx); gen_jump(ctx);
} }
...@@ -1824,7 +1827,7 @@ void gen_intermediate_code(CPUSH4State * env, struct TranslationBlock *tb) ...@@ -1824,7 +1827,7 @@ void gen_intermediate_code(CPUSH4State * env, struct TranslationBlock *tb)
pc_start = tb->pc; pc_start = tb->pc;
ctx.pc = pc_start; ctx.pc = pc_start;
ctx.tbflags = (uint32_t)tb->flags; ctx.tbflags = (uint32_t)tb->flags;
ctx.envflags = tb->flags & (DELAY_SLOT | DELAY_SLOT_CONDITIONAL); ctx.envflags = tb->flags & DELAY_SLOT_MASK;
ctx.bstate = BS_NONE; ctx.bstate = BS_NONE;
ctx.memidx = (ctx.tbflags & (1u << SR_MD)) == 0 ? 1 : 0; ctx.memidx = (ctx.tbflags & (1u << SR_MD)) == 0 ? 1 : 0;
/* We don't know if the delayed pc came from a dynamic or static branch, /* We don't know if the delayed pc came from a dynamic or static branch,
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册