aboutsummaryrefslogtreecommitdiff
path: root/cpu-exec.c
diff options
context:
space:
mode:
Diffstat (limited to 'cpu-exec.c')
-rw-r--r--cpu-exec.c326
1 files changed, 195 insertions, 131 deletions
diff --git a/cpu-exec.c b/cpu-exec.c
index c5e530c934..615915c13f 100644
--- a/cpu-exec.c
+++ b/cpu-exec.c
@@ -29,6 +29,8 @@
//#define DEBUG_EXEC
//#define DEBUG_SIGNAL
+/* enable it to have a fully working x86 emulator for ring 0 */
+//#define RING0_HACKS
#if defined(TARGET_ARM)
/* XXX: unify with i386 target */
@@ -140,146 +142,195 @@ int cpu_exec(CPUState *env1)
#error unsupported target CPU
#endif
env->interrupt_request = 0;
+ env->exception_index = -1;
/* prepare setjmp context for exception handling */
- if (setjmp(env->jmp_env) == 0) {
- T0 = 0; /* force lookup of first TB */
- for(;;) {
+ for(;;) {
+ if (setjmp(env->jmp_env) == 0) {
+ /* if an exception is pending, we execute it here */
+ if (env->exception_index >= 0) {
+ if (env->exception_index >= EXCP_INTERRUPT) {
+ /* exit request from the cpu execution loop */
+ ret = env->exception_index;
+ break;
+ } else if (env->user_mode_only) {
+ /* if user mode only, we simulate a fake exception
+ which will be hanlded outside the cpu execution
+ loop */
+ do_interrupt_user(env->exception_index,
+ env->exception_is_int,
+ env->error_code,
+ env->exception_next_eip);
+ ret = env->exception_index;
+ break;
+ } else {
+ /* simulate a real cpu exception. On i386, it can
+ trigger new exceptions, but we do not handle
+ double or triple faults yet. */
+ do_interrupt(env->exception_index,
+ env->exception_is_int,
+ env->error_code,
+ env->exception_next_eip);
+ }
+ env->exception_index = -1;
+ }
+#if defined(TARGET_I386)
+ /* if hardware interrupt pending, we execute it */
+ if (env->hard_interrupt_request &&
+ (env->eflags & IF_MASK)) {
+ int intno;
+ intno = cpu_x86_get_pic_interrupt(env);
+ if (loglevel) {
+ fprintf(logfile, "Servicing hardware INT=0x%02x\n", intno);
+ }
+ do_interrupt(intno, 0, 0, 0);
+ env->hard_interrupt_request = 0;
+ }
+#endif
+ T0 = 0; /* force lookup of first TB */
+ for(;;) {
#ifdef __sparc__
- /* g1 can be modified by some libc? functions */
- tmp_T0 = T0;
+ /* g1 can be modified by some libc? functions */
+ tmp_T0 = T0;
#endif
- if (env->interrupt_request) {
- env->exception_index = EXCP_INTERRUPT;
- cpu_loop_exit();
- }
+ if (env->interrupt_request) {
+ env->exception_index = EXCP_INTERRUPT;
+ cpu_loop_exit();
+ }
#ifdef DEBUG_EXEC
- if (loglevel) {
+ if (loglevel) {
#if defined(TARGET_I386)
- /* restore flags in standard format */
- env->regs[R_EAX] = EAX;
- env->regs[R_EBX] = EBX;
- env->regs[R_ECX] = ECX;
- env->regs[R_EDX] = EDX;
- env->regs[R_ESI] = ESI;
- env->regs[R_EDI] = EDI;
- env->regs[R_EBP] = EBP;
- env->regs[R_ESP] = ESP;
- env->eflags = env->eflags | cc_table[CC_OP].compute_all() | (DF & DF_MASK);
- cpu_x86_dump_state(env, logfile, 0);
- env->eflags &= ~(DF_MASK | CC_O | CC_S | CC_Z | CC_A | CC_P | CC_C);
+ /* restore flags in standard format */
+ env->regs[R_EAX] = EAX;
+ env->regs[R_EBX] = EBX;
+ env->regs[R_ECX] = ECX;
+ env->regs[R_EDX] = EDX;
+ env->regs[R_ESI] = ESI;
+ env->regs[R_EDI] = EDI;
+ env->regs[R_EBP] = EBP;
+ env->regs[R_ESP] = ESP;
+ env->eflags = env->eflags | cc_table[CC_OP].compute_all() | (DF & DF_MASK);
+ cpu_x86_dump_state(env, logfile, 0);
+ env->eflags &= ~(DF_MASK | CC_O | CC_S | CC_Z | CC_A | CC_P | CC_C);
#elif defined(TARGET_ARM)
- cpu_arm_dump_state(env, logfile, 0);
+ cpu_arm_dump_state(env, logfile, 0);
#else
#error unsupported target CPU
#endif
- }
+ }
#endif
- /* we compute the CPU state. We assume it will not
- change during the whole generated block. */
+ /* we compute the CPU state. We assume it will not
+ change during the whole generated block. */
#if defined(TARGET_I386)
- flags = env->segs[R_CS].seg_32bit << GEN_FLAG_CODE32_SHIFT;
- flags |= env->segs[R_SS].seg_32bit << GEN_FLAG_SS32_SHIFT;
- flags |= (((unsigned long)env->segs[R_DS].base |
- (unsigned long)env->segs[R_ES].base |
- (unsigned long)env->segs[R_SS].base) != 0) <<
- GEN_FLAG_ADDSEG_SHIFT;
- if (!(env->eflags & VM_MASK)) {
- flags |= (env->segs[R_CS].selector & 3) << GEN_FLAG_CPL_SHIFT;
- } else {
- /* NOTE: a dummy CPL is kept */
- flags |= (1 << GEN_FLAG_VM_SHIFT);
- flags |= (3 << GEN_FLAG_CPL_SHIFT);
- }
- flags |= (env->eflags & (IOPL_MASK | TF_MASK));
- cs_base = env->segs[R_CS].base;
- pc = cs_base + env->eip;
+ flags = (env->segs[R_CS].flags & DESC_B_MASK)
+ >> (DESC_B_SHIFT - GEN_FLAG_CODE32_SHIFT);
+ flags |= (env->segs[R_SS].flags & DESC_B_MASK)
+ >> (DESC_B_SHIFT - GEN_FLAG_SS32_SHIFT);
+ flags |= (((unsigned long)env->segs[R_DS].base |
+ (unsigned long)env->segs[R_ES].base |
+ (unsigned long)env->segs[R_SS].base) != 0) <<
+ GEN_FLAG_ADDSEG_SHIFT;
+ if (!(env->eflags & VM_MASK)) {
+ flags |= (env->segs[R_CS].selector & 3) << GEN_FLAG_CPL_SHIFT;
+ } else {
+ /* NOTE: a dummy CPL is kept */
+ flags |= (1 << GEN_FLAG_VM_SHIFT);
+ flags |= (3 << GEN_FLAG_CPL_SHIFT);
+ }
+ flags |= (env->eflags & (IOPL_MASK | TF_MASK));
+ cs_base = env->segs[R_CS].base;
+ pc = cs_base + env->eip;
#elif defined(TARGET_ARM)
- flags = 0;
- cs_base = 0;
- pc = (uint8_t *)env->regs[15];
+ flags = 0;
+ cs_base = 0;
+ pc = (uint8_t *)env->regs[15];
#else
#error unsupported CPU
#endif
- tb = tb_find(&ptb, (unsigned long)pc, (unsigned long)cs_base,
- flags);
- if (!tb) {
- spin_lock(&tb_lock);
- /* if no translated code available, then translate it now */
- tb = tb_alloc((unsigned long)pc);
+ tb = tb_find(&ptb, (unsigned long)pc, (unsigned long)cs_base,
+ flags);
if (!tb) {
- /* flush must be done */
- tb_flush();
- /* cannot fail at this point */
+ spin_lock(&tb_lock);
+ /* if no translated code available, then translate it now */
tb = tb_alloc((unsigned long)pc);
- /* don't forget to invalidate previous TB info */
- ptb = &tb_hash[tb_hash_func((unsigned long)pc)];
- T0 = 0;
- }
- tc_ptr = code_gen_ptr;
- tb->tc_ptr = tc_ptr;
- tb->cs_base = (unsigned long)cs_base;
- tb->flags = flags;
- ret = cpu_gen_code(tb, CODE_GEN_MAX_SIZE, &code_gen_size);
+ if (!tb) {
+ /* flush must be done */
+ tb_flush();
+ /* cannot fail at this point */
+ tb = tb_alloc((unsigned long)pc);
+ /* don't forget to invalidate previous TB info */
+ ptb = &tb_hash[tb_hash_func((unsigned long)pc)];
+ T0 = 0;
+ }
+ tc_ptr = code_gen_ptr;
+ tb->tc_ptr = tc_ptr;
+ tb->cs_base = (unsigned long)cs_base;
+ tb->flags = flags;
+ ret = cpu_gen_code(tb, CODE_GEN_MAX_SIZE, &code_gen_size);
#if defined(TARGET_I386)
- /* XXX: suppress that, this is incorrect */
- /* if invalid instruction, signal it */
- if (ret != 0) {
- /* NOTE: the tb is allocated but not linked, so we
- can leave it */
+ /* XXX: suppress that, this is incorrect */
+ /* if invalid instruction, signal it */
+ if (ret != 0) {
+ /* NOTE: the tb is allocated but not linked, so we
+ can leave it */
+ spin_unlock(&tb_lock);
+ raise_exception(EXCP06_ILLOP);
+ }
+#endif
+ *ptb = tb;
+ tb->hash_next = NULL;
+ tb_link(tb);
+ code_gen_ptr = (void *)(((unsigned long)code_gen_ptr + code_gen_size + CODE_GEN_ALIGN - 1) & ~(CODE_GEN_ALIGN - 1));
spin_unlock(&tb_lock);
- raise_exception(EXCP06_ILLOP);
}
-#endif
- *ptb = tb;
- tb->hash_next = NULL;
- tb_link(tb);
- code_gen_ptr = (void *)(((unsigned long)code_gen_ptr + code_gen_size + CODE_GEN_ALIGN - 1) & ~(CODE_GEN_ALIGN - 1));
- spin_unlock(&tb_lock);
- }
#ifdef DEBUG_EXEC
- if (loglevel) {
- fprintf(logfile, "Trace 0x%08lx [0x%08lx] %s\n",
- (long)tb->tc_ptr, (long)tb->pc,
- lookup_symbol((void *)tb->pc));
- }
+ if (loglevel) {
+ fprintf(logfile, "Trace 0x%08lx [0x%08lx] %s\n",
+ (long)tb->tc_ptr, (long)tb->pc,
+ lookup_symbol((void *)tb->pc));
+ }
#endif
#ifdef __sparc__
- T0 = tmp_T0;
+ T0 = tmp_T0;
#endif
- /* see if we can patch the calling TB. XXX: remove TF test */
- if (T0 != 0
+ /* see if we can patch the calling TB. XXX: remove TF test */
+#ifndef RING0_HACKS
+
+ if (T0 != 0
#if defined(TARGET_I386)
- && !(env->eflags & TF_MASK)
+ && !(env->eflags & TF_MASK)
#endif
- ) {
- spin_lock(&tb_lock);
- tb_add_jump((TranslationBlock *)(T0 & ~3), T0 & 3, tb);
- spin_unlock(&tb_lock);
- }
- tc_ptr = tb->tc_ptr;
-
- /* execute the generated code */
- gen_func = (void *)tc_ptr;
+ ) {
+ spin_lock(&tb_lock);
+ tb_add_jump((TranslationBlock *)(T0 & ~3), T0 & 3, tb);
+ spin_unlock(&tb_lock);
+ }
+#endif
+ tc_ptr = tb->tc_ptr;
+
+ /* execute the generated code */
+ gen_func = (void *)tc_ptr;
#if defined(__sparc__)
- __asm__ __volatile__("call %0\n\t"
- "mov %%o7,%%i0"
- : /* no outputs */
- : "r" (gen_func)
- : "i0", "i1", "i2", "i3", "i4", "i5");
+ __asm__ __volatile__("call %0\n\t"
+ "mov %%o7,%%i0"
+ : /* no outputs */
+ : "r" (gen_func)
+ : "i0", "i1", "i2", "i3", "i4", "i5");
#elif defined(__arm__)
- asm volatile ("mov pc, %0\n\t"
- ".global exec_loop\n\t"
- "exec_loop:\n\t"
- : /* no outputs */
- : "r" (gen_func)
- : "r1", "r2", "r3", "r8", "r9", "r10", "r12", "r14");
+ asm volatile ("mov pc, %0\n\t"
+ ".global exec_loop\n\t"
+ "exec_loop:\n\t"
+ : /* no outputs */
+ : "r" (gen_func)
+ : "r1", "r2", "r3", "r8", "r9", "r10", "r12", "r14");
#else
- gen_func();
+ gen_func();
#endif
+ }
+ } else {
}
- }
- ret = env->exception_index;
+ } /* for(;;) */
+
#if defined(TARGET_I386)
/* restore flags in standard format */
@@ -348,11 +399,11 @@ void cpu_x86_load_seg(CPUX86State *s, int seg_reg, int selector)
SegmentCache *sc;
selector &= 0xffff;
sc = &env->segs[seg_reg];
- /* NOTE: in VM86 mode, limit and seg_32bit are never reloaded,
+ /* NOTE: in VM86 mode, limit and flags are never reloaded,
so we must load them here */
sc->base = (void *)(selector << 4);
sc->limit = 0xffff;
- sc->seg_32bit = 0;
+ sc->flags = 0;
sc->selector = selector;
} else {
load_seg(seg_reg, selector, 0);
@@ -398,6 +449,8 @@ void cpu_x86_frstor(CPUX86State *s, uint8_t *ptr, int data32)
#include <signal.h>
#include <sys/ucontext.h>
+#if defined(TARGET_I386)
+
/* 'pc' is the host PC at which the exception was raised. 'address' is
the effective address of the memory exception. 'is_write' is 1 if a
write caused the exception and otherwise 0'. 'old_set' is the
@@ -407,42 +460,53 @@ static inline int handle_cpu_signal(unsigned long pc, unsigned long address,
{
TranslationBlock *tb;
int ret;
- uint32_t found_pc;
+#ifdef RING0_HACKS
+ env = global_env; /* XXX: find a better solution */
+#endif
#if defined(DEBUG_SIGNAL)
- printf("qemu: SIGSEGV pc=0x%08lx address=%08lx wr=%d oldset=0x%08lx\n",
+ printf("qemu: SIGSEGV pc=0x%08lx address=%08lx w=%d oldset=0x%08lx\n",
pc, address, is_write, *(unsigned long *)old_set);
#endif
/* XXX: locking issue */
if (is_write && page_unprotect(address)) {
return 1;
}
+ /* see if it is an MMU fault */
+ ret = cpu_x86_handle_mmu_fault(env, address, is_write);
+ if (ret < 0)
+ return 0; /* not an MMU fault */
+ if (ret == 0)
+ return 1; /* the MMU fault was handled without causing real CPU fault */
+ /* now we have a real cpu fault */
tb = tb_find_pc(pc);
if (tb) {
/* the PC is inside the translated code. It means that we have
a virtual CPU fault */
- ret = cpu_search_pc(tb, &found_pc, pc);
- if (ret < 0)
- return 0;
-#if defined(TARGET_I386)
- env->eip = found_pc - tb->cs_base;
- env->cr[2] = address;
- /* we restore the process signal mask as the sigreturn should
- do it (XXX: use sigsetjmp) */
- sigprocmask(SIG_SETMASK, old_set, NULL);
- raise_exception_err(EXCP0E_PAGE, 4 | (is_write << 1));
+ cpu_restore_state(tb, env, pc);
+ }
+#if 0
+ printf("PF exception: EIP=0x%08x CR2=0x%08x error=0x%x\n",
+ env->eip, env->cr[2], env->error_code);
+#endif
+ /* we restore the process signal mask as the sigreturn should
+ do it (XXX: use sigsetjmp) */
+ sigprocmask(SIG_SETMASK, old_set, NULL);
+ raise_exception_err(EXCP0E_PAGE, env->error_code);
+ /* never comes here */
+ return 1;
+}
+
#elif defined(TARGET_ARM)
- env->regs[15] = found_pc;
- /* XXX: do more */
+static inline int handle_cpu_signal(unsigned long pc, unsigned long address,
+ int is_write, sigset_t *old_set)
+{
+ /* XXX: do more */
+ return 0;
+}
#else
#error unsupported target CPU
#endif
- /* never comes here */
- return 1;
- } else {
- return 0;
- }
-}
#if defined(__i386__)
@@ -570,6 +634,6 @@ int cpu_signal_handler(int host_signum, struct siginfo *info,
#else
-#error CPU specific signal handler needed
+#error host CPU specific signal handler needed
#endif