cpu-exec.c 21.5 KB
Newer Older
B
bellard 已提交
1
/*
2
 *  emulator main execution loop
3
 *
B
bellard 已提交
4
 *  Copyright (c) 2003-2005 Fabrice Bellard
B
bellard 已提交
5
 *
B
bellard 已提交
6 7 8 9
 * This library is free software; you can redistribute it and/or
 * modify it under the terms of the GNU Lesser General Public
 * License as published by the Free Software Foundation; either
 * version 2 of the License, or (at your option) any later version.
B
bellard 已提交
10
 *
B
bellard 已提交
11 12 13 14
 * This library is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 * Lesser General Public License for more details.
B
bellard 已提交
15
 *
B
bellard 已提交
16
 * You should have received a copy of the GNU Lesser General Public
17
 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
B
bellard 已提交
18
 */
P
Peter Maydell 已提交
19
#include "qemu/osdep.h"
B
Blue Swirl 已提交
20
#include "cpu.h"
21
#include "trace.h"
22
#include "disas/disas.h"
23
#include "tcg.h"
24
#include "qemu/atomic.h"
25
#include "sysemu/qtest.h"
26
#include "qemu/timer.h"
P
Paolo Bonzini 已提交
27
#include "exec/address-spaces.h"
28
#include "qemu/rcu.h"
29
#include "exec/tb-hash.h"
30
#include "exec/log.h"
31 32 33
#if defined(TARGET_I386) && !defined(CONFIG_USER_ONLY)
#include "hw/i386/apic.h"
#endif
34
#include "sysemu/replay.h"
35 36 37 38 39 40

/* -icount align implementation. */

typedef struct SyncClocks {
    int64_t diff_clk;
    int64_t last_cpu_icount;
41
    int64_t realtime_clock;
42 43 44 45 46 47 48 49
} SyncClocks;

#if !defined(CONFIG_USER_ONLY)
/* Allow the guest to have a max 3ms advance.
 * The difference between the 2 clocks could therefore
 * oscillate around 0.
 */
#define VM_CLOCK_ADVANCE 3000000
50 51 52
#define THRESHOLD_REDUCE 1.5
#define MAX_DELAY_PRINT_RATE 2000000000LL
#define MAX_NB_PRINTS 100
53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71

static void align_clocks(SyncClocks *sc, const CPUState *cpu)
{
    int64_t cpu_icount;

    if (!icount_align_option) {
        return;
    }

    cpu_icount = cpu->icount_extra + cpu->icount_decr.u16.low;
    sc->diff_clk += cpu_icount_to_ns(sc->last_cpu_icount - cpu_icount);
    sc->last_cpu_icount = cpu_icount;

    if (sc->diff_clk > VM_CLOCK_ADVANCE) {
#ifndef _WIN32
        struct timespec sleep_delay, rem_delay;
        sleep_delay.tv_sec = sc->diff_clk / 1000000000LL;
        sleep_delay.tv_nsec = sc->diff_clk % 1000000000LL;
        if (nanosleep(&sleep_delay, &rem_delay) < 0) {
P
Paolo Bonzini 已提交
72
            sc->diff_clk = rem_delay.tv_sec * 1000000000LL + rem_delay.tv_nsec;
73 74 75 76 77 78 79 80 81 82
        } else {
            sc->diff_clk = 0;
        }
#else
        Sleep(sc->diff_clk / SCALE_MS);
        sc->diff_clk = 0;
#endif
    }
}

83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104
static void print_delay(const SyncClocks *sc)
{
    static float threshold_delay;
    static int64_t last_realtime_clock;
    static int nb_prints;

    if (icount_align_option &&
        sc->realtime_clock - last_realtime_clock >= MAX_DELAY_PRINT_RATE &&
        nb_prints < MAX_NB_PRINTS) {
        if ((-sc->diff_clk / (float)1000000000LL > threshold_delay) ||
            (-sc->diff_clk / (float)1000000000LL <
             (threshold_delay - THRESHOLD_REDUCE))) {
            threshold_delay = (-sc->diff_clk / 1000000000LL) + 1;
            printf("Warning: The guest is now late by %.1f to %.1f seconds\n",
                   threshold_delay - 1,
                   threshold_delay);
            nb_prints++;
            last_realtime_clock = sc->realtime_clock;
        }
    }
}

105 106 107 108 109 110
static void init_delay_params(SyncClocks *sc,
                              const CPUState *cpu)
{
    if (!icount_align_option) {
        return;
    }
111 112
    sc->realtime_clock = qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL_RT);
    sc->diff_clk = qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) - sc->realtime_clock;
113
    sc->last_cpu_icount = cpu->icount_extra + cpu->icount_decr.u16.low;
114 115 116 117 118 119
    if (sc->diff_clk < max_delay) {
        max_delay = sc->diff_clk;
    }
    if (sc->diff_clk > max_advance) {
        max_advance = sc->diff_clk;
    }
120 121 122 123

    /* Print every 2s max if the guest is late. We limit the number
       of printed messages to NB_PRINT_MAX(currently 100) */
    print_delay(sc);
124 125 126 127 128 129 130 131 132 133
}
#else
static void align_clocks(SyncClocks *sc, const CPUState *cpu)
{
}

static void init_delay_params(SyncClocks *sc, const CPUState *cpu)
{
}
#endif /* CONFIG USER ONLY */
B
bellard 已提交
134

135
/* Execute a TB, and fix up the CPU state afterwards if necessary */
136
static inline tcg_target_ulong cpu_tb_exec(CPUState *cpu, TranslationBlock *itb)
137 138
{
    CPUArchState *env = cpu->env_ptr;
S
Sergey Fedorov 已提交
139 140 141
    uintptr_t ret;
    TranslationBlock *last_tb;
    int tb_exit;
142 143
    uint8_t *tb_ptr = itb->tc_ptr;

144 145 146
    qemu_log_mask_and_addr(CPU_LOG_EXEC, itb->pc,
                           "Trace %p [" TARGET_FMT_lx "] %s\n",
                           itb->tc_ptr, itb->pc, lookup_symbol(itb->pc));
147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163

#if defined(DEBUG_DISAS)
    if (qemu_loglevel_mask(CPU_LOG_TB_CPU)) {
#if defined(TARGET_I386)
        log_cpu_state(cpu, CPU_DUMP_CCOP);
#elif defined(TARGET_M68K)
        /* ??? Should not modify env state for dumping.  */
        cpu_m68k_flush_flags(env, env->cc_op);
        env->cc_op = CC_OP_FLAGS;
        env->sr = (env->sr & 0xffe0) | env->cc_dest | (env->cc_x << 4);
        log_cpu_state(cpu, 0);
#else
        log_cpu_state(cpu, 0);
#endif
    }
#endif /* DEBUG_DISAS */

164
    cpu->can_do_io = !use_icount;
S
Sergey Fedorov 已提交
165
    ret = tcg_qemu_tb_exec(env, tb_ptr);
166
    cpu->can_do_io = 1;
S
Sergey Fedorov 已提交
167 168 169
    last_tb = (TranslationBlock *)(ret & ~TB_EXIT_MASK);
    tb_exit = ret & TB_EXIT_MASK;
    trace_exec_tb_exit(last_tb, tb_exit);
170

S
Sergey Fedorov 已提交
171
    if (tb_exit > TB_EXIT_IDX1) {
172 173 174 175
        /* We didn't start executing this TB (eg because the instruction
         * counter hit zero); we must restore the guest PC to the address
         * of the start of the TB.
         */
176
        CPUClass *cc = CPU_GET_CLASS(cpu);
S
Sergey Fedorov 已提交
177
        qemu_log_mask_and_addr(CPU_LOG_EXEC, last_tb->pc,
178 179
                               "Stopped execution of TB chain before %p ["
                               TARGET_FMT_lx "] %s\n",
S
Sergey Fedorov 已提交
180 181
                               last_tb->tc_ptr, last_tb->pc,
                               lookup_symbol(last_tb->pc));
182
        if (cc->synchronize_from_tb) {
S
Sergey Fedorov 已提交
183
            cc->synchronize_from_tb(cpu, last_tb);
184 185
        } else {
            assert(cc->set_pc);
S
Sergey Fedorov 已提交
186
            cc->set_pc(cpu, last_tb->pc);
187
        }
188
    }
S
Sergey Fedorov 已提交
189
    if (tb_exit == TB_EXIT_REQUESTED) {
190 191 192 193 194
        /* We were asked to stop executing TBs (probably a pending
         * interrupt. We've now stopped, so clear the flag.
         */
        cpu->tcg_exit_req = 0;
    }
S
Sergey Fedorov 已提交
195
    return ret;
196 197
}

198
#ifndef CONFIG_USER_ONLY
P
pbrook 已提交
199 200
/* Execute the code without caching the generated code. An interpreter
   could be used if available. */
201
static void cpu_exec_nocache(CPUState *cpu, int max_cycles,
202
                             TranslationBlock *orig_tb, bool ignore_icount)
P
pbrook 已提交
203 204
{
    TranslationBlock *tb;
S
Sergey Fedorov 已提交
205
    bool old_tb_flushed;
P
pbrook 已提交
206 207 208 209 210 211

    /* Should never happen.
       We only end up here when an existing TB is too long.  */
    if (max_cycles > CF_COUNT_MASK)
        max_cycles = CF_COUNT_MASK;

S
Sergey Fedorov 已提交
212 213
    old_tb_flushed = cpu->tb_flushed;
    cpu->tb_flushed = false;
214
    tb = tb_gen_code(cpu, orig_tb->pc, orig_tb->cs_base, orig_tb->flags,
215 216
                     max_cycles | CF_NOCACHE
                         | (ignore_icount ? CF_IGNORE_ICOUNT : 0));
S
Sergey Fedorov 已提交
217 218
    tb->orig_tb = cpu->tb_flushed ? NULL : orig_tb;
    cpu->tb_flushed |= old_tb_flushed;
P
pbrook 已提交
219
    /* execute the generated code */
220
    trace_exec_tb_nocache(tb, tb->pc);
221
    cpu_tb_exec(cpu, tb);
P
pbrook 已提交
222 223 224
    tb_phys_invalidate(tb, -1);
    tb_free(tb);
}
225
#endif
P
pbrook 已提交
226

227 228 229
static TranslationBlock *tb_find_physical(CPUState *cpu,
                                          target_ulong pc,
                                          target_ulong cs_base,
230
                                          uint32_t flags)
231
{
232
    CPUArchState *env = (CPUArchState *)cpu->env_ptr;
233
    TranslationBlock *tb, **tb_hash_head, **ptb1;
234
    unsigned int h;
235
    tb_page_addr_t phys_pc, phys_page1;
236

237
    /* find translated block using physical mappings */
P
Paul Brook 已提交
238
    phys_pc = get_page_addr_code(env, pc);
239 240
    phys_page1 = phys_pc & TARGET_PAGE_MASK;
    h = tb_phys_hash_func(phys_pc);
241 242 243 244 245 246

    /* Start at head of the hash entry */
    ptb1 = tb_hash_head = &tcg_ctx.tb_ctx.tb_phys_hash[h];
    tb = *ptb1;

    while (tb) {
247
        if (tb->pc == pc &&
248
            tb->page_addr[0] == phys_page1 &&
249
            tb->cs_base == cs_base &&
250
            tb->flags == flags) {
251

252 253 254 255 256 257 258 259 260
            if (tb->page_addr[1] == -1) {
                /* done, we have a match */
                break;
            } else {
                /* check next page if needed */
                target_ulong virt_page2 = (pc & TARGET_PAGE_MASK) +
                                          TARGET_PAGE_SIZE;
                tb_page_addr_t phys_page2 = get_page_addr_code(env, virt_page2);

261 262 263
                if (tb->page_addr[1] == phys_page2) {
                    break;
                }
264 265
            }
        }
266

267
        ptb1 = &tb->phys_hash_next;
268
        tb = *ptb1;
269
    }
270

271 272 273 274 275 276
    if (tb) {
        /* Move the TB to the head of the list */
        *ptb1 = tb->phys_hash_next;
        tb->phys_hash_next = *tb_hash_head;
        *tb_hash_head = tb;
    }
277 278 279 280 281 282
    return tb;
}

static TranslationBlock *tb_find_slow(CPUState *cpu,
                                      target_ulong pc,
                                      target_ulong cs_base,
283
                                      uint32_t flags)
284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304
{
    TranslationBlock *tb;

    tb = tb_find_physical(cpu, pc, cs_base, flags);
    if (tb) {
        goto found;
    }

#ifdef CONFIG_USER_ONLY
    /* mmap_lock is needed by tb_gen_code, and mmap_lock must be
     * taken outside tb_lock.  Since we're momentarily dropping
     * tb_lock, there's a chance that our desired tb has been
     * translated.
     */
    tb_unlock();
    mmap_lock();
    tb_lock();
    tb = tb_find_physical(cpu, pc, cs_base, flags);
    if (tb) {
        mmap_unlock();
        goto found;
305
    }
306 307 308 309 310 311 312 313 314 315
#endif

    /* if no translated code available, then translate it now */
    tb = tb_gen_code(cpu, pc, cs_base, flags, 0);

#ifdef CONFIG_USER_ONLY
    mmap_unlock();
#endif

found:
316
    /* we add the TB in the virtual pc hash table */
317
    cpu->tb_jmp_cache[tb_jmp_cache_hash_func(pc)] = tb;
318 319 320
    return tb;
}

321 322 323
static inline TranslationBlock *tb_find_fast(CPUState *cpu,
                                             TranslationBlock **last_tb,
                                             int tb_exit)
324
{
325
    CPUArchState *env = (CPUArchState *)cpu->env_ptr;
326 327
    TranslationBlock *tb;
    target_ulong cs_base, pc;
328
    uint32_t flags;
329 330 331 332

    /* we record a subset of the CPU state. It will
       always be the same before a given translated block
       is executed. */
333
    cpu_get_tb_cpu_state(env, &pc, &cs_base, &flags);
334
    tb_lock();
335
    tb = cpu->tb_jmp_cache[tb_jmp_cache_hash_func(pc)];
336 337
    if (unlikely(!tb || tb->pc != pc || tb->cs_base != cs_base ||
                 tb->flags != flags)) {
338
        tb = tb_find_slow(cpu, pc, cs_base, flags);
339
    }
340 341 342 343 344 345 346 347 348 349 350 351
    if (cpu->tb_flushed) {
        /* Ensure that no TB jump will be modified as the
         * translation buffer has been flushed.
         */
        *last_tb = NULL;
        cpu->tb_flushed = false;
    }
    /* See if we can patch the calling TB. */
    if (*last_tb && !qemu_loglevel_mask(CPU_LOG_TB_NOCHAIN)) {
        tb_add_jump(*last_tb, tb_exit, tb);
    }
    tb_unlock();
352 353 354
    return tb;
}

355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376
static inline bool cpu_handle_halt(CPUState *cpu)
{
    if (cpu->halted) {
#if defined(TARGET_I386) && !defined(CONFIG_USER_ONLY)
        if ((cpu->interrupt_request & CPU_INTERRUPT_POLL)
            && replay_interrupt()) {
            X86CPU *x86_cpu = X86_CPU(cpu);
            apic_poll_irq(x86_cpu->apic_state);
            cpu_reset_interrupt(cpu, CPU_INTERRUPT_POLL);
        }
#endif
        if (!cpu_has_work(cpu)) {
            current_cpu = NULL;
            return true;
        }

        cpu->halted = 0;
    }

    return false;
}

377
static inline void cpu_handle_debug_exception(CPUState *cpu)
378
{
379
    CPUClass *cc = CPU_GET_CLASS(cpu);
380 381
    CPUWatchpoint *wp;

382 383
    if (!cpu->watchpoint_hit) {
        QTAILQ_FOREACH(wp, &cpu->watchpoints, entry) {
384 385 386
            wp->flags &= ~BP_WATCHPOINT_HIT;
        }
    }
387 388

    cc->debug_excp_handler(cpu);
389 390
}

391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439
static inline bool cpu_handle_exception(CPUState *cpu, int *ret)
{
    if (cpu->exception_index >= 0) {
        if (cpu->exception_index >= EXCP_INTERRUPT) {
            /* exit request from the cpu execution loop */
            *ret = cpu->exception_index;
            if (*ret == EXCP_DEBUG) {
                cpu_handle_debug_exception(cpu);
            }
            cpu->exception_index = -1;
            return true;
        } else {
#if defined(CONFIG_USER_ONLY)
            /* if user mode only, we simulate a fake exception
               which will be handled outside the cpu execution
               loop */
#if defined(TARGET_I386)
            CPUClass *cc = CPU_GET_CLASS(cpu);
            cc->do_interrupt(cpu);
#endif
            *ret = cpu->exception_index;
            cpu->exception_index = -1;
            return true;
#else
            if (replay_exception()) {
                CPUClass *cc = CPU_GET_CLASS(cpu);
                cc->do_interrupt(cpu);
                cpu->exception_index = -1;
            } else if (!replay_has_interrupt()) {
                /* give a chance to iothread in replay mode */
                *ret = EXCP_INTERRUPT;
                return true;
            }
#endif
        }
#ifndef CONFIG_USER_ONLY
    } else if (replay_has_exception()
               && cpu->icount_decr.u16.low + cpu->icount_extra == 0) {
        /* try to cause an exception pending in the log */
        TranslationBlock *last_tb = NULL; /* Avoid chaining TBs */
        cpu_exec_nocache(cpu, 1, tb_find_fast(cpu, &last_tb, 0), true);
        *ret = -1;
        return true;
#endif
    }

    return false;
}

B
bellard 已提交
440 441
/* main execution loop */

442
int cpu_exec(CPUState *cpu)
B
bellard 已提交
443
{
444
    CPUClass *cc = CPU_GET_CLASS(cpu);
445 446
#ifdef TARGET_I386
    X86CPU *x86_cpu = X86_CPU(cpu);
447
    CPUArchState *env = &x86_cpu->env;
448
#endif
449
    int ret, interrupt_request;
S
Sergey Fedorov 已提交
450 451
    TranslationBlock *tb, *last_tb;
    int tb_exit = 0;
452 453
    SyncClocks sc;

454 455 456
    /* replay_interrupt may need current_cpu */
    current_cpu = cpu;

457 458
    if (cpu_handle_halt(cpu)) {
        return EXCP_HALTED;
459
    }
B
bellard 已提交
460

P
Paolo Bonzini 已提交
461
    atomic_mb_set(&tcg_current_cpu, cpu);
462 463
    rcu_read_lock();

464
    if (unlikely(atomic_mb_read(&exit_request))) {
465
        cpu->exit_request = 1;
466 467
    }

468
    cc->cpu_exec_enter(cpu);
469

470 471 472 473 474 475 476
    /* Calculate difference between guest clock and host clock.
     * This delay includes the delay of the last cycle, so
     * what we have to do is sleep until it is 0. As for the
     * advance/delay we gain here, we try to fix it next time.
     */
    init_delay_params(&sc, cpu);

477
    for(;;) {
478
        /* prepare setjmp context for exception handling */
479
        if (sigsetjmp(cpu->jmp_env, 0) == 0) {
480
            /* if an exception is pending, we execute it here */
481
            if (cpu_handle_exception(cpu, &ret)) {
482
                break;
483
            }
B
bellard 已提交
484

S
Sergey Fedorov 已提交
485
            last_tb = NULL; /* forget the last executed TB after exception */
S
Sergey Fedorov 已提交
486
            cpu->tb_flushed = false; /* reset before first TB lookup */
487
            for(;;) {
488
                interrupt_request = cpu->interrupt_request;
M
malc 已提交
489
                if (unlikely(interrupt_request)) {
490
                    if (unlikely(cpu->singlestep_enabled & SSTEP_NOIRQ)) {
M
malc 已提交
491
                        /* Mask out external interrupts for this step. */
492
                        interrupt_request &= ~CPU_INTERRUPT_SSTEP_MASK;
M
malc 已提交
493
                    }
494
                    if (interrupt_request & CPU_INTERRUPT_DEBUG) {
495
                        cpu->interrupt_request &= ~CPU_INTERRUPT_DEBUG;
496
                        cpu->exception_index = EXCP_DEBUG;
497
                        cpu_loop_exit(cpu);
498
                    }
499 500 501 502 503
                    if (replay_mode == REPLAY_MODE_PLAY
                        && !replay_has_interrupt()) {
                        /* Do nothing */
                    } else if (interrupt_request & CPU_INTERRUPT_HALT) {
                        replay_interrupt();
504 505
                        cpu->interrupt_request &= ~CPU_INTERRUPT_HALT;
                        cpu->halted = 1;
506
                        cpu->exception_index = EXCP_HLT;
507
                        cpu_loop_exit(cpu);
508
                    }
509
#if defined(TARGET_I386)
510 511
                    else if (interrupt_request & CPU_INTERRUPT_INIT) {
                        replay_interrupt();
512 513 514 515 516 517
                        cpu_svm_check_intercept_param(env, SVM_EXIT_INIT, 0);
                        do_cpu_init(x86_cpu);
                        cpu->exception_index = EXCP_HALTED;
                        cpu_loop_exit(cpu);
                    }
#else
518 519
                    else if (interrupt_request & CPU_INTERRUPT_RESET) {
                        replay_interrupt();
520
                        cpu_reset(cpu);
521
                        cpu_loop_exit(cpu);
522
                    }
B
bellard 已提交
523
#endif
524 525 526 527
                    /* The target hook has 3 exit conditions:
                       False when the interrupt isn't processed,
                       True when it is, and we should restart on a new TB,
                       and via longjmp via cpu_loop_exit.  */
528 529 530
                    else {
                        replay_interrupt();
                        if (cc->cpu_exec_interrupt(cpu, interrupt_request)) {
S
Sergey Fedorov 已提交
531
                            last_tb = NULL;
532
                        }
533 534 535
                    }
                    /* Don't use the cached interrupt_request value,
                       do_interrupt may have updated the EXITTB flag. */
536 537
                    if (cpu->interrupt_request & CPU_INTERRUPT_EXITTB) {
                        cpu->interrupt_request &= ~CPU_INTERRUPT_EXITTB;
538 539
                        /* ensure that no TB jump will be modified as
                           the program flow was changed */
S
Sergey Fedorov 已提交
540
                        last_tb = NULL;
541
                    }
542
                }
543 544
                if (unlikely(cpu->exit_request
                             || replay_has_interrupt())) {
545
                    cpu->exit_request = 0;
546
                    cpu->exception_index = EXCP_INTERRUPT;
547
                    cpu_loop_exit(cpu);
548
                }
549
                tb = tb_find_fast(cpu, &last_tb, tb_exit);
550
                if (likely(!cpu->exit_request)) {
S
Sergey Fedorov 已提交
551
                    uintptr_t ret;
552
                    trace_exec_tb(tb, tb->pc);
553
                    /* execute the generated code */
S
Sergey Fedorov 已提交
554 555 556 557
                    ret = cpu_tb_exec(cpu, tb);
                    last_tb = (TranslationBlock *)(ret & ~TB_EXIT_MASK);
                    tb_exit = ret & TB_EXIT_MASK;
                    switch (tb_exit) {
558 559 560 561 562 563
                    case TB_EXIT_REQUESTED:
                        /* Something asked us to stop executing
                         * chained TBs; just continue round the main
                         * loop. Whatever requested the exit will also
                         * have set something else (eg exit_request or
                         * interrupt_request) which we will handle
564 565 566 567
                         * next time around the loop.  But we need to
                         * ensure the tcg_exit_req read in generated code
                         * comes before the next read of cpu->exit_request
                         * or cpu->interrupt_request.
568
                         */
569
                        smp_rmb();
S
Sergey Fedorov 已提交
570
                        last_tb = NULL;
571 572 573
                        break;
                    case TB_EXIT_ICOUNT_EXPIRED:
                    {
T
ths 已提交
574
                        /* Instruction counter expired.  */
575 576 577
#ifdef CONFIG_USER_ONLY
                        abort();
#else
P
Paolo Bonzini 已提交
578
                        int insns_left = cpu->icount_decr.u32;
579
                        if (cpu->icount_extra && insns_left >= 0) {
P
pbrook 已提交
580
                            /* Refill decrementer and continue execution.  */
581
                            cpu->icount_extra += insns_left;
P
Paolo Bonzini 已提交
582
                            insns_left = MIN(0xffff, cpu->icount_extra);
583
                            cpu->icount_extra -= insns_left;
584
                            cpu->icount_decr.u16.low = insns_left;
P
pbrook 已提交
585 586 587
                        } else {
                            if (insns_left > 0) {
                                /* Execute remaining instructions.  */
S
Sergey Fedorov 已提交
588 589
                                cpu_exec_nocache(cpu, insns_left,
                                                 last_tb, false);
590
                                align_clocks(&sc, cpu);
P
pbrook 已提交
591
                            }
592
                            cpu->exception_index = EXCP_INTERRUPT;
S
Sergey Fedorov 已提交
593
                            last_tb = NULL;
594
                            cpu_loop_exit(cpu);
P
pbrook 已提交
595
                        }
596
                        break;
597
#endif
598 599 600
                    }
                    default:
                        break;
P
pbrook 已提交
601 602
                    }
                }
603 604 605
                /* Try to align the host and virtual clocks
                   if the guest is in advance */
                align_clocks(&sc, cpu);
T
ths 已提交
606
            } /* for(;;) */
607
        } else {
608 609 610 611 612
#if defined(__clang__) || !QEMU_GNUC_PREREQ(4, 6)
            /* Some compilers wrongly smash all local variables after
             * siglongjmp. There were bug reports for gcc 4.5.0 and clang.
             * Reload essential local variables here for those compilers.
             * Newer versions of gcc would complain about this code (-Wclobbered). */
613
            cpu = current_cpu;
614
            cc = CPU_GET_CLASS(cpu);
615 616
#ifdef TARGET_I386
            x86_cpu = X86_CPU(cpu);
617
            env = &x86_cpu->env;
618
#endif
619 620 621 622 623 624 625 626 627 628
#else /* buggy compiler */
            /* Assert that the compiler does not smash local variables. */
            g_assert(cpu == current_cpu);
            g_assert(cc == CPU_GET_CLASS(cpu));
#ifdef TARGET_I386
            g_assert(x86_cpu == X86_CPU(cpu));
            g_assert(env == &x86_cpu->env);
#endif
#endif /* buggy compiler */
            cpu->can_do_io = 1;
K
KONRAD Frederic 已提交
629
            tb_lock_reset();
B
bellard 已提交
630
        }
631 632
    } /* for(;;) */

633
    cc->cpu_exec_exit(cpu);
634
    rcu_read_unlock();
P
pbrook 已提交
635

636 637
    /* fail safe : never use current_cpu outside cpu_exec() */
    current_cpu = NULL;
P
Paolo Bonzini 已提交
638 639 640

    /* Does not need atomic_mb_set because a spurious wakeup is okay.  */
    atomic_set(&tcg_current_cpu, NULL);
B
bellard 已提交
641 642
    return ret;
}