exec.c 72.4 KB
Newer Older
B
bellard 已提交
1
/*
2
 *  Virtual page mapping
3
 *
B
bellard 已提交
4 5 6 7 8 9 10 11 12 13 14 15 16
 *  Copyright (c) 2003 Fabrice Bellard
 *
 * This library is free software; you can redistribute it and/or
 * modify it under the terms of the GNU Lesser General Public
 * License as published by the Free Software Foundation; either
 * version 2 of the License, or (at your option) any later version.
 *
 * This library is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 * Lesser General Public License for more details.
 *
 * You should have received a copy of the GNU Lesser General Public
17
 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
B
bellard 已提交
18
 */
B
bellard 已提交
19
#include "config.h"
B
bellard 已提交
20 21 22
#ifdef _WIN32
#include <windows.h>
#else
B
bellard 已提交
23
#include <sys/types.h>
B
bellard 已提交
24 25
#include <sys/mman.h>
#endif
B
bellard 已提交
26

27
#include "qemu-common.h"
B
bellard 已提交
28
#include "cpu.h"
B
bellard 已提交
29
#include "tcg.h"
30
#include "hw/hw.h"
31
#include "hw/qdev.h"
32
#include "qemu/osdep.h"
33
#include "sysemu/kvm.h"
P
Paolo Bonzini 已提交
34
#include "hw/xen/xen.h"
35 36
#include "qemu/timer.h"
#include "qemu/config-file.h"
37
#include "exec/memory.h"
38
#include "sysemu/dma.h"
39
#include "exec/address-spaces.h"
40 41
#if defined(CONFIG_USER_ONLY)
#include <qemu.h>
J
Jun Nakajima 已提交
42
#else /* !CONFIG_USER_ONLY */
43
#include "sysemu/xen-mapcache.h"
44
#include "trace.h"
45
#endif
46
#include "exec/cpu-all.h"
B
bellard 已提交
47

48
#include "exec/cputlb.h"
49
#include "translate-all.h"
50

51
#include "exec/memory-internal.h"
52

53
//#define DEBUG_SUBPAGE
T
ths 已提交
54

55
#if !defined(CONFIG_USER_ONLY)
56
int phys_ram_fd;
A
aliguori 已提交
57
static int in_migration;
P
pbrook 已提交
58

P
Paolo Bonzini 已提交
59
RAMList ram_list = { .blocks = QTAILQ_HEAD_INITIALIZER(ram_list.blocks) };
A
Avi Kivity 已提交
60 61

static MemoryRegion *system_memory;
62
static MemoryRegion *system_io;
A
Avi Kivity 已提交
63

64 65
AddressSpace address_space_io;
AddressSpace address_space_memory;
66
DMAContext dma_context_memory;
67

68
MemoryRegion io_mem_rom, io_mem_notdirty;
69
static MemoryRegion io_mem_unassigned;
70

71
#endif
72

73
CPUArchState *first_cpu;
B
bellard 已提交
74 75
/* current CPU in the current thread. It is only valid inside
   cpu_exec() */
76
DEFINE_TLS(CPUArchState *,cpu_single_env);
P
pbrook 已提交
77
/* 0 = Do not count executed instructions.
T
ths 已提交
78
   1 = Precise instruction counting.
P
pbrook 已提交
79
   2 = Adaptive rate instruction counting.  */
80
int use_icount;
B
bellard 已提交
81

82
#if !defined(CONFIG_USER_ONLY)
83

84 85 86 87 88 89 90 91 92 93 94 95 96 97
typedef struct PhysPageEntry PhysPageEntry;

struct PhysPageEntry {
    uint16_t is_leaf : 1;
     /* index into phys_sections (is_leaf) or phys_map_nodes (!is_leaf) */
    uint16_t ptr : 15;
};

struct AddressSpaceDispatch {
    /* This is a multi-level map on the physical address space.
     * The bottom level has pointers to MemoryRegionSections.
     */
    PhysPageEntry phys_map;
    MemoryListener listener;
98
    AddressSpace *as;
99 100
};

101 102 103
#define SUBPAGE_IDX(addr) ((addr) & ~TARGET_PAGE_MASK)
typedef struct subpage_t {
    MemoryRegion iomem;
104
    AddressSpace *as;
105 106 107 108
    hwaddr base;
    uint16_t sub_section[TARGET_PAGE_SIZE];
} subpage_t;

109 110 111
static MemoryRegionSection *phys_sections;
static unsigned phys_sections_nb, phys_sections_nb_alloc;
static uint16_t phys_section_unassigned;
112 113 114
static uint16_t phys_section_notdirty;
static uint16_t phys_section_rom;
static uint16_t phys_section_watch;
115

116 117 118 119
/* Simple allocator for PhysPageEntry nodes */
static PhysPageEntry (*phys_map_nodes)[L2_SIZE];
static unsigned phys_map_nodes_nb, phys_map_nodes_nb_alloc;

120
#define PHYS_MAP_NODE_NIL (((uint16_t)~0) >> 1)
121

122
static void io_mem_init(void);
A
Avi Kivity 已提交
123
static void memory_map_init(void);
B
Blue Swirl 已提交
124
static void *qemu_safe_ram_ptr(ram_addr_t addr);
125

126
static MemoryRegion io_mem_watch;
127
#endif
B
bellard 已提交
128

129
#if !defined(CONFIG_USER_ONLY)
130

131
static void phys_map_node_reserve(unsigned nodes)
132
{
133
    if (phys_map_nodes_nb + nodes > phys_map_nodes_nb_alloc) {
134 135
        typedef PhysPageEntry Node[L2_SIZE];
        phys_map_nodes_nb_alloc = MAX(phys_map_nodes_nb_alloc * 2, 16);
136 137
        phys_map_nodes_nb_alloc = MAX(phys_map_nodes_nb_alloc,
                                      phys_map_nodes_nb + nodes);
138 139 140
        phys_map_nodes = g_renew(Node, phys_map_nodes,
                                 phys_map_nodes_nb_alloc);
    }
141 142 143 144 145 146 147 148 149 150
}

static uint16_t phys_map_node_alloc(void)
{
    unsigned i;
    uint16_t ret;

    ret = phys_map_nodes_nb++;
    assert(ret != PHYS_MAP_NODE_NIL);
    assert(ret != phys_map_nodes_nb_alloc);
151
    for (i = 0; i < L2_SIZE; ++i) {
152
        phys_map_nodes[ret][i].is_leaf = 0;
153
        phys_map_nodes[ret][i].ptr = PHYS_MAP_NODE_NIL;
154
    }
155
    return ret;
156 157 158 159 160 161 162
}

static void phys_map_nodes_reset(void)
{
    phys_map_nodes_nb = 0;
}

B
bellard 已提交
163

A
Avi Kivity 已提交
164 165
static void phys_page_set_level(PhysPageEntry *lp, hwaddr *index,
                                hwaddr *nb, uint16_t leaf,
166
                                int level)
167 168 169
{
    PhysPageEntry *p;
    int i;
A
Avi Kivity 已提交
170
    hwaddr step = (hwaddr)1 << (level * L2_BITS);
171

172
    if (!lp->is_leaf && lp->ptr == PHYS_MAP_NODE_NIL) {
173 174
        lp->ptr = phys_map_node_alloc();
        p = phys_map_nodes[lp->ptr];
175 176
        if (level == 0) {
            for (i = 0; i < L2_SIZE; i++) {
177
                p[i].is_leaf = 1;
178
                p[i].ptr = phys_section_unassigned;
179
            }
P
pbrook 已提交
180
        }
181
    } else {
182
        p = phys_map_nodes[lp->ptr];
B
bellard 已提交
183
    }
184
    lp = &p[(*index >> (level * L2_BITS)) & (L2_SIZE - 1)];
185

186
    while (*nb && lp < &p[L2_SIZE]) {
187 188
        if ((*index & (step - 1)) == 0 && *nb >= step) {
            lp->is_leaf = true;
189
            lp->ptr = leaf;
190 191
            *index += step;
            *nb -= step;
192 193 194 195
        } else {
            phys_page_set_level(lp, index, nb, leaf, level - 1);
        }
        ++lp;
196 197 198
    }
}

A
Avi Kivity 已提交
199
static void phys_page_set(AddressSpaceDispatch *d,
A
Avi Kivity 已提交
200
                          hwaddr index, hwaddr nb,
201
                          uint16_t leaf)
202
{
203
    /* Wildly overreserve - it doesn't matter much. */
204
    phys_map_node_reserve(3 * P_L2_LEVELS);
205

A
Avi Kivity 已提交
206
    phys_page_set_level(&d->phys_map, &index, &nb, leaf, P_L2_LEVELS - 1);
B
bellard 已提交
207 208
}

209
static MemoryRegionSection *phys_page_find(AddressSpaceDispatch *d, hwaddr index)
B
bellard 已提交
210
{
A
Avi Kivity 已提交
211
    PhysPageEntry lp = d->phys_map;
212 213
    PhysPageEntry *p;
    int i;
214

215
    for (i = P_L2_LEVELS - 1; i >= 0 && !lp.is_leaf; i--) {
216
        if (lp.ptr == PHYS_MAP_NODE_NIL) {
P
Paolo Bonzini 已提交
217
            return &phys_sections[phys_section_unassigned];
218
        }
219
        p = phys_map_nodes[lp.ptr];
220
        lp = p[(index >> (i * L2_BITS)) & (L2_SIZE - 1)];
221
    }
P
Paolo Bonzini 已提交
222
    return &phys_sections[lp.ptr];
223 224
}

B
Blue Swirl 已提交
225 226
bool memory_region_is_unassigned(MemoryRegion *mr)
{
P
Paolo Bonzini 已提交
227
    return mr != &io_mem_rom && mr != &io_mem_notdirty && !mr->rom_device
228
        && mr != &io_mem_watch;
B
bellard 已提交
229
}
230

231
static MemoryRegionSection *address_space_lookup_region(AddressSpace *as,
232 233
                                                        hwaddr addr,
                                                        bool resolve_subpage)
234
{
235 236 237 238 239 240 241 242 243
    MemoryRegionSection *section;
    subpage_t *subpage;

    section = phys_page_find(as->dispatch, addr >> TARGET_PAGE_BITS);
    if (resolve_subpage && section->mr->subpage) {
        subpage = container_of(section->mr, subpage_t, iomem);
        section = &phys_sections[subpage->sub_section[SUBPAGE_IDX(addr)]];
    }
    return section;
244 245
}

246 247 248
static MemoryRegionSection *
address_space_translate_internal(AddressSpace *as, hwaddr addr, hwaddr *xlat,
                                 hwaddr *plen, bool resolve_subpage)
249 250 251 252
{
    MemoryRegionSection *section;
    Int128 diff;

253
    section = address_space_lookup_region(as, addr, resolve_subpage);
254 255 256 257 258 259 260
    /* Compute offset within MemoryRegionSection */
    addr -= section->offset_within_address_space;

    /* Compute offset within MemoryRegion */
    *xlat = addr + section->offset_within_region;

    diff = int128_sub(section->mr->size, int128_make64(addr));
261
    *plen = int128_get64(int128_min(diff, int128_make64(*plen)));
262 263
    return section;
}
264 265 266 267 268 269 270 271 272 273 274 275 276 277

MemoryRegionSection *address_space_translate(AddressSpace *as, hwaddr addr,
                                             hwaddr *xlat, hwaddr *plen,
                                             bool is_write)
{
    return address_space_translate_internal(as, addr, xlat, plen, true);
}

MemoryRegionSection *
address_space_translate_for_iotlb(AddressSpace *as, hwaddr addr, hwaddr *xlat,
                                  hwaddr *plen)
{
    return address_space_translate_internal(as, addr, xlat, plen, false);
}
278
#endif
B
bellard 已提交
279

280
void cpu_exec_init_all(void)
281
{
282
#if !defined(CONFIG_USER_ONLY)
283
    qemu_mutex_init(&ram_list.mutex);
284 285
    memory_map_init();
    io_mem_init();
286
#endif
287
}
288

289
#if !defined(CONFIG_USER_ONLY)
290 291

static int cpu_common_post_load(void *opaque, int version_id)
B
bellard 已提交
292
{
293
    CPUState *cpu = opaque;
B
bellard 已提交
294

295 296
    /* 0x01 was CPU_INTERRUPT_EXIT. This line can be removed when the
       version_id is increased. */
297 298
    cpu->interrupt_request &= ~0x01;
    tlb_flush(cpu->env_ptr, 1);
299 300

    return 0;
B
bellard 已提交
301
}
B
bellard 已提交
302

303 304 305 306 307 308 309
static const VMStateDescription vmstate_cpu_common = {
    .name = "cpu_common",
    .version_id = 1,
    .minimum_version_id = 1,
    .minimum_version_id_old = 1,
    .post_load = cpu_common_post_load,
    .fields      = (VMStateField []) {
310 311
        VMSTATE_UINT32(halted, CPUState),
        VMSTATE_UINT32(interrupt_request, CPUState),
312 313 314
        VMSTATE_END_OF_LIST()
    }
};
315 316
#else
#define vmstate_cpu_common vmstate_dummy
317
#endif
B
bellard 已提交
318

319
CPUState *qemu_get_cpu(int index)
B
bellard 已提交
320
{
321
    CPUArchState *env = first_cpu;
322
    CPUState *cpu = NULL;
B
bellard 已提交
323

324
    while (env) {
325 326
        cpu = ENV_GET_CPU(env);
        if (cpu->cpu_index == index) {
327
            break;
328
        }
329
        env = env->next_cpu;
B
bellard 已提交
330
    }
331

332
    return env ? cpu : NULL;
B
bellard 已提交
333 334
}

335 336 337 338 339 340 341 342 343 344
void qemu_for_each_cpu(void (*func)(CPUState *cpu, void *data), void *data)
{
    CPUArchState *env = first_cpu;

    while (env) {
        func(ENV_GET_CPU(env), data);
        env = env->next_cpu;
    }
}

345
void cpu_exec_init(CPUArchState *env)
B
bellard 已提交
346
{
347
    CPUState *cpu = ENV_GET_CPU(env);
348
    CPUClass *cc = CPU_GET_CLASS(cpu);
349 350 351 352 353 354 355 356 357 358 359 360 361
    CPUArchState **penv;
    int cpu_index;

#if defined(CONFIG_USER_ONLY)
    cpu_list_lock();
#endif
    env->next_cpu = NULL;
    penv = &first_cpu;
    cpu_index = 0;
    while (*penv != NULL) {
        penv = &(*penv)->next_cpu;
        cpu_index++;
    }
362
    cpu->cpu_index = cpu_index;
363
    cpu->numa_node = 0;
364 365 366 367 368 369 370 371 372
    QTAILQ_INIT(&env->breakpoints);
    QTAILQ_INIT(&env->watchpoints);
#ifndef CONFIG_USER_ONLY
    cpu->thread_id = qemu_get_thread_id();
#endif
    *penv = env;
#if defined(CONFIG_USER_ONLY)
    cpu_list_unlock();
#endif
373
    vmstate_register(NULL, cpu_index, &vmstate_cpu_common, cpu);
374 375 376
#if defined(CPU_SAVE_VERSION) && !defined(CONFIG_USER_ONLY)
    register_savevm(NULL, "cpu", cpu_index, CPU_SAVE_VERSION,
                    cpu_save, cpu_load, env);
377
    assert(cc->vmsd == NULL);
378
#endif
379 380 381
    if (cc->vmsd != NULL) {
        vmstate_register(NULL, cpu_index, cc->vmsd, cpu);
    }
B
bellard 已提交
382 383
}

B
bellard 已提交
384
#if defined(TARGET_HAS_ICE)
385
#if defined(CONFIG_USER_ONLY)
386
static void breakpoint_invalidate(CPUArchState *env, target_ulong pc)
387 388 389 390
{
    tb_invalidate_phys_page_range(pc, pc + 1, 0);
}
#else
391 392
static void breakpoint_invalidate(CPUArchState *env, target_ulong pc)
{
393 394
    tb_invalidate_phys_addr(cpu_get_phys_page_debug(env, pc) |
            (pc & ~TARGET_PAGE_MASK));
395
}
B
bellard 已提交
396
#endif
397
#endif /* TARGET_HAS_ICE */
B
bellard 已提交
398

399
#if defined(CONFIG_USER_ONLY)
400
void cpu_watchpoint_remove_all(CPUArchState *env, int mask)
401 402 403 404

{
}

405
int cpu_watchpoint_insert(CPUArchState *env, target_ulong addr, target_ulong len,
406 407 408 409 410
                          int flags, CPUWatchpoint **watchpoint)
{
    return -ENOSYS;
}
#else
411
/* Add a watchpoint.  */
412
int cpu_watchpoint_insert(CPUArchState *env, target_ulong addr, target_ulong len,
413
                          int flags, CPUWatchpoint **watchpoint)
414
{
415
    target_ulong len_mask = ~(len - 1);
416
    CPUWatchpoint *wp;
417

418
    /* sanity checks: allow power-of-2 lengths, deny unaligned watchpoints */
419 420
    if ((len & (len - 1)) || (addr & ~len_mask) ||
            len == 0 || len > TARGET_PAGE_SIZE) {
421 422 423 424
        fprintf(stderr, "qemu: tried to set invalid watchpoint at "
                TARGET_FMT_lx ", len=" TARGET_FMT_lu "\n", addr, len);
        return -EINVAL;
    }
425
    wp = g_malloc(sizeof(*wp));
426 427

    wp->vaddr = addr;
428
    wp->len_mask = len_mask;
429 430
    wp->flags = flags;

431
    /* keep all GDB-injected watchpoints in front */
432
    if (flags & BP_GDB)
B
Blue Swirl 已提交
433
        QTAILQ_INSERT_HEAD(&env->watchpoints, wp, entry);
434
    else
B
Blue Swirl 已提交
435
        QTAILQ_INSERT_TAIL(&env->watchpoints, wp, entry);
436 437

    tlb_flush_page(env, addr);
438 439 440 441

    if (watchpoint)
        *watchpoint = wp;
    return 0;
442 443
}

444
/* Remove a specific watchpoint.  */
445
int cpu_watchpoint_remove(CPUArchState *env, target_ulong addr, target_ulong len,
446
                          int flags)
447
{
448
    target_ulong len_mask = ~(len - 1);
449
    CPUWatchpoint *wp;
450

B
Blue Swirl 已提交
451
    QTAILQ_FOREACH(wp, &env->watchpoints, entry) {
452
        if (addr == wp->vaddr && len_mask == wp->len_mask
453
                && flags == (wp->flags & ~BP_WATCHPOINT_HIT)) {
454
            cpu_watchpoint_remove_by_ref(env, wp);
455 456 457
            return 0;
        }
    }
458
    return -ENOENT;
459 460
}

461
/* Remove a specific watchpoint by reference.  */
462
void cpu_watchpoint_remove_by_ref(CPUArchState *env, CPUWatchpoint *watchpoint)
463
{
B
Blue Swirl 已提交
464
    QTAILQ_REMOVE(&env->watchpoints, watchpoint, entry);
465

466 467
    tlb_flush_page(env, watchpoint->vaddr);

468
    g_free(watchpoint);
469 470 471
}

/* Remove all matching watchpoints.  */
472
void cpu_watchpoint_remove_all(CPUArchState *env, int mask)
473
{
474
    CPUWatchpoint *wp, *next;
475

B
Blue Swirl 已提交
476
    QTAILQ_FOREACH_SAFE(wp, &env->watchpoints, entry, next) {
477 478
        if (wp->flags & mask)
            cpu_watchpoint_remove_by_ref(env, wp);
479
    }
480
}
481
#endif
482

483
/* Add a breakpoint.  */
484
int cpu_breakpoint_insert(CPUArchState *env, target_ulong pc, int flags,
485
                          CPUBreakpoint **breakpoint)
B
bellard 已提交
486
{
B
bellard 已提交
487
#if defined(TARGET_HAS_ICE)
488
    CPUBreakpoint *bp;
489

490
    bp = g_malloc(sizeof(*bp));
B
bellard 已提交
491

492 493 494
    bp->pc = pc;
    bp->flags = flags;

495
    /* keep all GDB-injected breakpoints in front */
496
    if (flags & BP_GDB)
B
Blue Swirl 已提交
497
        QTAILQ_INSERT_HEAD(&env->breakpoints, bp, entry);
498
    else
B
Blue Swirl 已提交
499
        QTAILQ_INSERT_TAIL(&env->breakpoints, bp, entry);
500

B
bellard 已提交
501
    breakpoint_invalidate(env, pc);
502 503 504

    if (breakpoint)
        *breakpoint = bp;
B
bellard 已提交
505 506
    return 0;
#else
507
    return -ENOSYS;
B
bellard 已提交
508 509 510
#endif
}

511
/* Remove a specific breakpoint.  */
512
int cpu_breakpoint_remove(CPUArchState *env, target_ulong pc, int flags)
513
{
514
#if defined(TARGET_HAS_ICE)
515 516
    CPUBreakpoint *bp;

B
Blue Swirl 已提交
517
    QTAILQ_FOREACH(bp, &env->breakpoints, entry) {
518 519 520 521
        if (bp->pc == pc && bp->flags == flags) {
            cpu_breakpoint_remove_by_ref(env, bp);
            return 0;
        }
522
    }
523 524 525
    return -ENOENT;
#else
    return -ENOSYS;
526 527 528
#endif
}

529
/* Remove a specific breakpoint by reference.  */
530
void cpu_breakpoint_remove_by_ref(CPUArchState *env, CPUBreakpoint *breakpoint)
B
bellard 已提交
531
{
B
bellard 已提交
532
#if defined(TARGET_HAS_ICE)
B
Blue Swirl 已提交
533
    QTAILQ_REMOVE(&env->breakpoints, breakpoint, entry);
B
bellard 已提交
534

535 536
    breakpoint_invalidate(env, breakpoint->pc);

537
    g_free(breakpoint);
538 539 540 541
#endif
}

/* Remove all matching breakpoints. */
542
void cpu_breakpoint_remove_all(CPUArchState *env, int mask)
543 544
{
#if defined(TARGET_HAS_ICE)
545
    CPUBreakpoint *bp, *next;
546

B
Blue Swirl 已提交
547
    QTAILQ_FOREACH_SAFE(bp, &env->breakpoints, entry, next) {
548 549
        if (bp->flags & mask)
            cpu_breakpoint_remove_by_ref(env, bp);
550
    }
B
bellard 已提交
551 552 553
#endif
}

B
bellard 已提交
554 555
/* enable or disable single step mode. EXCP_DEBUG is returned by the
   CPU loop after each instruction */
556
void cpu_single_step(CPUArchState *env, int enabled)
B
bellard 已提交
557
{
B
bellard 已提交
558
#if defined(TARGET_HAS_ICE)
B
bellard 已提交
559 560
    if (env->singlestep_enabled != enabled) {
        env->singlestep_enabled = enabled;
561 562 563
        if (kvm_enabled())
            kvm_update_guest_debug(env, 0);
        else {
S
Stuart Brady 已提交
564
            /* must flush all the translated code to avoid inconsistencies */
565 566 567
            /* XXX: only flush what is necessary */
            tb_flush(env);
        }
B
bellard 已提交
568 569 570 571
    }
#endif
}

572
void cpu_exit(CPUArchState *env)
573
{
574 575 576
    CPUState *cpu = ENV_GET_CPU(env);

    cpu->exit_request = 1;
577
    cpu->tcg_exit_req = 1;
578 579
}

580
void cpu_abort(CPUArchState *env, const char *fmt, ...)
B
bellard 已提交
581 582
{
    va_list ap;
P
pbrook 已提交
583
    va_list ap2;
B
bellard 已提交
584 585

    va_start(ap, fmt);
P
pbrook 已提交
586
    va_copy(ap2, ap);
B
bellard 已提交
587 588 589
    fprintf(stderr, "qemu: fatal: ");
    vfprintf(stderr, fmt, ap);
    fprintf(stderr, "\n");
590
    cpu_dump_state(env, stderr, fprintf, CPU_DUMP_FPU | CPU_DUMP_CCOP);
591 592 593 594
    if (qemu_log_enabled()) {
        qemu_log("qemu: fatal: ");
        qemu_log_vprintf(fmt, ap2);
        qemu_log("\n");
595
        log_cpu_state(env, CPU_DUMP_FPU | CPU_DUMP_CCOP);
596
        qemu_log_flush();
597
        qemu_log_close();
598
    }
P
pbrook 已提交
599
    va_end(ap2);
600
    va_end(ap);
601 602 603 604 605 606 607 608
#if defined(CONFIG_USER_ONLY)
    {
        struct sigaction act;
        sigfillset(&act.sa_mask);
        act.sa_handler = SIG_DFL;
        sigaction(SIGABRT, &act, NULL);
    }
#endif
B
bellard 已提交
609 610 611
    abort();
}

612
CPUArchState *cpu_copy(CPUArchState *env)
613
{
614 615
    CPUArchState *new_env = cpu_init(env->cpu_model_str);
    CPUArchState *next_cpu = new_env->next_cpu;
616 617 618 619 620
#if defined(TARGET_HAS_ICE)
    CPUBreakpoint *bp;
    CPUWatchpoint *wp;
#endif

621
    memcpy(new_env, env, sizeof(CPUArchState));
622

623
    /* Preserve chaining. */
624
    new_env->next_cpu = next_cpu;
625 626 627 628

    /* Clone all break/watchpoints.
       Note: Once we support ptrace with hw-debug register access, make sure
       BP_CPU break/watchpoints are handled correctly on clone. */
B
Blue Swirl 已提交
629 630
    QTAILQ_INIT(&env->breakpoints);
    QTAILQ_INIT(&env->watchpoints);
631
#if defined(TARGET_HAS_ICE)
B
Blue Swirl 已提交
632
    QTAILQ_FOREACH(bp, &env->breakpoints, entry) {
633 634
        cpu_breakpoint_insert(new_env, bp->pc, bp->flags, NULL);
    }
B
Blue Swirl 已提交
635
    QTAILQ_FOREACH(wp, &env->watchpoints, entry) {
636 637 638 639 640
        cpu_watchpoint_insert(new_env, wp->vaddr, (~wp->len_mask) + 1,
                              wp->flags, NULL);
    }
#endif

641 642 643
    return new_env;
}

644
#if !defined(CONFIG_USER_ONLY)
J
Juan Quintela 已提交
645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662
static void tlb_reset_dirty_range_all(ram_addr_t start, ram_addr_t end,
                                      uintptr_t length)
{
    uintptr_t start1;

    /* we modify the TLB cache so that the dirty bit will be set again
       when accessing the range */
    start1 = (uintptr_t)qemu_safe_ram_ptr(start);
    /* Check that we don't span multiple blocks - this breaks the
       address comparisons below.  */
    if ((uintptr_t)qemu_safe_ram_ptr(end - 1) - start1
            != (end - 1) - start) {
        abort();
    }
    cpu_tlb_reset_dirty_all(start1, length);

}

P
pbrook 已提交
663
/* Note: start and end must be within the same ram block.  */
A
Anthony Liguori 已提交
664
void cpu_physical_memory_reset_dirty(ram_addr_t start, ram_addr_t end,
B
bellard 已提交
665
                                     int dirty_flags)
666
{
J
Juan Quintela 已提交
667
    uintptr_t length;
668 669 670 671 672 673 674

    start &= TARGET_PAGE_MASK;
    end = TARGET_PAGE_ALIGN(end);

    length = end - start;
    if (length == 0)
        return;
675
    cpu_physical_memory_mask_dirty_range(start, length, dirty_flags);
B
bellard 已提交
676

J
Juan Quintela 已提交
677 678
    if (tcg_enabled()) {
        tlb_reset_dirty_range_all(start, end, length);
P
pbrook 已提交
679
    }
680 681
}

B
Blue Swirl 已提交
682
static int cpu_physical_memory_set_dirty_tracking(int enable)
A
aliguori 已提交
683
{
M
Michael S. Tsirkin 已提交
684
    int ret = 0;
A
aliguori 已提交
685
    in_migration = enable;
M
Michael S. Tsirkin 已提交
686
    return ret;
A
aliguori 已提交
687 688
}

A
Avi Kivity 已提交
689
hwaddr memory_region_section_get_iotlb(CPUArchState *env,
690 691 692 693 694
                                       MemoryRegionSection *section,
                                       target_ulong vaddr,
                                       hwaddr paddr, hwaddr xlat,
                                       int prot,
                                       target_ulong *address)
B
Blue Swirl 已提交
695
{
A
Avi Kivity 已提交
696
    hwaddr iotlb;
B
Blue Swirl 已提交
697 698
    CPUWatchpoint *wp;

699
    if (memory_region_is_ram(section->mr)) {
B
Blue Swirl 已提交
700 701
        /* Normal RAM.  */
        iotlb = (memory_region_get_ram_addr(section->mr) & TARGET_PAGE_MASK)
702
            + xlat;
B
Blue Swirl 已提交
703 704 705 706 707 708 709
        if (!section->readonly) {
            iotlb |= phys_section_notdirty;
        } else {
            iotlb |= phys_section_rom;
        }
    } else {
        iotlb = section - phys_sections;
710
        iotlb += xlat;
B
Blue Swirl 已提交
711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727
    }

    /* Make accesses to pages with watchpoints go via the
       watchpoint trap routines.  */
    QTAILQ_FOREACH(wp, &env->watchpoints, entry) {
        if (vaddr == (wp->vaddr & TARGET_PAGE_MASK)) {
            /* Avoid trapping reads of pages with a write breakpoint. */
            if ((prot & PAGE_WRITE) || (wp->flags & BP_MEM_READ)) {
                iotlb = phys_section_watch + paddr;
                *address |= TLB_MMIO;
                break;
            }
        }
    }

    return iotlb;
}
728 729
#endif /* defined(CONFIG_USER_ONLY) */

730
#if !defined(CONFIG_USER_ONLY)
731

A
Anthony Liguori 已提交
732
static int subpage_register (subpage_t *mmio, uint32_t start, uint32_t end,
733
                             uint16_t section);
734
static subpage_t *subpage_init(AddressSpace *as, hwaddr base);
735
static void destroy_page_desc(uint16_t section_index)
736
{
737 738
    MemoryRegionSection *section = &phys_sections[section_index];
    MemoryRegion *mr = section->mr;
739 740 741 742 743 744 745 746

    if (mr->subpage) {
        subpage_t *subpage = container_of(mr, subpage_t, iomem);
        memory_region_destroy(&subpage->iomem);
        g_free(subpage);
    }
}

747
static void destroy_l2_mapping(PhysPageEntry *lp, unsigned level)
748 749
{
    unsigned i;
750
    PhysPageEntry *p;
751

752
    if (lp->ptr == PHYS_MAP_NODE_NIL) {
753 754 755
        return;
    }

756
    p = phys_map_nodes[lp->ptr];
757
    for (i = 0; i < L2_SIZE; ++i) {
758
        if (!p[i].is_leaf) {
759
            destroy_l2_mapping(&p[i], level - 1);
760
        } else {
761
            destroy_page_desc(p[i].ptr);
762 763
        }
    }
764
    lp->is_leaf = 0;
765
    lp->ptr = PHYS_MAP_NODE_NIL;
766 767
}

A
Avi Kivity 已提交
768
static void destroy_all_mappings(AddressSpaceDispatch *d)
769
{
A
Avi Kivity 已提交
770
    destroy_l2_mapping(&d->phys_map, P_L2_LEVELS - 1);
771
    phys_map_nodes_reset();
772 773
}

774 775
static uint16_t phys_section_add(MemoryRegionSection *section)
{
776 777 778 779 780 781
    /* The physical section number is ORed with a page-aligned
     * pointer to produce the iotlb entries.  Thus it should
     * never overflow into the page-aligned value.
     */
    assert(phys_sections_nb < TARGET_PAGE_SIZE);

782 783 784 785 786 787 788 789 790 791 792 793 794 795
    if (phys_sections_nb == phys_sections_nb_alloc) {
        phys_sections_nb_alloc = MAX(phys_sections_nb_alloc * 2, 16);
        phys_sections = g_renew(MemoryRegionSection, phys_sections,
                                phys_sections_nb_alloc);
    }
    phys_sections[phys_sections_nb] = *section;
    return phys_sections_nb++;
}

static void phys_sections_clear(void)
{
    phys_sections_nb = 0;
}

A
Avi Kivity 已提交
796
static void register_subpage(AddressSpaceDispatch *d, MemoryRegionSection *section)
797 798
{
    subpage_t *subpage;
A
Avi Kivity 已提交
799
    hwaddr base = section->offset_within_address_space
800
        & TARGET_PAGE_MASK;
A
Avi Kivity 已提交
801
    MemoryRegionSection *existing = phys_page_find(d, base >> TARGET_PAGE_BITS);
802 803 804 805
    MemoryRegionSection subsection = {
        .offset_within_address_space = base,
        .size = TARGET_PAGE_SIZE,
    };
A
Avi Kivity 已提交
806
    hwaddr start, end;
807

808
    assert(existing->mr->subpage || existing->mr == &io_mem_unassigned);
809

810
    if (!(existing->mr->subpage)) {
811
        subpage = subpage_init(d->as, base);
812
        subsection.mr = &subpage->iomem;
A
Avi Kivity 已提交
813
        phys_page_set(d, base >> TARGET_PAGE_BITS, 1,
814
                      phys_section_add(&subsection));
815
    } else {
816
        subpage = container_of(existing->mr, subpage_t, iomem);
817 818
    }
    start = section->offset_within_address_space & ~TARGET_PAGE_MASK;
819
    end = start + section->size - 1;
820 821 822 823
    subpage_register(subpage, start, end, phys_section_add(section));
}


A
Avi Kivity 已提交
824
static void register_multipage(AddressSpaceDispatch *d, MemoryRegionSection *section)
825
{
A
Avi Kivity 已提交
826
    hwaddr start_addr = section->offset_within_address_space;
827
    ram_addr_t size = section->size;
A
Avi Kivity 已提交
828
    hwaddr addr;
829
    uint16_t section_index = phys_section_add(section);
830

831
    assert(size);
M
Michael S. Tsirkin 已提交
832

833
    addr = start_addr;
A
Avi Kivity 已提交
834
    phys_page_set(d, addr >> TARGET_PAGE_BITS, size >> TARGET_PAGE_BITS,
835
                  section_index);
836 837
}

838 839 840 841 842 843 844 845 846 847 848
QEMU_BUILD_BUG_ON(TARGET_PHYS_ADDR_SPACE_BITS > MAX_PHYS_ADDR_SPACE_BITS)

static MemoryRegionSection limit(MemoryRegionSection section)
{
    section.size = MIN(section.offset_within_address_space + section.size,
                       MAX_PHYS_ADDR + 1)
                   - section.offset_within_address_space;

    return section;
}

A
Avi Kivity 已提交
849
static void mem_add(MemoryListener *listener, MemoryRegionSection *section)
850
{
A
Avi Kivity 已提交
851
    AddressSpaceDispatch *d = container_of(listener, AddressSpaceDispatch, listener);
852
    MemoryRegionSection now = limit(*section), remain = limit(*section);
853 854 855 856 857 858

    if ((now.offset_within_address_space & ~TARGET_PAGE_MASK)
        || (now.size < TARGET_PAGE_SIZE)) {
        now.size = MIN(TARGET_PAGE_ALIGN(now.offset_within_address_space)
                       - now.offset_within_address_space,
                       now.size);
A
Avi Kivity 已提交
859
        register_subpage(d, &now);
860 861 862 863
        remain.size -= now.size;
        remain.offset_within_address_space += now.size;
        remain.offset_within_region += now.size;
    }
864 865 866 867
    while (remain.size >= TARGET_PAGE_SIZE) {
        now = remain;
        if (remain.offset_within_region & ~TARGET_PAGE_MASK) {
            now.size = TARGET_PAGE_SIZE;
A
Avi Kivity 已提交
868
            register_subpage(d, &now);
869 870
        } else {
            now.size &= TARGET_PAGE_MASK;
A
Avi Kivity 已提交
871
            register_multipage(d, &now);
872
        }
873 874 875 876 877 878
        remain.size -= now.size;
        remain.offset_within_address_space += now.size;
        remain.offset_within_region += now.size;
    }
    now = remain;
    if (now.size) {
A
Avi Kivity 已提交
879
        register_subpage(d, &now);
880 881 882
    }
}

883 884 885 886 887 888
void qemu_flush_coalesced_mmio_buffer(void)
{
    if (kvm_enabled())
        kvm_flush_coalesced_mmio_buffer();
}

889 890 891 892 893 894 895 896 897 898
void qemu_mutex_lock_ramlist(void)
{
    qemu_mutex_lock(&ram_list.mutex);
}

void qemu_mutex_unlock_ramlist(void)
{
    qemu_mutex_unlock(&ram_list.mutex);
}

899 900 901 902 903 904 905 906 907 908 909 910
#if defined(__linux__) && !defined(TARGET_S390X)

#include <sys/vfs.h>

#define HUGETLBFS_MAGIC       0x958458f6

static long gethugepagesize(const char *path)
{
    struct statfs fs;
    int ret;

    do {
Y
Yoshiaki Tamura 已提交
911
        ret = statfs(path, &fs);
912 913 914
    } while (ret != 0 && errno == EINTR);

    if (ret != 0) {
Y
Yoshiaki Tamura 已提交
915 916
        perror(path);
        return 0;
917 918 919
    }

    if (fs.f_type != HUGETLBFS_MAGIC)
Y
Yoshiaki Tamura 已提交
920
        fprintf(stderr, "Warning: path not on HugeTLBFS: %s\n", path);
921 922 923 924

    return fs.f_bsize;
}

A
Alex Williamson 已提交
925 926 927
static void *file_ram_alloc(RAMBlock *block,
                            ram_addr_t memory,
                            const char *path)
928 929
{
    char *filename;
930 931
    char *sanitized_name;
    char *c;
932 933 934 935 936 937 938 939 940
    void *area;
    int fd;
#ifdef MAP_POPULATE
    int flags;
#endif
    unsigned long hpagesize;

    hpagesize = gethugepagesize(path);
    if (!hpagesize) {
Y
Yoshiaki Tamura 已提交
941
        return NULL;
942 943 944 945 946 947 948 949 950 951 952
    }

    if (memory < hpagesize) {
        return NULL;
    }

    if (kvm_enabled() && !kvm_has_sync_mmu()) {
        fprintf(stderr, "host lacks kvm mmu notifiers, -mem-path unsupported\n");
        return NULL;
    }

953 954 955 956 957 958 959 960 961 962
    /* Make name safe to use with mkstemp by replacing '/' with '_'. */
    sanitized_name = g_strdup(block->mr->name);
    for (c = sanitized_name; *c != '\0'; c++) {
        if (*c == '/')
            *c = '_';
    }

    filename = g_strdup_printf("%s/qemu_back_mem.%s.XXXXXX", path,
                               sanitized_name);
    g_free(sanitized_name);
963 964 965

    fd = mkstemp(filename);
    if (fd < 0) {
Y
Yoshiaki Tamura 已提交
966
        perror("unable to create backing store for hugepages");
967
        g_free(filename);
Y
Yoshiaki Tamura 已提交
968
        return NULL;
969 970
    }
    unlink(filename);
971
    g_free(filename);
972 973 974 975 976 977 978 979 980 981

    memory = (memory+hpagesize-1) & ~(hpagesize-1);

    /*
     * ftruncate is not supported by hugetlbfs in older
     * hosts, so don't bother bailing out on errors.
     * If anything goes wrong with it under other filesystems,
     * mmap will fail.
     */
    if (ftruncate(fd, memory))
Y
Yoshiaki Tamura 已提交
982
        perror("ftruncate");
983 984 985 986 987 988 989 990 991 992 993 994

#ifdef MAP_POPULATE
    /* NB: MAP_POPULATE won't exhaustively alloc all phys pages in the case
     * MAP_PRIVATE is requested.  For mem_prealloc we mmap as MAP_SHARED
     * to sidestep this quirk.
     */
    flags = mem_prealloc ? MAP_POPULATE | MAP_SHARED : MAP_PRIVATE;
    area = mmap(0, memory, PROT_READ | PROT_WRITE, flags, fd, 0);
#else
    area = mmap(0, memory, PROT_READ | PROT_WRITE, MAP_PRIVATE, fd, 0);
#endif
    if (area == MAP_FAILED) {
Y
Yoshiaki Tamura 已提交
995 996 997
        perror("file_ram_alloc: can't mmap RAM pages");
        close(fd);
        return (NULL);
998
    }
A
Alex Williamson 已提交
999
    block->fd = fd;
1000 1001 1002 1003
    return area;
}
#endif

1004
static ram_addr_t find_ram_offset(ram_addr_t size)
A
Alex Williamson 已提交
1005 1006
{
    RAMBlock *block, *next_block;
A
Alex Williamson 已提交
1007
    ram_addr_t offset = RAM_ADDR_MAX, mingap = RAM_ADDR_MAX;
A
Alex Williamson 已提交
1008

1009 1010
    assert(size != 0); /* it would hand out same offset multiple times */

P
Paolo Bonzini 已提交
1011
    if (QTAILQ_EMPTY(&ram_list.blocks))
A
Alex Williamson 已提交
1012 1013
        return 0;

P
Paolo Bonzini 已提交
1014
    QTAILQ_FOREACH(block, &ram_list.blocks, next) {
1015
        ram_addr_t end, next = RAM_ADDR_MAX;
A
Alex Williamson 已提交
1016 1017 1018

        end = block->offset + block->length;

P
Paolo Bonzini 已提交
1019
        QTAILQ_FOREACH(next_block, &ram_list.blocks, next) {
A
Alex Williamson 已提交
1020 1021 1022 1023 1024
            if (next_block->offset >= end) {
                next = MIN(next, next_block->offset);
            }
        }
        if (next - end >= size && next - end < mingap) {
A
Alex Williamson 已提交
1025
            offset = end;
A
Alex Williamson 已提交
1026 1027 1028
            mingap = next - end;
        }
    }
A
Alex Williamson 已提交
1029 1030 1031 1032 1033 1034 1035

    if (offset == RAM_ADDR_MAX) {
        fprintf(stderr, "Failed to find gap of requested size: %" PRIu64 "\n",
                (uint64_t)size);
        abort();
    }

A
Alex Williamson 已提交
1036 1037 1038
    return offset;
}

J
Juan Quintela 已提交
1039
ram_addr_t last_ram_offset(void)
1040 1041 1042 1043
{
    RAMBlock *block;
    ram_addr_t last = 0;

P
Paolo Bonzini 已提交
1044
    QTAILQ_FOREACH(block, &ram_list.blocks, next)
1045 1046 1047 1048 1049
        last = MAX(last, block->offset + block->length);

    return last;
}

1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067
static void qemu_ram_setup_dump(void *addr, ram_addr_t size)
{
    int ret;
    QemuOpts *machine_opts;

    /* Use MADV_DONTDUMP, if user doesn't want the guest memory in the core */
    machine_opts = qemu_opts_find(qemu_find_opts("machine"), 0);
    if (machine_opts &&
        !qemu_opt_get_bool(machine_opts, "dump-guest-core", true)) {
        ret = qemu_madvise(addr, size, QEMU_MADV_DONTDUMP);
        if (ret) {
            perror("qemu_madvise");
            fprintf(stderr, "madvise doesn't support MADV_DONTDUMP, "
                            "but dump_guest_core=off specified\n");
        }
    }
}

1068
void qemu_ram_set_idstr(ram_addr_t addr, const char *name, DeviceState *dev)
1069 1070 1071
{
    RAMBlock *new_block, *block;

1072
    new_block = NULL;
P
Paolo Bonzini 已提交
1073
    QTAILQ_FOREACH(block, &ram_list.blocks, next) {
1074 1075 1076 1077 1078 1079 1080
        if (block->offset == addr) {
            new_block = block;
            break;
        }
    }
    assert(new_block);
    assert(!new_block->idstr[0]);
1081

1082 1083
    if (dev) {
        char *id = qdev_get_dev_path(dev);
1084 1085
        if (id) {
            snprintf(new_block->idstr, sizeof(new_block->idstr), "%s/", id);
1086
            g_free(id);
1087 1088 1089 1090
        }
    }
    pstrcat(new_block->idstr, sizeof(new_block->idstr), name);

1091 1092
    /* This assumes the iothread lock is taken here too.  */
    qemu_mutex_lock_ramlist();
P
Paolo Bonzini 已提交
1093
    QTAILQ_FOREACH(block, &ram_list.blocks, next) {
1094
        if (block != new_block && !strcmp(block->idstr, new_block->idstr)) {
1095 1096 1097 1098 1099
            fprintf(stderr, "RAMBlock \"%s\" already registered, abort!\n",
                    new_block->idstr);
            abort();
        }
    }
1100
    qemu_mutex_unlock_ramlist();
1101 1102
}

1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115
static int memory_try_enable_merging(void *addr, size_t len)
{
    QemuOpts *opts;

    opts = qemu_opts_find(qemu_find_opts("machine"), 0);
    if (opts && !qemu_opt_get_bool(opts, "mem-merge", true)) {
        /* disabled by the user */
        return 0;
    }

    return qemu_madvise(addr, len, QEMU_MADV_MERGEABLE);
}

1116 1117 1118
ram_addr_t qemu_ram_alloc_from_ptr(ram_addr_t size, void *host,
                                   MemoryRegion *mr)
{
1119
    RAMBlock *block, *new_block;
1120 1121 1122

    size = TARGET_PAGE_ALIGN(size);
    new_block = g_malloc0(sizeof(*new_block));
1123

1124 1125
    /* This assumes the iothread lock is taken here too.  */
    qemu_mutex_lock_ramlist();
A
Avi Kivity 已提交
1126
    new_block->mr = mr;
J
Jun Nakajima 已提交
1127
    new_block->offset = find_ram_offset(size);
1128 1129
    if (host) {
        new_block->host = host;
H
Huang Ying 已提交
1130
        new_block->flags |= RAM_PREALLOC_MASK;
1131 1132
    } else {
        if (mem_path) {
1133
#if defined (__linux__) && !defined(TARGET_S390X)
1134 1135
            new_block->host = file_ram_alloc(new_block, size, mem_path);
            if (!new_block->host) {
1136
                new_block->host = qemu_anon_ram_alloc(size);
1137
                memory_try_enable_merging(new_block->host, size);
1138
            }
1139
#else
1140 1141
            fprintf(stderr, "-mem-path option unsupported\n");
            exit(1);
1142
#endif
1143
        } else {
1144
            if (xen_enabled()) {
1145
                xen_ram_alloc(new_block->offset, size, mr);
1146 1147
            } else if (kvm_enabled()) {
                /* some s390/kvm configurations have special constraints */
1148
                new_block->host = kvm_ram_alloc(size);
J
Jun Nakajima 已提交
1149
            } else {
1150
                new_block->host = qemu_anon_ram_alloc(size);
J
Jun Nakajima 已提交
1151
            }
1152
            memory_try_enable_merging(new_block->host, size);
1153
        }
1154
    }
P
pbrook 已提交
1155 1156
    new_block->length = size;

1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167
    /* Keep the list sorted from biggest to smallest block.  */
    QTAILQ_FOREACH(block, &ram_list.blocks, next) {
        if (block->length < new_block->length) {
            break;
        }
    }
    if (block) {
        QTAILQ_INSERT_BEFORE(block, new_block, next);
    } else {
        QTAILQ_INSERT_TAIL(&ram_list.blocks, new_block, next);
    }
1168
    ram_list.mru_block = NULL;
P
pbrook 已提交
1169

U
Umesh Deshpande 已提交
1170
    ram_list.version++;
1171
    qemu_mutex_unlock_ramlist();
U
Umesh Deshpande 已提交
1172

1173
    ram_list.phys_dirty = g_realloc(ram_list.phys_dirty,
A
Alex Williamson 已提交
1174
                                       last_ram_offset() >> TARGET_PAGE_BITS);
1175 1176
    memset(ram_list.phys_dirty + (new_block->offset >> TARGET_PAGE_BITS),
           0, size >> TARGET_PAGE_BITS);
J
Juan Quintela 已提交
1177
    cpu_physical_memory_set_dirty_range(new_block->offset, size, 0xff);
P
pbrook 已提交
1178

1179
    qemu_ram_setup_dump(new_block->host, size);
1180
    qemu_madvise(new_block->host, size, QEMU_MADV_HUGEPAGE);
1181

1182 1183 1184
    if (kvm_enabled())
        kvm_setup_guest_memory(new_block->host, size);

P
pbrook 已提交
1185 1186
    return new_block->offset;
}
B
bellard 已提交
1187

1188
ram_addr_t qemu_ram_alloc(ram_addr_t size, MemoryRegion *mr)
1189
{
1190
    return qemu_ram_alloc_from_ptr(size, NULL, mr);
1191 1192
}

1193 1194 1195 1196
void qemu_ram_free_from_ptr(ram_addr_t addr)
{
    RAMBlock *block;

1197 1198
    /* This assumes the iothread lock is taken here too.  */
    qemu_mutex_lock_ramlist();
P
Paolo Bonzini 已提交
1199
    QTAILQ_FOREACH(block, &ram_list.blocks, next) {
1200
        if (addr == block->offset) {
P
Paolo Bonzini 已提交
1201
            QTAILQ_REMOVE(&ram_list.blocks, block, next);
1202
            ram_list.mru_block = NULL;
U
Umesh Deshpande 已提交
1203
            ram_list.version++;
1204
            g_free(block);
1205
            break;
1206 1207
        }
    }
1208
    qemu_mutex_unlock_ramlist();
1209 1210
}

A
Anthony Liguori 已提交
1211
void qemu_ram_free(ram_addr_t addr)
B
bellard 已提交
1212
{
A
Alex Williamson 已提交
1213 1214
    RAMBlock *block;

1215 1216
    /* This assumes the iothread lock is taken here too.  */
    qemu_mutex_lock_ramlist();
P
Paolo Bonzini 已提交
1217
    QTAILQ_FOREACH(block, &ram_list.blocks, next) {
A
Alex Williamson 已提交
1218
        if (addr == block->offset) {
P
Paolo Bonzini 已提交
1219
            QTAILQ_REMOVE(&ram_list.blocks, block, next);
1220
            ram_list.mru_block = NULL;
U
Umesh Deshpande 已提交
1221
            ram_list.version++;
H
Huang Ying 已提交
1222 1223 1224
            if (block->flags & RAM_PREALLOC_MASK) {
                ;
            } else if (mem_path) {
A
Alex Williamson 已提交
1225 1226 1227 1228 1229
#if defined (__linux__) && !defined(TARGET_S390X)
                if (block->fd) {
                    munmap(block->host, block->length);
                    close(block->fd);
                } else {
1230
                    qemu_anon_ram_free(block->host, block->length);
A
Alex Williamson 已提交
1231
                }
1232 1233
#else
                abort();
A
Alex Williamson 已提交
1234 1235
#endif
            } else {
1236
                if (xen_enabled()) {
J
Jan Kiszka 已提交
1237
                    xen_invalidate_map_cache_entry(block->host);
J
Jun Nakajima 已提交
1238
                } else {
1239
                    qemu_anon_ram_free(block->host, block->length);
J
Jun Nakajima 已提交
1240
                }
A
Alex Williamson 已提交
1241
            }
1242
            g_free(block);
1243
            break;
A
Alex Williamson 已提交
1244 1245
        }
    }
1246
    qemu_mutex_unlock_ramlist();
A
Alex Williamson 已提交
1247

B
bellard 已提交
1248 1249
}

H
Huang Ying 已提交
1250 1251 1252 1253 1254 1255 1256 1257
#ifndef _WIN32
void qemu_ram_remap(ram_addr_t addr, ram_addr_t length)
{
    RAMBlock *block;
    ram_addr_t offset;
    int flags;
    void *area, *vaddr;

P
Paolo Bonzini 已提交
1258
    QTAILQ_FOREACH(block, &ram_list.blocks, next) {
H
Huang Ying 已提交
1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282
        offset = addr - block->offset;
        if (offset < block->length) {
            vaddr = block->host + offset;
            if (block->flags & RAM_PREALLOC_MASK) {
                ;
            } else {
                flags = MAP_FIXED;
                munmap(vaddr, length);
                if (mem_path) {
#if defined(__linux__) && !defined(TARGET_S390X)
                    if (block->fd) {
#ifdef MAP_POPULATE
                        flags |= mem_prealloc ? MAP_POPULATE | MAP_SHARED :
                            MAP_PRIVATE;
#else
                        flags |= MAP_PRIVATE;
#endif
                        area = mmap(vaddr, length, PROT_READ | PROT_WRITE,
                                    flags, block->fd, offset);
                    } else {
                        flags |= MAP_PRIVATE | MAP_ANONYMOUS;
                        area = mmap(vaddr, length, PROT_READ | PROT_WRITE,
                                    flags, -1, 0);
                    }
1283 1284
#else
                    abort();
H
Huang Ying 已提交
1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297
#endif
                } else {
#if defined(TARGET_S390X) && defined(CONFIG_KVM)
                    flags |= MAP_SHARED | MAP_ANONYMOUS;
                    area = mmap(vaddr, length, PROT_EXEC|PROT_READ|PROT_WRITE,
                                flags, -1, 0);
#else
                    flags |= MAP_PRIVATE | MAP_ANONYMOUS;
                    area = mmap(vaddr, length, PROT_READ | PROT_WRITE,
                                flags, -1, 0);
#endif
                }
                if (area != vaddr) {
1298 1299
                    fprintf(stderr, "Could not remap addr: "
                            RAM_ADDR_FMT "@" RAM_ADDR_FMT "\n",
H
Huang Ying 已提交
1300 1301 1302
                            length, addr);
                    exit(1);
                }
1303
                memory_try_enable_merging(vaddr, length);
1304
                qemu_ram_setup_dump(vaddr, length);
H
Huang Ying 已提交
1305 1306 1307 1308 1309 1310 1311
            }
            return;
        }
    }
}
#endif /* !_WIN32 */

1312
/* Return a host pointer to ram allocated with qemu_ram_alloc.
P
pbrook 已提交
1313 1314 1315 1316 1317 1318 1319
   With the exception of the softmmu code in this file, this should
   only be used for local memory (e.g. video ram) that the device owns,
   and knows it isn't going to access beyond the end of the block.

   It should not be used for general purpose DMA.
   Use cpu_physical_memory_map/cpu_physical_memory_rw instead.
 */
A
Anthony Liguori 已提交
1320
void *qemu_get_ram_ptr(ram_addr_t addr)
1321
{
P
pbrook 已提交
1322 1323
    RAMBlock *block;

1324
    /* The list is protected by the iothread lock here.  */
1325 1326 1327 1328
    block = ram_list.mru_block;
    if (block && addr - block->offset < block->length) {
        goto found;
    }
P
Paolo Bonzini 已提交
1329
    QTAILQ_FOREACH(block, &ram_list.blocks, next) {
A
Alex Williamson 已提交
1330
        if (addr - block->offset < block->length) {
1331
            goto found;
A
Alex Williamson 已提交
1332
        }
P
pbrook 已提交
1333
    }
A
Alex Williamson 已提交
1334 1335 1336 1337

    fprintf(stderr, "Bad ram offset %" PRIx64 "\n", (uint64_t)addr);
    abort();

1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352
found:
    ram_list.mru_block = block;
    if (xen_enabled()) {
        /* We need to check if the requested address is in the RAM
         * because we don't want to map the entire memory in QEMU.
         * In that case just map until the end of the page.
         */
        if (block->offset == 0) {
            return xen_map_cache(addr, 0, 0);
        } else if (block->host == NULL) {
            block->host =
                xen_map_cache(block->offset, block->length, 1);
        }
    }
    return block->host + (addr - block->offset);
1353 1354
}

1355 1356 1357 1358
/* Return a host pointer to ram allocated with qemu_ram_alloc.  Same as
 * qemu_get_ram_ptr but do not touch ram_list.mru_block.
 *
 * ??? Is this still necessary?
1359
 */
B
Blue Swirl 已提交
1360
static void *qemu_safe_ram_ptr(ram_addr_t addr)
1361 1362 1363
{
    RAMBlock *block;

1364
    /* The list is protected by the iothread lock here.  */
P
Paolo Bonzini 已提交
1365
    QTAILQ_FOREACH(block, &ram_list.blocks, next) {
1366
        if (addr - block->offset < block->length) {
1367
            if (xen_enabled()) {
J
Jun Nakajima 已提交
1368 1369
                /* We need to check if the requested address is in the RAM
                 * because we don't want to map the entire memory in QEMU.
1370
                 * In that case just map until the end of the page.
J
Jun Nakajima 已提交
1371 1372
                 */
                if (block->offset == 0) {
J
Jan Kiszka 已提交
1373
                    return xen_map_cache(addr, 0, 0);
J
Jun Nakajima 已提交
1374
                } else if (block->host == NULL) {
J
Jan Kiszka 已提交
1375 1376
                    block->host =
                        xen_map_cache(block->offset, block->length, 1);
J
Jun Nakajima 已提交
1377 1378
                }
            }
1379 1380 1381 1382 1383 1384 1385 1386 1387 1388
            return block->host + (addr - block->offset);
        }
    }

    fprintf(stderr, "Bad ram offset %" PRIx64 "\n", (uint64_t)addr);
    abort();

    return NULL;
}

1389 1390
/* Return a host pointer to guest's ram. Similar to qemu_get_ram_ptr
 * but takes a size argument */
B
Blue Swirl 已提交
1391
static void *qemu_ram_ptr_length(ram_addr_t addr, ram_addr_t *size)
1392
{
1393 1394 1395
    if (*size == 0) {
        return NULL;
    }
1396
    if (xen_enabled()) {
J
Jan Kiszka 已提交
1397
        return xen_map_cache(addr, *size, 1);
1398
    } else {
1399 1400
        RAMBlock *block;

P
Paolo Bonzini 已提交
1401
        QTAILQ_FOREACH(block, &ram_list.blocks, next) {
1402 1403 1404 1405 1406 1407 1408 1409 1410 1411 1412 1413
            if (addr - block->offset < block->length) {
                if (addr - block->offset + *size > block->length)
                    *size = block->length - addr + block->offset;
                return block->host + (addr - block->offset);
            }
        }

        fprintf(stderr, "Bad ram offset %" PRIx64 "\n", (uint64_t)addr);
        abort();
    }
}

M
Marcelo Tosatti 已提交
1414
int qemu_ram_addr_from_host(void *ptr, ram_addr_t *ram_addr)
P
pbrook 已提交
1415
{
P
pbrook 已提交
1416 1417 1418
    RAMBlock *block;
    uint8_t *host = ptr;

1419
    if (xen_enabled()) {
J
Jan Kiszka 已提交
1420
        *ram_addr = xen_ram_addr_from_mapcache(ptr);
1421 1422 1423
        return 0;
    }

P
Paolo Bonzini 已提交
1424
    QTAILQ_FOREACH(block, &ram_list.blocks, next) {
J
Jun Nakajima 已提交
1425 1426 1427 1428
        /* This case append when the block is not mapped. */
        if (block->host == NULL) {
            continue;
        }
A
Alex Williamson 已提交
1429
        if (host - block->host < block->length) {
M
Marcelo Tosatti 已提交
1430 1431
            *ram_addr = block->offset + (host - block->host);
            return 0;
A
Alex Williamson 已提交
1432
        }
P
pbrook 已提交
1433
    }
J
Jun Nakajima 已提交
1434

M
Marcelo Tosatti 已提交
1435 1436
    return -1;
}
A
Alex Williamson 已提交
1437

M
Marcelo Tosatti 已提交
1438 1439 1440 1441 1442
/* Some of the softmmu routines need to translate from a host pointer
   (typically a TLB entry) back to a ram offset.  */
ram_addr_t qemu_ram_addr_from_host_nofail(void *ptr)
{
    ram_addr_t ram_addr;
A
Alex Williamson 已提交
1443

M
Marcelo Tosatti 已提交
1444 1445 1446 1447 1448
    if (qemu_ram_addr_from_host(ptr, &ram_addr)) {
        fprintf(stderr, "Bad ram pointer %p\n", ptr);
        abort();
    }
    return ram_addr;
P
pbrook 已提交
1449 1450
}

A
Avi Kivity 已提交
1451
static void notdirty_mem_write(void *opaque, hwaddr ram_addr,
1452
                               uint64_t val, unsigned size)
1453
{
1454
    int dirty_flags;
1455
    dirty_flags = cpu_physical_memory_get_dirty_flags(ram_addr);
1456
    if (!(dirty_flags & CODE_DIRTY_FLAG)) {
1457
        tb_invalidate_phys_page_fast(ram_addr, size);
1458
        dirty_flags = cpu_physical_memory_get_dirty_flags(ram_addr);
1459
    }
1460 1461 1462 1463 1464 1465 1466 1467 1468 1469 1470 1471
    switch (size) {
    case 1:
        stb_p(qemu_get_ram_ptr(ram_addr), val);
        break;
    case 2:
        stw_p(qemu_get_ram_ptr(ram_addr), val);
        break;
    case 4:
        stl_p(qemu_get_ram_ptr(ram_addr), val);
        break;
    default:
        abort();
1472
    }
B
bellard 已提交
1473
    dirty_flags |= (0xff & ~CODE_DIRTY_FLAG);
1474
    cpu_physical_memory_set_dirty_flags(ram_addr, dirty_flags);
B
bellard 已提交
1475 1476 1477
    /* we remove the notdirty callback only if the code has been
       flushed */
    if (dirty_flags == 0xff)
P
pbrook 已提交
1478
        tlb_set_dirty(cpu_single_env, cpu_single_env->mem_io_vaddr);
1479 1480
}

1481 1482 1483 1484 1485 1486
static bool notdirty_mem_accepts(void *opaque, hwaddr addr,
                                 unsigned size, bool is_write)
{
    return is_write;
}

1487 1488
static const MemoryRegionOps notdirty_mem_ops = {
    .write = notdirty_mem_write,
1489
    .valid.accepts = notdirty_mem_accepts,
1490
    .endianness = DEVICE_NATIVE_ENDIAN,
1491 1492
};

P
pbrook 已提交
1493
/* Generate a debug exception if a watchpoint has been hit.  */
1494
static void check_watchpoint(int offset, int len_mask, int flags)
P
pbrook 已提交
1495
{
1496
    CPUArchState *env = cpu_single_env;
1497
    target_ulong pc, cs_base;
P
pbrook 已提交
1498
    target_ulong vaddr;
1499
    CPUWatchpoint *wp;
1500
    int cpu_flags;
P
pbrook 已提交
1501

1502 1503 1504 1505
    if (env->watchpoint_hit) {
        /* We re-entered the check after replacing the TB. Now raise
         * the debug interrupt so that is will trigger after the
         * current instruction. */
1506
        cpu_interrupt(ENV_GET_CPU(env), CPU_INTERRUPT_DEBUG);
1507 1508
        return;
    }
P
pbrook 已提交
1509
    vaddr = (env->mem_io_vaddr & TARGET_PAGE_MASK) + offset;
B
Blue Swirl 已提交
1510
    QTAILQ_FOREACH(wp, &env->watchpoints, entry) {
1511 1512
        if ((vaddr == (wp->vaddr & len_mask) ||
             (vaddr & wp->len_mask) == wp->vaddr) && (wp->flags & flags)) {
1513 1514 1515
            wp->flags |= BP_WATCHPOINT_HIT;
            if (!env->watchpoint_hit) {
                env->watchpoint_hit = wp;
B
Blue Swirl 已提交
1516
                tb_check_watchpoint(env);
1517 1518
                if (wp->flags & BP_STOP_BEFORE_ACCESS) {
                    env->exception_index = EXCP_DEBUG;
1519
                    cpu_loop_exit(env);
1520 1521 1522
                } else {
                    cpu_get_tb_cpu_state(env, &pc, &cs_base, &cpu_flags);
                    tb_gen_code(env, pc, cs_base, cpu_flags, 1);
1523
                    cpu_resume_from_signal(env, NULL);
1524
                }
1525
            }
1526 1527
        } else {
            wp->flags &= ~BP_WATCHPOINT_HIT;
P
pbrook 已提交
1528 1529 1530 1531
        }
    }
}

1532 1533 1534
/* Watchpoint access routines.  Watchpoints are inserted using TLB tricks,
   so these check for a hit then pass through to the normal out-of-line
   phys routines.  */
A
Avi Kivity 已提交
1535
static uint64_t watch_mem_read(void *opaque, hwaddr addr,
1536
                               unsigned size)
1537
{
1538 1539 1540 1541 1542 1543 1544
    check_watchpoint(addr & ~TARGET_PAGE_MASK, ~(size - 1), BP_MEM_READ);
    switch (size) {
    case 1: return ldub_phys(addr);
    case 2: return lduw_phys(addr);
    case 4: return ldl_phys(addr);
    default: abort();
    }
1545 1546
}

A
Avi Kivity 已提交
1547
static void watch_mem_write(void *opaque, hwaddr addr,
1548
                            uint64_t val, unsigned size)
1549
{
1550 1551
    check_watchpoint(addr & ~TARGET_PAGE_MASK, ~(size - 1), BP_MEM_WRITE);
    switch (size) {
1552 1553 1554 1555 1556 1557 1558 1559 1560
    case 1:
        stb_phys(addr, val);
        break;
    case 2:
        stw_phys(addr, val);
        break;
    case 4:
        stl_phys(addr, val);
        break;
1561 1562
    default: abort();
    }
1563 1564
}

1565 1566 1567 1568
static const MemoryRegionOps watch_mem_ops = {
    .read = watch_mem_read,
    .write = watch_mem_write,
    .endianness = DEVICE_NATIVE_ENDIAN,
1569 1570
};

A
Avi Kivity 已提交
1571
static uint64_t subpage_read(void *opaque, hwaddr addr,
1572
                             unsigned len)
1573
{
1574 1575
    subpage_t *subpage = opaque;
    uint8_t buf[4];
1576

1577
#if defined(DEBUG_SUBPAGE)
1578 1579
    printf("%s: subpage %p len %d addr " TARGET_FMT_plx "\n", __func__,
           subpage, len, addr);
1580
#endif
1581 1582 1583 1584 1585 1586 1587 1588 1589 1590 1591
    address_space_read(subpage->as, addr + subpage->base, buf, len);
    switch (len) {
    case 1:
        return ldub_p(buf);
    case 2:
        return lduw_p(buf);
    case 4:
        return ldl_p(buf);
    default:
        abort();
    }
1592 1593
}

A
Avi Kivity 已提交
1594
static void subpage_write(void *opaque, hwaddr addr,
1595
                          uint64_t value, unsigned len)
1596
{
1597 1598 1599
    subpage_t *subpage = opaque;
    uint8_t buf[4];

1600
#if defined(DEBUG_SUBPAGE)
1601
    printf("%s: subpage %p len %d addr " TARGET_FMT_plx
1602 1603
           " value %"PRIx64"\n",
           __func__, subpage, len, addr, value);
1604
#endif
1605 1606 1607 1608 1609 1610 1611 1612 1613 1614 1615 1616 1617 1618
    switch (len) {
    case 1:
        stb_p(buf, value);
        break;
    case 2:
        stw_p(buf, value);
        break;
    case 4:
        stl_p(buf, value);
        break;
    default:
        abort();
    }
    address_space_write(subpage->as, addr + subpage->base, buf, len);
1619 1620
}

1621 1622 1623
static bool subpage_accepts(void *opaque, hwaddr addr,
                            unsigned size, bool is_write)
{
1624
    subpage_t *subpage = opaque;
1625
#if defined(DEBUG_SUBPAGE)
1626 1627
    printf("%s: subpage %p %c len %d addr " TARGET_FMT_plx "\n",
           __func__, subpage, is_write ? 'w' : 'r', len, addr);
1628 1629
#endif

1630 1631
    return address_space_access_valid(subpage->as, addr + subpage->base,
                                      size, is_write);
1632 1633
}

1634 1635 1636
static const MemoryRegionOps subpage_ops = {
    .read = subpage_read,
    .write = subpage_write,
1637
    .valid.accepts = subpage_accepts,
1638
    .endianness = DEVICE_NATIVE_ENDIAN,
1639 1640
};

A
Anthony Liguori 已提交
1641
static int subpage_register (subpage_t *mmio, uint32_t start, uint32_t end,
1642
                             uint16_t section)
1643 1644 1645 1646 1647 1648 1649 1650
{
    int idx, eidx;

    if (start >= TARGET_PAGE_SIZE || end >= TARGET_PAGE_SIZE)
        return -1;
    idx = SUBPAGE_IDX(start);
    eidx = SUBPAGE_IDX(end);
#if defined(DEBUG_SUBPAGE)
1651
    printf("%s: %p start %08x end %08x idx %08x eidx %08x mem %ld\n", __func__,
1652 1653 1654
           mmio, start, end, idx, eidx, memory);
#endif
    for (; idx <= eidx; idx++) {
1655
        mmio->sub_section[idx] = section;
1656 1657 1658 1659 1660
    }

    return 0;
}

1661
static subpage_t *subpage_init(AddressSpace *as, hwaddr base)
1662
{
A
Anthony Liguori 已提交
1663
    subpage_t *mmio;
1664

1665
    mmio = g_malloc0(sizeof(subpage_t));
1666

1667
    mmio->as = as;
1668
    mmio->base = base;
1669 1670
    memory_region_init_io(&mmio->iomem, &subpage_ops, mmio,
                          "subpage", TARGET_PAGE_SIZE);
A
Avi Kivity 已提交
1671
    mmio->iomem.subpage = true;
1672
#if defined(DEBUG_SUBPAGE)
1673 1674
    printf("%s: %p base " TARGET_FMT_plx " len %08x %d\n", __func__,
           mmio, base, TARGET_PAGE_SIZE, subpage_memory);
1675
#endif
1676
    subpage_register(mmio, 0, TARGET_PAGE_SIZE-1, phys_section_unassigned);
1677 1678 1679 1680

    return mmio;
}

1681 1682 1683 1684 1685 1686 1687 1688 1689 1690 1691 1692
static uint16_t dummy_section(MemoryRegion *mr)
{
    MemoryRegionSection section = {
        .mr = mr,
        .offset_within_address_space = 0,
        .offset_within_region = 0,
        .size = UINT64_MAX,
    };

    return phys_section_add(&section);
}

A
Avi Kivity 已提交
1693
MemoryRegion *iotlb_to_region(hwaddr index)
1694
{
1695
    return phys_sections[index & ~TARGET_PAGE_MASK].mr;
1696 1697
}

A
Avi Kivity 已提交
1698 1699
static void io_mem_init(void)
{
P
Paolo Bonzini 已提交
1700
    memory_region_init_io(&io_mem_rom, &unassigned_mem_ops, NULL, "rom", UINT64_MAX);
1701 1702 1703 1704
    memory_region_init_io(&io_mem_unassigned, &unassigned_mem_ops, NULL,
                          "unassigned", UINT64_MAX);
    memory_region_init_io(&io_mem_notdirty, &notdirty_mem_ops, NULL,
                          "notdirty", UINT64_MAX);
1705 1706
    memory_region_init_io(&io_mem_watch, &watch_mem_ops, NULL,
                          "watch", UINT64_MAX);
A
Avi Kivity 已提交
1707 1708
}

A
Avi Kivity 已提交
1709 1710 1711 1712 1713 1714 1715 1716
static void mem_begin(MemoryListener *listener)
{
    AddressSpaceDispatch *d = container_of(listener, AddressSpaceDispatch, listener);

    destroy_all_mappings(d);
    d->phys_map.ptr = PHYS_MAP_NODE_NIL;
}

1717 1718
static void core_begin(MemoryListener *listener)
{
1719 1720
    phys_sections_clear();
    phys_section_unassigned = dummy_section(&io_mem_unassigned);
1721 1722 1723
    phys_section_notdirty = dummy_section(&io_mem_notdirty);
    phys_section_rom = dummy_section(&io_mem_rom);
    phys_section_watch = dummy_section(&io_mem_watch);
1724 1725
}

1726
static void tcg_commit(MemoryListener *listener)
1727
{
1728
    CPUArchState *env;
1729 1730 1731 1732 1733 1734 1735

    /* since each CPU stores ram addresses in its TLB cache, we must
       reset the modified entries */
    /* XXX: slow ! */
    for(env = first_cpu; env != NULL; env = env->next_cpu) {
        tlb_flush(env, 1);
    }
1736 1737
}

1738 1739 1740 1741 1742 1743 1744 1745 1746 1747
static void core_log_global_start(MemoryListener *listener)
{
    cpu_physical_memory_set_dirty_tracking(1);
}

static void core_log_global_stop(MemoryListener *listener)
{
    cpu_physical_memory_set_dirty_tracking(0);
}

1748 1749 1750
static void io_region_add(MemoryListener *listener,
                          MemoryRegionSection *section)
{
A
Avi Kivity 已提交
1751 1752 1753 1754 1755
    MemoryRegionIORange *mrio = g_new(MemoryRegionIORange, 1);

    mrio->mr = section->mr;
    mrio->offset = section->offset_within_region;
    iorange_init(&mrio->iorange, &memory_region_iorange_ops,
1756
                 section->offset_within_address_space, section->size);
A
Avi Kivity 已提交
1757
    ioport_register(&mrio->iorange);
1758 1759 1760 1761 1762 1763 1764 1765
}

static void io_region_del(MemoryListener *listener,
                          MemoryRegionSection *section)
{
    isa_unassign_ioport(section->offset_within_address_space, section->size);
}

1766
static MemoryListener core_memory_listener = {
1767
    .begin = core_begin,
1768 1769
    .log_global_start = core_log_global_start,
    .log_global_stop = core_log_global_stop,
A
Avi Kivity 已提交
1770
    .priority = 1,
1771 1772
};

1773 1774 1775 1776 1777 1778
static MemoryListener io_memory_listener = {
    .region_add = io_region_add,
    .region_del = io_region_del,
    .priority = 0,
};

1779 1780 1781 1782
static MemoryListener tcg_memory_listener = {
    .commit = tcg_commit,
};

A
Avi Kivity 已提交
1783 1784 1785 1786 1787 1788 1789 1790 1791 1792 1793
void address_space_init_dispatch(AddressSpace *as)
{
    AddressSpaceDispatch *d = g_new(AddressSpaceDispatch, 1);

    d->phys_map  = (PhysPageEntry) { .ptr = PHYS_MAP_NODE_NIL, .is_leaf = 0 };
    d->listener = (MemoryListener) {
        .begin = mem_begin,
        .region_add = mem_add,
        .region_nop = mem_add,
        .priority = 0,
    };
1794
    d->as = as;
A
Avi Kivity 已提交
1795 1796 1797 1798
    as->dispatch = d;
    memory_listener_register(&d->listener, as);
}

A
Avi Kivity 已提交
1799 1800 1801 1802 1803 1804 1805 1806 1807 1808
void address_space_destroy_dispatch(AddressSpace *as)
{
    AddressSpaceDispatch *d = as->dispatch;

    memory_listener_unregister(&d->listener);
    destroy_l2_mapping(&d->phys_map, P_L2_LEVELS - 1);
    g_free(d);
    as->dispatch = NULL;
}

A
Avi Kivity 已提交
1809 1810
static void memory_map_init(void)
{
1811
    system_memory = g_malloc(sizeof(*system_memory));
A
Avi Kivity 已提交
1812
    memory_region_init(system_memory, "system", INT64_MAX);
1813 1814
    address_space_init(&address_space_memory, system_memory);
    address_space_memory.name = "memory";
1815

1816
    system_io = g_malloc(sizeof(*system_io));
1817
    memory_region_init(system_io, "io", 65536);
1818 1819
    address_space_init(&address_space_io, system_io);
    address_space_io.name = "I/O";
1820

1821 1822 1823
    memory_listener_register(&core_memory_listener, &address_space_memory);
    memory_listener_register(&io_memory_listener, &address_space_io);
    memory_listener_register(&tcg_memory_listener, &address_space_memory);
1824 1825 1826

    dma_context_init(&dma_context_memory, &address_space_memory,
                     NULL, NULL, NULL);
A
Avi Kivity 已提交
1827 1828 1829 1830 1831 1832 1833
}

MemoryRegion *get_system_memory(void)
{
    return system_memory;
}

1834 1835 1836 1837 1838
MemoryRegion *get_system_io(void)
{
    return system_io;
}

1839 1840
#endif /* !defined(CONFIG_USER_ONLY) */

B
bellard 已提交
1841 1842
/* physical memory access (slow version, mainly for debug) */
#if defined(CONFIG_USER_ONLY)
1843
int cpu_memory_rw_debug(CPUArchState *env, target_ulong addr,
P
Paul Brook 已提交
1844
                        uint8_t *buf, int len, int is_write)
B
bellard 已提交
1845 1846 1847
{
    int l, flags;
    target_ulong page;
1848
    void * p;
B
bellard 已提交
1849 1850 1851 1852 1853 1854 1855 1856

    while (len > 0) {
        page = addr & TARGET_PAGE_MASK;
        l = (page + TARGET_PAGE_SIZE) - addr;
        if (l > len)
            l = len;
        flags = page_get_flags(page);
        if (!(flags & PAGE_VALID))
P
Paul Brook 已提交
1857
            return -1;
B
bellard 已提交
1858 1859
        if (is_write) {
            if (!(flags & PAGE_WRITE))
P
Paul Brook 已提交
1860
                return -1;
1861
            /* XXX: this code should not depend on lock_user */
A
aurel32 已提交
1862
            if (!(p = lock_user(VERIFY_WRITE, addr, l, 0)))
P
Paul Brook 已提交
1863
                return -1;
A
aurel32 已提交
1864 1865
            memcpy(p, buf, l);
            unlock_user(p, addr, l);
B
bellard 已提交
1866 1867
        } else {
            if (!(flags & PAGE_READ))
P
Paul Brook 已提交
1868
                return -1;
1869
            /* XXX: this code should not depend on lock_user */
A
aurel32 已提交
1870
            if (!(p = lock_user(VERIFY_READ, addr, l, 1)))
P
Paul Brook 已提交
1871
                return -1;
A
aurel32 已提交
1872
            memcpy(buf, p, l);
A
aurel32 已提交
1873
            unlock_user(p, addr, 0);
B
bellard 已提交
1874 1875 1876 1877 1878
        }
        len -= l;
        buf += l;
        addr += l;
    }
P
Paul Brook 已提交
1879
    return 0;
B
bellard 已提交
1880
}
B
bellard 已提交
1881

B
bellard 已提交
1882
#else
1883

A
Avi Kivity 已提交
1884 1885
static void invalidate_and_set_dirty(hwaddr addr,
                                     hwaddr length)
1886 1887 1888 1889 1890 1891 1892
{
    if (!cpu_physical_memory_is_dirty(addr)) {
        /* invalidate code */
        tb_invalidate_phys_page_range(addr, addr + length, 0);
        /* set dirty bit */
        cpu_physical_memory_set_dirty_flags(addr, (0xff & ~CODE_DIRTY_FLAG));
    }
1893
    xen_modified_memory(addr, length);
1894 1895
}

1896 1897 1898 1899 1900 1901 1902 1903 1904 1905 1906 1907
static inline bool memory_access_is_direct(MemoryRegion *mr, bool is_write)
{
    if (memory_region_is_ram(mr)) {
        return !(is_write && mr->readonly);
    }
    if (memory_region_is_romd(mr)) {
        return !is_write;
    }

    return false;
}

1908
static inline int memory_access_size(MemoryRegion *mr, int l, hwaddr addr)
1909
{
1910
    if (l >= 4 && (((addr & 3) == 0 || mr->ops->impl.unaligned))) {
1911 1912
        return 4;
    }
1913
    if (l >= 2 && (((addr & 1) == 0) || mr->ops->impl.unaligned)) {
1914 1915 1916 1917 1918
        return 2;
    }
    return 1;
}

1919
bool address_space_rw(AddressSpace *as, hwaddr addr, uint8_t *buf,
A
Avi Kivity 已提交
1920
                      int len, bool is_write)
B
bellard 已提交
1921
{
1922
    hwaddr l;
B
bellard 已提交
1923
    uint8_t *ptr;
1924
    uint64_t val;
1925
    hwaddr addr1;
1926
    MemoryRegionSection *section;
1927
    bool error = false;
1928

B
bellard 已提交
1929
    while (len > 0) {
1930 1931
        l = len;
        section = address_space_translate(as, addr, &addr1, &l, is_write);
1932

B
bellard 已提交
1933
        if (is_write) {
1934
            if (!memory_access_is_direct(section->mr, is_write)) {
1935
                l = memory_access_size(section->mr, l, addr1);
B
bellard 已提交
1936 1937
                /* XXX: could force cpu_single_env to NULL to avoid
                   potential bugs */
1938
                if (l == 4) {
B
bellard 已提交
1939
                    /* 32 bit write access */
B
bellard 已提交
1940
                    val = ldl_p(buf);
1941
                    error |= io_mem_write(section->mr, addr1, val, 4);
1942
                } else if (l == 2) {
B
bellard 已提交
1943
                    /* 16 bit write access */
B
bellard 已提交
1944
                    val = lduw_p(buf);
1945
                    error |= io_mem_write(section->mr, addr1, val, 2);
B
bellard 已提交
1946
                } else {
B
bellard 已提交
1947
                    /* 8 bit write access */
B
bellard 已提交
1948
                    val = ldub_p(buf);
1949
                    error |= io_mem_write(section->mr, addr1, val, 1);
B
bellard 已提交
1950
                }
1951
            } else {
1952
                addr1 += memory_region_get_ram_addr(section->mr);
B
bellard 已提交
1953
                /* RAM case */
P
pbrook 已提交
1954
                ptr = qemu_get_ram_ptr(addr1);
B
bellard 已提交
1955
                memcpy(ptr, buf, l);
1956
                invalidate_and_set_dirty(addr1, l);
B
bellard 已提交
1957 1958
            }
        } else {
1959
            if (!memory_access_is_direct(section->mr, is_write)) {
B
bellard 已提交
1960
                /* I/O case */
1961
                l = memory_access_size(section->mr, l, addr1);
1962
                if (l == 4) {
B
bellard 已提交
1963
                    /* 32 bit read access */
1964
                    error |= io_mem_read(section->mr, addr1, &val, 4);
B
bellard 已提交
1965
                    stl_p(buf, val);
1966
                } else if (l == 2) {
B
bellard 已提交
1967
                    /* 16 bit read access */
1968
                    error |= io_mem_read(section->mr, addr1, &val, 2);
B
bellard 已提交
1969
                    stw_p(buf, val);
B
bellard 已提交
1970
                } else {
B
bellard 已提交
1971
                    /* 8 bit read access */
1972
                    error |= io_mem_read(section->mr, addr1, &val, 1);
B
bellard 已提交
1973
                    stb_p(buf, val);
B
bellard 已提交
1974 1975 1976
                }
            } else {
                /* RAM case */
1977
                ptr = qemu_get_ram_ptr(section->mr->ram_addr + addr1);
1978
                memcpy(buf, ptr, l);
B
bellard 已提交
1979 1980 1981 1982 1983 1984
            }
        }
        len -= l;
        buf += l;
        addr += l;
    }
1985 1986

    return error;
B
bellard 已提交
1987
}
B
bellard 已提交
1988

1989
bool address_space_write(AddressSpace *as, hwaddr addr,
A
Avi Kivity 已提交
1990 1991
                         const uint8_t *buf, int len)
{
1992
    return address_space_rw(as, addr, (uint8_t *)buf, len, true);
A
Avi Kivity 已提交
1993 1994
}

1995
bool address_space_read(AddressSpace *as, hwaddr addr, uint8_t *buf, int len)
A
Avi Kivity 已提交
1996
{
1997
    return address_space_rw(as, addr, buf, len, false);
A
Avi Kivity 已提交
1998 1999 2000
}


A
Avi Kivity 已提交
2001
void cpu_physical_memory_rw(hwaddr addr, uint8_t *buf,
A
Avi Kivity 已提交
2002 2003
                            int len, int is_write)
{
2004
    address_space_rw(&address_space_memory, addr, buf, len, is_write);
A
Avi Kivity 已提交
2005 2006
}

B
bellard 已提交
2007
/* used for ROM loading : can write in RAM and ROM */
A
Avi Kivity 已提交
2008
void cpu_physical_memory_write_rom(hwaddr addr,
B
bellard 已提交
2009 2010
                                   const uint8_t *buf, int len)
{
2011
    hwaddr l;
B
bellard 已提交
2012
    uint8_t *ptr;
2013
    hwaddr addr1;
2014
    MemoryRegionSection *section;
2015

B
bellard 已提交
2016
    while (len > 0) {
2017 2018 2019
        l = len;
        section = address_space_translate(&address_space_memory,
                                          addr, &addr1, &l, true);
2020

2021 2022
        if (!(memory_region_is_ram(section->mr) ||
              memory_region_is_romd(section->mr))) {
B
bellard 已提交
2023 2024
            /* do nothing */
        } else {
2025
            addr1 += memory_region_get_ram_addr(section->mr);
B
bellard 已提交
2026
            /* ROM/RAM case */
P
pbrook 已提交
2027
            ptr = qemu_get_ram_ptr(addr1);
B
bellard 已提交
2028
            memcpy(ptr, buf, l);
2029
            invalidate_and_set_dirty(addr1, l);
B
bellard 已提交
2030 2031 2032 2033 2034 2035 2036
        }
        len -= l;
        buf += l;
        addr += l;
    }
}

2037 2038
typedef struct {
    void *buffer;
A
Avi Kivity 已提交
2039 2040
    hwaddr addr;
    hwaddr len;
2041 2042 2043 2044
} BounceBuffer;

static BounceBuffer bounce;

2045 2046 2047
typedef struct MapClient {
    void *opaque;
    void (*callback)(void *opaque);
B
Blue Swirl 已提交
2048
    QLIST_ENTRY(MapClient) link;
2049 2050
} MapClient;

B
Blue Swirl 已提交
2051 2052
static QLIST_HEAD(map_client_list, MapClient) map_client_list
    = QLIST_HEAD_INITIALIZER(map_client_list);
2053 2054 2055

void *cpu_register_map_client(void *opaque, void (*callback)(void *opaque))
{
2056
    MapClient *client = g_malloc(sizeof(*client));
2057 2058 2059

    client->opaque = opaque;
    client->callback = callback;
B
Blue Swirl 已提交
2060
    QLIST_INSERT_HEAD(&map_client_list, client, link);
2061 2062 2063
    return client;
}

B
Blue Swirl 已提交
2064
static void cpu_unregister_map_client(void *_client)
2065 2066 2067
{
    MapClient *client = (MapClient *)_client;

B
Blue Swirl 已提交
2068
    QLIST_REMOVE(client, link);
2069
    g_free(client);
2070 2071 2072 2073 2074 2075
}

static void cpu_notify_map_clients(void)
{
    MapClient *client;

B
Blue Swirl 已提交
2076 2077
    while (!QLIST_EMPTY(&map_client_list)) {
        client = QLIST_FIRST(&map_client_list);
2078
        client->callback(client->opaque);
2079
        cpu_unregister_map_client(client);
2080 2081 2082
    }
}

2083 2084 2085 2086 2087 2088 2089 2090 2091
bool address_space_access_valid(AddressSpace *as, hwaddr addr, int len, bool is_write)
{
    MemoryRegionSection *section;
    hwaddr l, xlat;

    while (len > 0) {
        l = len;
        section = address_space_translate(as, addr, &xlat, &l, is_write);
        if (!memory_access_is_direct(section->mr, is_write)) {
2092
            l = memory_access_size(section->mr, l, addr);
2093 2094 2095 2096 2097 2098 2099 2100 2101 2102 2103
            if (!memory_region_access_valid(section->mr, xlat, l, is_write)) {
                return false;
            }
        }

        len -= l;
        addr += l;
    }
    return true;
}

2104 2105 2106 2107
/* Map a physical memory region into a host virtual address.
 * May map a subset of the requested range, given by and returned in *plen.
 * May return NULL if resources needed to perform the mapping are exhausted.
 * Use only for reads OR writes - not for read-modify-write operations.
2108 2109
 * Use cpu_register_map_client() to know when retrying the map operation is
 * likely to succeed.
2110
 */
A
Avi Kivity 已提交
2111
void *address_space_map(AddressSpace *as,
A
Avi Kivity 已提交
2112 2113
                        hwaddr addr,
                        hwaddr *plen,
A
Avi Kivity 已提交
2114
                        bool is_write)
2115
{
A
Avi Kivity 已提交
2116 2117
    hwaddr len = *plen;
    hwaddr todo = 0;
2118
    hwaddr l, xlat;
2119
    MemoryRegionSection *section;
2120
    ram_addr_t raddr = RAM_ADDR_MAX;
2121 2122
    ram_addr_t rlen;
    void *ret;
2123 2124

    while (len > 0) {
2125 2126
        l = len;
        section = address_space_translate(as, addr, &xlat, &l, is_write);
2127

2128
        if (!memory_access_is_direct(section->mr, is_write)) {
2129
            if (todo || bounce.buffer) {
2130 2131 2132 2133 2134 2135
                break;
            }
            bounce.buffer = qemu_memalign(TARGET_PAGE_SIZE, TARGET_PAGE_SIZE);
            bounce.addr = addr;
            bounce.len = l;
            if (!is_write) {
A
Avi Kivity 已提交
2136
                address_space_read(as, addr, bounce.buffer, l);
2137
            }
2138 2139 2140

            *plen = l;
            return bounce.buffer;
2141
        }
2142
        if (!todo) {
2143 2144 2145 2146 2147
            raddr = memory_region_get_ram_addr(section->mr) + xlat;
        } else {
            if (memory_region_get_ram_addr(section->mr) + xlat != raddr + todo) {
                break;
            }
2148
        }
2149 2150 2151

        len -= l;
        addr += l;
2152
        todo += l;
2153
    }
2154 2155 2156 2157
    rlen = todo;
    ret = qemu_ram_ptr_length(raddr, &rlen);
    *plen = rlen;
    return ret;
2158 2159
}

A
Avi Kivity 已提交
2160
/* Unmaps a memory region previously mapped by address_space_map().
2161 2162 2163
 * Will also mark the memory as dirty if is_write == 1.  access_len gives
 * the amount of memory that was actually read or written by the caller.
 */
A
Avi Kivity 已提交
2164 2165
void address_space_unmap(AddressSpace *as, void *buffer, hwaddr len,
                         int is_write, hwaddr access_len)
2166 2167 2168
{
    if (buffer != bounce.buffer) {
        if (is_write) {
M
Marcelo Tosatti 已提交
2169
            ram_addr_t addr1 = qemu_ram_addr_from_host_nofail(buffer);
2170 2171 2172 2173 2174
            while (access_len) {
                unsigned l;
                l = TARGET_PAGE_SIZE;
                if (l > access_len)
                    l = access_len;
2175
                invalidate_and_set_dirty(addr1, l);
2176 2177 2178 2179
                addr1 += l;
                access_len -= l;
            }
        }
2180
        if (xen_enabled()) {
J
Jan Kiszka 已提交
2181
            xen_invalidate_map_cache_entry(buffer);
A
Anthony PERARD 已提交
2182
        }
2183 2184 2185
        return;
    }
    if (is_write) {
A
Avi Kivity 已提交
2186
        address_space_write(as, bounce.addr, bounce.buffer, access_len);
2187
    }
2188
    qemu_vfree(bounce.buffer);
2189
    bounce.buffer = NULL;
2190
    cpu_notify_map_clients();
2191
}
B
bellard 已提交
2192

A
Avi Kivity 已提交
2193 2194
void *cpu_physical_memory_map(hwaddr addr,
                              hwaddr *plen,
A
Avi Kivity 已提交
2195 2196 2197 2198 2199
                              int is_write)
{
    return address_space_map(&address_space_memory, addr, plen, is_write);
}

A
Avi Kivity 已提交
2200 2201
void cpu_physical_memory_unmap(void *buffer, hwaddr len,
                               int is_write, hwaddr access_len)
A
Avi Kivity 已提交
2202 2203 2204 2205
{
    return address_space_unmap(&address_space_memory, buffer, len, is_write, access_len);
}

B
bellard 已提交
2206
/* warning: addr must be aligned */
A
Avi Kivity 已提交
2207
static inline uint32_t ldl_phys_internal(hwaddr addr,
2208
                                         enum device_endian endian)
B
bellard 已提交
2209 2210
{
    uint8_t *ptr;
2211
    uint64_t val;
2212
    MemoryRegionSection *section;
2213 2214
    hwaddr l = 4;
    hwaddr addr1;
B
bellard 已提交
2215

2216 2217
    section = address_space_translate(&address_space_memory, addr, &addr1, &l,
                                      false);
2218
    if (l < 4 || !memory_access_is_direct(section->mr, false)) {
B
bellard 已提交
2219
        /* I/O case */
2220
        io_mem_read(section->mr, addr1, &val, 4);
2221 2222 2223 2224 2225 2226 2227 2228 2229
#if defined(TARGET_WORDS_BIGENDIAN)
        if (endian == DEVICE_LITTLE_ENDIAN) {
            val = bswap32(val);
        }
#else
        if (endian == DEVICE_BIG_ENDIAN) {
            val = bswap32(val);
        }
#endif
B
bellard 已提交
2230 2231
    } else {
        /* RAM case */
2232
        ptr = qemu_get_ram_ptr((memory_region_get_ram_addr(section->mr)
2233
                                & TARGET_PAGE_MASK)
2234
                               + addr1);
2235 2236 2237 2238 2239 2240 2241 2242 2243 2244 2245
        switch (endian) {
        case DEVICE_LITTLE_ENDIAN:
            val = ldl_le_p(ptr);
            break;
        case DEVICE_BIG_ENDIAN:
            val = ldl_be_p(ptr);
            break;
        default:
            val = ldl_p(ptr);
            break;
        }
B
bellard 已提交
2246 2247 2248 2249
    }
    return val;
}

A
Avi Kivity 已提交
2250
uint32_t ldl_phys(hwaddr addr)
2251 2252 2253 2254
{
    return ldl_phys_internal(addr, DEVICE_NATIVE_ENDIAN);
}

A
Avi Kivity 已提交
2255
uint32_t ldl_le_phys(hwaddr addr)
2256 2257 2258 2259
{
    return ldl_phys_internal(addr, DEVICE_LITTLE_ENDIAN);
}

A
Avi Kivity 已提交
2260
uint32_t ldl_be_phys(hwaddr addr)
2261 2262 2263 2264
{
    return ldl_phys_internal(addr, DEVICE_BIG_ENDIAN);
}

B
bellard 已提交
2265
/* warning: addr must be aligned */
A
Avi Kivity 已提交
2266
static inline uint64_t ldq_phys_internal(hwaddr addr,
2267
                                         enum device_endian endian)
B
bellard 已提交
2268 2269 2270
{
    uint8_t *ptr;
    uint64_t val;
2271
    MemoryRegionSection *section;
2272 2273
    hwaddr l = 8;
    hwaddr addr1;
B
bellard 已提交
2274

2275 2276
    section = address_space_translate(&address_space_memory, addr, &addr1, &l,
                                      false);
2277
    if (l < 8 || !memory_access_is_direct(section->mr, false)) {
B
bellard 已提交
2278
        /* I/O case */
2279
        io_mem_read(section->mr, addr1, &val, 8);
2280 2281 2282 2283 2284 2285 2286 2287
#if defined(TARGET_WORDS_BIGENDIAN)
        if (endian == DEVICE_LITTLE_ENDIAN) {
            val = bswap64(val);
        }
#else
        if (endian == DEVICE_BIG_ENDIAN) {
            val = bswap64(val);
        }
B
bellard 已提交
2288 2289 2290
#endif
    } else {
        /* RAM case */
2291
        ptr = qemu_get_ram_ptr((memory_region_get_ram_addr(section->mr)
2292
                                & TARGET_PAGE_MASK)
2293
                               + addr1);
2294 2295 2296 2297 2298 2299 2300 2301 2302 2303 2304
        switch (endian) {
        case DEVICE_LITTLE_ENDIAN:
            val = ldq_le_p(ptr);
            break;
        case DEVICE_BIG_ENDIAN:
            val = ldq_be_p(ptr);
            break;
        default:
            val = ldq_p(ptr);
            break;
        }
B
bellard 已提交
2305 2306 2307 2308
    }
    return val;
}

A
Avi Kivity 已提交
2309
uint64_t ldq_phys(hwaddr addr)
2310 2311 2312 2313
{
    return ldq_phys_internal(addr, DEVICE_NATIVE_ENDIAN);
}

A
Avi Kivity 已提交
2314
uint64_t ldq_le_phys(hwaddr addr)
2315 2316 2317 2318
{
    return ldq_phys_internal(addr, DEVICE_LITTLE_ENDIAN);
}

A
Avi Kivity 已提交
2319
uint64_t ldq_be_phys(hwaddr addr)
2320 2321 2322 2323
{
    return ldq_phys_internal(addr, DEVICE_BIG_ENDIAN);
}

B
bellard 已提交
2324
/* XXX: optimize */
A
Avi Kivity 已提交
2325
uint32_t ldub_phys(hwaddr addr)
B
bellard 已提交
2326 2327 2328 2329 2330 2331
{
    uint8_t val;
    cpu_physical_memory_read(addr, &val, 1);
    return val;
}

2332
/* warning: addr must be aligned */
A
Avi Kivity 已提交
2333
static inline uint32_t lduw_phys_internal(hwaddr addr,
2334
                                          enum device_endian endian)
B
bellard 已提交
2335
{
2336 2337
    uint8_t *ptr;
    uint64_t val;
2338
    MemoryRegionSection *section;
2339 2340
    hwaddr l = 2;
    hwaddr addr1;
2341

2342 2343
    section = address_space_translate(&address_space_memory, addr, &addr1, &l,
                                      false);
2344
    if (l < 2 || !memory_access_is_direct(section->mr, false)) {
2345
        /* I/O case */
2346
        io_mem_read(section->mr, addr1, &val, 2);
2347 2348 2349 2350 2351 2352 2353 2354 2355
#if defined(TARGET_WORDS_BIGENDIAN)
        if (endian == DEVICE_LITTLE_ENDIAN) {
            val = bswap16(val);
        }
#else
        if (endian == DEVICE_BIG_ENDIAN) {
            val = bswap16(val);
        }
#endif
2356 2357
    } else {
        /* RAM case */
2358
        ptr = qemu_get_ram_ptr((memory_region_get_ram_addr(section->mr)
2359
                                & TARGET_PAGE_MASK)
2360
                               + addr1);
2361 2362 2363 2364 2365 2366 2367 2368 2369 2370 2371
        switch (endian) {
        case DEVICE_LITTLE_ENDIAN:
            val = lduw_le_p(ptr);
            break;
        case DEVICE_BIG_ENDIAN:
            val = lduw_be_p(ptr);
            break;
        default:
            val = lduw_p(ptr);
            break;
        }
2372 2373
    }
    return val;
B
bellard 已提交
2374 2375
}

A
Avi Kivity 已提交
2376
uint32_t lduw_phys(hwaddr addr)
2377 2378 2379 2380
{
    return lduw_phys_internal(addr, DEVICE_NATIVE_ENDIAN);
}

A
Avi Kivity 已提交
2381
uint32_t lduw_le_phys(hwaddr addr)
2382 2383 2384 2385
{
    return lduw_phys_internal(addr, DEVICE_LITTLE_ENDIAN);
}

A
Avi Kivity 已提交
2386
uint32_t lduw_be_phys(hwaddr addr)
2387 2388 2389 2390
{
    return lduw_phys_internal(addr, DEVICE_BIG_ENDIAN);
}

B
bellard 已提交
2391 2392 2393
/* warning: addr must be aligned. The ram page is not masked as dirty
   and the code inside is not invalidated. It is useful if the dirty
   bits are used to track modified PTEs */
A
Avi Kivity 已提交
2394
void stl_phys_notdirty(hwaddr addr, uint32_t val)
B
bellard 已提交
2395 2396
{
    uint8_t *ptr;
2397
    MemoryRegionSection *section;
2398 2399
    hwaddr l = 4;
    hwaddr addr1;
B
bellard 已提交
2400

2401 2402
    section = address_space_translate(&address_space_memory, addr, &addr1, &l,
                                      true);
2403
    if (l < 4 || !memory_access_is_direct(section->mr, true)) {
2404
        io_mem_write(section->mr, addr1, val, 4);
B
bellard 已提交
2405
    } else {
2406
        addr1 += memory_region_get_ram_addr(section->mr) & TARGET_PAGE_MASK;
P
pbrook 已提交
2407
        ptr = qemu_get_ram_ptr(addr1);
B
bellard 已提交
2408
        stl_p(ptr, val);
A
aliguori 已提交
2409 2410 2411 2412 2413 2414

        if (unlikely(in_migration)) {
            if (!cpu_physical_memory_is_dirty(addr1)) {
                /* invalidate code */
                tb_invalidate_phys_page_range(addr1, addr1 + 4, 0);
                /* set dirty bit */
2415 2416
                cpu_physical_memory_set_dirty_flags(
                    addr1, (0xff & ~CODE_DIRTY_FLAG));
A
aliguori 已提交
2417 2418
            }
        }
B
bellard 已提交
2419 2420 2421 2422
    }
}

/* warning: addr must be aligned */
A
Avi Kivity 已提交
2423
static inline void stl_phys_internal(hwaddr addr, uint32_t val,
2424
                                     enum device_endian endian)
B
bellard 已提交
2425 2426
{
    uint8_t *ptr;
2427
    MemoryRegionSection *section;
2428 2429
    hwaddr l = 4;
    hwaddr addr1;
B
bellard 已提交
2430

2431 2432
    section = address_space_translate(&address_space_memory, addr, &addr1, &l,
                                      true);
2433
    if (l < 4 || !memory_access_is_direct(section->mr, true)) {
2434 2435 2436 2437 2438 2439 2440 2441 2442
#if defined(TARGET_WORDS_BIGENDIAN)
        if (endian == DEVICE_LITTLE_ENDIAN) {
            val = bswap32(val);
        }
#else
        if (endian == DEVICE_BIG_ENDIAN) {
            val = bswap32(val);
        }
#endif
2443
        io_mem_write(section->mr, addr1, val, 4);
B
bellard 已提交
2444 2445
    } else {
        /* RAM case */
2446
        addr1 += memory_region_get_ram_addr(section->mr) & TARGET_PAGE_MASK;
P
pbrook 已提交
2447
        ptr = qemu_get_ram_ptr(addr1);
2448 2449 2450 2451 2452 2453 2454 2455 2456 2457 2458
        switch (endian) {
        case DEVICE_LITTLE_ENDIAN:
            stl_le_p(ptr, val);
            break;
        case DEVICE_BIG_ENDIAN:
            stl_be_p(ptr, val);
            break;
        default:
            stl_p(ptr, val);
            break;
        }
2459
        invalidate_and_set_dirty(addr1, 4);
B
bellard 已提交
2460 2461 2462
    }
}

A
Avi Kivity 已提交
2463
void stl_phys(hwaddr addr, uint32_t val)
2464 2465 2466 2467
{
    stl_phys_internal(addr, val, DEVICE_NATIVE_ENDIAN);
}

A
Avi Kivity 已提交
2468
void stl_le_phys(hwaddr addr, uint32_t val)
2469 2470 2471 2472
{
    stl_phys_internal(addr, val, DEVICE_LITTLE_ENDIAN);
}

A
Avi Kivity 已提交
2473
void stl_be_phys(hwaddr addr, uint32_t val)
2474 2475 2476 2477
{
    stl_phys_internal(addr, val, DEVICE_BIG_ENDIAN);
}

B
bellard 已提交
2478
/* XXX: optimize */
A
Avi Kivity 已提交
2479
void stb_phys(hwaddr addr, uint32_t val)
B
bellard 已提交
2480 2481 2482 2483 2484
{
    uint8_t v = val;
    cpu_physical_memory_write(addr, &v, 1);
}

2485
/* warning: addr must be aligned */
A
Avi Kivity 已提交
2486
static inline void stw_phys_internal(hwaddr addr, uint32_t val,
2487
                                     enum device_endian endian)
B
bellard 已提交
2488
{
2489
    uint8_t *ptr;
2490
    MemoryRegionSection *section;
2491 2492
    hwaddr l = 2;
    hwaddr addr1;
2493

2494 2495
    section = address_space_translate(&address_space_memory, addr, &addr1, &l,
                                      true);
2496
    if (l < 2 || !memory_access_is_direct(section->mr, true)) {
2497 2498 2499 2500 2501 2502 2503 2504 2505
#if defined(TARGET_WORDS_BIGENDIAN)
        if (endian == DEVICE_LITTLE_ENDIAN) {
            val = bswap16(val);
        }
#else
        if (endian == DEVICE_BIG_ENDIAN) {
            val = bswap16(val);
        }
#endif
2506
        io_mem_write(section->mr, addr1, val, 2);
2507 2508
    } else {
        /* RAM case */
2509
        addr1 += memory_region_get_ram_addr(section->mr) & TARGET_PAGE_MASK;
2510
        ptr = qemu_get_ram_ptr(addr1);
2511 2512 2513 2514 2515 2516 2517 2518 2519 2520 2521
        switch (endian) {
        case DEVICE_LITTLE_ENDIAN:
            stw_le_p(ptr, val);
            break;
        case DEVICE_BIG_ENDIAN:
            stw_be_p(ptr, val);
            break;
        default:
            stw_p(ptr, val);
            break;
        }
2522
        invalidate_and_set_dirty(addr1, 2);
2523
    }
B
bellard 已提交
2524 2525
}

A
Avi Kivity 已提交
2526
void stw_phys(hwaddr addr, uint32_t val)
2527 2528 2529 2530
{
    stw_phys_internal(addr, val, DEVICE_NATIVE_ENDIAN);
}

A
Avi Kivity 已提交
2531
void stw_le_phys(hwaddr addr, uint32_t val)
2532 2533 2534 2535
{
    stw_phys_internal(addr, val, DEVICE_LITTLE_ENDIAN);
}

A
Avi Kivity 已提交
2536
void stw_be_phys(hwaddr addr, uint32_t val)
2537 2538 2539 2540
{
    stw_phys_internal(addr, val, DEVICE_BIG_ENDIAN);
}

B
bellard 已提交
2541
/* XXX: optimize */
A
Avi Kivity 已提交
2542
void stq_phys(hwaddr addr, uint64_t val)
B
bellard 已提交
2543 2544
{
    val = tswap64(val);
2545
    cpu_physical_memory_write(addr, &val, 8);
B
bellard 已提交
2546 2547
}

A
Avi Kivity 已提交
2548
void stq_le_phys(hwaddr addr, uint64_t val)
2549 2550 2551 2552 2553
{
    val = cpu_to_le64(val);
    cpu_physical_memory_write(addr, &val, 8);
}

A
Avi Kivity 已提交
2554
void stq_be_phys(hwaddr addr, uint64_t val)
2555 2556 2557 2558 2559
{
    val = cpu_to_be64(val);
    cpu_physical_memory_write(addr, &val, 8);
}

2560
/* virtual memory access for debug (includes writing to ROM) */
2561
int cpu_memory_rw_debug(CPUArchState *env, target_ulong addr,
2562
                        uint8_t *buf, int len, int is_write)
B
bellard 已提交
2563 2564
{
    int l;
A
Avi Kivity 已提交
2565
    hwaddr phys_addr;
2566
    target_ulong page;
B
bellard 已提交
2567 2568 2569 2570 2571 2572 2573 2574 2575 2576

    while (len > 0) {
        page = addr & TARGET_PAGE_MASK;
        phys_addr = cpu_get_phys_page_debug(env, page);
        /* if no physical page mapped, return an error */
        if (phys_addr == -1)
            return -1;
        l = (page + TARGET_PAGE_SIZE) - addr;
        if (l > len)
            l = len;
2577 2578 2579 2580 2581
        phys_addr += (addr & ~TARGET_PAGE_MASK);
        if (is_write)
            cpu_physical_memory_write_rom(phys_addr, buf, l);
        else
            cpu_physical_memory_rw(phys_addr, buf, l, is_write);
B
bellard 已提交
2582 2583 2584 2585 2586 2587
        len -= l;
        buf += l;
        addr += l;
    }
    return 0;
}
P
Paul Brook 已提交
2588
#endif
B
bellard 已提交
2589

2590 2591 2592 2593 2594 2595 2596 2597 2598 2599 2600 2601 2602 2603 2604 2605 2606 2607
#if !defined(CONFIG_USER_ONLY)

/*
 * A helper function for the _utterly broken_ virtio device model to find out if
 * it's running on a big endian machine. Don't do this at home kids!
 */
bool virtio_is_big_endian(void);
bool virtio_is_big_endian(void)
{
#if defined(TARGET_WORDS_BIGENDIAN)
    return true;
#else
    return false;
#endif
}

#endif

2608
#ifndef CONFIG_USER_ONLY
A
Avi Kivity 已提交
2609
bool cpu_physical_memory_is_io(hwaddr phys_addr)
2610 2611
{
    MemoryRegionSection *section;
2612
    hwaddr l = 1;
2613

2614 2615
    section = address_space_translate(&address_space_memory,
                                      phys_addr, &phys_addr, &l, false);
2616 2617 2618 2619 2620

    return !(memory_region_is_ram(section->mr) ||
             memory_region_is_romd(section->mr));
}
#endif