qemu_migration.c 187.9 KB
Newer Older
1 2 3
/*
 * qemu_migration.c: QEMU migration handling
 *
4
 * Copyright (C) 2006-2015 Red Hat, Inc.
5 6 7 8 9 10 11 12 13 14 15 16
 *
 * This library is free software; you can redistribute it and/or
 * modify it under the terms of the GNU Lesser General Public
 * License as published by the Free Software Foundation; either
 * version 2.1 of the License, or (at your option) any later version.
 *
 * This library is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 * Lesser General Public License for more details.
 *
 * You should have received a copy of the GNU Lesser General Public
17
 * License along with this library.  If not, see
O
Osier Yang 已提交
18
 * <http://www.gnu.org/licenses/>.
19 20 21 22 23
 *
 */

#include <config.h>

J
Ján Tomko 已提交
24 25
#include <netdb.h>
#include <sys/socket.h>
26
#include <sys/time.h>
27
#include <fcntl.h>
28
#include <poll.h>
29 30

#include "qemu_migration.h"
31
#include "qemu_migration_cookie.h"
32
#include "qemu_migration_params.h"
33 34 35 36
#include "qemu_monitor.h"
#include "qemu_domain.h"
#include "qemu_process.h"
#include "qemu_capabilities.h"
37
#include "qemu_alias.h"
38
#include "qemu_cgroup.h"
39
#include "qemu_hotplug.h"
40
#include "qemu_blockjob.h"
41
#include "qemu_security.h"
42
#include "qemu_block.h"
43

44
#include "domain_audit.h"
45
#include "virlog.h"
46
#include "virerror.h"
47
#include "viralloc.h"
E
Eric Blake 已提交
48
#include "virfile.h"
49
#include "virnetdevopenvswitch.h"
50
#include "datatypes.h"
51
#include "virfdstream.h"
52
#include "viruuid.h"
53
#include "virtime.h"
54
#include "locking/domain_lock.h"
55
#include "rpc/virnetsocket.h"
56
#include "virstoragefile.h"
M
Martin Kletzander 已提交
57
#include "viruri.h"
58
#include "virhook.h"
59
#include "virstring.h"
60
#include "virtypedparam.h"
M
Michael R. Hines 已提交
61
#include "virprocess.h"
62
#include "nwfilter_conf.h"
63 64 65

#define VIR_FROM_THIS VIR_FROM_QEMU

66 67
VIR_LOG_INIT("qemu.qemu_migration");

68 69 70 71 72 73 74 75 76 77 78 79 80
VIR_ENUM_IMPL(qemuMigrationJobPhase, QEMU_MIGRATION_PHASE_LAST,
              "none",
              "perform2",
              "begin3",
              "perform3",
              "perform3_done",
              "confirm3_cancelled",
              "confirm3",
              "prepare",
              "finish2",
              "finish3",
);

81 82 83
static int
qemuMigrationJobStart(virQEMUDriverPtr driver,
                      virDomainObjPtr vm,
84 85
                      qemuDomainAsyncJob job,
                      unsigned long apiFlags)
86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112
    ATTRIBUTE_NONNULL(1) ATTRIBUTE_NONNULL(2) ATTRIBUTE_RETURN_CHECK;

static void
qemuMigrationJobSetPhase(virQEMUDriverPtr driver,
                         virDomainObjPtr vm,
                         qemuMigrationJobPhase phase)
    ATTRIBUTE_NONNULL(1) ATTRIBUTE_NONNULL(2);

static void
qemuMigrationJobStartPhase(virQEMUDriverPtr driver,
                           virDomainObjPtr vm,
                           qemuMigrationJobPhase phase)
    ATTRIBUTE_NONNULL(1) ATTRIBUTE_NONNULL(2);

static void
qemuMigrationJobContinue(virDomainObjPtr obj)
    ATTRIBUTE_NONNULL(1);

static bool
qemuMigrationJobIsActive(virDomainObjPtr vm,
                         qemuDomainAsyncJob job)
    ATTRIBUTE_NONNULL(1);

static void
qemuMigrationJobFinish(virQEMUDriverPtr driver,
                       virDomainObjPtr obj)
    ATTRIBUTE_NONNULL(1) ATTRIBUTE_NONNULL(2);
113

114
static void
115
qemuMigrationSrcStoreDomainState(virDomainObjPtr vm)
116 117 118 119 120 121 122 123 124 125
{
    qemuDomainObjPrivatePtr priv = vm->privateData;
    priv->preMigrationState = virDomainObjGetState(vm, NULL);

    VIR_DEBUG("Storing pre-migration state=%d domain=%p",
              priv->preMigrationState, vm);
}

/* Returns true if the domain was resumed, false otherwise */
static bool
126
qemuMigrationSrcRestoreDomainState(virQEMUDriverPtr driver, virDomainObjPtr vm)
127 128
{
    qemuDomainObjPrivatePtr priv = vm->privateData;
129 130
    int reason;
    virDomainState state = virDomainObjGetState(vm, &reason);
131 132
    bool ret = false;

133 134 135 136 137
    VIR_DEBUG("driver=%p, vm=%p, pre-mig-state=%s, state=%s, reason=%s",
              driver, vm,
              virDomainStateTypeToString(priv->preMigrationState),
              virDomainStateTypeToString(state),
              virDomainStateReasonToString(state, reason));
138

139 140 141 142 143
    if (state != VIR_DOMAIN_PAUSED ||
        reason == VIR_DOMAIN_PAUSED_POSTCOPY_FAILED)
        goto cleanup;

    if (priv->preMigrationState == VIR_DOMAIN_RUNNING) {
144 145 146 147 148 149
        /* This is basically the only restore possibility that's safe
         * and we should attempt to do */

        VIR_DEBUG("Restoring pre-migration state due to migration error");

        /* we got here through some sort of failure; start the domain again */
150
        if (qemuProcessStartCPUs(driver, vm,
151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166
                                 VIR_DOMAIN_RUNNING_MIGRATION_CANCELED,
                                 QEMU_ASYNC_JOB_MIGRATION_OUT) < 0) {
            /* Hm, we already know we are in error here.  We don't want to
             * overwrite the previous error, though, so we just throw something
             * to the logs and hope for the best */
            VIR_ERROR(_("Failed to resume guest %s after failure"), vm->def->name);
            goto cleanup;
        }
        ret = true;
    }

 cleanup:
    priv->preMigrationState = VIR_DOMAIN_NOSTATE;
    return ret;
}

167 168

static int
169 170 171
qemuMigrationDstPrecreateDisk(virConnectPtr conn,
                              virDomainDiskDefPtr disk,
                              unsigned long long capacity)
172 173 174 175 176 177 178 179 180 181 182 183
{
    int ret = -1;
    virStoragePoolPtr pool = NULL;
    virStorageVolPtr vol = NULL;
    char *volName = NULL, *basePath = NULL;
    char *volStr = NULL;
    virBuffer buf = VIR_BUFFER_INITIALIZER;
    const char *format = NULL;
    unsigned int flags = 0;

    VIR_DEBUG("Precreate disk type=%s", virStorageTypeToString(disk->src->type));

184
    switch ((virStorageType)disk->src->type) {
185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204
    case VIR_STORAGE_TYPE_FILE:
        if (!virDomainDiskGetSource(disk)) {
            VIR_DEBUG("Dropping sourceless disk '%s'",
                      disk->dst);
            return 0;
        }

        if (VIR_STRDUP(basePath, disk->src->path) < 0)
            goto cleanup;

        if (!(volName = strrchr(basePath, '/'))) {
            virReportError(VIR_ERR_INVALID_ARG,
                           _("malformed disk path: %s"),
                           disk->src->path);
            goto cleanup;
        }

        *volName = '\0';
        volName++;

205
        if (!(pool = virStoragePoolLookupByTargetPath(conn, basePath)))
206 207 208 209 210 211 212 213 214 215 216 217 218 219 220
            goto cleanup;
        format = virStorageFileFormatTypeToString(disk->src->format);
        if (disk->src->format == VIR_STORAGE_FILE_QCOW2)
            flags |= VIR_STORAGE_VOL_CREATE_PREALLOC_METADATA;
        break;

    case VIR_STORAGE_TYPE_VOLUME:
        if (!(pool = virStoragePoolLookupByName(conn, disk->src->srcpool->pool)))
            goto cleanup;
        format = virStorageFileFormatTypeToString(disk->src->format);
        volName = disk->src->srcpool->volume;
        if (disk->src->format == VIR_STORAGE_FILE_QCOW2)
            flags |= VIR_STORAGE_VOL_CREATE_PREALLOC_METADATA;
        break;

221 222 223 224 225
    case VIR_STORAGE_TYPE_NETWORK:
        VIR_DEBUG("Skipping creation of network disk '%s'",
                  disk->dst);
        return 0;

226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272
    case VIR_STORAGE_TYPE_BLOCK:
    case VIR_STORAGE_TYPE_DIR:
    case VIR_STORAGE_TYPE_NONE:
    case VIR_STORAGE_TYPE_LAST:
        virReportError(VIR_ERR_INTERNAL_ERROR,
                       _("cannot precreate storage for disk type '%s'"),
                       virStorageTypeToString(disk->src->type));
        goto cleanup;
    }

    if ((vol = virStorageVolLookupByName(pool, volName))) {
        VIR_DEBUG("Skipping creation of already existing volume of name '%s'",
                  volName);
        ret = 0;
        goto cleanup;
    }

    virBufferAddLit(&buf, "<volume>\n");
    virBufferAdjustIndent(&buf, 2);
    virBufferEscapeString(&buf, "<name>%s</name>\n", volName);
    virBufferAsprintf(&buf, "<capacity>%llu</capacity>\n", capacity);
    virBufferAddLit(&buf, "<target>\n");
    virBufferAdjustIndent(&buf, 2);
    virBufferAsprintf(&buf, "<format type='%s'/>\n", format);
    virBufferAdjustIndent(&buf, -2);
    virBufferAddLit(&buf, "</target>\n");
    virBufferAdjustIndent(&buf, -2);
    virBufferAddLit(&buf, "</volume>\n");

    if (!(volStr = virBufferContentAndReset(&buf))) {
        virReportError(VIR_ERR_INTERNAL_ERROR, "%s",
                       _("unable to create volume XML"));
        goto cleanup;
    }

    if (!(vol = virStorageVolCreateXML(pool, volStr, flags)))
        goto cleanup;

    ret = 0;
 cleanup:
    VIR_FREE(basePath);
    VIR_FREE(volStr);
    virObjectUnref(vol);
    virObjectUnref(pool);
    return ret;
}

273
static bool
274 275
qemuMigrationAnyCopyDisk(virDomainDiskDef const *disk,
                         size_t nmigrate_disks, const char **migrate_disks)
276 277 278 279 280 281 282 283 284 285 286 287 288 289 290
{
    size_t i;

    /* Check if the disk alias is in the list */
    if (nmigrate_disks) {
        for (i = 0; i < nmigrate_disks; i++) {
            if (STREQ(disk->dst, migrate_disks[i]))
                return true;
        }
        return false;
    }

    /* Default is to migrate only non-shared non-readonly disks
     * with source */
    return !disk->src->shared && !disk->src->readonly &&
291
           !virStorageSourceIsEmpty(disk->src);
292 293
}

294 295

static int
296 297 298 299 300
qemuMigrationDstPrecreateStorage(virDomainObjPtr vm,
                                 qemuMigrationCookieNBDPtr nbd,
                                 size_t nmigrate_disks,
                                 const char **migrate_disks,
                                 bool incremental)
301 302 303
{
    int ret = -1;
    size_t i = 0;
304
    virConnectPtr conn;
305 306 307 308

    if (!nbd || !nbd->ndisks)
        return 0;

309 310 311
    if (!(conn = virGetConnectStorage()))
        return -1;

312 313 314 315
    for (i = 0; i < nbd->ndisks; i++) {
        virDomainDiskDefPtr disk;
        const char *diskSrcPath;

316
        VIR_DEBUG("Looking up disk target '%s' (capacity=%llu)",
317 318
                  nbd->disks[i].target, nbd->disks[i].capacity);

319 320
        if (!(disk = virDomainDiskByName(vm->def, nbd->disks[i].target,
                                         false))) {
321 322 323 324 325 326 327 328
            virReportError(VIR_ERR_INTERNAL_ERROR,
                           _("unable to find disk by target: %s"),
                           nbd->disks[i].target);
            goto cleanup;
        }

        diskSrcPath = virDomainDiskGetSource(disk);

329
        /* Skip disks we don't want to migrate and already existing disks. */
330
        if (!qemuMigrationAnyCopyDisk(disk, nmigrate_disks, migrate_disks) ||
331 332 333 334
            (diskSrcPath && virFileExists(diskSrcPath))) {
            continue;
        }

335 336 337 338 339 340 341
        if (incremental) {
            virReportError(VIR_ERR_OPERATION_UNSUPPORTED, "%s",
                           _("pre-creation of storage targets for incremental "
                             "storage migration is not supported"));
            goto cleanup;
        }

342 343
        VIR_DEBUG("Proceeding with disk source %s", NULLSTR(diskSrcPath));

344
        if (qemuMigrationDstPrecreateDisk(conn, disk, nbd->disks[i].capacity) < 0)
345 346 347 348 349
            goto cleanup;
    }

    ret = 0;
 cleanup:
350
    virObjectUnref(conn);
351 352 353 354
    return ret;
}


355
/**
356
 * qemuMigrationDstStartNBDServer:
357 358 359 360 361 362 363 364 365 366 367
 * @driver: qemu driver
 * @vm: domain
 *
 * Starts NBD server. This is a newer method to copy
 * storage during migration than using 'blk' and 'inc'
 * arguments in 'migrate' monitor command.
 * Error is reported here.
 *
 * Returns 0 on success, -1 otherwise.
 */
static int
368 369 370 371 372
qemuMigrationDstStartNBDServer(virQEMUDriverPtr driver,
                               virDomainObjPtr vm,
                               const char *listenAddr,
                               size_t nmigrate_disks,
                               const char **migrate_disks,
373 374
                               int nbdPort,
                               const char *tls_alias)
375 376 377 378 379 380 381
{
    int ret = -1;
    qemuDomainObjPrivatePtr priv = vm->privateData;
    unsigned short port = 0;
    char *diskAlias = NULL;
    size_t i;

382 383 384 385 386 387
    if (nbdPort < 0 || nbdPort > USHRT_MAX) {
        virReportError(VIR_ERR_INVALID_ARG, "%s",
                       _("nbd port must be in range 0-65535"));
        return -1;
    }

388 389 390
    for (i = 0; i < vm->def->ndisks; i++) {
        virDomainDiskDefPtr disk = vm->def->disks[i];

391
        /* check whether disk should be migrated */
392
        if (!qemuMigrationAnyCopyDisk(disk, nmigrate_disks, migrate_disks))
393 394
            continue;

395
        if (disk->src->readonly || virStorageSourceIsEmpty(disk->src)) {
396
            virReportError(VIR_ERR_OPERATION_UNSUPPORTED,
397 398
                           _("Cannot migrate empty or read-only disk %s"),
                           disk->dst);
399 400 401
            goto cleanup;
        }

402
        VIR_FREE(diskAlias);
403
        if (!(diskAlias = qemuAliasDiskDriveFromDisk(disk)))
404 405 406 407 408 409
            goto cleanup;

        if (qemuDomainObjEnterMonitorAsync(driver, vm,
                                           QEMU_ASYNC_JOB_MIGRATION_IN) < 0)
            goto cleanup;

410 411 412 413 414 415
        if (port == 0) {
            if (nbdPort)
                port = nbdPort;
            else if (virPortAllocatorAcquire(driver->migrationPorts, &port) < 0)
                goto exit_monitor;

416
            if (qemuMonitorNBDServerStart(priv->mon, listenAddr, port, tls_alias) < 0)
417
                goto exit_monitor;
418 419
        }

420 421 422
        if (qemuMonitorNBDServerAdd(priv->mon, diskAlias, true) < 0)
            goto exit_monitor;
        if (qemuDomainObjExitMonitor(driver, vm) < 0)
423 424 425 426 427 428
            goto cleanup;
    }

    priv->nbdPort = port;
    ret = 0;

429
 cleanup:
430
    VIR_FREE(diskAlias);
431
    if (ret < 0 && nbdPort == 0)
432
        virPortAllocatorRelease(port);
433
    return ret;
434 435 436 437

 exit_monitor:
    ignore_value(qemuDomainObjExitMonitor(driver, vm));
    goto cleanup;
438 439
}

440 441

static int
442 443 444
qemuMigrationDstStopNBDServer(virQEMUDriverPtr driver,
                              virDomainObjPtr vm,
                              qemuMigrationCookiePtr mig)
445 446 447 448 449 450 451 452 453 454 455 456 457 458 459
{
    qemuDomainObjPrivatePtr priv = vm->privateData;

    if (!mig->nbd)
        return 0;

    if (qemuDomainObjEnterMonitorAsync(driver, vm,
                                       QEMU_ASYNC_JOB_MIGRATION_IN) < 0)
        return -1;

    if (qemuMonitorNBDServerStop(priv->mon) < 0)
        VIR_WARN("Unable to stop NBD server");
    if (qemuDomainObjExitMonitor(driver, vm) < 0)
        return -1;

460
    virPortAllocatorRelease(priv->nbdPort);
461 462 463 464 465
    priv->nbdPort = 0;
    return 0;
}


466
static void
467 468
qemuMigrationNBDReportMirrorError(qemuBlockJobDataPtr job,
                                  const char *diskdst)
469
{
470
    if (job->errmsg) {
471 472
        virReportError(VIR_ERR_OPERATION_FAILED,
                       _("migration of disk %s failed: %s"),
473
                       diskdst, job->errmsg);
474 475
    } else {
        virReportError(VIR_ERR_OPERATION_FAILED,
476
                       _("migration of disk %s failed"), diskdst);
477 478 479 480
    }
}


481
/**
482
 * qemuMigrationSrcNBDStorageCopyReady:
483 484
 * @vm: domain
 *
485 486
 * Check the status of all drives copied via qemuMigrationSrcNBDStorageCopy.
 * Any pending block job events for the mirrored disks will be processed.
487 488 489 490 491 492
 *
 * Returns 1 if all mirrors are "ready",
 *         0 if some mirrors are still performing initial sync,
 *        -1 on error.
 */
static int
493 494
qemuMigrationSrcNBDStorageCopyReady(virDomainObjPtr vm,
                                    qemuDomainAsyncJob asyncJob)
495 496
{
    size_t i;
497 498
    size_t notReady = 0;
    int status;
499 500 501

    for (i = 0; i < vm->def->ndisks; i++) {
        virDomainDiskDefPtr disk = vm->def->disks[i];
502
        qemuDomainDiskPrivatePtr diskPriv = QEMU_DOMAIN_DISK_PRIVATE(disk);
503
        qemuBlockJobDataPtr job;
504

505
        if (!diskPriv->migrating)
506 507
            continue;

508 509 510 511 512 513 514
        if (!(job = qemuBlockJobDiskGetJob(disk))) {
            virReportError(VIR_ERR_INTERNAL_ERROR,
                           _("missing block job data for disk '%s'"), disk->dst);
            return -1;
        }

        status = qemuBlockJobUpdate(vm, job, asyncJob);
515
        if (status == VIR_DOMAIN_BLOCK_JOB_FAILED) {
516 517
            qemuMigrationNBDReportMirrorError(job, disk->dst);
            virObjectUnref(job);
518 519
            return -1;
        }
520

521 522
        virObjectUnref(job);

523 524
        if (disk->mirrorState != VIR_DOMAIN_DISK_MIRROR_STATE_READY)
            notReady++;
525 526
    }

527 528 529 530 531 532 533
    if (notReady) {
        VIR_DEBUG("Waiting for %zu disk mirrors to get ready", notReady);
        return 0;
    } else {
        VIR_DEBUG("All disk mirrors are ready");
        return 1;
    }
534 535 536
}


537 538 539 540
/*
 * If @check is true, the function will report an error and return a different
 * code in case a block job fails. This way we can properly abort migration in
 * case some block jobs failed once all memory has already been transferred.
541
 *
542 543 544 545
 * Returns 1 if all mirrors are gone,
 *         0 if some mirrors are still active,
 *         -1 some mirrors failed but some are still active,
 *         -2 all mirrors are gone but some of them failed.
546 547
 */
static int
548 549 550
qemuMigrationSrcNBDCopyCancelled(virDomainObjPtr vm,
                                 qemuDomainAsyncJob asyncJob,
                                 bool check)
551
{
552 553
    size_t i;
    size_t active = 0;
554
    size_t completed = 0;
555 556
    int status;
    bool failed = false;
557

558
 retry:
559 560 561
    for (i = 0; i < vm->def->ndisks; i++) {
        virDomainDiskDefPtr disk = vm->def->disks[i];
        qemuDomainDiskPrivatePtr diskPriv = QEMU_DOMAIN_DISK_PRIVATE(disk);
562
        qemuBlockJobDataPtr job;
563

564 565
        if (!diskPriv->migrating)
            continue;
566

567 568 569 570
        if (!(job = qemuBlockJobDiskGetJob(disk)))
            continue;

        status = qemuBlockJobUpdate(vm, job, asyncJob);
571 572 573
        switch (status) {
        case VIR_DOMAIN_BLOCK_JOB_FAILED:
            if (check) {
574
                qemuMigrationNBDReportMirrorError(job, disk->dst);
575
                failed = true;
576
            }
M
Marc Hartmayer 已提交
577
            ATTRIBUTE_FALLTHROUGH;
578 579 580 581
        case VIR_DOMAIN_BLOCK_JOB_CANCELED:
        case VIR_DOMAIN_BLOCK_JOB_COMPLETED:
            diskPriv->migrating = false;
            break;
582

583 584
        default:
            active++;
585
        }
586 587 588

        if (status == VIR_DOMAIN_BLOCK_JOB_COMPLETED)
            completed++;
589 590

        virObjectUnref(job);
591 592 593 594 595 596 597 598 599 600
    }

    /* Updating completed block job drops the lock thus we have to recheck
     * block jobs for disks that reside before the disk(s) with completed
     * block job.
     */
    if (completed > 0) {
        completed = 0;
        active = 0;
        goto retry;
601
    }
602

603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618
    if (failed) {
        if (active) {
            VIR_DEBUG("Some disk mirrors failed; still waiting for %zu "
                      "disk mirrors to finish", active);
            return -1;
        } else {
            VIR_DEBUG("All disk mirrors are gone; some of them failed");
            return -2;
        }
    } else {
        if (active) {
            VIR_DEBUG("Waiting for %zu disk mirrors to finish", active);
            return 0;
        } else {
            VIR_DEBUG("All disk mirrors are gone");
            return 1;
619
        }
620
    }
621
}
622 623


624 625 626 627 628 629
/*
 * Returns 0 on success,
 *         1 when job is already completed or it failed and failNoJob is false,
 *         -1 on error or when job failed and failNoJob is true.
 */
static int
630 631 632
qemuMigrationSrcNBDCopyCancelOne(virQEMUDriverPtr driver,
                                 virDomainObjPtr vm,
                                 virDomainDiskDefPtr disk,
633
                                 qemuBlockJobDataPtr job,
634 635
                                 bool failNoJob,
                                 qemuDomainAsyncJob asyncJob)
636 637 638 639 640 641 642
{
    qemuDomainObjPrivatePtr priv = vm->privateData;
    char *diskAlias = NULL;
    int ret = -1;
    int status;
    int rv;

643
    status = qemuBlockJobUpdate(vm, job, asyncJob);
644 645 646 647
    switch (status) {
    case VIR_DOMAIN_BLOCK_JOB_FAILED:
    case VIR_DOMAIN_BLOCK_JOB_CANCELED:
        if (failNoJob) {
648
            qemuMigrationNBDReportMirrorError(job, disk->dst);
649
            goto cleanup;
650
        }
651
        ATTRIBUTE_FALLTHROUGH;
652
    case VIR_DOMAIN_BLOCK_JOB_COMPLETED:
653 654
        ret = 1;
        goto cleanup;
655 656
    }

657
    if (!(diskAlias = qemuAliasDiskDriveFromDisk(disk)))
658 659
        return -1;

660
    if (qemuDomainObjEnterMonitorAsync(driver, vm, asyncJob) < 0)
661 662
        goto cleanup;

663
    rv = qemuMonitorBlockJobCancel(priv->mon, diskAlias);
664 665 666 667 668

    if (qemuDomainObjExitMonitor(driver, vm) < 0 || rv < 0)
        goto cleanup;

    ret = 0;
669 670 671 672 673 674 675 676

 cleanup:
    VIR_FREE(diskAlias);
    return ret;
}


/**
677
 * qemuMigrationSrcNBDCopyCancel:
678 679
 * @driver: qemu driver
 * @vm: domain
680
 * @check: if true report an error when some of the mirrors fails
681
 *
682
 * Cancel all drive-mirrors started by qemuMigrationSrcNBDStorageCopy.
683 684 685
 * Any pending block job events for the affected disks will be processed and
 * synchronous block job terminated regardless of return value unless qemu
 * has crashed.
686 687 688 689
 *
 * Returns 0 on success, -1 otherwise.
 */
static int
690 691 692 693 694
qemuMigrationSrcNBDCopyCancel(virQEMUDriverPtr driver,
                              virDomainObjPtr vm,
                              bool check,
                              qemuDomainAsyncJob asyncJob,
                              virConnectPtr dconn)
695
{
696
    virErrorPtr err = NULL;
697
    int ret = -1;
698
    size_t i;
699 700 701 702
    int rv;
    bool failed = false;

    VIR_DEBUG("Cancelling drive mirrors for domain %s", vm->def->name);
703 704 705

    for (i = 0; i < vm->def->ndisks; i++) {
        virDomainDiskDefPtr disk = vm->def->disks[i];
706
        qemuDomainDiskPrivatePtr diskPriv = QEMU_DOMAIN_DISK_PRIVATE(disk);
707
        qemuBlockJobDataPtr job;
708

709 710
        if (!(job = qemuBlockJobDiskGetJob(disk)) ||
            !qemuBlockJobIsRunning(job))
711 712
            diskPriv->migrating = false;

713 714
        if (!diskPriv->migrating) {
            virObjectUnref(job);
715
            continue;
716
        }
717

718
        rv = qemuMigrationSrcNBDCopyCancelOne(driver, vm, disk, job,
719
                                              check, asyncJob);
720 721 722 723 724 725
        if (rv != 0) {
            if (rv < 0) {
                if (!err)
                    err = virSaveLastError();
                failed = true;
            }
726
            qemuBlockJobSyncEnd(vm, job, asyncJob);
727 728
            diskPriv->migrating = false;
        }
729 730

        virObjectUnref(job);
731 732
    }

733
    while ((rv = qemuMigrationSrcNBDCopyCancelled(vm, asyncJob, check)) != 1) {
734 735 736 737 738 739 740
        if (check && !failed &&
            dconn && virConnectIsAlive(dconn) <= 0) {
            virReportError(VIR_ERR_OPERATION_FAILED, "%s",
                           _("Lost connection to destination host"));
            failed = true;
        }

741 742 743 744
        if (rv < 0) {
            failed = true;
            if (rv == -2)
                break;
745
        }
746

747 748 749 750 751
        if (failed && !err)
            err = virSaveLastError();

        if (virDomainObjWait(vm) < 0)
            goto cleanup;
752 753
    }

754 755 756 757 758 759 760 761 762 763 764 765 766
    for (i = 0; i < vm->def->ndisks; i++) {
        virDomainDiskDefPtr disk = vm->def->disks[i];
        qemuDomainDiskPrivatePtr diskPriv = QEMU_DOMAIN_DISK_PRIVATE(disk);

        if (!diskPriv->migrSource)
            continue;

        qemuBlockStorageSourceDetachOneBlockdev(driver, vm, asyncJob,
                                                diskPriv->migrSource);
        virStorageSourceFree(diskPriv->migrSource);
        diskPriv->migrSource = NULL;
    }

767 768 769
    ret = failed ? -1 : 0;

 cleanup:
770 771 772 773 774
    if (err) {
        virSetError(err);
        virFreeError(err);
    }
    return ret;
775 776 777
}


778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856
static int
qemuMigrationSrcNBDStorageCopyBlockdev(virQEMUDriverPtr driver,
                                       virDomainObjPtr vm,
                                       virDomainDiskDefPtr disk,
                                       const char *diskAlias,
                                       const char *host,
                                       int port,
                                       unsigned long long mirror_speed,
                                       unsigned int mirror_flags,
                                       const char *tlsAlias)
{
    qemuBlockStorageSourceAttachDataPtr data = NULL;
    qemuDomainDiskPrivatePtr diskPriv = QEMU_DOMAIN_DISK_PRIVATE(disk);
    virStorageSourcePtr copysrc = NULL;
    int mon_ret = 0;
    int ret = -1;

    VIR_DEBUG("starting blockdev mirror for disk=%s to host=%s", diskAlias, host);

    if (VIR_ALLOC(copysrc) < 0)
        goto cleanup;

    copysrc->type = VIR_STORAGE_TYPE_NETWORK;
    copysrc->protocol = VIR_STORAGE_NET_PROTOCOL_NBD;
    copysrc->format = VIR_STORAGE_FILE_RAW;

    if (VIR_ALLOC(copysrc->backingStore) < 0)
        goto cleanup;

    if (VIR_STRDUP(copysrc->path, diskAlias) < 0)
        goto cleanup;

    if (VIR_ALLOC_N(copysrc->hosts, 1) < 0)
        goto cleanup;

    copysrc->nhosts = 1;
    copysrc->hosts->transport = VIR_STORAGE_NET_HOST_TRANS_TCP;
    copysrc->hosts->port = port;
    if (VIR_STRDUP(copysrc->hosts->name, host) < 0)
        goto cleanup;

    if (VIR_STRDUP(copysrc->tlsAlias, tlsAlias) < 0)
        goto cleanup;

    if (virAsprintf(&copysrc->nodestorage, "migration-%s-storage", disk->dst) < 0 ||
        virAsprintf(&copysrc->nodeformat, "migration-%s-format", disk->dst) < 0)
        goto cleanup;

    if (!(data = qemuBlockStorageSourceAttachPrepareBlockdev(copysrc)))
        goto cleanup;

    if (qemuDomainObjEnterMonitorAsync(driver, vm,
                                       QEMU_ASYNC_JOB_MIGRATION_OUT) < 0)
        goto cleanup;

    mon_ret = qemuBlockStorageSourceAttachApply(qemuDomainGetMonitor(vm), data);

    if (mon_ret == 0)
        mon_ret = qemuMonitorBlockdevMirror(qemuDomainGetMonitor(vm), NULL,
                                            diskAlias, copysrc->nodeformat,
                                            mirror_speed, 0, 0, mirror_flags);

    if (mon_ret != 0)
        qemuBlockStorageSourceAttachRollback(qemuDomainGetMonitor(vm), data);

    if (qemuDomainObjExitMonitor(driver, vm) < 0 || mon_ret < 0)
        goto cleanup;

    VIR_STEAL_PTR(diskPriv->migrSource, copysrc);

    ret = 0;

 cleanup:
    qemuBlockStorageSourceAttachDataFree(data);
    virStorageSourceFree(copysrc);
    return ret;
}


857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898
static int
qemuMigrationSrcNBDStorageCopyDriveMirror(virQEMUDriverPtr driver,
                                          virDomainObjPtr vm,
                                          const char *diskAlias,
                                          const char *host,
                                          int port,
                                          unsigned long long mirror_speed,
                                          unsigned int mirror_flags)
{
    char *nbd_dest = NULL;
    int mon_ret;
    int ret = -1;

    if (strchr(host, ':')) {
        if (virAsprintf(&nbd_dest, "nbd:[%s]:%d:exportname=%s",
                        host, port, diskAlias) < 0)
            goto cleanup;
    } else {
        if (virAsprintf(&nbd_dest, "nbd:%s:%d:exportname=%s",
                        host, port, diskAlias) < 0)
            goto cleanup;
    }

    if (qemuDomainObjEnterMonitorAsync(driver, vm,
                                       QEMU_ASYNC_JOB_MIGRATION_OUT) < 0)
        goto cleanup;

    mon_ret = qemuMonitorDriveMirror(qemuDomainGetMonitor(vm),
                                     diskAlias, nbd_dest, "raw",
                                     mirror_speed, 0, 0, mirror_flags);

    if (qemuDomainObjExitMonitor(driver, vm) < 0 || mon_ret < 0)
        goto cleanup;

    ret = 0;

 cleanup:
    VIR_FREE(nbd_dest);
    return ret;
}


899 900 901 902 903 904 905 906 907 908 909 910
static int
qemuMigrationSrcNBDStorageCopyOne(virQEMUDriverPtr driver,
                                  virDomainObjPtr vm,
                                  virDomainDiskDefPtr disk,
                                  const char *host,
                                  int port,
                                  unsigned long long mirror_speed,
                                  unsigned int mirror_flags,
                                  const char *tlsAlias,
                                  unsigned int flags)
{
    qemuDomainDiskPrivatePtr diskPriv = QEMU_DOMAIN_DISK_PRIVATE(disk);
911
    qemuBlockJobDataPtr job = NULL;
912 913 914 915 916 917 918
    char *diskAlias = NULL;
    int rc;
    int ret = -1;

    if (!(diskAlias = qemuAliasDiskDriveFromDisk(disk)))
        goto cleanup;

919
    if (!(job = qemuBlockJobDiskNew(disk, QEMU_BLOCKJOB_TYPE_COPY)))
920 921
        goto cleanup;

922
    qemuBlockJobSyncBegin(job);
923 924 925 926 927 928 929 930 931 932 933 934 935 936 937

    if (flags & VIR_MIGRATE_TLS) {
        rc = qemuMigrationSrcNBDStorageCopyBlockdev(driver, vm,
                                                    disk, diskAlias,
                                                    host, port,
                                                    mirror_speed,
                                                    mirror_flags,
                                                    tlsAlias);
    } else {
        rc = qemuMigrationSrcNBDStorageCopyDriveMirror(driver, vm, diskAlias,
                                                       host, port,
                                                       mirror_speed,
                                                       mirror_flags);
    }

938
    if (rc < 0)
939 940 941
        goto cleanup;

    diskPriv->migrating = true;
942
    qemuBlockJobStarted(job);
943 944 945 946

    ret = 0;

 cleanup:
947
    qemuBlockJobStartupFinalize(job);
948 949 950 951 952
    VIR_FREE(diskAlias);
    return ret;
}


953
/**
954
 * qemuMigrationSrcNBDStorageCopy:
955 956 957 958
 * @driver: qemu driver
 * @vm: domain
 * @mig: migration cookie
 * @host: where are we migrating to
959
 * @speed: bandwidth limit in MiB/s
960 961
 * @migrate_flags: migrate monitor command flags
 *
962 963 964
 * Migrate non-shared storage using the NBD protocol to the server running
 * inside the qemu process on dst and wait until the copy converges.
 * On success update @migrate_flags so we don't tell 'migrate' command
965
 * to do the very same operation. On failure, the caller is
966 967
 * expected to call qemuMigrationSrcNBDCopyCancel to stop all
 * running copy operations.
968 969 970 971 972
 *
 * Returns 0 on success (@migrate_flags updated),
 *        -1 otherwise.
 */
static int
973 974 975 976 977 978 979 980
qemuMigrationSrcNBDStorageCopy(virQEMUDriverPtr driver,
                               virDomainObjPtr vm,
                               qemuMigrationCookiePtr mig,
                               const char *host,
                               unsigned long speed,
                               unsigned int *migrate_flags,
                               size_t nmigrate_disks,
                               const char **migrate_disks,
981 982 983
                               virConnectPtr dconn,
                               const char *tlsAlias,
                               unsigned int flags)
984 985 986 987
{
    qemuDomainObjPrivatePtr priv = vm->privateData;
    int ret = -1;
    int port;
988
    size_t i;
989
    unsigned long long mirror_speed = speed;
990
    unsigned int mirror_flags = VIR_DOMAIN_BLOCK_REBASE_REUSE_EXT;
991
    int rv;
992
    virQEMUDriverConfigPtr cfg = virQEMUDriverGetConfig(driver);
993 994

    VIR_DEBUG("Starting drive mirrors for domain %s", vm->def->name);
995

996 997 998 999 1000 1001 1002 1003
    if (mirror_speed > LLONG_MAX >> 20) {
        virReportError(VIR_ERR_OVERFLOW,
                       _("bandwidth must be less than %llu"),
                       LLONG_MAX >> 20);
        goto cleanup;
    }
    mirror_speed <<= 20;

1004 1005 1006 1007 1008 1009 1010 1011 1012 1013
    /* steal NBD port and thus prevent its propagation back to destination */
    port = mig->nbd->port;
    mig->nbd->port = 0;

    if (*migrate_flags & QEMU_MONITOR_MIGRATE_NON_SHARED_INC)
        mirror_flags |= VIR_DOMAIN_BLOCK_REBASE_SHALLOW;

    for (i = 0; i < vm->def->ndisks; i++) {
        virDomainDiskDefPtr disk = vm->def->disks[i];

1014
        /* check whether disk should be migrated */
1015
        if (!qemuMigrationAnyCopyDisk(disk, nmigrate_disks, migrate_disks))
1016 1017
            continue;

1018 1019 1020
        if (qemuMigrationSrcNBDStorageCopyOne(driver, vm, disk, host, port,
                                              mirror_speed, mirror_flags,
                                              tlsAlias, flags) < 0)
1021 1022
            goto cleanup;

1023
        if (virDomainSaveStatus(driver->xmlopt, cfg->stateDir, vm, driver->caps) < 0) {
1024 1025 1026
            VIR_WARN("Failed to save status on vm %s", vm->def->name);
            goto cleanup;
        }
1027
    }
1028

1029
    while ((rv = qemuMigrationSrcNBDStorageCopyReady(vm, QEMU_ASYNC_JOB_MIGRATION_OUT)) != 1) {
1030 1031
        if (rv < 0)
            goto cleanup;
1032

1033
        if (priv->job.abortJob) {
1034
            priv->job.current->status = QEMU_DOMAIN_JOB_STATUS_CANCELED;
1035 1036 1037 1038
            virReportError(VIR_ERR_OPERATION_ABORTED, _("%s: %s"),
                           qemuDomainAsyncJobTypeToString(priv->job.asyncJob),
                           _("canceled by client"));
            goto cleanup;
1039
        }
1040

1041 1042 1043 1044 1045 1046
        if (dconn && virConnectIsAlive(dconn) <= 0) {
            virReportError(VIR_ERR_OPERATION_FAILED, "%s",
                           _("Lost connection to destination host"));
            goto cleanup;
        }

1047
        if (virDomainObjWait(vm) < 0)
1048
            goto cleanup;
1049 1050
    }

1051 1052
    qemuMigrationSrcFetchMirrorStats(driver, vm, QEMU_ASYNC_JOB_MIGRATION_OUT,
                                     priv->job.current);
1053

1054
    /* Okay, all disks are ready. Modify migrate_flags */
1055 1056 1057 1058
    *migrate_flags &= ~(QEMU_MONITOR_MIGRATE_NON_SHARED_DISK |
                        QEMU_MONITOR_MIGRATE_NON_SHARED_INC);
    ret = 0;

1059
 cleanup:
1060
    virObjectUnref(cfg);
1061 1062
    return ret;
}
1063

1064

1065
/**
1066
 * qemuMigrationSrcIsAllowedHostdev:
1067 1068
 * @def: domain definition
 *
1069
 * Checks that @def does not contain any host devices unsupported across
1070 1071 1072
 * migrations. Returns true if the vm is allowed to migrate.
 */
static bool
1073
qemuMigrationSrcIsAllowedHostdev(const virDomainDef *def)
1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092
{
    size_t i;

    /* Migration with USB host devices is allowed, all other devices are
     * forbidden. */
    for (i = 0; i < def->nhostdevs; i++) {
        virDomainHostdevDefPtr hostdev = def->hostdevs[i];
        if (hostdev->mode != VIR_DOMAIN_HOSTDEV_MODE_SUBSYS ||
            hostdev->source.subsys.type != VIR_DOMAIN_HOSTDEV_SUBSYS_TYPE_USB) {
            virReportError(VIR_ERR_OPERATION_INVALID, "%s",
                           _("domain has assigned non-USB host devices"));
            return false;
        }
    }

    return true;
}


1093
/**
1094
 * qemuMigrationSrcIsAllowed:
1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106
 * @driver: qemu driver struct
 * @vm: domain object
 * @remote: migration is remote
 * @flags: migration flags (see struct virDomainMigrateFlags)
 *
 * Validates that the configuration of @vm can be migrated in various
 * situations. If @remote is true, the migration happens to remote host. @flags
 * is used to check various special migration types according to the request.
 *
 * Returns true if migration is supported. Reports libvirt error and returns
 * false otherwise.
 */
1107
bool
1108 1109 1110 1111
qemuMigrationSrcIsAllowed(virQEMUDriverPtr driver,
                          virDomainObjPtr vm,
                          bool remote,
                          unsigned int flags)
1112
{
1113
    int nsnapshots;
1114
    int pauseReason;
1115
    size_t i;
1116

1117 1118 1119 1120 1121
    /* perform these checks only when migrating to remote hosts */
    if (remote) {
        nsnapshots = virDomainSnapshotObjListNum(vm->snapshots, NULL, 0);
        if (nsnapshots < 0)
            return false;
1122

1123 1124 1125 1126 1127
        if (nsnapshots > 0) {
            virReportError(VIR_ERR_OPERATION_INVALID,
                           _("cannot migrate domain with %d snapshots"),
                           nsnapshots);
            return false;
1128
        }
1129

1130
        /* cancel migration if disk I/O error is emitted while migrating */
1131
        if (flags & VIR_MIGRATE_ABORT_ON_ERROR &&
1132
            !(flags & VIR_MIGRATE_OFFLINE) &&
1133 1134
            virDomainObjGetState(vm, &pauseReason) == VIR_DOMAIN_PAUSED &&
            pauseReason == VIR_DOMAIN_PAUSED_IOERROR) {
E
Eric Blake 已提交
1135
            virReportError(VIR_ERR_OPERATION_INVALID, "%s",
1136
                           _("cannot migrate domain with I/O error"));
E
Eric Blake 已提交
1137 1138
            return false;
        }
1139 1140

    }
1141

1142 1143 1144 1145 1146 1147 1148
    /* following checks don't make sense for offline migration */
    if (!(flags & VIR_MIGRATE_OFFLINE)) {
        if (qemuProcessAutoDestroyActive(driver, vm)) {
            virReportError(VIR_ERR_OPERATION_INVALID,
                           "%s", _("domain is marked for auto destroy"));
            return false;
        }
1149

1150

1151 1152
        if (qemuDomainHasBlockjob(vm, false)) {
            virReportError(VIR_ERR_OPERATION_INVALID, "%s",
1153
                           _("domain has active block job"));
1154 1155 1156
            return false;
        }

1157
        if (!qemuMigrationSrcIsAllowedHostdev(vm->def))
1158 1159
            return false;

1160
        if (vm->def->cpu) {
1161 1162 1163
            /* QEMU blocks migration and save with invariant TSC enabled
             * unless TSC frequency is explicitly set.
             */
1164 1165
            if (virCPUCheckFeature(vm->def->os.arch, vm->def->cpu,
                                   "invtsc") == 1) {
1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183
                bool block = true;

                for (i = 0; i < vm->def->clock.ntimers; i++) {
                    virDomainTimerDefPtr timer = vm->def->clock.timers[i];

                    if (timer->name == VIR_DOMAIN_TIMER_NAME_TSC &&
                        timer->frequency > 0) {
                        block = false;
                        break;
                    }
                }

                if (block) {
                    virReportError(VIR_ERR_OPERATION_INVALID, "%s",
                                   _("domain has 'invtsc' CPU feature but "
                                     "TSC frequency is not specified"));
                    return false;
                }
1184
            }
J
Ján Tomko 已提交
1185 1186
        }

1187 1188 1189
        /* Verify that memory device config can be transferred reliably */
        for (i = 0; i < vm->def->nmems; i++) {
            virDomainMemoryDefPtr mem = vm->def->mems[i];
1190

1191 1192 1193 1194 1195
            if (mem->model == VIR_DOMAIN_MEMORY_MODEL_DIMM &&
                mem->info.type != VIR_DOMAIN_DEVICE_ADDRESS_TYPE_DIMM) {
                virReportError(VIR_ERR_OPERATION_INVALID, "%s",
                               _("domain's dimm info lacks slot ID "
                                 "or base address"));
1196

1197 1198
                return false;
            }
1199
        }
1200 1201 1202 1203 1204 1205

        if (vm->def->nshmems) {
            virReportError(VIR_ERR_OPERATION_INVALID, "%s",
                           _("migration with shmem device is not supported"));
            return false;
        }
1206 1207
    }

1208 1209 1210
    return true;
}

1211
static bool
1212 1213 1214 1215
qemuMigrationSrcIsSafe(virDomainDefPtr def,
                       size_t nmigrate_disks,
                       const char **migrate_disks,
                       unsigned int flags)
1216

1217
{
1218 1219
    bool storagemigration = flags & (VIR_MIGRATE_NON_SHARED_DISK |
                                     VIR_MIGRATE_NON_SHARED_INC);
1220
    size_t i;
1221
    int rc;
1222

1223
    for (i = 0; i < def->ndisks; i++) {
1224
        virDomainDiskDefPtr disk = def->disks[i];
1225
        const char *src = virDomainDiskGetSource(disk);
1226

1227 1228
        /* Disks without any source (i.e. floppies and CD-ROMs)
         * OR readonly are safe. */
1229
        if (virStorageSourceIsEmpty(disk->src) ||
1230
            disk->src->readonly)
1231
            continue;
1232

1233
        /* Disks which are migrated by qemu are safe too. */
1234
        if (storagemigration &&
1235
            qemuMigrationAnyCopyDisk(disk, nmigrate_disks, migrate_disks))
1236
            continue;
1237

1238
        /* However, disks on local FS (e.g. ext4) are not safe. */
1239
        if (virDomainDiskGetType(disk) == VIR_STORAGE_TYPE_FILE) {
1240
            if ((rc = virFileIsSharedFS(src)) < 0) {
1241
                return false;
1242 1243 1244 1245 1246
            } else if (rc == 0) {
                virReportError(VIR_ERR_MIGRATE_UNSAFE, "%s",
                               _("Migration without shared storage is unsafe"));
                return false;
            }
1247 1248 1249 1250 1251 1252 1253
            if ((rc = virStorageFileIsClusterFS(src)) < 0)
                return false;
            else if (rc == 1)
                continue;
        } else if (disk->src->type == VIR_STORAGE_TYPE_NETWORK &&
                   disk->src->protocol == VIR_STORAGE_NET_PROTOCOL_RBD) {
            continue;
1254
        }
1255

1256 1257 1258 1259 1260 1261 1262
        /* Our code elsewhere guarantees shared disks are either readonly (in
         * which case cache mode doesn't matter) or used with cache=none or used with cache=directsync */
        if (disk->src->shared ||
            disk->cachemode == VIR_DOMAIN_DISK_CACHE_DISABLE ||
            disk->cachemode == VIR_DOMAIN_DISK_CACHE_DIRECTSYNC)
            continue;

1263 1264
        virReportError(VIR_ERR_MIGRATE_UNSAFE, "%s",
                       _("Migration may lead to data corruption if disks"
1265
                         " use cache != none or cache != directsync"));
1266
        return false;
1267 1268 1269 1270 1271
    }

    return true;
}

1272
/** qemuMigrationSrcSetOffline
1273 1274 1275
 * Pause domain for non-live migration.
 */
int
1276 1277
qemuMigrationSrcSetOffline(virQEMUDriverPtr driver,
                           virDomainObjPtr vm)
1278 1279
{
    int ret;
1280
    VIR_DEBUG("driver=%p vm=%p", driver, vm);
1281 1282
    ret = qemuProcessStopCPUs(driver, vm, VIR_DOMAIN_PAUSED_MIGRATION,
                              QEMU_ASYNC_JOB_MIGRATION_OUT);
1283
    if (ret == 0) {
1284
        virObjectEventPtr event;
1285

1286
        event = virDomainEventLifecycleNewFromObj(vm,
1287 1288
                                         VIR_DOMAIN_EVENT_SUSPENDED,
                                         VIR_DOMAIN_EVENT_SUSPENDED_MIGRATED);
1289
        virObjectEventStateQueue(driver->domainEventState, event);
1290 1291 1292 1293 1294
    }

    return ret;
}

1295 1296

void
1297 1298
qemuMigrationAnyPostcopyFailed(virQEMUDriverPtr driver,
                               virDomainObjPtr vm)
1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328
{
    virDomainState state;
    int reason;

    state = virDomainObjGetState(vm, &reason);

    if (state != VIR_DOMAIN_PAUSED &&
        state != VIR_DOMAIN_RUNNING)
        return;

    if (state == VIR_DOMAIN_PAUSED &&
        reason == VIR_DOMAIN_PAUSED_POSTCOPY_FAILED)
        return;

    VIR_WARN("Migration of domain %s failed during post-copy; "
             "leaving the domain paused", vm->def->name);

    if (state == VIR_DOMAIN_RUNNING) {
        virObjectEventPtr event;

        if (qemuProcessStopCPUs(driver, vm,
                                VIR_DOMAIN_PAUSED_POSTCOPY_FAILED,
                                QEMU_ASYNC_JOB_MIGRATION_IN) < 0) {
            VIR_WARN("Unable to pause guest CPUs for %s", vm->def->name);
            return;
        }

        event = virDomainEventLifecycleNewFromObj(vm,
                                VIR_DOMAIN_EVENT_SUSPENDED,
                                VIR_DOMAIN_EVENT_SUSPENDED_POSTCOPY_FAILED);
1329
        virObjectEventStateQueue(driver->domainEventState, event);
1330 1331 1332 1333 1334 1335 1336
    } else {
        virDomainObjSetState(vm, VIR_DOMAIN_PAUSED,
                             VIR_DOMAIN_PAUSED_POSTCOPY_FAILED);
    }
}


1337
static int
1338
qemuMigrationSrcWaitForSpice(virDomainObjPtr vm)
1339 1340 1341
{
    qemuDomainObjPrivatePtr priv = vm->privateData;

1342 1343
    if (!virQEMUCapsGet(priv->qemuCaps, QEMU_CAPS_SEAMLESS_MIGRATION) ||
        !priv->job.spiceMigration)
1344 1345
        return 0;

1346
    VIR_DEBUG("Waiting for SPICE to finish migration");
1347 1348
    while (!priv->job.spiceMigrated && !priv->job.abortJob) {
        if (virDomainObjWait(vm) < 0)
1349 1350 1351 1352
            return -1;
    }
    return 0;
}
1353

1354 1355 1356 1357

static void
qemuMigrationUpdateJobType(qemuDomainJobInfoPtr jobInfo)
{
1358
    switch ((qemuMonitorMigrationStatus) jobInfo->stats.mig.status) {
1359 1360 1361 1362
    case QEMU_MONITOR_MIGRATION_STATUS_POSTCOPY:
        jobInfo->status = QEMU_DOMAIN_JOB_STATUS_POSTCOPY;
        break;

1363
    case QEMU_MONITOR_MIGRATION_STATUS_COMPLETED:
1364
        jobInfo->status = QEMU_DOMAIN_JOB_STATUS_QEMU_COMPLETED;
1365 1366 1367
        break;

    case QEMU_MONITOR_MIGRATION_STATUS_INACTIVE:
1368
        jobInfo->status = QEMU_DOMAIN_JOB_STATUS_NONE;
1369 1370 1371
        break;

    case QEMU_MONITOR_MIGRATION_STATUS_ERROR:
1372
        jobInfo->status = QEMU_DOMAIN_JOB_STATUS_FAILED;
1373 1374 1375
        break;

    case QEMU_MONITOR_MIGRATION_STATUS_CANCELLED:
1376
        jobInfo->status = QEMU_DOMAIN_JOB_STATUS_CANCELED;
1377 1378
        break;

1379 1380 1381 1382 1383 1384 1385 1386
    case QEMU_MONITOR_MIGRATION_STATUS_PRE_SWITCHOVER:
        jobInfo->status = QEMU_DOMAIN_JOB_STATUS_PAUSED;
        break;

    case QEMU_MONITOR_MIGRATION_STATUS_DEVICE:
        jobInfo->status = QEMU_DOMAIN_JOB_STATUS_MIGRATING;
        break;

1387 1388 1389
    case QEMU_MONITOR_MIGRATION_STATUS_SETUP:
    case QEMU_MONITOR_MIGRATION_STATUS_ACTIVE:
    case QEMU_MONITOR_MIGRATION_STATUS_CANCELLING:
1390
    case QEMU_MONITOR_MIGRATION_STATUS_LAST:
1391 1392 1393 1394 1395 1396
        break;
    }
}


int
1397 1398 1399 1400 1401
qemuMigrationAnyFetchStats(virQEMUDriverPtr driver,
                           virDomainObjPtr vm,
                           qemuDomainAsyncJob asyncJob,
                           qemuDomainJobInfoPtr jobInfo,
                           char **error)
1402 1403
{
    qemuDomainObjPrivatePtr priv = vm->privateData;
1404
    qemuMonitorMigrationStats stats;
1405 1406 1407 1408 1409
    int rv;

    if (qemuDomainObjEnterMonitorAsync(driver, vm, asyncJob) < 0)
        return -1;

1410
    rv = qemuMonitorGetMigrationStats(priv->mon, &stats, error);
1411 1412 1413 1414

    if (qemuDomainObjExitMonitor(driver, vm) < 0 || rv < 0)
        return -1;

1415
    jobInfo->stats.mig = stats;
1416 1417

    return 0;
1418 1419 1420
}


1421 1422 1423 1424 1425 1426 1427
static const char *
qemuMigrationJobName(virDomainObjPtr vm)
{
    qemuDomainObjPrivatePtr priv = vm->privateData;

    switch (priv->job.asyncJob) {
    case QEMU_ASYNC_JOB_MIGRATION_OUT:
1428
        return _("migration out job");
1429 1430 1431 1432
    case QEMU_ASYNC_JOB_SAVE:
        return _("domain save job");
    case QEMU_ASYNC_JOB_DUMP:
        return _("domain core dump job");
1433 1434 1435 1436 1437 1438 1439 1440 1441
    case QEMU_ASYNC_JOB_NONE:
        return _("undefined");
    case QEMU_ASYNC_JOB_MIGRATION_IN:
        return _("migration in job");
    case QEMU_ASYNC_JOB_SNAPSHOT:
        return _("snapshot job");
    case QEMU_ASYNC_JOB_START:
        return _("start job");
    case QEMU_ASYNC_JOB_LAST:
1442 1443 1444 1445 1446 1447
    default:
        return _("job");
    }
}


1448
static int
1449
qemuMigrationJobCheckStatus(virQEMUDriverPtr driver,
1450
                            virDomainObjPtr vm,
1451
                            qemuDomainAsyncJob asyncJob)
1452 1453 1454
{
    qemuDomainObjPrivatePtr priv = vm->privateData;
    qemuDomainJobInfoPtr jobInfo = priv->job.current;
1455
    char *error = NULL;
1456
    bool events = virQEMUCapsGet(priv->qemuCaps, QEMU_CAPS_MIGRATION_EVENT);
1457
    int ret = -1;
1458

1459
    if (!events ||
1460
        jobInfo->stats.mig.status == QEMU_MONITOR_MIGRATION_STATUS_ERROR) {
1461
        if (qemuMigrationAnyFetchStats(driver, vm, asyncJob, jobInfo, &error) < 0)
1462 1463
            return -1;
    }
1464

1465 1466
    qemuMigrationUpdateJobType(jobInfo);

1467 1468
    switch (jobInfo->status) {
    case QEMU_DOMAIN_JOB_STATUS_NONE:
1469 1470
        virReportError(VIR_ERR_OPERATION_FAILED, _("%s: %s"),
                       qemuMigrationJobName(vm), _("is not active"));
1471
        goto cleanup;
1472

1473
    case QEMU_DOMAIN_JOB_STATUS_FAILED:
1474
        virReportError(VIR_ERR_OPERATION_FAILED, _("%s: %s"),
1475 1476 1477
                       qemuMigrationJobName(vm),
                       error ? error : _("unexpectedly failed"));
        goto cleanup;
1478

1479
    case QEMU_DOMAIN_JOB_STATUS_CANCELED:
1480 1481
        virReportError(VIR_ERR_OPERATION_ABORTED, _("%s: %s"),
                       qemuMigrationJobName(vm), _("canceled by client"));
1482
        goto cleanup;
1483

1484 1485
    case QEMU_DOMAIN_JOB_STATUS_COMPLETED:
    case QEMU_DOMAIN_JOB_STATUS_ACTIVE:
1486
    case QEMU_DOMAIN_JOB_STATUS_MIGRATING:
1487
    case QEMU_DOMAIN_JOB_STATUS_QEMU_COMPLETED:
1488
    case QEMU_DOMAIN_JOB_STATUS_POSTCOPY:
1489
    case QEMU_DOMAIN_JOB_STATUS_PAUSED:
1490 1491
        break;
    }
1492 1493 1494 1495 1496 1497

    ret = 0;

 cleanup:
    VIR_FREE(error);
    return ret;
1498 1499 1500
}


1501 1502
enum qemuMigrationCompletedFlags {
    QEMU_MIGRATION_COMPLETED_ABORT_ON_ERROR = (1 << 0),
1503
    /* This flag should only be set when run on src host */
1504
    QEMU_MIGRATION_COMPLETED_CHECK_STORAGE  = (1 << 1),
1505
    QEMU_MIGRATION_COMPLETED_POSTCOPY       = (1 << 2),
1506
    QEMU_MIGRATION_COMPLETED_PRE_SWITCHOVER = (1 << 3),
1507 1508
};

1509

1510 1511 1512 1513 1514 1515 1516
/**
 * Returns 1 if migration completed successfully,
 *         0 if the domain is still being migrated,
 *         -1 migration failed,
 *         -2 something else failed, we need to cancel migration.
 */
static int
1517 1518 1519 1520 1521
qemuMigrationAnyCompleted(virQEMUDriverPtr driver,
                          virDomainObjPtr vm,
                          qemuDomainAsyncJob asyncJob,
                          virConnectPtr dconn,
                          unsigned int flags)
1522 1523 1524 1525 1526
{
    qemuDomainObjPrivatePtr priv = vm->privateData;
    qemuDomainJobInfoPtr jobInfo = priv->job.current;
    int pauseReason;

1527
    if (qemuMigrationJobCheckStatus(driver, vm, asyncJob) < 0)
1528 1529
        goto error;

1530
    /* This flag should only be set when run on src host */
1531
    if (flags & QEMU_MIGRATION_COMPLETED_CHECK_STORAGE &&
1532
        qemuMigrationSrcNBDStorageCopyReady(vm, asyncJob) < 0)
1533 1534
        goto error;

1535
    if (flags & QEMU_MIGRATION_COMPLETED_ABORT_ON_ERROR &&
1536 1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548
        virDomainObjGetState(vm, &pauseReason) == VIR_DOMAIN_PAUSED &&
        pauseReason == VIR_DOMAIN_PAUSED_IOERROR) {
        virReportError(VIR_ERR_OPERATION_FAILED, _("%s: %s"),
                       qemuMigrationJobName(vm), _("failed due to I/O error"));
        goto error;
    }

    if (dconn && virConnectIsAlive(dconn) <= 0) {
        virReportError(VIR_ERR_OPERATION_FAILED, "%s",
                       _("Lost connection to destination host"));
        goto error;
    }

1549 1550 1551 1552 1553 1554 1555 1556 1557 1558
    /* Migration was paused before serializing device state, let's return to
     * the caller so that it can finish all block jobs, resume migration, and
     * wait again for the real end of the migration.
     */
    if (flags & QEMU_MIGRATION_COMPLETED_PRE_SWITCHOVER &&
        jobInfo->status == QEMU_DOMAIN_JOB_STATUS_PAUSED) {
        VIR_DEBUG("Migration paused before switchover");
        return 1;
    }

1559 1560 1561 1562 1563
    /* In case of postcopy the source considers migration completed at the
     * moment it switched from active to postcopy-active state. The destination
     * will continue waiting until the migrate state changes to completed.
     */
    if (flags & QEMU_MIGRATION_COMPLETED_POSTCOPY &&
1564
        jobInfo->status == QEMU_DOMAIN_JOB_STATUS_POSTCOPY) {
1565 1566 1567 1568
        VIR_DEBUG("Migration switched to post-copy");
        return 1;
    }

1569
    if (jobInfo->status == QEMU_DOMAIN_JOB_STATUS_QEMU_COMPLETED)
1570 1571 1572 1573 1574
        return 1;
    else
        return 0;

 error:
1575 1576 1577
    switch (jobInfo->status) {
    case QEMU_DOMAIN_JOB_STATUS_MIGRATING:
    case QEMU_DOMAIN_JOB_STATUS_POSTCOPY:
1578
    case QEMU_DOMAIN_JOB_STATUS_PAUSED:
1579
        /* The migration was aborted by us rather than QEMU itself. */
1580
        jobInfo->status = QEMU_DOMAIN_JOB_STATUS_FAILED;
1581
        return -2;
1582 1583 1584

    case QEMU_DOMAIN_JOB_STATUS_QEMU_COMPLETED:
        /* Something failed after QEMU already finished the migration. */
1585
        jobInfo->status = QEMU_DOMAIN_JOB_STATUS_FAILED;
1586
        return -1;
1587 1588 1589 1590

    case QEMU_DOMAIN_JOB_STATUS_FAILED:
    case QEMU_DOMAIN_JOB_STATUS_CANCELED:
        /* QEMU aborted the migration. */
1591
        return -1;
1592 1593 1594 1595 1596 1597

    case QEMU_DOMAIN_JOB_STATUS_ACTIVE:
    case QEMU_DOMAIN_JOB_STATUS_COMPLETED:
    case QEMU_DOMAIN_JOB_STATUS_NONE:
        /* Impossible. */
        break;
1598
    }
1599 1600

    return -1;
1601 1602 1603
}


1604 1605 1606
/* Returns 0 on success, -2 when migration needs to be cancelled, or -1 when
 * QEMU reports failed migration.
 */
1607
static int
1608 1609 1610 1611 1612
qemuMigrationSrcWaitForCompletion(virQEMUDriverPtr driver,
                                  virDomainObjPtr vm,
                                  qemuDomainAsyncJob asyncJob,
                                  virConnectPtr dconn,
                                  unsigned int flags)
1613
{
1614
    qemuDomainObjPrivatePtr priv = vm->privateData;
J
Jiri Denemark 已提交
1615
    qemuDomainJobInfoPtr jobInfo = priv->job.current;
1616
    bool events = virQEMUCapsGet(priv->qemuCaps, QEMU_CAPS_MIGRATION_EVENT);
1617
    int rv;
1618

1619 1620
    jobInfo->status = QEMU_DOMAIN_JOB_STATUS_MIGRATING;

1621 1622
    while ((rv = qemuMigrationAnyCompleted(driver, vm, asyncJob,
                                           dconn, flags)) != 1) {
1623 1624
        if (rv < 0)
            return rv;
1625

1626 1627
        if (events) {
            if (virDomainObjWait(vm) < 0) {
1628 1629
                if (virDomainObjIsActive(vm))
                    jobInfo->status = QEMU_DOMAIN_JOB_STATUS_FAILED;
1630 1631 1632 1633 1634 1635 1636 1637 1638 1639
                return -2;
            }
        } else {
            /* Poll every 50ms for progress & to allow cancellation */
            struct timespec ts = { .tv_sec = 0, .tv_nsec = 50 * 1000 * 1000ull };

            virObjectUnlock(vm);
            nanosleep(&ts, NULL);
            virObjectLock(vm);
        }
1640 1641
    }

1642
    if (events)
1643
        ignore_value(qemuMigrationAnyFetchStats(driver, vm, asyncJob, jobInfo, NULL));
1644

1645
    qemuDomainJobInfoUpdateTime(jobInfo);
1646 1647
    qemuDomainJobInfoUpdateDowntime(jobInfo);
    VIR_FREE(priv->job.completed);
1648
    if (VIR_ALLOC(priv->job.completed) == 0) {
1649
        *priv->job.completed = *jobInfo;
1650 1651
        priv->job.completed->status = QEMU_DOMAIN_JOB_STATUS_COMPLETED;
    }
1652

1653 1654 1655 1656
    if (asyncJob != QEMU_ASYNC_JOB_MIGRATION_OUT &&
        jobInfo->status == QEMU_DOMAIN_JOB_STATUS_QEMU_COMPLETED)
        jobInfo->status = QEMU_DOMAIN_JOB_STATUS_COMPLETED;

1657
    return 0;
1658 1659 1660
}


1661
static int
1662 1663 1664 1665
qemuMigrationDstWaitForCompletion(virQEMUDriverPtr driver,
                                  virDomainObjPtr vm,
                                  qemuDomainAsyncJob asyncJob,
                                  bool postcopy)
1666 1667
{
    qemuDomainObjPrivatePtr priv = vm->privateData;
1668
    unsigned int flags = 0;
1669 1670 1671 1672 1673 1674 1675
    int rv;

    if (!virQEMUCapsGet(priv->qemuCaps, QEMU_CAPS_MIGRATION_EVENT))
        return 0;

    VIR_DEBUG("Waiting for incoming migration to complete");

1676 1677 1678
    if (postcopy)
        flags = QEMU_MIGRATION_COMPLETED_POSTCOPY;

1679 1680
    while ((rv = qemuMigrationAnyCompleted(driver, vm, asyncJob,
                                           NULL, flags)) != 1) {
1681 1682 1683 1684 1685 1686 1687 1688
        if (rv < 0 || virDomainObjWait(vm) < 0)
            return -1;
    }

    return 0;
}


1689
static int
1690 1691 1692 1693
qemuMigrationSrcGraphicsRelocate(virQEMUDriverPtr driver,
                                 virDomainObjPtr vm,
                                 qemuMigrationCookiePtr cookie,
                                 const char *graphicsuri)
1694 1695
{
    qemuDomainObjPrivatePtr priv = vm->privateData;
1696 1697
    int ret = -1;
    const char *listenAddress = NULL;
1698
    virSocketAddr addr;
1699 1700 1701 1702 1703
    virURIPtr uri = NULL;
    int type = -1;
    int port = -1;
    int tlsPort = -1;
    const char *tlsSubject = NULL;
1704

1705
    if (!cookie || (!cookie->graphics && !graphicsuri))
1706 1707
        return 0;

1708 1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720 1721 1722 1723 1724 1725 1726
    if (graphicsuri && !(uri = virURIParse(graphicsuri)))
        goto cleanup;

    if (cookie->graphics) {
        type = cookie->graphics->type;

        listenAddress = cookie->graphics->listen;

        if (!listenAddress ||
            (virSocketAddrParse(&addr, listenAddress, AF_UNSPEC) > 0 &&
             virSocketAddrIsWildcard(&addr)))
            listenAddress = cookie->remoteHostname;

        port = cookie->graphics->port;
        tlsPort = cookie->graphics->tlsPort;
        tlsSubject = cookie->graphics->tlsSubject;
    }

    if (uri) {
1727
        size_t i;
1728 1729 1730 1731 1732 1733 1734 1735 1736 1737 1738 1739 1740 1741 1742 1743 1744 1745 1746 1747 1748 1749 1750 1751 1752 1753 1754

        if ((type = virDomainGraphicsTypeFromString(uri->scheme)) < 0) {
            virReportError(VIR_ERR_INVALID_ARG,
                           _("unknown graphics type %s"), uri->scheme);
            goto cleanup;
        }

        if (uri->server)
            listenAddress = uri->server;
        if (uri->port > 0)
            port = uri->port;

        for (i = 0; i < uri->paramsCount; i++) {
            virURIParamPtr param = uri->params + i;

            if (STRCASEEQ(param->name, "tlsPort")) {
                if (virStrToLong_i(param->value, NULL, 10, &tlsPort) < 0) {
                    virReportError(VIR_ERR_INVALID_ARG,
                                   _("invalid tlsPort number: %s"),
                                   param->value);
                    goto cleanup;
                }
            } else if (STRCASEEQ(param->name, "tlsSubject")) {
                tlsSubject = param->value;
            }
        }
    }
1755 1756 1757 1758

    /* QEMU doesn't support VNC relocation yet, so
     * skip it to avoid generating an error
     */
1759 1760 1761 1762
    if (type != VIR_DOMAIN_GRAPHICS_TYPE_SPICE) {
        ret = 0;
        goto cleanup;
    }
1763

1764 1765 1766 1767 1768 1769 1770 1771 1772
    /* Older libvirt sends port == 0 for listen type='none' graphics. It's
     * safe to ignore such requests since relocation to unknown port does
     * not make sense in general.
     */
    if (port <= 0 && tlsPort <= 0) {
        ret = 0;
        goto cleanup;
    }

1773 1774 1775 1776
    if (qemuDomainObjEnterMonitorAsync(driver, vm,
                                       QEMU_ASYNC_JOB_MIGRATION_OUT) == 0) {
        ret = qemuMonitorGraphicsRelocate(priv->mon, type, listenAddress,
                                          port, tlsPort, tlsSubject);
1777
        priv->job.spiceMigration = !ret;
1778 1779
        if (qemuDomainObjExitMonitor(driver, vm) < 0)
            ret = -1;
1780
    }
1781

1782
 cleanup:
1783
    virURIFree(uri);
1784 1785 1786 1787
    return ret;
}


1788
static int
1789 1790 1791
qemuMigrationDstOPDRelocate(virQEMUDriverPtr driver ATTRIBUTE_UNUSED,
                            virDomainObjPtr vm,
                            qemuMigrationCookiePtr cookie)
1792
{
1793 1794
    virDomainNetDefPtr netptr;
    int ret = -1;
1795
    size_t i;
1796 1797 1798 1799 1800 1801 1802 1803

    for (i = 0; i < cookie->network->nnets; i++) {
        netptr = vm->def->nets[i];

        switch (cookie->network->net[i].vporttype) {
        case VIR_NETDEV_VPORT_PROFILE_NONE:
        case VIR_NETDEV_VPORT_PROFILE_8021QBG:
        case VIR_NETDEV_VPORT_PROFILE_8021QBH:
1804
           break;
1805
        case VIR_NETDEV_VPORT_PROFILE_OPENVSWITCH:
1806 1807
            if (virNetDevOpenvswitchSetMigrateData(cookie->network->net[i].portdata,
                                                   netptr->ifname) != 0) {
J
Jiri Denemark 已提交
1808 1809 1810
                virReportError(VIR_ERR_INTERNAL_ERROR,
                               _("Unable to run command to set OVS port data for "
                                 "interface %s"), netptr->ifname);
1811 1812 1813
                goto cleanup;
            }
            break;
1814 1815 1816 1817 1818
        default:
            break;
        }
    }

1819
    ret = 0;
1820
 cleanup:
1821 1822 1823 1824
    return ret;
}


1825
int
1826 1827
qemuMigrationDstCheckProtocol(virQEMUCapsPtr qemuCaps,
                              const char *migrateFrom)
1828 1829 1830 1831 1832 1833 1834 1835 1836 1837 1838 1839 1840 1841 1842 1843 1844 1845 1846 1847 1848 1849 1850
{
    if (STRPREFIX(migrateFrom, "rdma")) {
        if (!virQEMUCapsGet(qemuCaps, QEMU_CAPS_MIGRATE_RDMA)) {
            virReportError(VIR_ERR_OPERATION_UNSUPPORTED, "%s",
                           _("incoming RDMA migration is not supported "
                             "with this QEMU binary"));
            return -1;
        }
    } else if (!STRPREFIX(migrateFrom, "tcp") &&
               !STRPREFIX(migrateFrom, "exec") &&
               !STRPREFIX(migrateFrom, "fd") &&
               !STRPREFIX(migrateFrom, "unix") &&
               STRNEQ(migrateFrom, "stdio")) {
        virReportError(VIR_ERR_OPERATION_UNSUPPORTED, "%s",
                       _("unknown migration protocol"));
        return -1;
    }

    return 0;
}


char *
1851 1852
qemuMigrationDstGetURI(const char *migrateFrom,
                       int migrateFd)
1853 1854 1855 1856 1857 1858 1859 1860 1861 1862 1863 1864
{
    char *uri = NULL;

    if (STREQ(migrateFrom, "stdio"))
        ignore_value(virAsprintf(&uri, "fd:%d", migrateFd));
    else
        ignore_value(VIR_STRDUP(uri, migrateFrom));

    return uri;
}


1865
int
1866 1867 1868 1869
qemuMigrationDstRun(virQEMUDriverPtr driver,
                    virDomainObjPtr vm,
                    const char *uri,
                    qemuDomainAsyncJob asyncJob)
1870 1871 1872 1873 1874 1875 1876 1877 1878 1879 1880 1881 1882 1883 1884 1885
{
    qemuDomainObjPrivatePtr priv = vm->privateData;
    int ret = -1;
    int rv;

    VIR_DEBUG("Setting up incoming migration with URI %s", uri);

    if (qemuDomainObjEnterMonitorAsync(driver, vm, asyncJob) < 0)
        return -1;

    rv = qemuMonitorMigrateIncoming(priv->mon, uri);

    if (qemuDomainObjExitMonitor(driver, vm) < 0 || rv < 0)
        goto cleanup;

    if (asyncJob == QEMU_ASYNC_JOB_MIGRATION_IN) {
1886
        /* qemuMigrationDstWaitForCompletion is called from the Finish phase */
1887 1888 1889 1890
        ret = 0;
        goto cleanup;
    }

1891
    if (qemuMigrationDstWaitForCompletion(driver, vm, asyncJob, false) < 0)
1892 1893 1894 1895 1896 1897 1898 1899 1900
        goto cleanup;

    ret = 0;

 cleanup:
    return ret;
}


1901 1902 1903 1904 1905 1906
/* This is called for outgoing non-p2p migrations when a connection to the
 * client which initiated the migration was closed but we were waiting for it
 * to follow up with the next phase, that is, in between
 * qemuDomainMigrateBegin3 and qemuDomainMigratePerform3 or
 * qemuDomainMigratePerform3 and qemuDomainMigrateConfirm3.
 */
1907
static void
1908 1909 1910
qemuMigrationSrcCleanup(virDomainObjPtr vm,
                        virConnectPtr conn,
                        void *opaque)
1911
{
1912
    virQEMUDriverPtr driver = opaque;
1913 1914 1915 1916 1917 1918 1919 1920 1921
    qemuDomainObjPrivatePtr priv = vm->privateData;

    VIR_DEBUG("vm=%s, conn=%p, asyncJob=%s, phase=%s",
              vm->def->name, conn,
              qemuDomainAsyncJobTypeToString(priv->job.asyncJob),
              qemuDomainAsyncJobPhaseToString(priv->job.asyncJob,
                                              priv->job.phase));

    if (!qemuMigrationJobIsActive(vm, QEMU_ASYNC_JOB_MIGRATION_OUT))
1922
        return;
1923 1924 1925 1926 1927

    VIR_DEBUG("The connection which started outgoing migration of domain %s"
              " was closed; canceling the migration",
              vm->def->name);

1928
    switch ((qemuMigrationJobPhase) priv->job.phase) {
1929 1930 1931 1932 1933 1934 1935 1936 1937
    case QEMU_MIGRATION_PHASE_BEGIN3:
        /* just forget we were about to migrate */
        qemuDomainObjDiscardAsyncJob(driver, vm);
        break;

    case QEMU_MIGRATION_PHASE_PERFORM3_DONE:
        VIR_WARN("Migration of domain %s finished but we don't know if the"
                 " domain was successfully started on destination or not",
                 vm->def->name);
1938
        qemuMigrationParamsReset(driver, vm, QEMU_ASYNC_JOB_MIGRATION_OUT,
1939
                                 priv->job.migParams, priv->job.apiFlags);
1940 1941 1942 1943 1944 1945 1946 1947 1948 1949 1950 1951 1952 1953 1954 1955 1956 1957 1958 1959 1960 1961
        /* clear the job and let higher levels decide what to do */
        qemuDomainObjDiscardAsyncJob(driver, vm);
        break;

    case QEMU_MIGRATION_PHASE_PERFORM3:
        /* cannot be seen without an active migration API; unreachable */
    case QEMU_MIGRATION_PHASE_CONFIRM3:
    case QEMU_MIGRATION_PHASE_CONFIRM3_CANCELLED:
        /* all done; unreachable */
    case QEMU_MIGRATION_PHASE_PREPARE:
    case QEMU_MIGRATION_PHASE_FINISH2:
    case QEMU_MIGRATION_PHASE_FINISH3:
        /* incoming migration; unreachable */
    case QEMU_MIGRATION_PHASE_PERFORM2:
        /* single phase outgoing migration; unreachable */
    case QEMU_MIGRATION_PHASE_NONE:
    case QEMU_MIGRATION_PHASE_LAST:
        /* unreachable */
        ;
    }
}

1962

1963
/* The caller is supposed to lock the vm and start a migration job. */
1964
static char *
1965 1966 1967 1968 1969 1970 1971 1972 1973
qemuMigrationSrcBeginPhase(virQEMUDriverPtr driver,
                           virDomainObjPtr vm,
                           const char *xmlin,
                           const char *dname,
                           char **cookieout,
                           int *cookieoutlen,
                           size_t nmigrate_disks,
                           const char **migrate_disks,
                           unsigned long flags)
1974 1975 1976
{
    char *rv = NULL;
    qemuMigrationCookiePtr mig = NULL;
1977
    virDomainDefPtr def = NULL;
1978
    qemuDomainObjPrivatePtr priv = vm->privateData;
1979
    virCapsPtr caps = NULL;
1980
    unsigned int cookieFlags = QEMU_MIGRATION_COOKIE_LOCKSTATE;
1981

1982
    VIR_DEBUG("driver=%p, vm=%p, xmlin=%s, dname=%s,"
1983
              " cookieout=%p, cookieoutlen=%p,"
1984
              " nmigrate_disks=%zu, migrate_disks=%p, flags=0x%lx",
1985
              driver, vm, NULLSTR(xmlin), NULLSTR(dname),
1986 1987
              cookieout, cookieoutlen, nmigrate_disks,
              migrate_disks, flags);
1988

1989 1990 1991
    if (!(caps = virQEMUDriverGetCapabilities(driver, false)))
        goto cleanup;

1992 1993 1994 1995 1996 1997
    /* Only set the phase if we are inside QEMU_ASYNC_JOB_MIGRATION_OUT.
     * Otherwise we will start the async job later in the perform phase losing
     * change protection.
     */
    if (priv->job.asyncJob == QEMU_ASYNC_JOB_MIGRATION_OUT)
        qemuMigrationJobSetPhase(driver, vm, QEMU_MIGRATION_PHASE_BEGIN3);
1998

1999
    if (!qemuMigrationSrcIsAllowed(driver, vm, true, flags))
2000 2001
        goto cleanup;

2002
    if (!(flags & (VIR_MIGRATE_UNSAFE | VIR_MIGRATE_OFFLINE)) &&
2003
        !qemuMigrationSrcIsSafe(vm->def, nmigrate_disks, migrate_disks, flags))
2004 2005
        goto cleanup;

2006 2007 2008 2009 2010 2011 2012 2013 2014
    if (flags & VIR_MIGRATE_POSTCOPY &&
        (!(flags & VIR_MIGRATE_LIVE) ||
         flags & VIR_MIGRATE_PAUSED)) {
        virReportError(VIR_ERR_ARGUMENT_UNSUPPORTED, "%s",
                       _("post-copy migration is not supported with non-live "
                         "or paused migration"));
        goto cleanup;
    }

2015 2016 2017 2018 2019 2020
    if (flags & VIR_MIGRATE_POSTCOPY && flags & VIR_MIGRATE_TUNNELLED) {
        virReportError(VIR_ERR_ARGUMENT_UNSUPPORTED, "%s",
                       _("post-copy is not supported with tunnelled migration"));
        goto cleanup;
    }

2021 2022 2023 2024 2025 2026 2027 2028 2029 2030 2031 2032 2033 2034 2035 2036 2037 2038 2039 2040 2041 2042 2043 2044 2045 2046 2047 2048 2049 2050 2051 2052 2053 2054 2055 2056 2057 2058 2059 2060 2061 2062 2063
    if (flags & (VIR_MIGRATE_NON_SHARED_DISK | VIR_MIGRATE_NON_SHARED_INC)) {
        bool has_drive_mirror =  virQEMUCapsGet(priv->qemuCaps,
                                                QEMU_CAPS_DRIVE_MIRROR);

        if (nmigrate_disks) {
            if (has_drive_mirror) {
                size_t i, j;
                /* Check user requested only known disk targets. */
                for (i = 0; i < nmigrate_disks; i++) {
                    for (j = 0; j < vm->def->ndisks; j++) {
                        if (STREQ(vm->def->disks[j]->dst, migrate_disks[i]))
                            break;
                    }

                    if (j == vm->def->ndisks) {
                        virReportError(VIR_ERR_INVALID_ARG,
                                       _("disk target %s not found"),
                                       migrate_disks[i]);
                        goto cleanup;
                    }
                }

                if (flags & VIR_MIGRATE_TUNNELLED) {
                    virReportError(VIR_ERR_OPERATION_UNSUPPORTED, "%s",
                                   _("Selecting disks to migrate is not "
                                     "implemented for tunnelled migration"));
                    goto cleanup;
                }
            } else {
                virReportError(VIR_ERR_OPERATION_UNSUPPORTED, "%s",
                               _("qemu does not support drive-mirror command"));
                goto cleanup;
            }
        }

        if (has_drive_mirror) {
            /* TODO support NBD for TUNNELLED migration */
            if (flags & VIR_MIGRATE_TUNNELLED) {
                VIR_WARN("NBD in tunnelled migration is currently not supported");
            } else {
                cookieFlags |= QEMU_MIGRATION_COOKIE_NBD;
                priv->nbdPort = 0;
            }
2064 2065 2066
        }
    }

2067
    if (virDomainDefHasMemoryHotplug(vm->def) ||
2068
        ((flags & VIR_MIGRATE_PERSIST_DEST) &&
2069
         vm->newDef && virDomainDefHasMemoryHotplug(vm->newDef)))
2070 2071
        cookieFlags |= QEMU_MIGRATION_COOKIE_MEMORY_HOTPLUG;

2072 2073 2074 2075 2076
    if (!qemuDomainVcpuHotplugIsInOrder(vm->def) ||
        ((flags & VIR_MIGRATE_PERSIST_DEST) &&
         vm->newDef && !qemuDomainVcpuHotplugIsInOrder(vm->newDef)))
        cookieFlags |= QEMU_MIGRATION_COOKIE_CPU_HOTPLUG;

2077 2078 2079
    if (priv->origCPU)
        cookieFlags |= QEMU_MIGRATION_COOKIE_CPU;

2080 2081
    cookieFlags |= QEMU_MIGRATION_COOKIE_ALLOW_REBOOT;

2082 2083 2084
    if (!(flags & VIR_MIGRATE_OFFLINE))
        cookieFlags |= QEMU_MIGRATION_COOKIE_CAPS;

2085 2086
    if (!(mig = qemuMigrationEatCookie(driver, vm->def,
                                       priv->origname, priv, NULL, 0, 0)))
2087 2088 2089
        goto cleanup;

    if (qemuMigrationBakeCookie(mig, driver, vm,
2090
                                QEMU_MIGRATION_SOURCE,
2091
                                cookieout, cookieoutlen,
2092
                                cookieFlags) < 0)
2093 2094
        goto cleanup;

L
liguang 已提交
2095 2096 2097 2098 2099 2100 2101 2102 2103 2104 2105 2106 2107 2108 2109 2110 2111 2112 2113 2114 2115 2116
    if (flags & VIR_MIGRATE_OFFLINE) {
        if (flags & (VIR_MIGRATE_NON_SHARED_DISK |
                     VIR_MIGRATE_NON_SHARED_INC)) {
            virReportError(VIR_ERR_OPERATION_INVALID, "%s",
                           _("offline migration cannot handle "
                             "non-shared storage"));
            goto cleanup;
        }
        if (!(flags & VIR_MIGRATE_PERSIST_DEST)) {
            virReportError(VIR_ERR_OPERATION_INVALID, "%s",
                           _("offline migration must be specified with "
                             "the persistent flag set"));
            goto cleanup;
        }
        if (flags & VIR_MIGRATE_TUNNELLED) {
            virReportError(VIR_ERR_OPERATION_INVALID, "%s",
                           _("tunnelled offline migration does not "
                             "make sense"));
            goto cleanup;
        }
    }

2117
    if (xmlin) {
2118
        if (!(def = virDomainDefParseString(xmlin, caps, driver->xmlopt, priv->qemuCaps,
2119 2120
                                            VIR_DOMAIN_DEF_PARSE_INACTIVE |
                                            VIR_DOMAIN_DEF_PARSE_SKIP_VALIDATE)))
2121 2122
            goto cleanup;

2123
        if (!qemuDomainCheckABIStability(driver, vm, def))
2124 2125
            goto cleanup;

2126
        rv = qemuDomainDefFormatLive(driver, def, NULL, false, true);
2127
    } else {
2128 2129
        rv = qemuDomainDefFormatLive(driver, vm->def, priv->origCPU,
                                     false, true);
2130
    }
2131

2132
 cleanup:
2133
    qemuMigrationCookieFree(mig);
2134
    virObjectUnref(caps);
2135
    virDomainDefFree(def);
2136 2137 2138
    return rv;
}

2139
char *
2140 2141 2142 2143 2144 2145 2146 2147 2148
qemuMigrationSrcBegin(virConnectPtr conn,
                      virDomainObjPtr vm,
                      const char *xmlin,
                      const char *dname,
                      char **cookieout,
                      int *cookieoutlen,
                      size_t nmigrate_disks,
                      const char **migrate_disks,
                      unsigned long flags)
2149 2150 2151
{
    virQEMUDriverPtr driver = conn->privateData;
    char *xml = NULL;
2152
    qemuDomainAsyncJob asyncJob;
2153 2154

    if ((flags & VIR_MIGRATE_CHANGE_PROTECTION)) {
2155 2156
        if (qemuMigrationJobStart(driver, vm, QEMU_ASYNC_JOB_MIGRATION_OUT,
                                  flags) < 0)
2157 2158 2159 2160 2161 2162 2163 2164
            goto cleanup;
        asyncJob = QEMU_ASYNC_JOB_MIGRATION_OUT;
    } else {
        if (qemuDomainObjBeginJob(driver, vm, QEMU_JOB_MODIFY) < 0)
            goto cleanup;
        asyncJob = QEMU_ASYNC_JOB_NONE;
    }

2165
    qemuMigrationSrcStoreDomainState(vm);
2166

2167
    if (!(flags & VIR_MIGRATE_OFFLINE) && virDomainObjCheckActive(vm) < 0)
2168 2169 2170 2171 2172 2173
        goto endjob;

    /* Check if there is any ejected media.
     * We don't want to require them on the destination.
     */
    if (!(flags & VIR_MIGRATE_OFFLINE) &&
2174
        qemuProcessRefreshDisks(driver, vm, asyncJob) < 0)
2175 2176
        goto endjob;

2177 2178 2179
    if (!(xml = qemuMigrationSrcBeginPhase(driver, vm, xmlin, dname,
                                           cookieout, cookieoutlen,
                                           nmigrate_disks, migrate_disks, flags)))
2180 2181 2182 2183 2184 2185 2186
        goto endjob;

    if ((flags & VIR_MIGRATE_CHANGE_PROTECTION)) {
        /* We keep the job active across API calls until the confirm() call.
         * This prevents any other APIs being invoked while migration is taking
         * place.
         */
2187
        if (virCloseCallbacksSet(driver->closeCallbacks, vm, conn,
2188
                                 qemuMigrationSrcCleanup) < 0) {
2189
            VIR_FREE(xml);
2190
            goto endjob;
2191
        }
2192
        qemuMigrationJobContinue(vm);
2193 2194 2195 2196
    } else {
        goto endjob;
    }

2197
 cleanup:
M
Michal Privoznik 已提交
2198
    virDomainObjEndAPI(&vm);
2199 2200
    return xml;

2201
 endjob:
2202 2203 2204 2205
    if (flags & VIR_MIGRATE_CHANGE_PROTECTION)
        qemuMigrationJobFinish(driver, vm);
    else
        qemuDomainObjEndJob(driver, vm);
2206 2207 2208
    goto cleanup;
}

2209

2210 2211
/* Prepare is the first step, and it runs on the destination host.
 */
2212

2213
static void
2214 2215
qemuMigrationDstPrepareCleanup(virQEMUDriverPtr driver,
                               virDomainObjPtr vm)
2216 2217 2218 2219 2220 2221 2222 2223 2224
{
    qemuDomainObjPrivatePtr priv = vm->privateData;

    VIR_DEBUG("driver=%p, vm=%s, job=%s, asyncJob=%s",
              driver,
              vm->def->name,
              qemuDomainJobTypeToString(priv->job.active),
              qemuDomainAsyncJobTypeToString(priv->job.asyncJob));

2225
    virPortAllocatorRelease(priv->migrationPort);
2226 2227
    priv->migrationPort = 0;

2228 2229 2230 2231 2232
    if (!qemuMigrationJobIsActive(vm, QEMU_ASYNC_JOB_MIGRATION_IN))
        return;
    qemuDomainObjDiscardAsyncJob(driver, vm);
}

2233
static qemuProcessIncomingDefPtr
2234 2235 2236 2237 2238 2239
qemuMigrationDstPrepare(virDomainObjPtr vm,
                        bool tunnel,
                        const char *protocol,
                        const char *listenAddress,
                        unsigned short port,
                        int fd)
2240 2241
{
    qemuDomainObjPrivatePtr priv = vm->privateData;
2242
    qemuProcessIncomingDefPtr inc = NULL;
2243 2244 2245 2246 2247 2248 2249 2250 2251 2252 2253 2254 2255 2256 2257 2258 2259 2260 2261 2262 2263 2264 2265 2266 2267 2268 2269 2270 2271 2272
    char *migrateFrom = NULL;

    if (tunnel) {
        if (VIR_STRDUP(migrateFrom, "stdio") < 0)
            goto cleanup;
    } else {
        bool encloseAddress = false;
        bool hostIPv6Capable = false;
        struct addrinfo *info = NULL;
        struct addrinfo hints = { .ai_flags = AI_ADDRCONFIG,
                                  .ai_socktype = SOCK_STREAM };
        const char *incFormat;

        if (getaddrinfo("::", NULL, &hints, &info) == 0) {
            freeaddrinfo(info);
            hostIPv6Capable = true;
        }

        if (listenAddress) {
            if (virSocketAddrNumericFamily(listenAddress) == AF_INET6) {
                if (!hostIPv6Capable) {
                    virReportError(VIR_ERR_ARGUMENT_UNSUPPORTED, "%s",
                                   _("host isn't capable of IPv6"));
                    goto cleanup;
                }
                /* IPv6 address must be escaped in brackets on the cmd line */
                encloseAddress = true;
            } else {
                /* listenAddress is a hostname or IPv4 */
            }
2273
        } else if (hostIPv6Capable) {
2274 2275 2276 2277 2278 2279 2280 2281 2282 2283 2284 2285 2286 2287 2288 2289 2290 2291 2292 2293 2294 2295 2296
            /* Listen on :: instead of 0.0.0.0 if QEMU understands it
             * and there is at least one IPv6 address configured
             */
            listenAddress = "::";
            encloseAddress = true;
        } else {
            listenAddress = "0.0.0.0";
        }

        /* QEMU will be started with
         *   -incoming protocol:[<IPv6 addr>]:port,
         *   -incoming protocol:<IPv4 addr>:port, or
         *   -incoming protocol:<hostname>:port
         */
        if (encloseAddress)
            incFormat = "%s:[%s]:%d";
        else
            incFormat = "%s:%s:%d";
        if (virAsprintf(&migrateFrom, incFormat,
                        protocol, listenAddress, port) < 0)
            goto cleanup;
    }

2297 2298
    inc = qemuProcessIncomingDefNew(priv->qemuCaps, listenAddress,
                                    migrateFrom, fd, NULL);
2299

2300
 cleanup:
2301 2302
    VIR_FREE(migrateFrom);
    return inc;
2303 2304
}

2305
static int
2306
qemuMigrationDstPrepareAny(virQEMUDriverPtr driver,
2307
                           virConnectPtr dconn,
2308 2309 2310 2311 2312 2313 2314 2315 2316 2317 2318 2319 2320 2321
                           const char *cookiein,
                           int cookieinlen,
                           char **cookieout,
                           int *cookieoutlen,
                           virDomainDefPtr *def,
                           const char *origname,
                           virStreamPtr st,
                           const char *protocol,
                           unsigned short port,
                           bool autoPort,
                           const char *listenAddress,
                           size_t nmigrate_disks,
                           const char **migrate_disks,
                           int nbdPort,
2322
                           qemuMigrationParamsPtr migParams,
2323
                           unsigned long flags)
2324 2325
{
    virDomainObjPtr vm = NULL;
2326
    virObjectEventPtr event = NULL;
2327
    virErrorPtr origErr;
2328
    int ret = -1;
2329
    int dataFD[2] = { -1, -1 };
2330
    qemuDomainObjPrivatePtr priv = NULL;
2331
    qemuMigrationCookiePtr mig = NULL;
2332
    bool tunnel = !!st;
J
Jiri Denemark 已提交
2333
    char *xmlout = NULL;
L
liguang 已提交
2334
    unsigned int cookieFlags;
2335
    unsigned int startFlags;
2336
    virCapsPtr caps = NULL;
2337
    qemuProcessIncomingDefPtr incoming = NULL;
2338
    bool taint_hook = false;
2339 2340 2341
    bool stopProcess = false;
    bool relabel = false;
    int rv;
2342
    char *tlsAlias = NULL;
2343

2344 2345
    virNWFilterReadLockFilterUpdates();

L
liguang 已提交
2346 2347 2348 2349 2350 2351 2352 2353 2354 2355 2356 2357 2358 2359 2360 2361 2362 2363 2364 2365
    if (flags & VIR_MIGRATE_OFFLINE) {
        if (flags & (VIR_MIGRATE_NON_SHARED_DISK |
                     VIR_MIGRATE_NON_SHARED_INC)) {
            virReportError(VIR_ERR_OPERATION_INVALID, "%s",
                           _("offline migration cannot handle "
                             "non-shared storage"));
            goto cleanup;
        }
        if (!(flags & VIR_MIGRATE_PERSIST_DEST)) {
            virReportError(VIR_ERR_OPERATION_INVALID, "%s",
                           _("offline migration must be specified with "
                             "the persistent flag set"));
            goto cleanup;
        }
        if (tunnel) {
            virReportError(VIR_ERR_OPERATION_INVALID, "%s",
                           _("tunnelled offline migration does not "
                             "make sense"));
            goto cleanup;
        }
2366 2367
        cookieFlags = 0;
    } else {
2368 2369
        cookieFlags = QEMU_MIGRATION_COOKIE_GRAPHICS |
                      QEMU_MIGRATION_COOKIE_CAPS;
L
liguang 已提交
2370 2371
    }

2372 2373 2374 2375 2376 2377 2378 2379 2380
    if (flags & VIR_MIGRATE_POSTCOPY &&
        (!(flags & VIR_MIGRATE_LIVE) ||
         flags & VIR_MIGRATE_PAUSED)) {
        virReportError(VIR_ERR_ARGUMENT_UNSUPPORTED, "%s",
                       _("post-copy migration is not supported with non-live "
                         "or paused migration"));
        goto cleanup;
    }

2381 2382 2383 2384 2385 2386
    if (flags & VIR_MIGRATE_POSTCOPY && flags & VIR_MIGRATE_TUNNELLED) {
        virReportError(VIR_ERR_ARGUMENT_UNSUPPORTED, "%s",
                       _("post-copy is not supported with tunnelled migration"));
        goto cleanup;
    }

2387 2388 2389
    if (!(caps = virQEMUDriverGetCapabilities(driver, false)))
        goto cleanup;

2390
    if (!qemuMigrationSrcIsAllowedHostdev(*def))
2391 2392
        goto cleanup;

J
Jiri Denemark 已提交
2393 2394 2395 2396 2397
    /* Let migration hook filter domain XML */
    if (virHookPresent(VIR_HOOK_DRIVER_QEMU)) {
        char *xml;
        int hookret;

2398
        if (!(xml = qemuDomainDefFormatXML(driver, *def,
2399 2400
                                           VIR_DOMAIN_XML_SECURE |
                                           VIR_DOMAIN_XML_MIGRATABLE)))
J
Jiri Denemark 已提交
2401 2402
            goto cleanup;

2403
        hookret = virHookCall(VIR_HOOK_DRIVER_QEMU, (*def)->name,
J
Jiri Denemark 已提交
2404 2405 2406 2407 2408 2409 2410
                              VIR_HOOK_QEMU_OP_MIGRATE, VIR_HOOK_SUBOP_BEGIN,
                              NULL, xml, &xmlout);
        VIR_FREE(xml);

        if (hookret < 0) {
            goto cleanup;
        } else if (hookret == 0) {
2411
            if (virStringIsEmpty(xmlout)) {
J
Jiri Denemark 已提交
2412 2413 2414 2415 2416 2417
                VIR_DEBUG("Migrate hook filter returned nothing; using the"
                          " original XML");
            } else {
                virDomainDefPtr newdef;

                VIR_DEBUG("Using hook-filtered domain XML: %s", xmlout);
2418
                newdef = virDomainDefParseString(xmlout, caps, driver->xmlopt, NULL,
2419 2420
                                                 VIR_DOMAIN_DEF_PARSE_INACTIVE |
                                                 VIR_DOMAIN_DEF_PARSE_SKIP_VALIDATE);
J
Jiri Denemark 已提交
2421 2422 2423
                if (!newdef)
                    goto cleanup;

2424
                if (!qemuDomainDefCheckABIStability(driver, *def, newdef)) {
J
Jiri Denemark 已提交
2425 2426 2427 2428
                    virDomainDefFree(newdef);
                    goto cleanup;
                }

2429 2430
                virDomainDefFree(*def);
                *def = newdef;
2431 2432 2433 2434
                /* We should taint the domain here. However, @vm and therefore
                 * privateData too are still NULL, so just notice the fact and
                 * taint it later. */
                taint_hook = true;
J
Jiri Denemark 已提交
2435 2436 2437 2438
            }
        }
    }

2439 2440 2441 2442 2443 2444 2445 2446 2447 2448 2449 2450 2451 2452
    /* Parse cookie earlier than adding the domain onto the
     * domain list. Parsing/validation may fail and there's no
     * point in having the domain in the list at that point. */
    if (!(mig = qemuMigrationEatCookie(driver, *def, origname, NULL,
                                       cookiein, cookieinlen,
                                       QEMU_MIGRATION_COOKIE_LOCKSTATE |
                                       QEMU_MIGRATION_COOKIE_NBD |
                                       QEMU_MIGRATION_COOKIE_MEMORY_HOTPLUG |
                                       QEMU_MIGRATION_COOKIE_CPU_HOTPLUG |
                                       QEMU_MIGRATION_COOKIE_CPU |
                                       QEMU_MIGRATION_COOKIE_ALLOW_REBOOT |
                                       QEMU_MIGRATION_COOKIE_CAPS)))
        goto cleanup;

2453
    if (!(vm = virDomainObjListAdd(driver->domains, *def,
2454
                                   driver->xmlopt,
2455 2456 2457
                                   VIR_DOMAIN_OBJ_LIST_ADD_LIVE |
                                   VIR_DOMAIN_OBJ_LIST_ADD_CHECK_LIVE,
                                   NULL)))
2458
        goto cleanup;
2459
    *def = NULL;
2460

2461
    priv = vm->privateData;
2462 2463
    if (VIR_STRDUP(priv->origname, origname) < 0)
        goto cleanup;
2464

2465 2466 2467 2468 2469
    if (taint_hook) {
        /* Domain XML has been altered by a hook script. */
        priv->hookRun = true;
    }

2470 2471
    if (STREQ_NULLABLE(protocol, "rdma") &&
        !virMemoryLimitIsSet(vm->def->mem.hard_limit)) {
M
Michael R. Hines 已提交
2472 2473 2474 2475 2476 2477
        virReportError(VIR_ERR_OPERATION_INVALID, "%s",
                       _("cannot start RDMA migration with no memory hard "
                         "limit set"));
        goto cleanup;
    }

2478 2479 2480
    if (qemuMigrationDstPrecreateStorage(vm, mig->nbd,
                                         nmigrate_disks, migrate_disks,
                                         !!(flags & VIR_MIGRATE_NON_SHARED_INC)) < 0)
2481 2482
        goto cleanup;

2483 2484
    if (qemuMigrationJobStart(driver, vm, QEMU_ASYNC_JOB_MIGRATION_IN,
                              flags) < 0)
2485
        goto cleanup;
2486
    qemuMigrationJobSetPhase(driver, vm, QEMU_MIGRATION_PHASE_PREPARE);
2487 2488 2489 2490

    /* Domain starts inactive, even if the domain XML had an id field. */
    vm->def->id = -1;

L
liguang 已提交
2491 2492 2493
    if (flags & VIR_MIGRATE_OFFLINE)
        goto done;

2494 2495
    if (tunnel &&
        (pipe(dataFD) < 0 || virSetCloseExec(dataFD[1]) < 0)) {
2496 2497
        virReportSystemError(errno, "%s",
                             _("cannot create pipe for tunnelled migration"));
2498
        goto stopjob;
2499 2500
    }

2501 2502
    startFlags = VIR_QEMU_PROCESS_START_AUTODESTROY;

2503
    if (qemuProcessInit(driver, vm, mig->cpu, QEMU_ASYNC_JOB_MIGRATION_IN,
2504
                        true, startFlags) < 0)
2505
        goto stopjob;
2506
    stopProcess = true;
2507

2508 2509
    priv->allowReboot = mig->allowReboot;

2510 2511 2512
    if (!(incoming = qemuMigrationDstPrepare(vm, tunnel, protocol,
                                             listenAddress, port,
                                             dataFD[0])))
2513
        goto stopjob;
2514

2515
    if (qemuProcessPrepareDomain(driver, vm, startFlags) < 0)
2516 2517
        goto stopjob;

2518
    if (qemuProcessPrepareHost(driver, vm, startFlags) < 0)
2519 2520
        goto stopjob;

2521
    rv = qemuProcessLaunch(dconn, driver, vm, QEMU_ASYNC_JOB_MIGRATION_IN,
2522 2523
                           incoming, NULL,
                           VIR_NETDEV_VPORT_PROFILE_OP_MIGRATE_IN_START,
2524
                           startFlags);
2525 2526 2527
    if (rv < 0) {
        if (rv == -2)
            relabel = true;
2528
        goto stopjob;
2529
    }
2530
    relabel = true;
2531

2532 2533 2534 2535
    if (tunnel) {
        if (virFDStreamOpen(st, dataFD[1]) < 0) {
            virReportSystemError(errno, "%s",
                                 _("cannot pass pipe for tunnelled migration"));
2536
            goto stopjob;
2537
        }
2538
        dataFD[1] = -1; /* 'st' owns the FD now & will close it */
2539 2540
    }

2541
    if (STREQ_NULLABLE(protocol, "rdma") &&
M
Michael R. Hines 已提交
2542
        virProcessSetMaxMemLock(vm->pid, vm->def->mem.hard_limit << 10) < 0) {
2543
        goto stopjob;
M
Michael R. Hines 已提交
2544 2545
    }

2546
    if (qemuMigrationParamsCheck(driver, vm, QEMU_ASYNC_JOB_MIGRATION_IN,
2547
                                 migParams, mig->caps->automatic) < 0)
2548 2549 2550 2551 2552 2553 2554
        goto stopjob;

    /* Migrations using TLS need to add the "tls-creds-x509" object and
     * set the migration TLS parameters */
    if (flags & VIR_MIGRATE_TLS) {
        if (qemuMigrationParamsEnableTLS(driver, vm, true,
                                         QEMU_ASYNC_JOB_MIGRATION_IN,
2555
                                         &tlsAlias, NULL,
2556 2557 2558 2559 2560 2561 2562
                                         migParams) < 0)
            goto stopjob;
    } else {
        if (qemuMigrationParamsDisableTLS(vm, migParams) < 0)
            goto stopjob;
    }

2563 2564
    if (qemuMigrationParamsApply(driver, vm, QEMU_ASYNC_JOB_MIGRATION_IN,
                                 migParams) < 0)
2565 2566
        goto stopjob;

2567 2568 2569
    if (mig->nbd &&
        flags & (VIR_MIGRATE_NON_SHARED_DISK | VIR_MIGRATE_NON_SHARED_INC) &&
        virQEMUCapsGet(priv->qemuCaps, QEMU_CAPS_NBD_SERVER)) {
2570 2571 2572 2573 2574 2575 2576 2577 2578 2579 2580 2581
        const char *nbdTLSAlias = NULL;

        if (flags & VIR_MIGRATE_TLS) {
            if (!virQEMUCapsGet(priv->qemuCaps, QEMU_CAPS_NBD_TLS)) {
                virReportError(VIR_ERR_OPERATION_UNSUPPORTED, "%s",
                               _("QEMU NBD server does not support TLS transport"));
                goto stopjob;
            }

            nbdTLSAlias = tlsAlias;
        }

2582 2583
        if (qemuMigrationDstStartNBDServer(driver, vm, incoming->address,
                                           nmigrate_disks, migrate_disks,
2584
                                           nbdPort, nbdTLSAlias) < 0) {
2585
            goto stopjob;
2586
        }
2587
        cookieFlags |= QEMU_MIGRATION_COOKIE_NBD;
2588 2589
    }

2590 2591 2592 2593 2594 2595 2596 2597 2598
    if (mig->lockState) {
        VIR_DEBUG("Received lockstate %s", mig->lockState);
        VIR_FREE(priv->lockState);
        priv->lockState = mig->lockState;
        mig->lockState = NULL;
    } else {
        VIR_DEBUG("Received no lockstate");
    }

2599
    if (incoming->deferredURI &&
2600 2601
        qemuMigrationDstRun(driver, vm, incoming->deferredURI,
                            QEMU_ASYNC_JOB_MIGRATION_IN) < 0)
2602 2603
        goto stopjob;

2604
    if (qemuProcessFinishStartup(driver, vm, QEMU_ASYNC_JOB_MIGRATION_IN,
2605 2606 2607
                                 false, VIR_DOMAIN_PAUSED_MIGRATION) < 0)
        goto stopjob;

2608
 done:
2609 2610 2611
    if (qemuMigrationBakeCookie(mig, driver, vm,
                                QEMU_MIGRATION_DESTINATION,
                                cookieout, cookieoutlen, cookieFlags) < 0) {
2612 2613 2614 2615 2616 2617 2618
        /* We could tear down the whole guest here, but
         * cookie data is (so far) non-critical, so that
         * seems a little harsh. We'll just warn for now.
         */
        VIR_WARN("Unable to encode migration cookie");
    }

2619
    if (qemuDomainCleanupAdd(vm, qemuMigrationDstPrepareCleanup) < 0)
2620
        goto stopjob;
2621

L
liguang 已提交
2622 2623
    if (!(flags & VIR_MIGRATE_OFFLINE)) {
        virDomainAuditStart(vm, "migrated", true);
2624
        event = virDomainEventLifecycleNewFromObj(vm,
L
liguang 已提交
2625 2626 2627
                                         VIR_DOMAIN_EVENT_STARTED,
                                         VIR_DOMAIN_EVENT_STARTED_MIGRATED);
    }
2628

2629 2630 2631 2632
    /* We keep the job active across API calls until the finish() call.
     * This prevents any other APIs being invoked while incoming
     * migration is taking place.
     */
2633
    qemuMigrationJobContinue(vm);
2634

2635 2636
    if (autoPort)
        priv->migrationPort = port;
2637 2638 2639 2640 2641
    /* in this case port is not auto selected and we don't need to manage it
     * anymore after cookie is baked
     */
    if (nbdPort != 0)
        priv->nbdPort = 0;
2642
    ret = 0;
2643

2644
 cleanup:
2645
    virErrorPreserveLast(&origErr);
2646
    VIR_FREE(tlsAlias);
2647
    qemuProcessIncomingDefFree(incoming);
J
Jiri Denemark 已提交
2648
    VIR_FREE(xmlout);
2649 2650
    VIR_FORCE_CLOSE(dataFD[0]);
    VIR_FORCE_CLOSE(dataFD[1]);
2651 2652 2653
    if (ret < 0 && priv) {
        /* priv is set right after vm is added to the list of domains
         * and there is no 'goto cleanup;' in the middle of those */
2654
        VIR_FREE(priv->origname);
2655 2656 2657 2658
        /* release if port is auto selected which is not the case if
         * it is given in parameters
         */
        if (nbdPort == 0)
2659
            virPortAllocatorRelease(priv->nbdPort);
2660
        priv->nbdPort = 0;
2661
        virDomainObjRemoveTransientDef(vm);
2662
        qemuDomainRemoveInactiveJob(driver, vm);
2663
    }
M
Michal Privoznik 已提交
2664
    virDomainObjEndAPI(&vm);
2665
    virObjectEventStateQueue(driver->domainEventState, event);
2666
    qemuMigrationCookieFree(mig);
2667
    virObjectUnref(caps);
2668
    virNWFilterUnlockFilterUpdates();
2669
    virErrorRestore(&origErr);
2670
    return ret;
2671

2672
 stopjob:
2673
    qemuMigrationParamsReset(driver, vm, QEMU_ASYNC_JOB_MIGRATION_IN,
2674
                             priv->job.migParams, priv->job.apiFlags);
2675

2676 2677 2678 2679
    if (stopProcess) {
        unsigned int stopFlags = VIR_QEMU_PROCESS_STOP_MIGRATED;
        if (!relabel)
            stopFlags |= VIR_QEMU_PROCESS_STOP_NO_RELABEL;
2680
        virDomainAuditStart(vm, "migrated", false);
2681 2682
        qemuProcessStop(driver, vm, VIR_DOMAIN_SHUTOFF_FAILED,
                        QEMU_ASYNC_JOB_MIGRATION_IN, stopFlags);
2683
    }
2684

2685
    qemuMigrationJobFinish(driver, vm);
2686
    goto cleanup;
2687 2688 2689
}


2690 2691 2692 2693 2694
/*
 * This version starts an empty VM listening on a localhost TCP port, and
 * sets up the corresponding virStream to handle the incoming data.
 */
int
2695
qemuMigrationDstPrepareTunnel(virQEMUDriverPtr driver,
2696
                              virConnectPtr dconn,
2697 2698 2699 2700 2701 2702 2703
                              const char *cookiein,
                              int cookieinlen,
                              char **cookieout,
                              int *cookieoutlen,
                              virStreamPtr st,
                              virDomainDefPtr *def,
                              const char *origname,
2704
                              qemuMigrationParamsPtr migParams,
2705
                              unsigned long flags)
2706
{
2707
    VIR_DEBUG("driver=%p, dconn=%p, cookiein=%s, cookieinlen=%d, "
2708
              "cookieout=%p, cookieoutlen=%p, st=%p, def=%p, "
2709
              "origname=%s, flags=0x%lx",
2710
              driver, dconn, NULLSTR(cookiein), cookieinlen,
2711
              cookieout, cookieoutlen, st, *def, origname, flags);
2712

2713 2714 2715 2716 2717 2718
    if (st == NULL) {
        virReportError(VIR_ERR_INTERNAL_ERROR, "%s",
                       _("tunnelled migration requested but NULL stream passed"));
        return -1;
    }

2719 2720 2721
    return qemuMigrationDstPrepareAny(driver, dconn, cookiein, cookieinlen,
                                      cookieout, cookieoutlen, def, origname,
                                      st, NULL, 0, false, NULL, 0, NULL, 0,
2722
                                      migParams, flags);
2723 2724 2725
}


2726
static virURIPtr
2727
qemuMigrationAnyParseURI(const char *uri, bool *wellFormed)
2728 2729 2730 2731 2732 2733 2734 2735 2736 2737 2738 2739 2740 2741 2742 2743 2744 2745 2746 2747 2748
{
    char *tmp = NULL;
    virURIPtr parsed;

    /* For compatibility reasons tcp://... URIs are sent as tcp:...
     * We need to transform them to a well-formed URI before parsing. */
    if (STRPREFIX(uri, "tcp:") && !STRPREFIX(uri + 4, "//")) {
        if (virAsprintf(&tmp, "tcp://%s", uri + 4) < 0)
            return NULL;
        uri = tmp;
    }

    parsed = virURIParse(uri);
    if (parsed && wellFormed)
        *wellFormed = !tmp;
    VIR_FREE(tmp);

    return parsed;
}


2749
int
2750
qemuMigrationDstPrepareDirect(virQEMUDriverPtr driver,
2751
                              virConnectPtr dconn,
2752 2753 2754 2755 2756 2757 2758 2759 2760 2761 2762 2763
                              const char *cookiein,
                              int cookieinlen,
                              char **cookieout,
                              int *cookieoutlen,
                              const char *uri_in,
                              char **uri_out,
                              virDomainDefPtr *def,
                              const char *origname,
                              const char *listenAddress,
                              size_t nmigrate_disks,
                              const char **migrate_disks,
                              int nbdPort,
2764
                              qemuMigrationParamsPtr migParams,
2765
                              unsigned long flags)
2766
{
2767 2768
    unsigned short port = 0;
    bool autoPort = true;
2769 2770
    char *hostname = NULL;
    int ret = -1;
2771
    virURIPtr uri = NULL;
2772 2773
    virQEMUDriverConfigPtr cfg = virQEMUDriverGetConfig(driver);
    const char *migrateHost = cfg->migrateHost;
J
Jiri Denemark 已提交
2774

2775
    VIR_DEBUG("driver=%p, dconn=%p, cookiein=%s, cookieinlen=%d, "
2776
              "cookieout=%p, cookieoutlen=%p, uri_in=%s, uri_out=%p, "
2777
              "def=%p, origname=%s, listenAddress=%s, "
2778
              "nmigrate_disks=%zu, migrate_disks=%p, nbdPort=%d, flags=0x%lx",
2779
              driver, dconn, NULLSTR(cookiein), cookieinlen,
2780
              cookieout, cookieoutlen, NULLSTR(uri_in), uri_out,
2781
              *def, origname, NULLSTR(listenAddress),
2782
              nmigrate_disks, migrate_disks, nbdPort, flags);
2783

2784 2785
    *uri_out = NULL;

2786 2787 2788
    /* The URI passed in may be NULL or a string "tcp://somehostname:port".
     *
     * If the URI passed in is NULL then we allocate a port number
2789 2790 2791
     * from our pool of port numbers, and if the migrateHost is configured,
     * we return a URI of "tcp://migrateHost:port", otherwise return a URI
     * of "tcp://ourhostname:port".
2792 2793 2794 2795 2796 2797
     *
     * If the URI passed in is not NULL then we try to parse out the
     * port number and use that (note that the hostname is assumed
     * to be a correct hostname which refers to the target machine).
     */
    if (uri_in == NULL) {
2798 2799 2800
        bool encloseAddress = false;
        const char *incFormat;

2801
        if (virPortAllocatorAcquire(driver->migrationPorts, &port) < 0)
2802
            goto cleanup;
2803

2804
        if (migrateHost != NULL) {
2805 2806
            if (virSocketAddrNumericFamily(migrateHost) == AF_INET6)
                encloseAddress = true;
2807

2808
            if (VIR_STRDUP(hostname, migrateHost) < 0)
2809 2810 2811 2812 2813
                goto cleanup;
        } else {
            if ((hostname = virGetHostname()) == NULL)
                goto cleanup;
        }
2814 2815

        if (STRPREFIX(hostname, "localhost")) {
2816 2817 2818
            virReportError(VIR_ERR_INTERNAL_ERROR, "%s",
                           _("hostname on destination resolved to localhost,"
                             " but migration requires an FQDN"));
2819 2820 2821 2822 2823
            goto cleanup;
        }

        /* XXX this really should have been a properly well-formed
         * URI, but we can't add in tcp:// now without breaking
2824
         * compatibility with old targets. We at least make the
2825 2826
         * new targets accept both syntaxes though.
         */
2827 2828 2829 2830 2831 2832
        if (encloseAddress)
            incFormat = "%s:[%s]:%d";
        else
            incFormat = "%s:%s:%d";

        if (virAsprintf(uri_out, incFormat, "tcp", hostname, port) < 0)
2833 2834
            goto cleanup;
    } else {
2835
        bool well_formed_uri;
J
Ján Tomko 已提交
2836

2837
        if (!(uri = qemuMigrationAnyParseURI(uri_in, &well_formed_uri)))
2838
            goto cleanup;
J
Ján Tomko 已提交
2839

2840 2841 2842 2843 2844 2845 2846
        if (uri->scheme == NULL) {
            virReportError(VIR_ERR_INVALID_ARG,
                           _("missing scheme in migration URI: %s"),
                           uri_in);
            goto cleanup;
        }

M
Michael R. Hines 已提交
2847 2848
        if (STRNEQ(uri->scheme, "tcp") &&
            STRNEQ(uri->scheme, "rdma")) {
2849 2850 2851
            virReportError(VIR_ERR_ARGUMENT_UNSUPPORTED,
                           _("unsupported scheme %s in migration URI %s"),
                           uri->scheme, uri_in);
J
Ján Tomko 已提交
2852 2853 2854 2855 2856 2857 2858 2859 2860 2861
            goto cleanup;
        }

        if (uri->server == NULL) {
            virReportError(VIR_ERR_INVALID_ARG, _("missing host in migration"
                                                  " URI: %s"), uri_in);
            goto cleanup;
        }

        if (uri->port == 0) {
2862
            if (virPortAllocatorAcquire(driver->migrationPorts, &port) < 0)
2863
                goto cleanup;
2864

2865
            /* Send well-formed URI only if uri_in was well-formed */
2866 2867 2868 2869 2870 2871 2872 2873
            if (well_formed_uri) {
                uri->port = port;
                if (!(*uri_out = virURIFormat(uri)))
                    goto cleanup;
            } else {
                if (virAsprintf(uri_out, "%s:%d", uri_in, port) < 0)
                    goto cleanup;
            }
2874
        } else {
2875 2876
            port = uri->port;
            autoPort = false;
2877 2878 2879 2880 2881 2882
        }
    }

    if (*uri_out)
        VIR_DEBUG("Generated uri_out=%s", *uri_out);

2883
    ret = qemuMigrationDstPrepareAny(driver, dconn, cookiein, cookieinlen,
2884 2885 2886 2887
                                     cookieout, cookieoutlen, def, origname,
                                     NULL, uri ? uri->scheme : "tcp",
                                     port, autoPort, listenAddress,
                                     nmigrate_disks, migrate_disks, nbdPort,
2888
                                     migParams, flags);
2889
 cleanup:
2890
    virURIFree(uri);
2891
    VIR_FREE(hostname);
2892
    virObjectUnref(cfg);
2893
    if (ret != 0) {
2894
        VIR_FREE(*uri_out);
2895
        if (autoPort)
2896
            virPortAllocatorRelease(port);
2897
    }
2898 2899 2900 2901
    return ret;
}


2902
virDomainDefPtr
2903 2904 2905 2906
qemuMigrationAnyPrepareDef(virQEMUDriverPtr driver,
                           const char *dom_xml,
                           const char *dname,
                           char **origname)
2907 2908 2909
{
    virCapsPtr caps = NULL;
    virDomainDefPtr def;
2910
    char *name = NULL;
2911 2912 2913 2914 2915 2916 2917 2918 2919 2920

    if (!dom_xml) {
        virReportError(VIR_ERR_INTERNAL_ERROR, "%s",
                       _("no domain XML passed"));
        return NULL;
    }

    if (!(caps = virQEMUDriverGetCapabilities(driver, false)))
        return NULL;

2921
    if (!(def = virDomainDefParseString(dom_xml, caps, driver->xmlopt, NULL,
2922 2923
                                        VIR_DOMAIN_DEF_PARSE_INACTIVE |
                                        VIR_DOMAIN_DEF_PARSE_SKIP_VALIDATE)))
2924 2925 2926
        goto cleanup;

    if (dname) {
2927
        name = def->name;
2928 2929 2930 2931 2932 2933
        if (VIR_STRDUP(def->name, dname) < 0) {
            virDomainDefFree(def);
            def = NULL;
        }
    }

2934
 cleanup:
2935
    virObjectUnref(caps);
2936 2937 2938 2939
    if (def && origname)
        *origname = name;
    else
        VIR_FREE(name);
2940 2941 2942 2943
    return def;
}


2944
static int
2945 2946 2947 2948 2949 2950
qemuMigrationSrcConfirmPhase(virQEMUDriverPtr driver,
                             virDomainObjPtr vm,
                             const char *cookiein,
                             int cookieinlen,
                             unsigned int flags,
                             int retcode)
2951 2952
{
    qemuMigrationCookiePtr mig;
2953
    virObjectEventPtr event;
2954 2955
    int rv = -1;
    virQEMUDriverConfigPtr cfg = virQEMUDriverGetConfig(driver);
2956 2957
    qemuDomainObjPrivatePtr priv = vm->privateData;
    qemuDomainJobInfoPtr jobInfo = NULL;
2958

2959
    VIR_DEBUG("driver=%p, vm=%p, cookiein=%s, cookieinlen=%d, "
2960
              "flags=0x%x, retcode=%d",
2961
              driver, vm, NULLSTR(cookiein), cookieinlen,
2962 2963 2964 2965 2966 2967 2968 2969 2970
              flags, retcode);

    virCheckFlags(QEMU_MIGRATION_FLAGS, -1);

    qemuMigrationJobSetPhase(driver, vm,
                             retcode == 0
                             ? QEMU_MIGRATION_PHASE_CONFIRM3
                             : QEMU_MIGRATION_PHASE_CONFIRM3_CANCELLED);

2971 2972
    if (!(mig = qemuMigrationEatCookie(driver, vm->def, priv->origname, priv,
                                       cookiein, cookieinlen,
2973
                                       QEMU_MIGRATION_COOKIE_STATS)))
2974 2975
        goto cleanup;

2976 2977 2978
    if (retcode == 0)
        jobInfo = priv->job.completed;
    else
2979
        VIR_FREE(priv->job.completed);
2980 2981 2982

    /* Update times with the values sent by the destination daemon */
    if (mig->jobInfo && jobInfo) {
2983 2984 2985 2986 2987 2988 2989 2990
        int reason;

        /* We need to refresh migration statistics after a completed post-copy
         * migration since priv->job.completed contains obsolete data from the
         * time we switched to post-copy mode.
         */
        if (virDomainObjGetState(vm, &reason) == VIR_DOMAIN_PAUSED &&
            reason == VIR_DOMAIN_PAUSED_POSTCOPY &&
2991 2992
            qemuMigrationAnyFetchStats(driver, vm, QEMU_ASYNC_JOB_MIGRATION_OUT,
                                       jobInfo, NULL) < 0)
2993 2994
            VIR_WARN("Could not refresh migration statistics");

2995 2996 2997
        qemuDomainJobInfoUpdateTime(jobInfo);
        jobInfo->timeDeltaSet = mig->jobInfo->timeDeltaSet;
        jobInfo->timeDelta = mig->jobInfo->timeDelta;
2998 2999
        jobInfo->stats.mig.downtime_set = mig->jobInfo->stats.mig.downtime_set;
        jobInfo->stats.mig.downtime = mig->jobInfo->stats.mig.downtime;
3000 3001
    }

3002 3003 3004
    if (flags & VIR_MIGRATE_OFFLINE)
        goto done;

3005 3006
    /* Did the migration go as planned?  If yes, kill off the domain object.
     * If something failed, resume CPUs, but only if we didn't use post-copy.
3007 3008 3009 3010
     */
    if (retcode == 0) {
        /* If guest uses SPICE and supports seamless migration we have to hold
         * up domain shutdown until SPICE server transfers its data */
3011
        qemuMigrationSrcWaitForSpice(vm);
3012 3013

        qemuProcessStop(driver, vm, VIR_DOMAIN_SHUTOFF_MIGRATED,
3014
                        QEMU_ASYNC_JOB_MIGRATION_OUT,
3015 3016 3017
                        VIR_QEMU_PROCESS_STOP_MIGRATED);
        virDomainAuditStop(vm, "migrated");

3018
        event = virDomainEventLifecycleNewFromObj(vm,
3019 3020
                                         VIR_DOMAIN_EVENT_STOPPED,
                                         VIR_DOMAIN_EVENT_STOPPED_MIGRATED);
3021
        virObjectEventStateQueue(driver->domainEventState, event);
3022
        qemuDomainEventEmitJobCompleted(driver, vm);
3023
    } else {
3024
        virErrorPtr orig_err = virSaveLastError();
3025
        int reason;
3026 3027

        /* cancel any outstanding NBD jobs */
3028 3029
        qemuMigrationSrcNBDCopyCancel(driver, vm, false,
                                      QEMU_ASYNC_JOB_MIGRATION_OUT, NULL);
3030 3031 3032

        virSetError(orig_err);
        virFreeError(orig_err);
3033

3034
        if (virDomainObjGetState(vm, &reason) == VIR_DOMAIN_PAUSED &&
3035
            reason == VIR_DOMAIN_PAUSED_POSTCOPY)
3036
            qemuMigrationAnyPostcopyFailed(driver, vm);
3037 3038
        else
            qemuMigrationSrcRestoreDomainState(driver, vm);
3039

3040
        qemuMigrationParamsReset(driver, vm, QEMU_ASYNC_JOB_MIGRATION_OUT,
3041
                                 priv->job.migParams, priv->job.apiFlags);
3042

3043
        if (virDomainSaveStatus(driver->xmlopt, cfg->stateDir, vm, driver->caps) < 0)
3044 3045 3046
            VIR_WARN("Failed to save status on vm %s", vm->def->name);
    }

3047
 done:
3048 3049 3050
    qemuMigrationCookieFree(mig);
    rv = 0;

3051
 cleanup:
3052 3053 3054 3055 3056
    virObjectUnref(cfg);
    return rv;
}

int
3057 3058 3059 3060 3061 3062
qemuMigrationSrcConfirm(virQEMUDriverPtr driver,
                        virDomainObjPtr vm,
                        const char *cookiein,
                        int cookieinlen,
                        unsigned int flags,
                        int cancelled)
3063
{
3064
    qemuMigrationJobPhase phase;
3065 3066 3067 3068 3069 3070 3071 3072 3073 3074 3075 3076 3077 3078
    virQEMUDriverConfigPtr cfg = NULL;
    int ret = -1;

    cfg = virQEMUDriverGetConfig(driver);

    if (!qemuMigrationJobIsActive(vm, QEMU_ASYNC_JOB_MIGRATION_OUT))
        goto cleanup;

    if (cancelled)
        phase = QEMU_MIGRATION_PHASE_CONFIRM3_CANCELLED;
    else
        phase = QEMU_MIGRATION_PHASE_CONFIRM3;

    qemuMigrationJobStartPhase(driver, vm, phase);
3079
    virCloseCallbacksUnset(driver->closeCallbacks, vm,
3080
                           qemuMigrationSrcCleanup);
3081

3082 3083 3084
    ret = qemuMigrationSrcConfirmPhase(driver, vm,
                                       cookiein, cookieinlen,
                                       flags, cancelled);
3085

3086
    qemuMigrationJobFinish(driver, vm);
3087
    if (!virDomainObjIsActive(vm)) {
3088
        if (!cancelled && ret == 0 && flags & VIR_MIGRATE_UNDEFINE_SOURCE) {
3089
            virDomainDeleteConfig(cfg->configDir, cfg->autostartDir, vm);
3090 3091
            vm->persistent = 0;
        }
3092
        qemuDomainRemoveInactiveJob(driver, vm);
3093 3094
    }

3095
 cleanup:
M
Michal Privoznik 已提交
3096
    virDomainObjEndAPI(&vm);
3097 3098 3099 3100 3101
    virObjectUnref(cfg);
    return ret;
}


3102 3103
enum qemuMigrationDestinationType {
    MIGRATION_DEST_HOST,
3104
    MIGRATION_DEST_CONNECT_HOST,
3105
    MIGRATION_DEST_FD,
3106
};
3107

3108 3109 3110 3111
enum qemuMigrationForwardType {
    MIGRATION_FWD_DIRECT,
    MIGRATION_FWD_STREAM,
};
3112

3113 3114 3115 3116 3117 3118
typedef struct _qemuMigrationSpec qemuMigrationSpec;
typedef qemuMigrationSpec *qemuMigrationSpecPtr;
struct _qemuMigrationSpec {
    enum qemuMigrationDestinationType destType;
    union {
        struct {
3119
            const char *protocol;
3120 3121 3122 3123
            const char *name;
            int port;
        } host;

3124 3125 3126 3127
        struct {
            int qemu;
            int local;
        } fd;
3128 3129 3130 3131 3132 3133 3134
    } dest;

    enum qemuMigrationForwardType fwdType;
    union {
        virStreamPtr stream;
    } fwd;
};
3135 3136 3137

#define TUNNEL_SEND_BUF_SIZE 65536

3138 3139 3140 3141 3142 3143 3144
typedef struct _qemuMigrationIOThread qemuMigrationIOThread;
typedef qemuMigrationIOThread *qemuMigrationIOThreadPtr;
struct _qemuMigrationIOThread {
    virThread thread;
    virStreamPtr st;
    int sock;
    virError err;
3145 3146
    int wakeupRecvFD;
    int wakeupSendFD;
3147 3148
};

3149
static void qemuMigrationSrcIOFunc(void *arg)
3150
{
3151
    qemuMigrationIOThreadPtr data = arg;
3152 3153 3154 3155 3156 3157 3158
    char *buffer = NULL;
    struct pollfd fds[2];
    int timeout = -1;
    virErrorPtr err = NULL;

    VIR_DEBUG("Running migration tunnel; stream=%p, sock=%d",
              data->st, data->sock);
3159

3160
    if (VIR_ALLOC_N(buffer, TUNNEL_SEND_BUF_SIZE) < 0)
3161
        goto abrt;
3162

3163 3164 3165
    fds[0].fd = data->sock;
    fds[1].fd = data->wakeupRecvFD;

3166
    for (;;) {
3167 3168 3169 3170 3171 3172 3173 3174 3175 3176
        int ret;

        fds[0].events = fds[1].events = POLLIN;
        fds[0].revents = fds[1].revents = 0;

        ret = poll(fds, ARRAY_CARDINALITY(fds), timeout);

        if (ret < 0) {
            if (errno == EAGAIN || errno == EINTR)
                continue;
3177
            virReportSystemError(errno, "%s",
3178 3179
                                 _("poll failed in migration tunnel"));
            goto abrt;
3180
        }
3181 3182 3183 3184 3185 3186 3187

        if (ret == 0) {
            /* We were asked to gracefully stop but reading would block. This
             * can only happen if qemu told us migration finished but didn't
             * close the migration fd. We handle this in the same way as EOF.
             */
            VIR_DEBUG("QEMU forgot to close migration fd");
3188
            break;
3189
        }
3190

3191 3192 3193 3194 3195 3196 3197 3198 3199 3200 3201 3202 3203 3204 3205 3206
        if (fds[1].revents & (POLLIN | POLLERR | POLLHUP)) {
            char stop = 0;

            if (saferead(data->wakeupRecvFD, &stop, 1) != 1) {
                virReportSystemError(errno, "%s",
                                     _("failed to read from wakeup fd"));
                goto abrt;
            }

            VIR_DEBUG("Migration tunnel was asked to %s",
                      stop ? "abort" : "finish");
            if (stop) {
                goto abrt;
            } else {
                timeout = 0;
            }
3207 3208
        }

3209 3210 3211 3212 3213 3214 3215 3216 3217 3218 3219 3220 3221 3222 3223 3224 3225
        if (fds[0].revents & (POLLIN | POLLERR | POLLHUP)) {
            int nbytes;

            nbytes = saferead(data->sock, buffer, TUNNEL_SEND_BUF_SIZE);
            if (nbytes > 0) {
                if (virStreamSend(data->st, buffer, nbytes) < 0)
                    goto error;
            } else if (nbytes < 0) {
                virReportSystemError(errno, "%s",
                        _("tunnelled migration failed to read from qemu"));
                goto abrt;
            } else {
                /* EOF; get out of here */
                break;
            }
        }
    }
3226

3227 3228
    if (virStreamFinish(data->st) < 0)
        goto error;
3229

3230
    VIR_FORCE_CLOSE(data->sock);
3231 3232
    VIR_FREE(buffer);

3233 3234
    return;

3235
 abrt:
3236 3237 3238 3239 3240 3241 3242 3243 3244 3245 3246
    err = virSaveLastError();
    if (err && err->code == VIR_ERR_OK) {
        virFreeError(err);
        err = NULL;
    }
    virStreamAbort(data->st);
    if (err) {
        virSetError(err);
        virFreeError(err);
    }

3247
 error:
3248 3249 3250 3251 3252
    /* Let the source qemu know that the transfer cant continue anymore.
     * Don't copy the error for EPIPE as destination has the actual error. */
    VIR_FORCE_CLOSE(data->sock);
    if (!virLastErrorIsSystemErrno(EPIPE))
        virCopyLastError(&data->err);
3253
    virResetLastError();
3254
    VIR_FREE(buffer);
3255 3256 3257 3258
}


static qemuMigrationIOThreadPtr
3259 3260
qemuMigrationSrcStartTunnel(virStreamPtr st,
                            int sock)
3261
{
3262 3263
    qemuMigrationIOThreadPtr io = NULL;
    int wakeupFD[2] = { -1, -1 };
3264

3265 3266 3267 3268
    if (pipe2(wakeupFD, O_CLOEXEC) < 0) {
        virReportSystemError(errno, "%s",
                             _("Unable to make pipe"));
        goto error;
3269 3270
    }

3271
    if (VIR_ALLOC(io) < 0)
3272
        goto error;
3273

3274 3275
    io->st = st;
    io->sock = sock;
3276 3277
    io->wakeupRecvFD = wakeupFD[0];
    io->wakeupSendFD = wakeupFD[1];
3278 3279

    if (virThreadCreate(&io->thread, true,
3280
                        qemuMigrationSrcIOFunc,
3281 3282 3283
                        io) < 0) {
        virReportSystemError(errno, "%s",
                             _("Unable to create migration thread"));
3284
        goto error;
3285 3286 3287
    }

    return io;
3288

3289
 error:
3290 3291 3292 3293
    VIR_FORCE_CLOSE(wakeupFD[0]);
    VIR_FORCE_CLOSE(wakeupFD[1]);
    VIR_FREE(io);
    return NULL;
3294 3295 3296
}

static int
3297
qemuMigrationSrcStopTunnel(qemuMigrationIOThreadPtr io, bool error)
3298 3299
{
    int rv = -1;
3300 3301 3302 3303 3304 3305 3306 3307 3308
    char stop = error ? 1 : 0;

    /* make sure the thread finishes its job and is joinable */
    if (safewrite(io->wakeupSendFD, &stop, 1) != 1) {
        virReportSystemError(errno, "%s",
                             _("failed to wakeup migration tunnel"));
        goto cleanup;
    }

3309 3310 3311 3312
    virThreadJoin(&io->thread);

    /* Forward error from the IO thread, to this thread */
    if (io->err.code != VIR_ERR_OK) {
3313 3314 3315 3316
        if (error)
            rv = 0;
        else
            virSetError(&io->err);
3317 3318 3319 3320 3321 3322
        virResetError(&io->err);
        goto cleanup;
    }

    rv = 0;

3323
 cleanup:
3324 3325
    VIR_FORCE_CLOSE(io->wakeupSendFD);
    VIR_FORCE_CLOSE(io->wakeupRecvFD);
3326 3327
    VIR_FREE(io);
    return rv;
3328 3329
}

3330
static int
3331 3332 3333
qemuMigrationSrcConnect(virQEMUDriverPtr driver,
                        virDomainObjPtr vm,
                        qemuMigrationSpecPtr spec)
3334 3335 3336 3337 3338 3339 3340
{
    virNetSocketPtr sock;
    const char *host;
    char *port = NULL;
    int ret = -1;

    host = spec->dest.host.name;
3341
    if (virAsprintf(&port, "%d", spec->dest.host.port) < 0)
3342 3343 3344 3345 3346
        return -1;

    spec->destType = MIGRATION_DEST_FD;
    spec->dest.fd.qemu = -1;

3347
    if (qemuSecuritySetSocketLabel(driver->securityManager, vm->def) < 0)
3348
        goto cleanup;
3349 3350 3351
    if (virNetSocketNewConnectTCP(host, port,
                                  AF_UNSPEC,
                                  &sock) == 0) {
3352
        spec->dest.fd.qemu = virNetSocketDupFD(sock, true);
3353
        virObjectUnref(sock);
3354
    }
3355
    if (qemuSecurityClearSocketLabel(driver->securityManager, vm->def) < 0 ||
3356 3357 3358
        spec->dest.fd.qemu == -1)
        goto cleanup;

3359 3360 3361 3362 3363 3364 3365
    /* Migration expects a blocking FD */
    if (virSetBlocking(spec->dest.fd.qemu, true) < 0) {
        virReportSystemError(errno, _("Unable to set FD %d blocking"),
                             spec->dest.fd.qemu);
        goto cleanup;
    }

3366 3367
    ret = 0;

3368
 cleanup:
3369 3370 3371 3372 3373 3374
    VIR_FREE(port);
    if (ret < 0)
        VIR_FORCE_CLOSE(spec->dest.fd.qemu);
    return ret;
}

3375 3376

static int
3377 3378 3379 3380
qemuMigrationSrcContinue(virQEMUDriverPtr driver,
                         virDomainObjPtr vm,
                         qemuMonitorMigrationStatus status,
                         qemuDomainAsyncJob asyncJob)
3381 3382 3383 3384 3385 3386 3387 3388 3389 3390 3391 3392 3393 3394 3395 3396
{
    qemuDomainObjPrivatePtr priv = vm->privateData;
    int ret;

    if (qemuDomainObjEnterMonitorAsync(driver, vm, asyncJob) < 0)
        return -1;

    ret = qemuMonitorMigrateContinue(priv->mon, status);

    if (qemuDomainObjExitMonitor(driver, vm) < 0)
        ret = -1;

    return ret;
}


3397
static int
3398 3399 3400 3401 3402 3403 3404 3405 3406 3407 3408 3409 3410 3411
qemuMigrationSrcRun(virQEMUDriverPtr driver,
                    virDomainObjPtr vm,
                    const char *persist_xml,
                    const char *cookiein,
                    int cookieinlen,
                    char **cookieout,
                    int *cookieoutlen,
                    unsigned long flags,
                    unsigned long resource,
                    qemuMigrationSpecPtr spec,
                    virConnectPtr dconn,
                    const char *graphicsuri,
                    size_t nmigrate_disks,
                    const char **migrate_disks,
3412
                    qemuMigrationParamsPtr migParams)
3413
{
3414
    int ret = -1;
3415 3416
    unsigned int migrate_flags = QEMU_MONITOR_MIGRATE_BACKGROUND;
    qemuDomainObjPrivatePtr priv = vm->privateData;
3417
    qemuMigrationCookiePtr mig = NULL;
3418
    char *tlsAlias = NULL;
3419
    qemuMigrationIOThreadPtr iothread = NULL;
3420
    int fd = -1;
3421
    unsigned long migrate_speed = resource ? resource : priv->migMaxBandwidth;
3422
    virErrorPtr orig_err = NULL;
3423
    unsigned int cookieFlags = 0;
3424
    bool abort_on_error = !!(flags & VIR_MIGRATE_ABORT_ON_ERROR);
3425
    bool events = virQEMUCapsGet(priv->qemuCaps, QEMU_CAPS_MIGRATION_EVENT);
3426
    bool cancel = false;
3427
    unsigned int waitFlags;
3428
    virDomainDefPtr persistDef = NULL;
3429
    char *timestamp;
3430
    int rc;
3431 3432

    VIR_DEBUG("driver=%p, vm=%p, cookiein=%s, cookieinlen=%d, "
3433
              "cookieout=%p, cookieoutlen=%p, flags=0x%lx, resource=%lu, "
3434 3435
              "spec=%p (dest=%d, fwd=%d), dconn=%p, graphicsuri=%s, "
              "nmigrate_disks=%zu, migrate_disks=%p",
3436 3437
              driver, vm, NULLSTR(cookiein), cookieinlen,
              cookieout, cookieoutlen, flags, resource,
3438
              spec, spec->destType, spec->fwdType, dconn,
3439
              NULLSTR(graphicsuri), nmigrate_disks, migrate_disks);
3440

3441 3442 3443 3444 3445 3446 3447 3448 3449 3450
    if (flags & VIR_MIGRATE_NON_SHARED_DISK) {
        migrate_flags |= QEMU_MONITOR_MIGRATE_NON_SHARED_DISK;
        cookieFlags |= QEMU_MIGRATION_COOKIE_NBD;
    }

    if (flags & VIR_MIGRATE_NON_SHARED_INC) {
        migrate_flags |= QEMU_MONITOR_MIGRATE_NON_SHARED_INC;
        cookieFlags |= QEMU_MIGRATION_COOKIE_NBD;
    }

3451 3452
    if (virLockManagerPluginUsesState(driver->lockManager) &&
        !cookieout) {
3453 3454 3455 3456
        virReportError(VIR_ERR_INTERNAL_ERROR,
                       _("Migration with lock driver %s requires"
                         " cookie support"),
                       virLockManagerPluginGetName(driver->lockManager));
3457 3458 3459
        return -1;
    }

3460 3461 3462
    if (events)
        priv->signalIOError = abort_on_error;

3463 3464
    if (flags & VIR_MIGRATE_PERSIST_DEST) {
        if (persist_xml) {
3465 3466
            if (!(persistDef = qemuMigrationAnyPrepareDef(driver, persist_xml,
                                                          NULL, NULL)))
3467
                goto error;
3468 3469 3470
        } else {
            virDomainDefPtr def = vm->newDef ? vm->newDef : vm->def;
            if (!(persistDef = qemuDomainDefCopy(driver, def,
3471 3472
                                                 VIR_DOMAIN_XML_SECURE |
                                                 VIR_DOMAIN_XML_MIGRATABLE)))
3473
                goto error;
3474 3475 3476
        }
    }

3477 3478
    mig = qemuMigrationEatCookie(driver, vm->def, priv->origname, priv,
                                 cookiein, cookieinlen,
3479 3480 3481
                                 cookieFlags |
                                 QEMU_MIGRATION_COOKIE_GRAPHICS |
                                 QEMU_MIGRATION_COOKIE_CAPS);
3482
    if (!mig)
3483
        goto error;
3484

3485
    if (qemuMigrationSrcGraphicsRelocate(driver, vm, mig, graphicsuri) < 0)
3486 3487
        VIR_WARN("unable to provide data for graphics client relocation");

3488
    if (qemuMigrationParamsCheck(driver, vm, QEMU_ASYNC_JOB_MIGRATION_OUT,
3489
                                 migParams, mig->caps->automatic) < 0)
3490 3491
        goto error;

3492
    if (flags & VIR_MIGRATE_TLS) {
3493
        const char *hostname = NULL;
3494 3495 3496 3497

        /* We need to add tls-hostname whenever QEMU itself does not
         * connect directly to the destination. */
        if (spec->destType == MIGRATION_DEST_CONNECT_HOST ||
3498 3499 3500
            spec->destType == MIGRATION_DEST_FD)
            hostname = spec->dest.host.name;

3501
        if (qemuMigrationParamsEnableTLS(driver, vm, false,
3502
                                         QEMU_ASYNC_JOB_MIGRATION_OUT,
3503
                                         &tlsAlias, hostname,
3504 3505
                                         migParams) < 0)
            goto error;
3506
    } else {
3507
        if (qemuMigrationParamsDisableTLS(vm, migParams) < 0)
3508
            goto error;
3509 3510
    }

3511 3512 3513 3514
    if (qemuMigrationParamsApply(driver, vm, QEMU_ASYNC_JOB_MIGRATION_OUT,
                                 migParams) < 0)
        goto error;

3515 3516 3517
    if (migrate_flags & (QEMU_MONITOR_MIGRATE_NON_SHARED_DISK |
                         QEMU_MONITOR_MIGRATE_NON_SHARED_INC)) {
        if (mig->nbd) {
3518 3519 3520 3521
            /* Currently libvirt does not support setting up of the NBD
             * non-shared storage migration with TLS. As we need to honour the
             * VIR_MIGRATE_TLS flag, we need to reject such migration until
             * we implement TLS for NBD. */
3522 3523
            if (flags & VIR_MIGRATE_TLS &&
                !virQEMUCapsGet(priv->qemuCaps, QEMU_CAPS_BLOCKDEV_DEL)) {
3524 3525 3526 3527 3528
                virReportError(VIR_ERR_OPERATION_UNSUPPORTED, "%s",
                               _("NBD migration with TLS is not supported"));
                goto error;
            }

3529
            /* This will update migrate_flags on success */
3530 3531 3532 3533 3534 3535
            if (qemuMigrationSrcNBDStorageCopy(driver, vm, mig,
                                               spec->dest.host.name,
                                               migrate_speed,
                                               &migrate_flags,
                                               nmigrate_disks,
                                               migrate_disks,
3536
                                               dconn, tlsAlias, flags) < 0) {
3537
                goto error;
3538 3539 3540 3541 3542 3543 3544
            }
        } else {
            /* Destination doesn't support NBD server.
             * Fall back to previous implementation. */
            VIR_DEBUG("Destination doesn't support NBD server "
                      "Falling back to previous implementation.");
        }
3545 3546
    }

3547
    /* Before EnterMonitor, since qemuMigrationSetOffline already does that */
3548 3549
    if (!(flags & VIR_MIGRATE_LIVE) &&
        virDomainObjGetState(vm, NULL) == VIR_DOMAIN_RUNNING) {
3550
        if (qemuMigrationSrcSetOffline(driver, vm) < 0)
3551
            goto error;
3552 3553
    }

3554 3555
    if (qemuDomainObjEnterMonitorAsync(driver, vm,
                                       QEMU_ASYNC_JOB_MIGRATION_OUT) < 0)
3556
        goto error;
3557

3558
    if (priv->job.abortJob) {
3559 3560
        /* explicitly do this *after* we entered the monitor,
         * as this is a critical section so we are guaranteed
3561
         * priv->job.abortJob will not change */
3562
        priv->job.current->status = QEMU_DOMAIN_JOB_STATUS_CANCELED;
3563 3564 3565
        virReportError(VIR_ERR_OPERATION_ABORTED, _("%s: %s"),
                       qemuDomainAsyncJobTypeToString(priv->job.asyncJob),
                       _("canceled by client"));
3566
        goto exit_monitor;
3567 3568
    }

3569 3570
    if (qemuMonitorSetMigrationSpeed(priv->mon, migrate_speed) < 0)
        goto exit_monitor;
3571

3572 3573
    /* connect to the destination qemu if needed */
    if (spec->destType == MIGRATION_DEST_CONNECT_HOST &&
3574
        qemuMigrationSrcConnect(driver, vm, spec) < 0) {
3575
        goto exit_monitor;
3576
    }
3577

3578 3579 3580 3581 3582 3583
    /* log start of migration */
    if ((timestamp = virTimeStringNow()) != NULL) {
        qemuDomainLogAppendMessage(driver, vm, "%s: initiating migration\n", timestamp);
        VIR_FREE(timestamp);
    }

3584
    rc = -1;
3585 3586
    switch (spec->destType) {
    case MIGRATION_DEST_HOST:
M
Michael R. Hines 已提交
3587 3588
        if (STREQ(spec->dest.host.protocol, "rdma") &&
            virProcessSetMaxMemLock(vm->pid, vm->def->mem.hard_limit << 10) < 0) {
3589
            goto exit_monitor;
M
Michael R. Hines 已提交
3590
        }
3591 3592 3593 3594
        rc = qemuMonitorMigrateToHost(priv->mon, migrate_flags,
                                      spec->dest.host.protocol,
                                      spec->dest.host.name,
                                      spec->dest.host.port);
3595 3596
        break;

3597 3598 3599 3600
    case MIGRATION_DEST_CONNECT_HOST:
        /* handled above and transformed into MIGRATION_DEST_FD */
        break;

3601
    case MIGRATION_DEST_FD:
3602
        if (spec->fwdType != MIGRATION_FWD_DIRECT) {
3603
            fd = spec->dest.fd.local;
3604 3605
            spec->dest.fd.local = -1;
        }
3606 3607
        rc = qemuMonitorMigrateToFd(priv->mon, migrate_flags,
                                    spec->dest.fd.qemu);
3608 3609
        VIR_FORCE_CLOSE(spec->dest.fd.qemu);
        break;
3610
    }
3611 3612

    if (qemuDomainObjExitMonitor(driver, vm) < 0 || rc < 0)
3613
        goto error;
3614 3615 3616

    /* From this point onwards we *must* call cancel to abort the
     * migration on source if anything goes wrong */
3617
    cancel = true;
3618

3619
    if (spec->fwdType != MIGRATION_FWD_DIRECT) {
3620
        if (!(iothread = qemuMigrationSrcStartTunnel(spec->fwd.stream, fd)))
3621
            goto error;
3622 3623 3624 3625 3626
        /* If we've created a tunnel, then the 'fd' will be closed in the
         * qemuMigrationIOFunc as data->sock.
         */
        fd = -1;
    }
3627

3628
    waitFlags = QEMU_MIGRATION_COMPLETED_PRE_SWITCHOVER;
3629 3630 3631 3632 3633 3634 3635
    if (abort_on_error)
        waitFlags |= QEMU_MIGRATION_COMPLETED_ABORT_ON_ERROR;
    if (mig->nbd)
        waitFlags |= QEMU_MIGRATION_COMPLETED_CHECK_STORAGE;
    if (flags & VIR_MIGRATE_POSTCOPY)
        waitFlags |= QEMU_MIGRATION_COMPLETED_POSTCOPY;

3636 3637 3638
    rc = qemuMigrationSrcWaitForCompletion(driver, vm,
                                           QEMU_ASYNC_JOB_MIGRATION_OUT,
                                           dconn, waitFlags);
3639
    if (rc == -2) {
3640
        goto error;
3641 3642 3643 3644 3645
    } else if (rc == -1) {
        /* QEMU reported failed migration, nothing to cancel anymore */
        cancel = false;
        goto error;
    }
3646

3647 3648 3649
    /* When migration completed, QEMU will have paused the CPUs for us.
     * Wait for the STOP event to be processed or explicitly stop CPUs
     * (for old QEMU which does not send events) to release the lock state.
3650
     */
3651 3652 3653 3654 3655
    if (priv->monJSON) {
        while (virDomainObjGetState(vm, NULL) == VIR_DOMAIN_RUNNING) {
            priv->signalStop = true;
            rc = virDomainObjWait(vm);
            priv->signalStop = false;
3656
            if (rc < 0)
3657
                goto error;
3658
        }
3659
    } else if (virDomainObjGetState(vm, NULL) == VIR_DOMAIN_RUNNING &&
3660
               qemuMigrationSrcSetOffline(driver, vm) < 0) {
3661
        goto error;
3662
    }
3663

J
Ján Tomko 已提交
3664
    if (mig->nbd &&
3665 3666 3667
        qemuMigrationSrcNBDCopyCancel(driver, vm, true,
                                      QEMU_ASYNC_JOB_MIGRATION_OUT,
                                      dconn) < 0)
3668
        goto error;
3669

3670 3671 3672 3673 3674
    /* When migration was paused before serializing device state we need to
     * resume it now once we finished all block jobs and wait for the real
     * end of the migration.
     */
    if (priv->job.current->status == QEMU_DOMAIN_JOB_STATUS_PAUSED) {
3675 3676 3677
        if (qemuMigrationSrcContinue(driver, vm,
                                     QEMU_MONITOR_MIGRATION_STATUS_PRE_SWITCHOVER,
                                     QEMU_ASYNC_JOB_MIGRATION_OUT) < 0)
3678 3679 3680 3681
            goto error;

        waitFlags ^= QEMU_MIGRATION_COMPLETED_PRE_SWITCHOVER;

3682 3683 3684
        rc = qemuMigrationSrcWaitForCompletion(driver, vm,
                                               QEMU_ASYNC_JOB_MIGRATION_OUT,
                                               dconn, waitFlags);
3685 3686 3687 3688 3689 3690 3691 3692 3693
        if (rc == -2) {
            goto error;
        } else if (rc == -1) {
            /* QEMU reported failed migration, nothing to cancel anymore */
            cancel = false;
            goto error;
        }
    }

3694 3695 3696 3697
    if (iothread) {
        qemuMigrationIOThreadPtr io;

        VIR_STEAL_PTR(io, iothread);
3698
        if (qemuMigrationSrcStopTunnel(io, false) < 0)
3699
            goto error;
3700 3701 3702
    }

    if (priv->job.completed) {
3703
        priv->job.completed->stopped = priv->job.current->stopped;
3704 3705 3706 3707 3708 3709 3710 3711 3712
        qemuDomainJobInfoUpdateTime(priv->job.completed);
        qemuDomainJobInfoUpdateDowntime(priv->job.completed);
        ignore_value(virTimeMillisNow(&priv->job.completed->sent));
    }

    cookieFlags |= QEMU_MIGRATION_COOKIE_NETWORK |
                   QEMU_MIGRATION_COOKIE_STATS;

    if (qemuMigrationCookieAddPersistent(mig, &persistDef) < 0 ||
3713 3714 3715
        qemuMigrationBakeCookie(mig, driver, vm,
                                QEMU_MIGRATION_SOURCE,
                                cookieout, cookieoutlen, cookieFlags) < 0) {
3716 3717
        VIR_WARN("Unable to encode migration cookie");
    }
3718

3719
    ret = 0;
3720

3721
 cleanup:
3722
    VIR_FREE(tlsAlias);
3723
    VIR_FORCE_CLOSE(fd);
3724
    virDomainDefFree(persistDef);
3725 3726
    qemuMigrationCookieFree(mig);

3727 3728 3729
    if (events)
        priv->signalIOError = false;

3730 3731 3732 3733 3734
    if (orig_err) {
        virSetError(orig_err);
        virFreeError(orig_err);
    }

3735 3736
    return ret;

3737
 error:
3738 3739
    orig_err = virSaveLastError();

3740 3741 3742 3743 3744 3745 3746 3747
    if (virDomainObjIsActive(vm)) {
        if (cancel &&
            priv->job.current->status != QEMU_DOMAIN_JOB_STATUS_QEMU_COMPLETED &&
            qemuDomainObjEnterMonitorAsync(driver, vm,
                                           QEMU_ASYNC_JOB_MIGRATION_OUT) == 0) {
            qemuMonitorMigrateCancel(priv->mon);
            ignore_value(qemuDomainObjExitMonitor(driver, vm));
        }
3748

3749 3750
        /* cancel any outstanding NBD jobs */
        if (mig && mig->nbd)
3751 3752 3753
            qemuMigrationSrcNBDCopyCancel(driver, vm, false,
                                          QEMU_ASYNC_JOB_MIGRATION_OUT,
                                          dconn);
3754 3755 3756 3757

        if (priv->job.current->status != QEMU_DOMAIN_JOB_STATUS_CANCELED)
            priv->job.current->status = QEMU_DOMAIN_JOB_STATUS_FAILED;
    }
3758 3759

    if (iothread)
3760
        qemuMigrationSrcStopTunnel(iothread, true);
3761 3762 3763

    goto cleanup;

3764 3765
 exit_monitor:
    ignore_value(qemuDomainObjExitMonitor(driver, vm));
3766
    goto error;
3767 3768
}

3769
/* Perform migration using QEMU's native migrate support,
3770 3771
 * not encrypted obviously
 */
3772 3773 3774 3775 3776 3777 3778 3779 3780 3781 3782 3783 3784 3785 3786
static int
qemuMigrationSrcPerformNative(virQEMUDriverPtr driver,
                              virDomainObjPtr vm,
                              const char *persist_xml,
                              const char *uri,
                              const char *cookiein,
                              int cookieinlen,
                              char **cookieout,
                              int *cookieoutlen,
                              unsigned long flags,
                              unsigned long resource,
                              virConnectPtr dconn,
                              const char *graphicsuri,
                              size_t nmigrate_disks,
                              const char **migrate_disks,
3787
                              qemuMigrationParamsPtr migParams)
3788
{
3789
    qemuDomainObjPrivatePtr priv = vm->privateData;
M
Martin Kletzander 已提交
3790
    virURIPtr uribits = NULL;
3791
    int ret = -1;
3792 3793 3794
    qemuMigrationSpec spec;

    VIR_DEBUG("driver=%p, vm=%p, uri=%s, cookiein=%s, cookieinlen=%d, "
3795
              "cookieout=%p, cookieoutlen=%p, flags=0x%lx, resource=%lu, "
3796
              "graphicsuri=%s, nmigrate_disks=%zu migrate_disks=%p",
3797
              driver, vm, uri, NULLSTR(cookiein), cookieinlen,
3798
              cookieout, cookieoutlen, flags, resource,
3799
              NULLSTR(graphicsuri), nmigrate_disks, migrate_disks);
3800

3801
    if (!(uribits = qemuMigrationAnyParseURI(uri, NULL)))
3802 3803
        return -1;

3804 3805 3806 3807 3808 3809 3810
    if (uribits->scheme == NULL) {
        virReportError(VIR_ERR_INTERNAL_ERROR,
                       _("missing scheme in migration URI: %s"),
                       uri);
        goto cleanup;
    }

M
Michael R. Hines 已提交
3811 3812 3813 3814 3815
    if (STREQ(uribits->scheme, "rdma")) {
        if (!virQEMUCapsGet(priv->qemuCaps, QEMU_CAPS_MIGRATE_RDMA)) {
            virReportError(VIR_ERR_OPERATION_UNSUPPORTED, "%s",
                           _("outgoing RDMA migration is not supported "
                             "with this QEMU binary"));
3816
            goto cleanup;
M
Michael R. Hines 已提交
3817
        }
3818
        if (!virMemoryLimitIsSet(vm->def->mem.hard_limit)) {
M
Michael R. Hines 已提交
3819 3820 3821
            virReportError(VIR_ERR_OPERATION_INVALID, "%s",
                           _("cannot start RDMA migration with no memory hard "
                             "limit set"));
3822
            goto cleanup;
M
Michael R. Hines 已提交
3823 3824 3825
        }
    }

3826
    if (STRNEQ(uribits->scheme, "rdma"))
3827 3828
        spec.destType = MIGRATION_DEST_CONNECT_HOST;
    else
3829
        spec.destType = MIGRATION_DEST_HOST;
3830
    spec.dest.host.protocol = uribits->scheme;
3831 3832 3833
    spec.dest.host.name = uribits->server;
    spec.dest.host.port = uribits->port;
    spec.fwdType = MIGRATION_FWD_DIRECT;
3834

3835 3836 3837
    ret = qemuMigrationSrcRun(driver, vm, persist_xml, cookiein, cookieinlen, cookieout,
                              cookieoutlen, flags, resource, &spec, dconn,
                              graphicsuri, nmigrate_disks, migrate_disks,
3838
                              migParams);
3839 3840 3841 3842

    if (spec.destType == MIGRATION_DEST_FD)
        VIR_FORCE_CLOSE(spec.dest.fd.qemu);

3843
 cleanup:
3844
    virURIFree(uribits);
3845 3846 3847 3848 3849

    return ret;
}


3850 3851 3852 3853 3854 3855 3856 3857 3858 3859 3860 3861 3862 3863 3864
static int
qemuMigrationSrcPerformTunnel(virQEMUDriverPtr driver,
                              virDomainObjPtr vm,
                              virStreamPtr st,
                              const char *persist_xml,
                              const char *cookiein,
                              int cookieinlen,
                              char **cookieout,
                              int *cookieoutlen,
                              unsigned long flags,
                              unsigned long resource,
                              virConnectPtr dconn,
                              const char *graphicsuri,
                              size_t nmigrate_disks,
                              const char **migrate_disks,
3865
                              qemuMigrationParamsPtr migParams)
3866 3867 3868
{
    int ret = -1;
    qemuMigrationSpec spec;
3869
    virQEMUDriverConfigPtr cfg = virQEMUDriverGetConfig(driver);
3870
    int fds[2] = { -1, -1 };
3871 3872

    VIR_DEBUG("driver=%p, vm=%p, st=%p, cookiein=%s, cookieinlen=%d, "
3873
              "cookieout=%p, cookieoutlen=%p, flags=0x%lx, resource=%lu, "
3874
              "graphicsuri=%s, nmigrate_disks=%zu, migrate_disks=%p",
3875
              driver, vm, st, NULLSTR(cookiein), cookieinlen,
3876
              cookieout, cookieoutlen, flags, resource,
3877
              NULLSTR(graphicsuri), nmigrate_disks, migrate_disks);
3878 3879 3880 3881

    spec.fwdType = MIGRATION_FWD_STREAM;
    spec.fwd.stream = st;

3882

3883 3884 3885
    spec.destType = MIGRATION_DEST_FD;
    spec.dest.fd.qemu = -1;
    spec.dest.fd.local = -1;
3886

3887 3888 3889 3890 3891
    if (pipe2(fds, O_CLOEXEC) == 0) {
        spec.dest.fd.qemu = fds[1];
        spec.dest.fd.local = fds[0];
    }
    if (spec.dest.fd.qemu == -1 ||
3892 3893
        qemuSecuritySetImageFDLabel(driver->securityManager, vm->def,
                                    spec.dest.fd.qemu) < 0) {
3894 3895 3896
        virReportSystemError(errno, "%s",
                             _("cannot create pipe for tunnelled migration"));
        goto cleanup;
3897 3898
    }

3899 3900 3901
    ret = qemuMigrationSrcRun(driver, vm, persist_xml, cookiein, cookieinlen,
                              cookieout, cookieoutlen, flags, resource, &spec,
                              dconn, graphicsuri, nmigrate_disks, migrate_disks,
3902
                              migParams);
3903

3904
 cleanup:
3905 3906
    VIR_FORCE_CLOSE(spec.dest.fd.qemu);
    VIR_FORCE_CLOSE(spec.dest.fd.local);
3907

3908
    virObjectUnref(cfg);
3909 3910 3911 3912
    return ret;
}


3913 3914 3915 3916
/* This is essentially a re-impl of virDomainMigrateVersion2
 * from libvirt.c, but running in source libvirtd context,
 * instead of client app context & also adding in tunnel
 * handling */
3917 3918 3919 3920 3921 3922 3923 3924
static int
qemuMigrationSrcPerformPeer2Peer2(virQEMUDriverPtr driver,
                                  virConnectPtr sconn,
                                  virConnectPtr dconn,
                                  virDomainObjPtr vm,
                                  const char *dconnuri,
                                  unsigned long flags,
                                  const char *dname,
3925 3926
                                  unsigned long resource,
                                  qemuMigrationParamsPtr migParams)
3927 3928 3929
{
    virDomainPtr ddomain = NULL;
    char *uri_out = NULL;
3930
    char *cookie = NULL;
3931 3932 3933
    char *dom_xml = NULL;
    int cookielen = 0, ret;
    virErrorPtr orig_err = NULL;
3934
    bool cancelled;
3935
    virStreamPtr st = NULL;
3936
    unsigned long destflags;
3937

3938
    VIR_DEBUG("driver=%p, sconn=%p, dconn=%p, vm=%p, dconnuri=%s, "
3939
              "flags=0x%lx, dname=%s, resource=%lu",
3940 3941
              driver, sconn, dconn, vm, NULLSTR(dconnuri),
              flags, NULLSTR(dname), resource);
3942

3943 3944 3945 3946 3947
    /* In version 2 of the protocol, the prepare step is slightly
     * different.  We fetch the domain XML of the source domain
     * and pass it to Prepare2.
     */
    if (!(dom_xml = qemuDomainFormatXML(driver, vm,
3948 3949
                                        QEMU_DOMAIN_FORMAT_LIVE_FLAGS |
                                        VIR_DOMAIN_XML_MIGRATABLE)))
3950 3951 3952 3953 3954
        return -1;

    if (virDomainObjGetState(vm, NULL) == VIR_DOMAIN_PAUSED)
        flags |= VIR_MIGRATE_PAUSED;

3955 3956
    destflags = flags & ~(VIR_MIGRATE_ABORT_ON_ERROR |
                          VIR_MIGRATE_AUTO_CONVERGE);
3957

3958 3959 3960 3961 3962 3963 3964 3965 3966 3967
    VIR_DEBUG("Prepare2 %p", dconn);
    if (flags & VIR_MIGRATE_TUNNELLED) {
        /*
         * Tunnelled Migrate Version 2 does not support cookies
         * due to missing parameters in the prepareTunnel() API.
         */

        if (!(st = virStreamNew(dconn, 0)))
            goto cleanup;

3968
        qemuDomainObjEnterRemote(vm);
3969
        ret = dconn->driver->domainMigratePrepareTunnel
3970
            (dconn, st, destflags, dname, resource, dom_xml);
3971 3972
        if (qemuDomainObjExitRemote(vm, true) < 0)
            goto cleanup;
3973
    } else {
3974
        qemuDomainObjEnterRemote(vm);
3975 3976
        ret = dconn->driver->domainMigratePrepare2
            (dconn, &cookie, &cookielen, NULL, &uri_out,
3977
             destflags, dname, resource, dom_xml);
3978 3979
        if (qemuDomainObjExitRemote(vm, true) < 0)
            goto cleanup;
3980 3981 3982
    }
    VIR_FREE(dom_xml);
    if (ret == -1)
3983 3984
        goto cleanup;

3985 3986
    if (!(flags & VIR_MIGRATE_TUNNELLED) &&
        (uri_out == NULL)) {
3987 3988
        virReportError(VIR_ERR_INTERNAL_ERROR, "%s",
                       _("domainMigratePrepare2 did not set uri"));
3989
        cancelled = true;
3990
        orig_err = virSaveLastError();
3991
        goto finish;
3992 3993
    }

3994 3995 3996 3997
    /* Perform the migration.  The driver isn't supposed to return
     * until the migration is complete.
     */
    VIR_DEBUG("Perform %p", sconn);
3998
    qemuMigrationJobSetPhase(driver, vm, QEMU_MIGRATION_PHASE_PERFORM2);
3999
    if (flags & VIR_MIGRATE_TUNNELLED)
4000 4001 4002
        ret = qemuMigrationSrcPerformTunnel(driver, vm, st, NULL,
                                            NULL, 0, NULL, NULL,
                                            flags, resource, dconn,
4003
                                            NULL, 0, NULL, migParams);
4004
    else
4005 4006 4007 4008
        ret = qemuMigrationSrcPerformNative(driver, vm, NULL, uri_out,
                                            cookie, cookielen,
                                            NULL, NULL, /* No out cookie with v2 migration */
                                            flags, resource, dconn, NULL, 0, NULL,
4009
                                            migParams);
4010 4011 4012 4013

    /* Perform failed. Make sure Finish doesn't overwrite the error */
    if (ret < 0)
        orig_err = virSaveLastError();
4014

4015 4016 4017
    /* If Perform returns < 0, then we need to cancel the VM
     * startup on the destination
     */
4018
    cancelled = ret < 0;
4019

4020
 finish:
4021 4022 4023 4024
    /* In version 2 of the migration protocol, we pass the
     * status code from the sender to the destination host,
     * so it can do any cleanup if the migration failed.
     */
4025
    dname = dname ? dname : vm->def->name;
4026
    VIR_DEBUG("Finish2 %p ret=%d", dconn, ret);
4027
    qemuDomainObjEnterRemote(vm);
4028
    ddomain = dconn->driver->domainMigrateFinish2
4029
        (dconn, dname, cookie, cookielen,
4030
         uri_out ? uri_out : dconnuri, destflags, cancelled);
4031 4032
    /* The domain is already gone at this point */
    ignore_value(qemuDomainObjExitRemote(vm, false));
4033 4034
    if (cancelled && ddomain)
        VIR_ERROR(_("finish step ignored that migration was cancelled"));
4035

4036
 cleanup:
4037
    if (ddomain) {
4038
        virObjectUnref(ddomain);
4039 4040 4041 4042
        ret = 0;
    } else {
        ret = -1;
    }
4043

4044
    virObjectUnref(st);
4045 4046 4047 4048 4049 4050

    if (orig_err) {
        virSetError(orig_err);
        virFreeError(orig_err);
    }
    VIR_FREE(uri_out);
4051
    VIR_FREE(cookie);
4052 4053

    return ret;
4054 4055 4056
}


4057 4058 4059 4060
/* This is essentially a re-impl of virDomainMigrateVersion3
 * from libvirt.c, but running in source libvirtd context,
 * instead of client app context & also adding in tunnel
 * handling */
4061
static int
4062 4063 4064 4065 4066 4067 4068 4069 4070 4071 4072 4073 4074 4075
qemuMigrationSrcPerformPeer2Peer3(virQEMUDriverPtr driver,
                                  virConnectPtr sconn,
                                  virConnectPtr dconn,
                                  const char *dconnuri,
                                  virDomainObjPtr vm,
                                  const char *xmlin,
                                  const char *persist_xml,
                                  const char *dname,
                                  const char *uri,
                                  const char *graphicsuri,
                                  const char *listenAddress,
                                  size_t nmigrate_disks,
                                  const char **migrate_disks,
                                  int nbdPort,
4076
                                  qemuMigrationParamsPtr migParams,
4077 4078 4079
                                  unsigned long long bandwidth,
                                  bool useParams,
                                  unsigned long flags)
4080 4081 4082 4083 4084 4085 4086 4087 4088 4089
{
    virDomainPtr ddomain = NULL;
    char *uri_out = NULL;
    char *cookiein = NULL;
    char *cookieout = NULL;
    char *dom_xml = NULL;
    int cookieinlen = 0;
    int cookieoutlen = 0;
    int ret = -1;
    virErrorPtr orig_err = NULL;
4090
    bool cancelled = true;
4091
    virStreamPtr st = NULL;
4092
    unsigned long destflags;
4093 4094 4095
    virTypedParameterPtr params = NULL;
    int nparams = 0;
    int maxparams = 0;
4096
    size_t i;
4097
    bool offline = !!(flags & VIR_MIGRATE_OFFLINE);
4098 4099

    VIR_DEBUG("driver=%p, sconn=%p, dconn=%p, dconnuri=%s, vm=%p, xmlin=%s, "
4100
              "dname=%s, uri=%s, graphicsuri=%s, listenAddress=%s, "
4101
              "nmigrate_disks=%zu, migrate_disks=%p, nbdPort=%d, "
4102
              "bandwidth=%llu, useParams=%d, flags=0x%lx",
4103
              driver, sconn, dconn, NULLSTR(dconnuri), vm, NULLSTR(xmlin),
4104
              NULLSTR(dname), NULLSTR(uri), NULLSTR(graphicsuri),
4105
              NULLSTR(listenAddress), nmigrate_disks, migrate_disks, nbdPort,
4106
              bandwidth, useParams, flags);
4107

4108 4109 4110 4111 4112
    /* Unlike the virDomainMigrateVersion3 counterpart, we don't need
     * to worry about auto-setting the VIR_MIGRATE_CHANGE_PROTECTION
     * bit here, because we are already running inside the context of
     * a single job.  */

4113 4114 4115
    dom_xml = qemuMigrationSrcBeginPhase(driver, vm, xmlin, dname,
                                         &cookieout, &cookieoutlen,
                                         nmigrate_disks, migrate_disks, flags);
4116 4117 4118
    if (!dom_xml)
        goto cleanup;

4119 4120 4121 4122 4123 4124 4125 4126 4127 4128 4129 4130 4131 4132 4133 4134 4135 4136 4137 4138
    if (useParams) {
        if (virTypedParamsAddString(&params, &nparams, &maxparams,
                                    VIR_MIGRATE_PARAM_DEST_XML, dom_xml) < 0)
            goto cleanup;

        if (dname &&
            virTypedParamsAddString(&params, &nparams, &maxparams,
                                    VIR_MIGRATE_PARAM_DEST_NAME, dname) < 0)
            goto cleanup;

        if (uri &&
            virTypedParamsAddString(&params, &nparams, &maxparams,
                                    VIR_MIGRATE_PARAM_URI, uri) < 0)
            goto cleanup;

        if (bandwidth &&
            virTypedParamsAddULLong(&params, &nparams, &maxparams,
                                    VIR_MIGRATE_PARAM_BANDWIDTH,
                                    bandwidth) < 0)
            goto cleanup;
4139 4140 4141 4142 4143 4144

        if (graphicsuri &&
            virTypedParamsAddString(&params, &nparams, &maxparams,
                                    VIR_MIGRATE_PARAM_GRAPHICS_URI,
                                    graphicsuri) < 0)
            goto cleanup;
4145 4146 4147 4148 4149
        if (listenAddress &&
            virTypedParamsAddString(&params, &nparams, &maxparams,
                                    VIR_MIGRATE_PARAM_LISTEN_ADDRESS,
                                    listenAddress) < 0)
            goto cleanup;
4150 4151 4152 4153 4154
        for (i = 0; i < nmigrate_disks; i++)
            if (virTypedParamsAddString(&params, &nparams, &maxparams,
                                        VIR_MIGRATE_PARAM_MIGRATE_DISKS,
                                        migrate_disks[i]) < 0)
                goto cleanup;
4155 4156 4157 4158 4159
        if (nbdPort &&
            virTypedParamsAddInt(&params, &nparams, &maxparams,
                                 VIR_MIGRATE_PARAM_DISKS_PORT,
                                 nbdPort) < 0)
            goto cleanup;
4160

4161 4162
        if (qemuMigrationParamsDump(migParams, &params, &nparams,
                                    &maxparams, &flags) < 0)
4163
            goto cleanup;
4164 4165
    }

4166 4167 4168
    if (virDomainObjGetState(vm, NULL) == VIR_DOMAIN_PAUSED)
        flags |= VIR_MIGRATE_PAUSED;

4169 4170
    destflags = flags & ~(VIR_MIGRATE_ABORT_ON_ERROR |
                          VIR_MIGRATE_AUTO_CONVERGE);
4171

4172 4173 4174 4175 4176 4177 4178 4179 4180
    VIR_DEBUG("Prepare3 %p", dconn);
    cookiein = cookieout;
    cookieinlen = cookieoutlen;
    cookieout = NULL;
    cookieoutlen = 0;
    if (flags & VIR_MIGRATE_TUNNELLED) {
        if (!(st = virStreamNew(dconn, 0)))
            goto cleanup;

4181
        qemuDomainObjEnterRemote(vm);
4182 4183 4184 4185 4186 4187 4188 4189 4190
        if (useParams) {
            ret = dconn->driver->domainMigratePrepareTunnel3Params
                (dconn, st, params, nparams, cookiein, cookieinlen,
                 &cookieout, &cookieoutlen, destflags);
        } else {
            ret = dconn->driver->domainMigratePrepareTunnel3
                (dconn, st, cookiein, cookieinlen, &cookieout, &cookieoutlen,
                 destflags, dname, bandwidth, dom_xml);
        }
4191 4192
        if (qemuDomainObjExitRemote(vm, !offline) < 0)
            goto cleanup;
4193
    } else {
4194
        qemuDomainObjEnterRemote(vm);
4195 4196 4197 4198 4199 4200 4201 4202 4203
        if (useParams) {
            ret = dconn->driver->domainMigratePrepare3Params
                (dconn, params, nparams, cookiein, cookieinlen,
                 &cookieout, &cookieoutlen, &uri_out, destflags);
        } else {
            ret = dconn->driver->domainMigratePrepare3
                (dconn, cookiein, cookieinlen, &cookieout, &cookieoutlen,
                 uri, &uri_out, destflags, dname, bandwidth, dom_xml);
        }
4204 4205
        if (qemuDomainObjExitRemote(vm, !offline) < 0)
            goto cleanup;
4206 4207 4208 4209 4210
    }
    VIR_FREE(dom_xml);
    if (ret == -1)
        goto cleanup;

4211
    if (offline) {
L
liguang 已提交
4212 4213 4214
        VIR_DEBUG("Offline migration, skipping Perform phase");
        VIR_FREE(cookieout);
        cookieoutlen = 0;
4215
        cancelled = false;
L
liguang 已提交
4216 4217 4218
        goto finish;
    }

4219 4220 4221 4222
    if (uri_out) {
        uri = uri_out;
        if (useParams &&
            virTypedParamsReplaceString(&params, &nparams,
4223 4224
                                        VIR_MIGRATE_PARAM_URI, uri_out) < 0) {
            orig_err = virSaveLastError();
4225
            goto finish;
4226
        }
4227
    } else if (!uri && !(flags & VIR_MIGRATE_TUNNELLED)) {
4228 4229
        virReportError(VIR_ERR_INTERNAL_ERROR, "%s",
                       _("domainMigratePrepare3 did not set uri"));
4230
        orig_err = virSaveLastError();
4231 4232 4233 4234 4235 4236 4237 4238
        goto finish;
    }

    /* Perform the migration.  The driver isn't supposed to return
     * until the migration is complete. The src VM should remain
     * running, but in paused state until the destination can
     * confirm migration completion.
     */
4239
    VIR_DEBUG("Perform3 %p uri=%s", sconn, NULLSTR(uri));
4240
    qemuMigrationJobSetPhase(driver, vm, QEMU_MIGRATION_PHASE_PERFORM3);
4241 4242 4243 4244 4245
    VIR_FREE(cookiein);
    cookiein = cookieout;
    cookieinlen = cookieoutlen;
    cookieout = NULL;
    cookieoutlen = 0;
4246
    if (flags & VIR_MIGRATE_TUNNELLED) {
4247 4248 4249 4250
        ret = qemuMigrationSrcPerformTunnel(driver, vm, st, persist_xml,
                                            cookiein, cookieinlen,
                                            &cookieout, &cookieoutlen,
                                            flags, bandwidth, dconn, graphicsuri,
4251
                                            nmigrate_disks, migrate_disks,
4252
                                            migParams);
4253
    } else {
4254 4255 4256 4257
        ret = qemuMigrationSrcPerformNative(driver, vm, persist_xml, uri,
                                            cookiein, cookieinlen,
                                            &cookieout, &cookieoutlen,
                                            flags, bandwidth, dconn, graphicsuri,
4258
                                            nmigrate_disks, migrate_disks,
4259
                                            migParams);
4260
    }
4261 4262

    /* Perform failed. Make sure Finish doesn't overwrite the error */
4263
    if (ret < 0) {
4264
        orig_err = virSaveLastError();
4265 4266 4267 4268
    } else {
        qemuMigrationJobSetPhase(driver, vm,
                                 QEMU_MIGRATION_PHASE_PERFORM3_DONE);
    }
4269 4270 4271 4272

    /* If Perform returns < 0, then we need to cancel the VM
     * startup on the destination
     */
4273
    cancelled = ret < 0;
4274

4275
 finish:
4276 4277 4278 4279 4280 4281 4282 4283 4284 4285 4286 4287
    /*
     * The status code from the source is passed to the destination.
     * The dest can cleanup in the source indicated it failed to
     * send all migration data. Returns NULL for ddomain if
     * the dest was unable to complete migration.
     */
    VIR_DEBUG("Finish3 %p ret=%d", dconn, ret);
    VIR_FREE(cookiein);
    cookiein = cookieout;
    cookieinlen = cookieoutlen;
    cookieout = NULL;
    cookieoutlen = 0;
4288 4289 4290 4291 4292 4293 4294 4295 4296 4297 4298 4299 4300

    if (useParams) {
        if (virTypedParamsGetString(params, nparams,
                                    VIR_MIGRATE_PARAM_DEST_NAME, NULL) <= 0 &&
            virTypedParamsReplaceString(&params, &nparams,
                                        VIR_MIGRATE_PARAM_DEST_NAME,
                                        vm->def->name) < 0) {
            ddomain = NULL;
        } else {
            qemuDomainObjEnterRemote(vm);
            ddomain = dconn->driver->domainMigrateFinish3Params
                (dconn, params, nparams, cookiein, cookieinlen,
                 &cookieout, &cookieoutlen, destflags, cancelled);
4301 4302
            if (qemuDomainObjExitRemote(vm, !offline) < 0)
                goto cleanup;
4303 4304 4305 4306 4307 4308 4309
        }
    } else {
        dname = dname ? dname : vm->def->name;
        qemuDomainObjEnterRemote(vm);
        ddomain = dconn->driver->domainMigrateFinish3
            (dconn, dname, cookiein, cookieinlen, &cookieout, &cookieoutlen,
             dconnuri, uri, destflags, cancelled);
4310 4311
        if (qemuDomainObjExitRemote(vm, !offline) < 0)
            goto cleanup;
4312
    }
4313 4314 4315 4316 4317 4318 4319 4320 4321 4322 4323 4324 4325 4326 4327 4328 4329 4330 4331 4332

    if (cancelled) {
        if (ddomain) {
            VIR_ERROR(_("finish step ignored that migration was cancelled"));
        } else {
            /* If Finish reported a useful error, use it instead of the
             * original "migration unexpectedly failed" error.
             *
             * This is ugly but we can't do better with the APIs we have. We
             * only replace the error if Finish was called with cancelled == 1
             * and reported a real error (old libvirt would report an error
             * from RPC instead of MIGRATE_FINISH_OK), which only happens when
             * the domain died on destination. To further reduce a possibility
             * of false positives we also check that Perform returned
             * VIR_ERR_OPERATION_FAILED.
             */
            if (orig_err &&
                orig_err->domain == VIR_FROM_QEMU &&
                orig_err->code == VIR_ERR_OPERATION_FAILED) {
                virErrorPtr err = virGetLastError();
4333 4334
                if (err &&
                    err->domain == VIR_FROM_QEMU &&
4335 4336 4337 4338 4339 4340 4341
                    err->code != VIR_ERR_MIGRATE_FINISH_OK) {
                    virFreeError(orig_err);
                    orig_err = NULL;
                }
            }
        }
    }
4342

4343 4344 4345 4346 4347 4348 4349
    /* If ddomain is NULL, then we were unable to start
     * the guest on the target, and must restart on the
     * source. There is a small chance that the ddomain
     * is NULL due to an RPC failure, in which case
     * ddomain could in fact be running on the dest.
     * The lock manager plugins should take care of
     * safety in this scenario.
4350
     */
4351
    cancelled = ddomain == NULL;
4352

4353 4354 4355 4356 4357 4358
    /* If finish3 set an error, and we don't have an earlier
     * one we need to preserve it in case confirm3 overwrites
     */
    if (!orig_err)
        orig_err = virSaveLastError();

4359 4360 4361 4362
    /*
     * If cancelled, then src VM will be restarted, else
     * it will be killed
     */
4363
    VIR_DEBUG("Confirm3 %p cancelled=%d vm=%p", sconn, cancelled, vm);
4364 4365 4366 4367 4368
    VIR_FREE(cookiein);
    cookiein = cookieout;
    cookieinlen = cookieoutlen;
    cookieout = NULL;
    cookieoutlen = 0;
4369 4370 4371
    ret = qemuMigrationSrcConfirmPhase(driver, vm,
                                       cookiein, cookieinlen,
                                       flags, cancelled);
4372 4373 4374 4375
    /* If Confirm3 returns -1, there's nothing more we can
     * do, but fortunately worst case is that there is a
     * domain left in 'paused' state on source.
     */
4376 4377 4378
    if (ret < 0)
        VIR_WARN("Guest %s probably left in 'paused' state on source",
                 vm->def->name);
4379 4380 4381

 cleanup:
    if (ddomain) {
4382
        virObjectUnref(ddomain);
4383 4384 4385 4386 4387
        ret = 0;
    } else {
        ret = -1;
    }

4388
    virObjectUnref(st);
4389 4390 4391 4392 4393 4394 4395 4396

    if (orig_err) {
        virSetError(orig_err);
        virFreeError(orig_err);
    }
    VIR_FREE(uri_out);
    VIR_FREE(cookiein);
    VIR_FREE(cookieout);
4397
    virTypedParamsFree(params, nparams);
4398 4399 4400 4401
    return ret;
}


4402
static void
4403 4404 4405
qemuMigrationSrcConnectionClosed(virConnectPtr conn,
                                 int reason,
                                 void *opaque)
4406 4407 4408 4409 4410 4411 4412 4413
{
    virDomainObjPtr vm = opaque;

    VIR_DEBUG("conn=%p, reason=%d, vm=%s", conn, reason, vm->def->name);
    virDomainObjBroadcast(vm);
}


4414 4415 4416 4417 4418 4419 4420 4421 4422 4423 4424 4425
static int virConnectCredType[] = {
    VIR_CRED_AUTHNAME,
    VIR_CRED_PASSPHRASE,
};


static virConnectAuth virConnectAuthConfig = {
    .credtype = virConnectCredType,
    .ncredtype = ARRAY_CARDINALITY(virConnectCredType),
};


4426 4427 4428 4429 4430 4431 4432 4433 4434 4435 4436 4437 4438
static int
qemuMigrationSrcPerformPeer2Peer(virQEMUDriverPtr driver,
                                 virConnectPtr sconn,
                                 virDomainObjPtr vm,
                                 const char *xmlin,
                                 const char *persist_xml,
                                 const char *dconnuri,
                                 const char *uri,
                                 const char *graphicsuri,
                                 const char *listenAddress,
                                 size_t nmigrate_disks,
                                 const char **migrate_disks,
                                 int nbdPort,
4439
                                 qemuMigrationParamsPtr migParams,
4440 4441 4442 4443
                                 unsigned long flags,
                                 const char *dname,
                                 unsigned long resource,
                                 bool *v3proto)
4444 4445 4446 4447
{
    int ret = -1;
    virConnectPtr dconn = NULL;
    bool p2p;
4448
    virErrorPtr orig_err = NULL;
4449
    bool offline = !!(flags & VIR_MIGRATE_OFFLINE);
4450
    bool dstOffline = false;
4451
    virQEMUDriverConfigPtr cfg = virQEMUDriverGetConfig(driver);
4452
    bool useParams;
4453

4454 4455
    VIR_DEBUG("driver=%p, sconn=%p, vm=%p, xmlin=%s, dconnuri=%s, uri=%s, "
              "graphicsuri=%s, listenAddress=%s, nmigrate_disks=%zu, "
4456
              "migrate_disks=%p, nbdPort=%d, flags=0x%lx, dname=%s, "
4457
              "resource=%lu",
4458
              driver, sconn, vm, NULLSTR(xmlin), NULLSTR(dconnuri),
4459
              NULLSTR(uri), NULLSTR(graphicsuri), NULLSTR(listenAddress),
4460 4461
              nmigrate_disks, migrate_disks, nbdPort, flags, NULLSTR(dname),
              resource);
4462

4463 4464 4465 4466 4467 4468 4469 4470 4471 4472 4473 4474 4475 4476
    if (flags & VIR_MIGRATE_TUNNELLED && uri) {
        virReportError(VIR_ERR_ARGUMENT_UNSUPPORTED, "%s",
                       _("migration URI is not supported by tunnelled "
                         "migration"));
        goto cleanup;
    }

    if (flags & VIR_MIGRATE_TUNNELLED && listenAddress) {
        virReportError(VIR_ERR_ARGUMENT_UNSUPPORTED, "%s",
                       _("listen address is not supported by tunnelled "
                         "migration"));
        goto cleanup;
    }

4477 4478 4479 4480 4481 4482 4483
    if (flags & VIR_MIGRATE_TUNNELLED && nbdPort) {
        virReportError(VIR_ERR_ARGUMENT_UNSUPPORTED, "%s",
                       _("disk port address is not supported by tunnelled "
                         "migration"));
        goto cleanup;
    }

4484 4485 4486 4487
    /* the order of operations is important here; we make sure the
     * destination side is completely setup before we touch the source
     */

4488
    qemuDomainObjEnterRemote(vm);
4489
    dconn = virConnectOpenAuth(dconnuri, &virConnectAuthConfig, 0);
4490 4491 4492
    if (qemuDomainObjExitRemote(vm, !offline) < 0)
        goto cleanup;

4493
    if (dconn == NULL) {
4494
        virReportError(VIR_ERR_OPERATION_FAILED,
4495 4496
                       _("Failed to connect to remote libvirt URI %s: %s"),
                       dconnuri, virGetLastErrorMessage());
4497
        virObjectUnref(cfg);
4498 4499 4500
        return -1;
    }

4501 4502
    if (virConnectSetKeepAlive(dconn, cfg->keepAliveInterval,
                               cfg->keepAliveCount) < 0)
4503 4504
        goto cleanup;

4505
    if (virConnectRegisterCloseCallback(dconn, qemuMigrationSrcConnectionClosed,
4506 4507 4508 4509
                                        vm, NULL) < 0) {
        goto cleanup;
    }

4510
    qemuDomainObjEnterRemote(vm);
4511 4512
    p2p = VIR_DRV_SUPPORTS_FEATURE(dconn->driver, dconn,
                                   VIR_DRV_FEATURE_MIGRATION_P2P);
4513
        /* v3proto reflects whether the caller used Perform3, but with
4514
         * p2p migrate, regardless of whether Perform2 or Perform3
4515 4516 4517 4518
         * were used, we decide protocol based on what target supports
         */
    *v3proto = VIR_DRV_SUPPORTS_FEATURE(dconn->driver, dconn,
                                        VIR_DRV_FEATURE_MIGRATION_V3);
4519 4520
    useParams = VIR_DRV_SUPPORTS_FEATURE(dconn->driver, dconn,
                                         VIR_DRV_FEATURE_MIGRATION_PARAMS);
4521
    if (offline)
4522 4523
        dstOffline = VIR_DRV_SUPPORTS_FEATURE(dconn->driver, dconn,
                                              VIR_DRV_FEATURE_MIGRATION_OFFLINE);
4524 4525
    if (qemuDomainObjExitRemote(vm, !offline) < 0)
        goto cleanup;
4526

4527
    if (!p2p) {
4528 4529
        virReportError(VIR_ERR_OPERATION_FAILED, "%s",
                       _("Destination libvirt does not support peer-to-peer migration protocol"));
4530 4531 4532
        goto cleanup;
    }

4533 4534
    /* Only xmlin, dname, uri, and bandwidth parameters can be used with
     * old-style APIs. */
4535
    if (!useParams && (graphicsuri || listenAddress || nmigrate_disks)) {
4536 4537 4538 4539 4540 4541
        virReportError(VIR_ERR_ARGUMENT_UNSUPPORTED, "%s",
                       _("Migration APIs with extensible parameters are not "
                         "supported but extended parameters were passed"));
        goto cleanup;
    }

4542
    if (offline && !dstOffline) {
L
liguang 已提交
4543 4544 4545 4546 4547 4548
        virReportError(VIR_ERR_ARGUMENT_UNSUPPORTED, "%s",
                       _("offline migration is not supported by "
                         "the destination host"));
        goto cleanup;
    }

4549 4550 4551 4552 4553 4554 4555
    /* Change protection is only required on the source side (us), and
     * only for v3 migration when begin and perform are separate jobs.
     * But peer-2-peer is already a single job, and we still want to
     * talk to older destinations that would reject the flag.
     * Therefore it is safe to clear the bit here.  */
    flags &= ~VIR_MIGRATE_CHANGE_PROTECTION;

4556
    if (*v3proto) {
4557 4558 4559
        ret = qemuMigrationSrcPerformPeer2Peer3(driver, sconn, dconn, dconnuri, vm, xmlin,
                                                persist_xml, dname, uri, graphicsuri,
                                                listenAddress, nmigrate_disks, migrate_disks,
4560
                                                nbdPort, migParams, resource,
4561
                                                useParams, flags);
4562
    } else {
4563
        ret = qemuMigrationSrcPerformPeer2Peer2(driver, sconn, dconn, vm,
4564
                                                dconnuri, flags, dname, resource,
4565
                                                migParams);
4566
    }
4567

4568
 cleanup:
4569
    orig_err = virSaveLastError();
4570
    qemuDomainObjEnterRemote(vm);
4571
    virConnectUnregisterCloseCallback(dconn, qemuMigrationSrcConnectionClosed);
4572
    virObjectUnref(dconn);
4573
    ignore_value(qemuDomainObjExitRemote(vm, false));
4574 4575 4576 4577
    if (orig_err) {
        virSetError(orig_err);
        virFreeError(orig_err);
    }
4578
    virObjectUnref(cfg);
4579 4580 4581 4582
    return ret;
}


4583 4584 4585 4586 4587 4588
/*
 * This implements perform part of the migration protocol when migration job
 * does not need to be active across several APIs, i.e., peer2peer migration or
 * perform phase of v2 non-peer2peer migration.
 */
static int
4589 4590 4591 4592 4593 4594 4595 4596 4597 4598 4599 4600
qemuMigrationSrcPerformJob(virQEMUDriverPtr driver,
                           virConnectPtr conn,
                           virDomainObjPtr vm,
                           const char *xmlin,
                           const char *persist_xml,
                           const char *dconnuri,
                           const char *uri,
                           const char *graphicsuri,
                           const char *listenAddress,
                           size_t nmigrate_disks,
                           const char **migrate_disks,
                           int nbdPort,
4601
                           qemuMigrationParamsPtr migParams,
4602 4603 4604 4605 4606 4607 4608 4609
                           const char *cookiein,
                           int cookieinlen,
                           char **cookieout,
                           int *cookieoutlen,
                           unsigned long flags,
                           const char *dname,
                           unsigned long resource,
                           bool v3proto)
4610
{
4611
    virObjectEventPtr event = NULL;
4612
    int ret = -1;
4613
    virErrorPtr orig_err = NULL;
4614
    virQEMUDriverConfigPtr cfg = virQEMUDriverGetConfig(driver);
4615
    qemuDomainObjPrivatePtr priv = vm->privateData;
4616

4617 4618
    if (qemuMigrationJobStart(driver, vm, QEMU_ASYNC_JOB_MIGRATION_OUT,
                              flags) < 0)
4619 4620
        goto cleanup;

4621
    if (!(flags & VIR_MIGRATE_OFFLINE) && virDomainObjCheckActive(vm) < 0)
4622 4623
        goto endjob;

4624
    if (!qemuMigrationSrcIsAllowed(driver, vm, true, flags))
4625
        goto endjob;
4626

4627
    if (!(flags & (VIR_MIGRATE_UNSAFE | VIR_MIGRATE_OFFLINE)) &&
4628
        !qemuMigrationSrcIsSafe(vm->def, nmigrate_disks, migrate_disks, flags))
4629
        goto endjob;
4630

4631
    qemuMigrationSrcStoreDomainState(vm);
4632 4633

    if ((flags & (VIR_MIGRATE_TUNNELLED | VIR_MIGRATE_PEER2PEER))) {
4634 4635 4636
        ret = qemuMigrationSrcPerformPeer2Peer(driver, conn, vm, xmlin, persist_xml,
                                               dconnuri, uri, graphicsuri, listenAddress,
                                               nmigrate_disks, migrate_disks, nbdPort,
4637
                                               migParams, flags, dname, resource,
4638
                                               &v3proto);
4639
    } else {
4640
        qemuMigrationJobSetPhase(driver, vm, QEMU_MIGRATION_PHASE_PERFORM2);
4641 4642 4643
        ret = qemuMigrationSrcPerformNative(driver, vm, persist_xml, uri, cookiein, cookieinlen,
                                            cookieout, cookieoutlen,
                                            flags, resource, NULL, NULL, 0, NULL,
4644
                                            migParams);
4645
    }
4646 4647
    if (ret < 0)
        goto endjob;
4648

4649 4650 4651 4652
    /*
     * In v3 protocol, the source VM is not killed off until the
     * confirm step.
     */
4653
    if (!v3proto) {
4654
        qemuProcessStop(driver, vm, VIR_DOMAIN_SHUTOFF_MIGRATED,
4655
                        QEMU_ASYNC_JOB_MIGRATION_OUT,
4656
                        VIR_QEMU_PROCESS_STOP_MIGRATED);
4657
        virDomainAuditStop(vm, "migrated");
4658
        event = virDomainEventLifecycleNewFromObj(vm,
4659 4660
                                         VIR_DOMAIN_EVENT_STOPPED,
                                         VIR_DOMAIN_EVENT_STOPPED_MIGRATED);
4661 4662
    }

4663
 endjob:
4664 4665 4666
    if (ret < 0)
        orig_err = virSaveLastError();

4667 4668 4669 4670
    /* v2 proto has no confirm phase so we need to reset migration parameters
     * here
     */
    if (!v3proto && ret < 0)
4671
        qemuMigrationParamsReset(driver, vm, QEMU_ASYNC_JOB_MIGRATION_OUT,
4672
                                 priv->job.migParams, priv->job.apiFlags);
4673

4674
    qemuMigrationSrcRestoreDomainState(driver, vm);
4675

4676
    qemuMigrationJobFinish(driver, vm);
4677
    if (!virDomainObjIsActive(vm) && ret == 0) {
4678
        if (flags & VIR_MIGRATE_UNDEFINE_SOURCE) {
4679
            virDomainDeleteConfig(cfg->configDir, cfg->autostartDir, vm);
4680 4681
            vm->persistent = 0;
        }
4682
        qemuDomainRemoveInactiveJob(driver, vm);
4683 4684
    }

4685 4686 4687 4688 4689
    if (orig_err) {
        virSetError(orig_err);
        virFreeError(orig_err);
    }

4690
 cleanup:
M
Michal Privoznik 已提交
4691
    virDomainObjEndAPI(&vm);
4692
    virObjectEventStateQueue(driver->domainEventState, event);
4693
    virObjectUnref(cfg);
4694 4695 4696 4697 4698 4699 4700
    return ret;
}

/*
 * This implements perform phase of v3 migration protocol.
 */
static int
4701 4702 4703 4704 4705 4706 4707 4708
qemuMigrationSrcPerformPhase(virQEMUDriverPtr driver,
                             virConnectPtr conn,
                             virDomainObjPtr vm,
                             const char *persist_xml,
                             const char *uri,
                             const char *graphicsuri,
                             size_t nmigrate_disks,
                             const char **migrate_disks,
4709
                             qemuMigrationParamsPtr migParams,
4710 4711 4712 4713 4714 4715
                             const char *cookiein,
                             int cookieinlen,
                             char **cookieout,
                             int *cookieoutlen,
                             unsigned long flags,
                             unsigned long resource)
4716
{
4717
    qemuDomainObjPrivatePtr priv = vm->privateData;
4718 4719 4720 4721
    int ret = -1;

    /* If we didn't start the job in the begin phase, start it now. */
    if (!(flags & VIR_MIGRATE_CHANGE_PROTECTION)) {
4722 4723
        if (qemuMigrationJobStart(driver, vm, QEMU_ASYNC_JOB_MIGRATION_OUT,
                                  flags) < 0)
4724 4725 4726 4727 4728 4729
            goto cleanup;
    } else if (!qemuMigrationJobIsActive(vm, QEMU_ASYNC_JOB_MIGRATION_OUT)) {
        goto cleanup;
    }

    qemuMigrationJobStartPhase(driver, vm, QEMU_MIGRATION_PHASE_PERFORM3);
4730
    virCloseCallbacksUnset(driver->closeCallbacks, vm,
4731
                           qemuMigrationSrcCleanup);
4732

4733 4734 4735
    ret = qemuMigrationSrcPerformNative(driver, vm, persist_xml, uri, cookiein, cookieinlen,
                                        cookieout, cookieoutlen,
                                        flags, resource, NULL, graphicsuri,
4736
                                        nmigrate_disks, migrate_disks, migParams);
4737

4738
    if (ret < 0) {
4739
        qemuMigrationSrcRestoreDomainState(driver, vm);
4740
        goto endjob;
4741
    }
4742 4743 4744

    qemuMigrationJobSetPhase(driver, vm, QEMU_MIGRATION_PHASE_PERFORM3_DONE);

4745
    if (virCloseCallbacksSet(driver->closeCallbacks, vm, conn,
4746
                             qemuMigrationSrcCleanup) < 0)
4747 4748
        goto endjob;

4749
 endjob:
4750
    if (ret < 0) {
4751
        qemuMigrationParamsReset(driver, vm, QEMU_ASYNC_JOB_MIGRATION_OUT,
4752
                                 priv->job.migParams, priv->job.apiFlags);
4753
        qemuMigrationJobFinish(driver, vm);
4754
    } else {
4755
        qemuMigrationJobContinue(vm);
4756 4757
    }

4758
    if (!virDomainObjIsActive(vm))
4759
        qemuDomainRemoveInactiveJob(driver, vm);
4760

4761
 cleanup:
M
Michal Privoznik 已提交
4762
    virDomainObjEndAPI(&vm);
4763 4764 4765
    return ret;
}

4766
int
4767 4768 4769 4770 4771 4772 4773 4774 4775 4776 4777 4778
qemuMigrationSrcPerform(virQEMUDriverPtr driver,
                        virConnectPtr conn,
                        virDomainObjPtr vm,
                        const char *xmlin,
                        const char *persist_xml,
                        const char *dconnuri,
                        const char *uri,
                        const char *graphicsuri,
                        const char *listenAddress,
                        size_t nmigrate_disks,
                        const char **migrate_disks,
                        int nbdPort,
4779
                        qemuMigrationParamsPtr migParams,
4780 4781 4782 4783 4784 4785 4786 4787
                        const char *cookiein,
                        int cookieinlen,
                        char **cookieout,
                        int *cookieoutlen,
                        unsigned long flags,
                        const char *dname,
                        unsigned long resource,
                        bool v3proto)
4788 4789
{
    VIR_DEBUG("driver=%p, conn=%p, vm=%p, xmlin=%s, dconnuri=%s, "
4790
              "uri=%s, graphicsuri=%s, listenAddress=%s, "
4791
              "nmigrate_disks=%zu, migrate_disks=%p, nbdPort=%d, "
4792
              "cookiein=%s, cookieinlen=%d, cookieout=%p, cookieoutlen=%p, "
4793
              "flags=0x%lx, dname=%s, resource=%lu, v3proto=%d",
4794
              driver, conn, vm, NULLSTR(xmlin), NULLSTR(dconnuri),
4795
              NULLSTR(uri), NULLSTR(graphicsuri), NULLSTR(listenAddress),
4796 4797 4798
              nmigrate_disks, migrate_disks, nbdPort,
              NULLSTR(cookiein), cookieinlen, cookieout, cookieoutlen,
              flags, NULLSTR(dname), resource, v3proto);
4799 4800 4801

    if ((flags & (VIR_MIGRATE_TUNNELLED | VIR_MIGRATE_PEER2PEER))) {
        if (cookieinlen) {
4802 4803
            virReportError(VIR_ERR_OPERATION_INVALID,
                           "%s", _("received unexpected cookie with P2P migration"));
4804 4805 4806
            return -1;
        }

4807 4808 4809
        return qemuMigrationSrcPerformJob(driver, conn, vm, xmlin, persist_xml, dconnuri, uri,
                                          graphicsuri, listenAddress,
                                          nmigrate_disks, migrate_disks, nbdPort,
4810
                                          migParams,
4811 4812 4813
                                          cookiein, cookieinlen,
                                          cookieout, cookieoutlen,
                                          flags, dname, resource, v3proto);
4814 4815
    } else {
        if (dconnuri) {
4816 4817
            virReportError(VIR_ERR_INTERNAL_ERROR,
                           "%s", _("Unexpected dconnuri parameter with non-peer2peer migration"));
4818 4819 4820 4821
            return -1;
        }

        if (v3proto) {
4822 4823 4824
            return qemuMigrationSrcPerformPhase(driver, conn, vm, persist_xml, uri,
                                                graphicsuri,
                                                nmigrate_disks, migrate_disks,
4825
                                                migParams,
4826 4827 4828
                                                cookiein, cookieinlen,
                                                cookieout, cookieoutlen,
                                                flags, resource);
4829
        } else {
4830 4831 4832
            return qemuMigrationSrcPerformJob(driver, conn, vm, xmlin, persist_xml, NULL,
                                              uri, graphicsuri, listenAddress,
                                              nmigrate_disks, migrate_disks, nbdPort,
4833
                                              migParams,
4834 4835 4836
                                              cookiein, cookieinlen,
                                              cookieout, cookieoutlen, flags,
                                              dname, resource, v3proto);
4837 4838 4839
        }
    }
}
4840

4841
static int
4842
qemuMigrationDstVPAssociatePortProfiles(virDomainDefPtr def)
4843
{
4844
    size_t i;
4845 4846 4847 4848 4849
    int last_good_net = -1;
    virDomainNetDefPtr net;

    for (i = 0; i < def->nnets; i++) {
        net = def->nets[i];
4850
        if (virDomainNetGetActualType(net) == VIR_DOMAIN_NET_TYPE_DIRECT) {
4851
            if (virNetDevVPortProfileAssociate(net->ifname,
4852
                                               virDomainNetGetActualVirtPortProfile(net),
4853
                                               &net->mac,
4854
                                               virDomainNetGetActualDirectDev(net),
4855
                                               -1,
4856
                                               def->uuid,
4857 4858
                                               VIR_NETDEV_VPORT_PROFILE_OP_MIGRATE_IN_FINISH,
                                               false) < 0) {
4859 4860 4861
                virReportError(VIR_ERR_OPERATION_FAILED,
                               _("Port profile Associate failed for %s"),
                               net->ifname);
4862
                goto err_exit;
4863
            }
4864
            last_good_net = i;
4865
            VIR_DEBUG("Port profile Associate succeeded for %s", net->ifname);
4866

4867
            if (virNetDevMacVLanVPortProfileRegisterCallback(net->ifname, &net->mac,
4868 4869 4870 4871
                                                             virDomainNetGetActualDirectDev(net), def->uuid,
                                                             virDomainNetGetActualVirtPortProfile(net),
                                                             VIR_NETDEV_VPORT_PROFILE_OP_CREATE))
                goto err_exit;
4872 4873 4874
        }
    }

4875
    return 0;
4876

4877
 err_exit:
4878
    for (i = 0; last_good_net != -1 && i <= last_good_net; i++) {
4879
        net = def->nets[i];
4880
        if (virDomainNetGetActualType(net) == VIR_DOMAIN_NET_TYPE_DIRECT) {
4881
            ignore_value(virNetDevVPortProfileDisassociate(net->ifname,
4882
                                                           virDomainNetGetActualVirtPortProfile(net),
4883
                                                           &net->mac,
4884
                                                           virDomainNetGetActualDirectDev(net),
4885
                                                           -1,
4886
                                                           VIR_NETDEV_VPORT_PROFILE_OP_MIGRATE_IN_FINISH));
4887 4888
        }
    }
4889
    return -1;
4890 4891 4892
}


J
Jiri Denemark 已提交
4893
static int
4894 4895 4896 4897
qemuMigrationDstPersist(virQEMUDriverPtr driver,
                        virDomainObjPtr vm,
                        qemuMigrationCookiePtr mig,
                        bool ignoreSaveError)
J
Jiri Denemark 已提交
4898 4899 4900 4901
{
    virQEMUDriverConfigPtr cfg = virQEMUDriverGetConfig(driver);
    virCapsPtr caps = NULL;
    virDomainDefPtr vmdef;
4902 4903
    virDomainDefPtr oldDef = NULL;
    unsigned int oldPersist = vm->persistent;
J
Jiri Denemark 已提交
4904 4905 4906 4907 4908 4909 4910
    virObjectEventPtr event;
    int ret = -1;

    if (!(caps = virQEMUDriverGetCapabilities(driver, false)))
        goto cleanup;

    vm->persistent = 1;
4911 4912
    oldDef = vm->newDef;
    vm->newDef = qemuMigrationCookieGetPersistent(mig);
J
Jiri Denemark 已提交
4913

4914 4915
    if (!(vmdef = virDomainObjGetPersistentDef(caps, driver->xmlopt, vm)))
        goto error;
J
Jiri Denemark 已提交
4916

4917 4918
    if (virDomainSaveConfig(cfg->configDir, driver->caps, vmdef) < 0 &&
        !ignoreSaveError)
4919
        goto error;
J
Jiri Denemark 已提交
4920 4921 4922

    event = virDomainEventLifecycleNewFromObj(vm,
                                              VIR_DOMAIN_EVENT_DEFINED,
4923 4924 4925
                                              oldPersist ?
                                              VIR_DOMAIN_EVENT_DEFINED_UPDATED :
                                              VIR_DOMAIN_EVENT_DEFINED_ADDED);
4926
    virObjectEventStateQueue(driver->domainEventState, event);
J
Jiri Denemark 已提交
4927 4928 4929 4930

    ret = 0;

 cleanup:
4931
    virDomainDefFree(oldDef);
J
Jiri Denemark 已提交
4932 4933 4934
    virObjectUnref(caps);
    virObjectUnref(cfg);
    return ret;
4935 4936 4937 4938 4939 4940 4941

 error:
    virDomainDefFree(vm->newDef);
    vm->persistent = oldPersist;
    vm->newDef = oldDef;
    oldDef = NULL;
    goto cleanup;
J
Jiri Denemark 已提交
4942 4943 4944
}


4945
virDomainPtr
4946 4947 4948 4949 4950 4951 4952 4953 4954 4955
qemuMigrationDstFinish(virQEMUDriverPtr driver,
                       virConnectPtr dconn,
                       virDomainObjPtr vm,
                       const char *cookiein,
                       int cookieinlen,
                       char **cookieout,
                       int *cookieoutlen,
                       unsigned long flags,
                       int retcode,
                       bool v3proto)
4956 4957
{
    virDomainPtr dom = NULL;
4958
    qemuMigrationCookiePtr mig = NULL;
4959
    virErrorPtr orig_err = NULL;
4960
    int cookie_flags = 0;
J
Jiri Denemark 已提交
4961
    qemuDomainObjPrivatePtr priv = vm->privateData;
4962
    virQEMUDriverConfigPtr cfg = virQEMUDriverGetConfig(driver);
4963
    unsigned short port;
4964 4965
    unsigned long long timeReceived = 0;
    virObjectEventPtr event;
4966
    qemuDomainJobInfoPtr jobInfo = NULL;
4967
    bool inPostCopy = false;
4968
    bool doKill = true;
4969

4970
    VIR_DEBUG("driver=%p, dconn=%p, vm=%p, cookiein=%s, cookieinlen=%d, "
4971
              "cookieout=%p, cookieoutlen=%p, flags=0x%lx, retcode=%d",
4972 4973
              driver, dconn, vm, NULLSTR(cookiein), cookieinlen,
              cookieout, cookieoutlen, flags, retcode);
4974

4975 4976 4977
    port = priv->migrationPort;
    priv->migrationPort = 0;

4978
    if (!qemuMigrationJobIsActive(vm, QEMU_ASYNC_JOB_MIGRATION_IN)) {
4979
        qemuMigrationDstErrorReport(driver, vm->def->name);
4980
        goto cleanup;
4981
    }
4982

4983 4984
    ignore_value(virTimeMillisNow(&timeReceived));

4985 4986 4987
    qemuMigrationJobStartPhase(driver, vm,
                               v3proto ? QEMU_MIGRATION_PHASE_FINISH3
                                       : QEMU_MIGRATION_PHASE_FINISH2);
4988

4989
    qemuDomainCleanupRemove(vm, qemuMigrationDstPrepareCleanup);
4990
    VIR_FREE(priv->job.completed);
4991

4992
    cookie_flags = QEMU_MIGRATION_COOKIE_NETWORK |
4993 4994
                   QEMU_MIGRATION_COOKIE_STATS |
                   QEMU_MIGRATION_COOKIE_NBD;
4995 4996 4997
    /* Some older versions of libvirt always send persistent XML in the cookie
     * even though VIR_MIGRATE_PERSIST_DEST was not used. */
    cookie_flags |= QEMU_MIGRATION_COOKIE_PERSISTENT;
4998

4999 5000
    if (!(mig = qemuMigrationEatCookie(driver, vm->def, priv->origname, priv,
                                       cookiein, cookieinlen, cookie_flags)))
5001
        goto endjob;
5002

5003
    if (flags & VIR_MIGRATE_OFFLINE) {
5004
        if (retcode == 0 &&
5005
            qemuMigrationDstPersist(driver, vm, mig, false) == 0)
5006
            dom = virGetDomain(dconn, vm->def->name, vm->def->uuid, -1);
5007 5008
        goto endjob;
    }
5009

5010 5011 5012 5013
    if (retcode != 0) {
        /* Check for a possible error on the monitor in case Finish was called
         * earlier than monitor EOF handler got a chance to process the error
         */
5014
        qemuDomainCheckMonitor(driver, vm, QEMU_ASYNC_JOB_MIGRATION_IN);
5015 5016
        goto endjob;
    }
5017

5018 5019 5020
    if (!virDomainObjIsActive(vm)) {
        virReportError(VIR_ERR_INTERNAL_ERROR, "%s",
                       _("guest unexpectedly quit"));
5021
        qemuMigrationDstErrorReport(driver, vm->def->name);
5022 5023
        goto endjob;
    }
5024

5025
    if (qemuMigrationDstVPAssociatePortProfiles(vm->def) < 0)
5026
        goto endjob;
5027

5028
    if (mig->network && qemuMigrationDstOPDRelocate(driver, vm, mig) < 0)
5029
        VIR_WARN("unable to provide network data for relocation");
5030

5031
    if (qemuMigrationDstStopNBDServer(driver, vm, mig) < 0)
5032
        goto endjob;
5033

5034 5035
    if (qemuRefreshVirtioChannelState(driver, vm,
                                      QEMU_ASYNC_JOB_MIGRATION_IN) < 0)
5036 5037
        goto endjob;

5038 5039
    if (qemuConnectAgent(driver, vm) < 0)
        goto endjob;
5040

5041
    if (flags & VIR_MIGRATE_PERSIST_DEST) {
5042
        if (qemuMigrationDstPersist(driver, vm, mig, !v3proto) < 0) {
5043 5044 5045 5046 5047 5048 5049 5050 5051 5052 5053
            /* Hmpf.  Migration was successful, but making it persistent
             * was not.  If we report successful, then when this domain
             * shuts down, management tools are in for a surprise.  On the
             * other hand, if we report failure, then the management tools
             * might try to restart the domain on the source side, even
             * though the domain is actually running on the destination.
             * Pretend success and hope that this is a rare situation and
             * management tools are smart.
             *
             * However, in v3 protocol, the source VM is still available
             * to restart during confirm() step, so we kill it off now.
5054 5055 5056
             */
            if (v3proto)
                goto endjob;
5057
        }
5058
    }
5059

5060 5061 5062
    /* We need to wait for QEMU to process all data sent by the source
     * before starting guest CPUs.
     */
5063 5064 5065
    if (qemuMigrationDstWaitForCompletion(driver, vm,
                                          QEMU_ASYNC_JOB_MIGRATION_IN,
                                          !!(flags & VIR_MIGRATE_POSTCOPY)) < 0) {
5066 5067 5068 5069 5070 5071
        /* There's not much we can do for v2 protocol since the
         * original domain on the source host is already gone.
         */
        if (v3proto)
            goto endjob;
    }
5072

5073 5074 5075 5076 5077 5078 5079 5080 5081
    /* Now that the state data was transferred we can refresh the actual state
     * of the devices */
    if (qemuProcessRefreshState(driver, vm, QEMU_ASYNC_JOB_MIGRATION_IN) < 0) {
        /* Similarly to the case above v2 protocol will not be able to recover
         * from this. Let's ignore this and perhaps stuff will not break. */
        if (v3proto)
            goto endjob;
    }

5082
    if (priv->job.current->status == QEMU_DOMAIN_JOB_STATUS_POSTCOPY)
5083 5084
        inPostCopy = true;

5085 5086 5087 5088 5089
    if (!(flags & VIR_MIGRATE_PAUSED)) {
        /* run 'cont' on the destination, which allows migration on qemu
         * >= 0.10.6 to work properly.  This isn't strictly necessary on
         * older qemu's, but it also doesn't hurt anything there
         */
5090
        if (qemuProcessStartCPUs(driver, vm,
5091 5092
                                 inPostCopy ? VIR_DOMAIN_RUNNING_POSTCOPY
                                            : VIR_DOMAIN_RUNNING_MIGRATED,
5093
                                 QEMU_ASYNC_JOB_MIGRATION_IN) < 0) {
5094
            if (virGetLastErrorCode() == VIR_ERR_OK)
5095 5096 5097 5098 5099 5100
                virReportError(VIR_ERR_INTERNAL_ERROR,
                               "%s", _("resume operation failed"));
            /* Need to save the current error, in case shutting
             * down the process overwrites it
             */
            orig_err = virSaveLastError();
5101

5102 5103 5104 5105 5106 5107 5108 5109 5110 5111
            /*
             * In v3 protocol, the source VM is still available to
             * restart during confirm() step, so we kill it off
             * now.
             * In v2 protocol, the source is dead, so we leave
             * target in paused state, in case admin can fix
             * things up.
             */
            if (v3proto)
                goto endjob;
5112
        }
5113

5114
        if (inPostCopy)
5115
            doKill = false;
5116
    }
5117

5118
    if (mig->jobInfo) {
5119
        jobInfo = mig->jobInfo;
5120 5121 5122 5123 5124 5125
        mig->jobInfo = NULL;

        if (jobInfo->sent && timeReceived) {
            jobInfo->timeDelta = timeReceived - jobInfo->sent;
            jobInfo->received = timeReceived;
            jobInfo->timeDeltaSet = true;
5126
        }
5127 5128
        qemuDomainJobInfoUpdateTime(jobInfo);
        qemuDomainJobInfoUpdateDowntime(jobInfo);
5129
    }
L
liguang 已提交
5130

5131
    if (inPostCopy) {
5132 5133 5134
        if (qemuMigrationDstWaitForCompletion(driver, vm,
                                              QEMU_ASYNC_JOB_MIGRATION_IN,
                                              false) < 0) {
5135 5136 5137 5138 5139 5140 5141 5142 5143
            goto endjob;
        }
        if (virDomainObjGetState(vm, NULL) == VIR_DOMAIN_RUNNING) {
            virDomainObjSetState(vm,
                                 VIR_DOMAIN_RUNNING,
                                 VIR_DOMAIN_RUNNING_MIGRATED);
        }
    }

5144
    dom = virGetDomain(dconn, vm->def->name, vm->def->uuid, vm->def->id);
5145

5146 5147 5148 5149 5150 5151 5152 5153 5154 5155 5156 5157 5158 5159
    if (inPostCopy) {
        /* The only RESUME event during post-copy migration is triggered by
         * QEMU when the running domain moves from the source to the
         * destination host, but then the migration keeps running until all
         * modified memory is transferred from the source host. This will
         * result in VIR_DOMAIN_EVENT_RESUMED with RESUMED_POSTCOPY detail.
         * However, our API documentation says we need to fire another RESUMED
         * event at the very end of migration with RESUMED_MIGRATED detail.
         */
        event = virDomainEventLifecycleNewFromObj(vm,
                                                  VIR_DOMAIN_EVENT_RESUMED,
                                                  VIR_DOMAIN_EVENT_RESUMED_MIGRATED);
        virObjectEventStateQueue(driver->domainEventState, event);
    }
5160

5161 5162 5163 5164 5165
    if (virDomainObjGetState(vm, NULL) == VIR_DOMAIN_PAUSED) {
        virDomainObjSetState(vm, VIR_DOMAIN_PAUSED, VIR_DOMAIN_PAUSED_USER);
        event = virDomainEventLifecycleNewFromObj(vm,
                                                  VIR_DOMAIN_EVENT_SUSPENDED,
                                                  VIR_DOMAIN_EVENT_SUSPENDED_PAUSED);
5166
        virObjectEventStateQueue(driver->domainEventState, event);
5167
    }
5168

5169
    if (virDomainObjIsActive(vm) &&
5170
        virDomainSaveStatus(driver->xmlopt, cfg->stateDir, vm, driver->caps) < 0)
5171 5172 5173 5174 5175
        VIR_WARN("Failed to save status on vm %s", vm->def->name);

    /* Guest is successfully running, so cancel previous auto destroy */
    qemuProcessAutoDestroyRemove(driver, vm);

5176
 endjob:
5177
    if (!dom &&
5178 5179
        !(flags & VIR_MIGRATE_OFFLINE) &&
        virDomainObjIsActive(vm)) {
5180
        if (doKill) {
5181 5182 5183 5184 5185 5186 5187
            qemuProcessStop(driver, vm, VIR_DOMAIN_SHUTOFF_FAILED,
                            QEMU_ASYNC_JOB_MIGRATION_IN,
                            VIR_QEMU_PROCESS_STOP_MIGRATED);
            virDomainAuditStop(vm, "failed");
            event = virDomainEventLifecycleNewFromObj(vm,
                                VIR_DOMAIN_EVENT_STOPPED,
                                VIR_DOMAIN_EVENT_STOPPED_FAILED);
5188
            virObjectEventStateQueue(driver->domainEventState, event);
5189
        } else {
5190
            qemuMigrationAnyPostcopyFailed(driver, vm);
5191
        }
5192 5193
    }

5194
    if (dom) {
5195 5196 5197
        if (jobInfo) {
            VIR_STEAL_PTR(priv->job.completed, jobInfo);
            priv->job.completed->status = QEMU_DOMAIN_JOB_STATUS_COMPLETED;
5198
            priv->job.completed->statsType = QEMU_DOMAIN_JOB_STATS_TYPE_MIGRATION;
5199
        }
5200

5201 5202 5203
        if (qemuMigrationBakeCookie(mig, driver, vm,
                                    QEMU_MIGRATION_DESTINATION,
                                    cookieout, cookieoutlen,
5204 5205
                                    QEMU_MIGRATION_COOKIE_STATS) < 0)
            VIR_WARN("Unable to encode migration cookie");
5206 5207 5208 5209 5210 5211

        /* Remove completed stats for post-copy, everything but timing fields
         * is obsolete anyway.
         */
        if (inPostCopy)
            VIR_FREE(priv->job.completed);
5212
    }
5213

5214
    qemuMigrationParamsReset(driver, vm, QEMU_ASYNC_JOB_MIGRATION_IN,
5215
                             priv->job.migParams, priv->job.apiFlags);
5216

5217
    qemuMigrationJobFinish(driver, vm);
5218
    if (!virDomainObjIsActive(vm))
5219
        qemuDomainRemoveInactiveJob(driver, vm);
5220

5221
 cleanup:
5222
    VIR_FREE(jobInfo);
5223
    virPortAllocatorRelease(port);
5224
    if (priv->mon)
5225
        qemuMonitorSetDomainLog(priv->mon, NULL, NULL, NULL);
5226
    VIR_FREE(priv->origname);
M
Michal Privoznik 已提交
5227
    virDomainObjEndAPI(&vm);
5228
    qemuMigrationCookieFree(mig);
5229 5230 5231 5232
    if (orig_err) {
        virSetError(orig_err);
        virFreeError(orig_err);
    }
5233
    virObjectUnref(cfg);
5234 5235 5236 5237

    /* Set a special error if Finish is expected to return NULL as a result of
     * successful call with retcode != 0
     */
5238
    if (retcode != 0 && !dom && virGetLastErrorCode() == VIR_ERR_OK)
5239
        virReportError(VIR_ERR_MIGRATE_FINISH_OK, NULL);
5240 5241
    return dom;
}
5242

5243

5244
/* Helper function called while vm is active.  */
5245
int
5246 5247 5248 5249
qemuMigrationSrcToFile(virQEMUDriverPtr driver, virDomainObjPtr vm,
                       int fd,
                       const char *compressor,
                       qemuDomainAsyncJob asyncJob)
5250 5251 5252
{
    qemuDomainObjPrivatePtr priv = vm->privateData;
    int rc;
5253
    int ret = -1;
5254 5255
    virCommandPtr cmd = NULL;
    int pipeFD[2] = { -1, -1 };
5256
    unsigned long saveMigBandwidth = priv->migMaxBandwidth;
5257
    char *errbuf = NULL;
5258
    virErrorPtr orig_err = NULL;
5259 5260 5261 5262 5263

    /* Increase migration bandwidth to unlimited since target is a file.
     * Failure to change migration speed is not fatal. */
    if (qemuDomainObjEnterMonitorAsync(driver, vm, asyncJob) == 0) {
        qemuMonitorSetMigrationSpeed(priv->mon,
5264 5265
                                     QEMU_DOMAIN_MIG_BANDWIDTH_MAX);
        priv->migMaxBandwidth = QEMU_DOMAIN_MIG_BANDWIDTH_MAX;
5266 5267
        if (qemuDomainObjExitMonitor(driver, vm) < 0)
            return -1;
5268
    }
5269

5270 5271 5272 5273 5274 5275 5276
    if (!virDomainObjIsActive(vm)) {
        virReportError(VIR_ERR_INTERNAL_ERROR, "%s",
                       _("guest unexpectedly quit"));
        /* nothing to tear down */
        return -1;
    }

5277 5278 5279 5280
    if (compressor && pipe(pipeFD) < 0) {
        virReportSystemError(errno, "%s",
                             _("Failed to create pipe for migration"));
        return -1;
5281 5282
    }

5283 5284 5285 5286
    /* All right! We can use fd migration, which means that qemu
     * doesn't have to open() the file, so while we still have to
     * grant SELinux access, we can do it on fd and avoid cleanup
     * later, as well as skip futzing with cgroup.  */
5287 5288
    if (qemuSecuritySetImageFDLabel(driver->securityManager, vm->def,
                                    compressor ? pipeFD[1] : fd) < 0)
5289 5290
        goto cleanup;

5291
    if (qemuDomainObjEnterMonitorAsync(driver, vm, asyncJob) < 0)
5292 5293
        goto cleanup;

5294
    if (!compressor) {
5295 5296 5297
        rc = qemuMonitorMigrateToFd(priv->mon,
                                    QEMU_MONITOR_MIGRATE_BACKGROUND,
                                    fd);
5298 5299 5300 5301 5302 5303 5304
    } else {
        const char *prog = compressor;
        const char *args[] = {
            prog,
            "-c",
            NULL
        };
5305 5306 5307 5308 5309 5310 5311 5312 5313 5314 5315

        cmd = virCommandNewArgs(args);
        virCommandSetInputFD(cmd, pipeFD[0]);
        virCommandSetOutputFD(cmd, &fd);
        virCommandSetErrorBuffer(cmd, &errbuf);
        virCommandDoAsyncIO(cmd);
        if (virSetCloseExec(pipeFD[1]) < 0) {
            virReportSystemError(errno, "%s",
                                 _("Unable to set cloexec flag"));
            ignore_value(qemuDomainObjExitMonitor(driver, vm));
            goto cleanup;
5316
        }
5317 5318 5319 5320 5321 5322 5323 5324 5325 5326
        if (virCommandRunAsync(cmd, NULL) < 0) {
            ignore_value(qemuDomainObjExitMonitor(driver, vm));
            goto cleanup;
        }
        rc = qemuMonitorMigrateToFd(priv->mon,
                                    QEMU_MONITOR_MIGRATE_BACKGROUND,
                                    pipeFD[1]);
        if (VIR_CLOSE(pipeFD[0]) < 0 ||
            VIR_CLOSE(pipeFD[1]) < 0)
            VIR_WARN("failed to close intermediate pipe");
5327
    }
5328
    if (qemuDomainObjExitMonitor(driver, vm) < 0)
5329
        goto cleanup;
5330 5331 5332
    if (rc < 0)
        goto cleanup;

5333
    rc = qemuMigrationSrcWaitForCompletion(driver, vm, asyncJob, NULL, 0);
5334

5335 5336 5337 5338
    if (rc < 0) {
        if (rc == -2) {
            orig_err = virSaveLastError();
            virCommandAbort(cmd);
5339 5340
            if (virDomainObjIsActive(vm) &&
                qemuDomainObjEnterMonitorAsync(driver, vm, asyncJob) == 0) {
5341
                qemuMonitorMigrateCancel(priv->mon);
5342
                ignore_value(qemuDomainObjExitMonitor(driver, vm));
5343 5344
            }
        }
5345
        goto cleanup;
5346
    }
5347

5348 5349 5350
    if (cmd && virCommandWait(cmd, NULL) < 0)
        goto cleanup;

5351
    qemuDomainEventEmitJobCompleted(driver, vm);
5352 5353
    ret = 0;

5354
 cleanup:
5355 5356 5357
    if (ret < 0 && !orig_err)
        orig_err = virSaveLastError();

5358
    /* Restore max migration bandwidth */
5359 5360
    if (virDomainObjIsActive(vm) &&
        qemuDomainObjEnterMonitorAsync(driver, vm, asyncJob) == 0) {
5361 5362
        qemuMonitorSetMigrationSpeed(priv->mon, saveMigBandwidth);
        priv->migMaxBandwidth = saveMigBandwidth;
5363
        ignore_value(qemuDomainObjExitMonitor(driver, vm));
5364 5365
    }

5366 5367
    VIR_FORCE_CLOSE(pipeFD[0]);
    VIR_FORCE_CLOSE(pipeFD[1]);
5368 5369 5370 5371 5372
    if (cmd) {
        VIR_DEBUG("Compression binary stderr: %s", NULLSTR(errbuf));
        VIR_FREE(errbuf);
        virCommandFree(cmd);
    }
5373 5374 5375 5376 5377 5378

    if (orig_err) {
        virSetError(orig_err);
        virFreeError(orig_err);
    }

5379 5380
    return ret;
}
5381

5382 5383

int
5384 5385
qemuMigrationSrcCancel(virQEMUDriverPtr driver,
                       virDomainObjPtr vm)
5386 5387 5388 5389 5390 5391 5392 5393 5394 5395 5396
{
    qemuDomainObjPrivatePtr priv = vm->privateData;
    bool storage = false;
    size_t i;
    int ret = -1;

    VIR_DEBUG("Canceling unfinished outgoing migration of domain %s",
              vm->def->name);

    qemuDomainObjEnterMonitor(driver, vm);
    ignore_value(qemuMonitorMigrateCancel(priv->mon));
5397
    if (qemuDomainObjExitMonitor(driver, vm) < 0)
5398 5399 5400 5401 5402
        goto cleanup;

    for (i = 0; i < vm->def->ndisks; i++) {
        virDomainDiskDefPtr disk = vm->def->disks[i];
        qemuDomainDiskPrivatePtr diskPriv = QEMU_DOMAIN_DISK_PRIVATE(disk);
5403
        qemuBlockJobDataPtr job;
5404

5405
        if (!(job = qemuBlockJobDiskGetJob(disk)) ||
5406
            !qemuBlockJobIsRunning(job))
5407
            diskPriv->migrating = false;
5408 5409

        if (diskPriv->migrating) {
5410
            qemuBlockJobSyncBegin(job);
5411
            storage = true;
5412 5413
        }

5414 5415
        virObjectUnref(job);
    }
5416 5417 5418

    if (storage &&
        qemuMigrationSrcNBDCopyCancel(driver, vm, false,
5419
                                      QEMU_ASYNC_JOB_NONE, NULL) < 0)
5420
        goto cleanup;
5421 5422 5423 5424 5425 5426 5427 5428

    ret = 0;

 cleanup:
    return ret;
}


5429
static int
5430
qemuMigrationJobStart(virQEMUDriverPtr driver,
5431
                      virDomainObjPtr vm,
5432 5433
                      qemuDomainAsyncJob job,
                      unsigned long apiFlags)
5434
{
5435
    qemuDomainObjPrivatePtr priv = vm->privateData;
5436 5437
    virDomainJobOperation op;
    unsigned long long mask;
5438

5439
    if (job == QEMU_ASYNC_JOB_MIGRATION_IN) {
5440 5441
        op = VIR_DOMAIN_JOB_OPERATION_MIGRATION_IN;
        mask = QEMU_JOB_NONE;
5442
    } else {
5443 5444 5445 5446
        op = VIR_DOMAIN_JOB_OPERATION_MIGRATION_OUT;
        mask = QEMU_JOB_DEFAULT_MASK |
               JOB_MASK(QEMU_JOB_SUSPEND) |
               JOB_MASK(QEMU_JOB_MIGRATION_OP);
5447
    }
5448

5449
    if (qemuDomainObjBeginAsyncJob(driver, vm, job, op, apiFlags) < 0)
5450 5451
        return -1;

5452 5453
    priv->job.current->statsType = QEMU_DOMAIN_JOB_STATS_TYPE_MIGRATION;

5454
    qemuDomainObjSetAsyncJobMask(vm, mask);
5455 5456 5457
    return 0;
}

5458
static void
5459
qemuMigrationJobSetPhase(virQEMUDriverPtr driver,
5460
                         virDomainObjPtr vm,
5461
                         qemuMigrationJobPhase phase)
5462 5463 5464 5465 5466 5467 5468 5469 5470 5471 5472 5473 5474
{
    qemuDomainObjPrivatePtr priv = vm->privateData;

    if (phase < priv->job.phase) {
        VIR_ERROR(_("migration protocol going backwards %s => %s"),
                  qemuMigrationJobPhaseTypeToString(priv->job.phase),
                  qemuMigrationJobPhaseTypeToString(phase));
        return;
    }

    qemuDomainObjSetJobPhase(driver, vm, phase);
}

5475
static void
5476
qemuMigrationJobStartPhase(virQEMUDriverPtr driver,
5477
                           virDomainObjPtr vm,
5478
                           qemuMigrationJobPhase phase)
5479 5480 5481 5482
{
    qemuMigrationJobSetPhase(driver, vm, phase);
}

5483
static void
5484 5485
qemuMigrationJobContinue(virDomainObjPtr vm)
{
5486
    qemuDomainObjReleaseAsyncJob(vm);
5487 5488
}

5489
static bool
5490
qemuMigrationJobIsActive(virDomainObjPtr vm,
5491
                         qemuDomainAsyncJob job)
5492 5493 5494 5495 5496 5497 5498 5499 5500 5501 5502
{
    qemuDomainObjPrivatePtr priv = vm->privateData;

    if (priv->job.asyncJob != job) {
        const char *msg;

        if (job == QEMU_ASYNC_JOB_MIGRATION_IN)
            msg = _("domain '%s' is not processing incoming migration");
        else
            msg = _("domain '%s' is not being migrated");

5503
        virReportError(VIR_ERR_OPERATION_INVALID, msg, vm->def->name);
5504 5505 5506 5507 5508
        return false;
    }
    return true;
}

5509
static void
5510
qemuMigrationJobFinish(virQEMUDriverPtr driver, virDomainObjPtr vm)
5511
{
5512
    qemuDomainObjEndAsyncJob(driver, vm);
5513
}
5514 5515 5516


static void
5517
qemuMigrationDstErrorFree(void *data,
5518 5519 5520 5521 5522 5523 5524
                       const void *name ATTRIBUTE_UNUSED)
{
    virErrorPtr err = data;
    virFreeError(err);
}

int
5525
qemuMigrationDstErrorInit(virQEMUDriverPtr driver)
5526
{
5527
    driver->migrationErrors = virHashAtomicNew(64, qemuMigrationDstErrorFree);
5528 5529 5530 5531 5532 5533 5534 5535 5536 5537 5538
    if (driver->migrationErrors)
        return 0;
    else
        return -1;
}

/**
 * This function consumes @err; the caller should consider the @err pointer
 * invalid after calling this function.
 */
void
5539 5540 5541
qemuMigrationDstErrorSave(virQEMUDriverPtr driver,
                          const char *name,
                          virErrorPtr err)
5542 5543 5544 5545 5546 5547 5548 5549 5550 5551 5552 5553 5554
{
    if (!err)
        return;

    VIR_DEBUG("Saving incoming migration error for domain %s: %s",
              name, err->message);
    if (virHashAtomicUpdate(driver->migrationErrors, name, err) < 0) {
        VIR_WARN("Failed to save migration error for domain '%s'", name);
        virFreeError(err);
    }
}

void
5555 5556
qemuMigrationDstErrorReport(virQEMUDriverPtr driver,
                            const char *name)
5557 5558 5559 5560 5561 5562 5563 5564 5565 5566 5567
{
    virErrorPtr err;

    if (!(err = virHashAtomicSteal(driver->migrationErrors, name)))
        return;

    VIR_DEBUG("Restoring saved incoming migration error for domain %s: %s",
              name, err->message);
    virSetError(err);
    virFreeError(err);
}
5568 5569


5570
int
5571 5572 5573 5574
qemuMigrationSrcFetchMirrorStats(virQEMUDriverPtr driver,
                                 virDomainObjPtr vm,
                                 qemuDomainAsyncJob asyncJob,
                                 qemuDomainJobInfoPtr jobInfo)
5575 5576 5577 5578 5579 5580 5581 5582 5583 5584 5585 5586 5587 5588 5589 5590 5591 5592 5593 5594 5595 5596 5597 5598 5599 5600 5601 5602 5603 5604 5605 5606 5607 5608 5609 5610 5611 5612 5613 5614 5615 5616 5617 5618
{
    size_t i;
    qemuDomainObjPrivatePtr priv = vm->privateData;
    bool nbd = false;
    virHashTablePtr blockinfo = NULL;
    qemuDomainMirrorStatsPtr stats = &jobInfo->mirrorStats;

    for (i = 0; i < vm->def->ndisks; i++) {
        virDomainDiskDefPtr disk = vm->def->disks[i];
        if (QEMU_DOMAIN_DISK_PRIVATE(disk)->migrating) {
            nbd = true;
            break;
        }
    }

    if (!nbd)
        return 0;

    if (qemuDomainObjEnterMonitorAsync(driver, vm, asyncJob) < 0)
        return -1;

    blockinfo = qemuMonitorGetAllBlockJobInfo(priv->mon);

    if (qemuDomainObjExitMonitor(driver, vm) < 0 || !blockinfo)
        return -1;

    memset(stats, 0, sizeof(*stats));

    for (i = 0; i < vm->def->ndisks; i++) {
        virDomainDiskDefPtr disk = vm->def->disks[i];
        qemuDomainDiskPrivatePtr diskPriv = QEMU_DOMAIN_DISK_PRIVATE(disk);
        qemuMonitorBlockJobInfoPtr data;

        if (!diskPriv->migrating ||
            !(data = virHashLookup(blockinfo, disk->info.alias)))
            continue;

        stats->transferred += data->cur;
        stats->total += data->end;
    }

    virHashFree(blockinfo);
    return 0;
}