virtio-net.c 68.0 KB
Newer Older
A
aliguori 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13
/*
 * Virtio Network Device
 *
 * Copyright IBM, Corp. 2007
 *
 * Authors:
 *  Anthony Liguori   <aliguori@us.ibm.com>
 *
 * This work is licensed under the terms of the GNU GPL, version 2.  See
 * the COPYING file in the top-level directory.
 *
 */

P
Peter Maydell 已提交
14
#include "qemu/osdep.h"
15
#include "qemu/iov.h"
P
Paolo Bonzini 已提交
16
#include "hw/virtio/virtio.h"
P
Paolo Bonzini 已提交
17
#include "net/net.h"
18
#include "net/checksum.h"
19
#include "net/tap.h"
20 21
#include "qemu/error-report.h"
#include "qemu/timer.h"
P
Paolo Bonzini 已提交
22 23
#include "hw/virtio/virtio-net.h"
#include "net/vhost_net.h"
24
#include "hw/virtio/virtio-bus.h"
25
#include "qapi/error.h"
26
#include "qapi/qapi-events-net.h"
27
#include "hw/virtio/virtio-access.h"
28
#include "migration/misc.h"
29
#include "standard-headers/linux/ethtool.h"
A
aliguori 已提交
30

31
#define VIRTIO_NET_VM_VERSION    11
32

33
#define MAC_TABLE_ENTRIES    64
34
#define MAX_VLAN    (1 << 12)   /* Per 802.1Q definition */
35

36 37
/* previously fixed value */
#define VIRTIO_NET_RX_QUEUE_DEFAULT_SIZE 256
38 39
#define VIRTIO_NET_TX_QUEUE_DEFAULT_SIZE 256

40 41
/* for now, only allow larger queues; with virtio-1, guest can downsize */
#define VIRTIO_NET_RX_QUEUE_MIN_SIZE VIRTIO_NET_RX_QUEUE_DEFAULT_SIZE
42
#define VIRTIO_NET_TX_QUEUE_MIN_SIZE VIRTIO_NET_TX_QUEUE_DEFAULT_SIZE
43

44 45 46 47 48 49 50 51
/*
 * Calculate the number of bytes up to and including the given 'field' of
 * 'container'.
 */
#define endof(container, field) \
    (offsetof(container, field) + sizeof(((container *)0)->field))

typedef struct VirtIOFeature {
52
    uint64_t flags;
53 54 55 56
    size_t end;
} VirtIOFeature;

static VirtIOFeature feature_sizes[] = {
57
    {.flags = 1ULL << VIRTIO_NET_F_MAC,
58
     .end = endof(struct virtio_net_config, mac)},
59
    {.flags = 1ULL << VIRTIO_NET_F_STATUS,
60
     .end = endof(struct virtio_net_config, status)},
61
    {.flags = 1ULL << VIRTIO_NET_F_MQ,
62
     .end = endof(struct virtio_net_config, max_virtqueue_pairs)},
63
    {.flags = 1ULL << VIRTIO_NET_F_MTU,
64
     .end = endof(struct virtio_net_config, mtu)},
65 66
    {.flags = 1ULL << VIRTIO_NET_F_SPEED_DUPLEX,
     .end = endof(struct virtio_net_config, duplex)},
67 68 69
    {}
};

J
Jason Wang 已提交
70
static VirtIONetQueue *virtio_net_get_subqueue(NetClientState *nc)
71 72 73
{
    VirtIONet *n = qemu_get_nic_opaque(nc);

J
Jason Wang 已提交
74
    return &n->vqs[nc->queue_index];
75
}
J
Jason Wang 已提交
76 77 78 79 80 81

static int vq2q(int queue_index)
{
    return queue_index / 2;
}

A
aliguori 已提交
82 83 84 85
/* TODO
 * - we could suppress RX interrupt if we were so inclined.
 */

86
static void virtio_net_get_config(VirtIODevice *vdev, uint8_t *config)
A
aliguori 已提交
87
{
88
    VirtIONet *n = VIRTIO_NET(vdev);
A
aliguori 已提交
89 90
    struct virtio_net_config netcfg;

91 92
    virtio_stw_p(vdev, &netcfg.status, n->status);
    virtio_stw_p(vdev, &netcfg.max_virtqueue_pairs, n->max_queues);
93
    virtio_stw_p(vdev, &netcfg.mtu, n->net_conf.mtu);
94
    memcpy(netcfg.mac, n->mac, ETH_ALEN);
95 96
    virtio_stl_p(vdev, &netcfg.speed, n->net_conf.speed);
    netcfg.duplex = n->net_conf.duplex;
97
    memcpy(config, &netcfg, n->config_size);
A
aliguori 已提交
98 99
}

100 101
static void virtio_net_set_config(VirtIODevice *vdev, const uint8_t *config)
{
102
    VirtIONet *n = VIRTIO_NET(vdev);
103
    struct virtio_net_config netcfg = {};
104

105
    memcpy(&netcfg, config, n->config_size);
106

107 108
    if (!virtio_vdev_has_feature(vdev, VIRTIO_NET_F_CTRL_MAC_ADDR) &&
        !virtio_vdev_has_feature(vdev, VIRTIO_F_VERSION_1) &&
109
        memcmp(netcfg.mac, n->mac, ETH_ALEN)) {
110
        memcpy(n->mac, netcfg.mac, ETH_ALEN);
J
Jason Wang 已提交
111
        qemu_format_nic_info_str(qemu_get_queue(n->nic), n->mac);
112 113 114
    }
}

115 116
static bool virtio_net_started(VirtIONet *n, uint8_t status)
{
117
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
118
    return (status & VIRTIO_CONFIG_S_DRIVER_OK) &&
119
        (n->status & VIRTIO_NET_S_LINK_UP) && vdev->vm_running;
120 121
}

J
Jason Wang 已提交
122 123 124 125 126 127 128 129 130 131
static void virtio_net_announce_timer(void *opaque)
{
    VirtIONet *n = opaque;
    VirtIODevice *vdev = VIRTIO_DEVICE(n);

    n->announce_counter--;
    n->status |= VIRTIO_NET_S_ANNOUNCE;
    virtio_notify_config(vdev);
}

132
static void virtio_net_vhost_status(VirtIONet *n, uint8_t status)
133
{
134
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
J
Jason Wang 已提交
135
    NetClientState *nc = qemu_get_queue(n->nic);
J
Jason Wang 已提交
136
    int queues = n->multiqueue ? n->max_queues : 1;
J
Jason Wang 已提交
137

138
    if (!get_vhost_net(nc->peer)) {
139 140
        return;
    }
J
Jason Wang 已提交
141

142 143
    if ((virtio_net_started(n, status) && !nc->peer->link_down) ==
        !!n->vhost_started) {
144 145 146
        return;
    }
    if (!n->vhost_started) {
147 148
        int r, i;

149 150 151 152 153 154 155
        if (n->needs_vnet_hdr_swap) {
            error_report("backend does not support %s vnet headers; "
                         "falling back on userspace virtio",
                         virtio_is_big_endian(vdev) ? "BE" : "LE");
            return;
        }

156 157 158 159 160 161 162 163 164 165 166
        /* Any packets outstanding? Purge them to avoid touching rings
         * when vhost is running.
         */
        for (i = 0;  i < queues; i++) {
            NetClientState *qnc = qemu_get_subqueue(n->nic, i);

            /* Purge both directions: TX and RX. */
            qemu_net_queue_purge(qnc->peer->incoming_queue, qnc);
            qemu_net_queue_purge(qnc->incoming_queue, qnc->peer);
        }

167 168 169 170 171 172 173 174 175 176
        if (virtio_has_feature(vdev->guest_features, VIRTIO_NET_F_MTU)) {
            r = vhost_net_set_mtu(get_vhost_net(nc->peer), n->net_conf.mtu);
            if (r < 0) {
                error_report("%uBytes MTU not supported by the backend",
                             n->net_conf.mtu);

                return;
            }
        }

177
        n->vhost_started = 1;
178
        r = vhost_net_start(vdev, n->nic->ncs, queues);
179
        if (r < 0) {
180 181
            error_report("unable to start vhost net: %d: "
                         "falling back on userspace virtio", -r);
182
            n->vhost_started = 0;
183 184
        }
    } else {
185
        vhost_net_stop(vdev, n->nic->ncs, queues);
186 187 188 189
        n->vhost_started = 0;
    }
}

190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242
static int virtio_net_set_vnet_endian_one(VirtIODevice *vdev,
                                          NetClientState *peer,
                                          bool enable)
{
    if (virtio_is_big_endian(vdev)) {
        return qemu_set_vnet_be(peer, enable);
    } else {
        return qemu_set_vnet_le(peer, enable);
    }
}

static bool virtio_net_set_vnet_endian(VirtIODevice *vdev, NetClientState *ncs,
                                       int queues, bool enable)
{
    int i;

    for (i = 0; i < queues; i++) {
        if (virtio_net_set_vnet_endian_one(vdev, ncs[i].peer, enable) < 0 &&
            enable) {
            while (--i >= 0) {
                virtio_net_set_vnet_endian_one(vdev, ncs[i].peer, false);
            }

            return true;
        }
    }

    return false;
}

static void virtio_net_vnet_endian_status(VirtIONet *n, uint8_t status)
{
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
    int queues = n->multiqueue ? n->max_queues : 1;

    if (virtio_net_started(n, status)) {
        /* Before using the device, we tell the network backend about the
         * endianness to use when parsing vnet headers. If the backend
         * can't do it, we fallback onto fixing the headers in the core
         * virtio-net code.
         */
        n->needs_vnet_hdr_swap = virtio_net_set_vnet_endian(vdev, n->nic->ncs,
                                                            queues, true);
    } else if (virtio_net_started(n, vdev->status)) {
        /* After using the device, we need to reset the network backend to
         * the default (guest native endianness), otherwise the guest may
         * lose network connectivity if it is rebooted into a different
         * endianness.
         */
        virtio_net_set_vnet_endian(vdev, n->nic->ncs, queues, false);
    }
}

243 244 245 246 247 248 249 250
static void virtio_net_drop_tx_queue_data(VirtIODevice *vdev, VirtQueue *vq)
{
    unsigned int dropped = virtqueue_drop_all(vq);
    if (dropped) {
        virtio_notify(vdev, vq);
    }
}

251 252
static void virtio_net_set_status(struct VirtIODevice *vdev, uint8_t status)
{
253
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
254 255 256
    VirtIONetQueue *q;
    int i;
    uint8_t queue_status;
257

258
    virtio_net_vnet_endian_status(n, status);
259 260
    virtio_net_vhost_status(n, status);

J
Jason Wang 已提交
261
    for (i = 0; i < n->max_queues; i++) {
262 263
        NetClientState *ncs = qemu_get_subqueue(n->nic, i);
        bool queue_started;
J
Jason Wang 已提交
264
        q = &n->vqs[i];
265

J
Jason Wang 已提交
266 267
        if ((!n->multiqueue && i != 0) || i >= n->curr_queues) {
            queue_status = 0;
268
        } else {
J
Jason Wang 已提交
269
            queue_status = status;
270
        }
271 272 273 274 275 276
        queue_started =
            virtio_net_started(n, queue_status) && !n->vhost_started;

        if (queue_started) {
            qemu_flush_queued_packets(ncs);
        }
J
Jason Wang 已提交
277 278 279 280 281

        if (!q->tx_waiting) {
            continue;
        }

282
        if (queue_started) {
J
Jason Wang 已提交
283
            if (q->tx_timer) {
284 285
                timer_mod(q->tx_timer,
                               qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + n->tx_timeout);
J
Jason Wang 已提交
286 287 288
            } else {
                qemu_bh_schedule(q->tx_bh);
            }
289
        } else {
J
Jason Wang 已提交
290
            if (q->tx_timer) {
291
                timer_del(q->tx_timer);
J
Jason Wang 已提交
292 293 294
            } else {
                qemu_bh_cancel(q->tx_bh);
            }
295
            if ((n->status & VIRTIO_NET_S_LINK_UP) == 0 &&
296 297
                (queue_status & VIRTIO_CONFIG_S_DRIVER_OK) &&
                vdev->vm_running) {
298 299 300 301 302 303
                /* if tx is waiting we are likely have some packets in tx queue
                 * and disabled notification */
                q->tx_waiting = 0;
                virtio_queue_set_notification(q->tx_vq, 1);
                virtio_net_drop_tx_queue_data(vdev, q->tx_vq);
            }
304 305 306 307
        }
    }
}

308
static void virtio_net_set_link_status(NetClientState *nc)
309
{
J
Jason Wang 已提交
310
    VirtIONet *n = qemu_get_nic_opaque(nc);
311
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
312 313
    uint16_t old_status = n->status;

M
Mark McLoughlin 已提交
314
    if (nc->link_down)
315 316 317 318 319
        n->status &= ~VIRTIO_NET_S_LINK_UP;
    else
        n->status |= VIRTIO_NET_S_LINK_UP;

    if (n->status != old_status)
320
        virtio_notify_config(vdev);
321

322
    virtio_net_set_status(vdev, vdev->status);
323 324
}

325 326 327 328 329
static void rxfilter_notify(NetClientState *nc)
{
    VirtIONet *n = qemu_get_nic_opaque(nc);

    if (nc->rxfilter_notify_enabled) {
330
        gchar *path = object_get_canonical_path(OBJECT(n->qdev));
331 332
        qapi_event_send_nic_rx_filter_changed(!!n->netclient_name,
                                              n->netclient_name, path, &error_abort);
333
        g_free(path);
334 335 336 337 338 339

        /* disable event notification to avoid events flooding */
        nc->rxfilter_notify_enabled = 0;
    }
}

340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359
static intList *get_vlan_table(VirtIONet *n)
{
    intList *list, *entry;
    int i, j;

    list = NULL;
    for (i = 0; i < MAX_VLAN >> 5; i++) {
        for (j = 0; n->vlans[i] && j <= 0x1f; j++) {
            if (n->vlans[i] & (1U << j)) {
                entry = g_malloc0(sizeof(*entry));
                entry->value = (i << 5) + j;
                entry->next = list;
                list = entry;
            }
        }
    }

    return list;
}

360 361 362
static RxFilterInfo *virtio_net_query_rxfilter(NetClientState *nc)
{
    VirtIONet *n = qemu_get_nic_opaque(nc);
363
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
364 365
    RxFilterInfo *info;
    strList *str_list, *entry;
366
    int i;
367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391

    info = g_malloc0(sizeof(*info));
    info->name = g_strdup(nc->name);
    info->promiscuous = n->promisc;

    if (n->nouni) {
        info->unicast = RX_STATE_NONE;
    } else if (n->alluni) {
        info->unicast = RX_STATE_ALL;
    } else {
        info->unicast = RX_STATE_NORMAL;
    }

    if (n->nomulti) {
        info->multicast = RX_STATE_NONE;
    } else if (n->allmulti) {
        info->multicast = RX_STATE_ALL;
    } else {
        info->multicast = RX_STATE_NORMAL;
    }

    info->broadcast_allowed = n->nobcast;
    info->multicast_overflow = n->mac_table.multi_overflow;
    info->unicast_overflow = n->mac_table.uni_overflow;

392
    info->main_mac = qemu_mac_strdup_printf(n->mac);
393 394 395 396

    str_list = NULL;
    for (i = 0; i < n->mac_table.first_multi; i++) {
        entry = g_malloc0(sizeof(*entry));
397
        entry->value = qemu_mac_strdup_printf(n->mac_table.macs + i * ETH_ALEN);
398 399 400 401 402 403 404 405
        entry->next = str_list;
        str_list = entry;
    }
    info->unicast_table = str_list;

    str_list = NULL;
    for (i = n->mac_table.first_multi; i < n->mac_table.in_use; i++) {
        entry = g_malloc0(sizeof(*entry));
406
        entry->value = qemu_mac_strdup_printf(n->mac_table.macs + i * ETH_ALEN);
407 408 409 410
        entry->next = str_list;
        str_list = entry;
    }
    info->multicast_table = str_list;
411
    info->vlan_table = get_vlan_table(n);
412

413
    if (!virtio_vdev_has_feature(vdev, VIRTIO_NET_F_CTRL_VLAN)) {
414 415 416 417 418
        info->vlan = RX_STATE_ALL;
    } else if (!info->vlan_table) {
        info->vlan = RX_STATE_NONE;
    } else {
        info->vlan = RX_STATE_NORMAL;
419 420 421 422 423 424 425 426
    }

    /* enable event notification after query */
    nc->rxfilter_notify_enabled = 1;

    return info;
}

427 428
static void virtio_net_reset(VirtIODevice *vdev)
{
429
    VirtIONet *n = VIRTIO_NET(vdev);
430 431 432 433

    /* Reset back to compatibility mode */
    n->promisc = 1;
    n->allmulti = 0;
434 435 436 437
    n->alluni = 0;
    n->nomulti = 0;
    n->nouni = 0;
    n->nobcast = 0;
J
Jason Wang 已提交
438 439
    /* multiqueue is disabled by default */
    n->curr_queues = 1;
J
Jason Wang 已提交
440 441 442
    timer_del(n->announce_timer);
    n->announce_counter = 0;
    n->status &= ~VIRTIO_NET_S_ANNOUNCE;
443

444
    /* Flush any MAC and VLAN filter table state */
445
    n->mac_table.in_use = 0;
446
    n->mac_table.first_multi = 0;
447 448
    n->mac_table.multi_overflow = 0;
    n->mac_table.uni_overflow = 0;
449
    memset(n->mac_table.macs, 0, MAC_TABLE_ENTRIES * ETH_ALEN);
450
    memcpy(&n->mac[0], &n->nic->conf->macaddr, sizeof(n->mac));
451
    qemu_format_nic_info_str(qemu_get_queue(n->nic), n->mac);
452
    memset(n->vlans, 0, MAX_VLAN >> 3);
453 454
}

455
static void peer_test_vnet_hdr(VirtIONet *n)
M
Mark McLoughlin 已提交
456
{
J
Jason Wang 已提交
457 458
    NetClientState *nc = qemu_get_queue(n->nic);
    if (!nc->peer) {
459
        return;
J
Jason Wang 已提交
460
    }
M
Mark McLoughlin 已提交
461

462
    n->has_vnet_hdr = qemu_has_vnet_hdr(nc->peer);
463
}
M
Mark McLoughlin 已提交
464

465 466
static int peer_has_vnet_hdr(VirtIONet *n)
{
M
Mark McLoughlin 已提交
467 468 469
    return n->has_vnet_hdr;
}

470 471 472 473 474
static int peer_has_ufo(VirtIONet *n)
{
    if (!peer_has_vnet_hdr(n))
        return 0;

475
    n->has_ufo = qemu_has_ufo(qemu_get_queue(n->nic)->peer);
476 477 478 479

    return n->has_ufo;
}

480 481
static void virtio_net_set_mrg_rx_bufs(VirtIONet *n, int mergeable_rx_bufs,
                                       int version_1)
482
{
J
Jason Wang 已提交
483 484 485
    int i;
    NetClientState *nc;

486 487
    n->mergeable_rx_bufs = mergeable_rx_bufs;

488 489 490 491 492 493 494
    if (version_1) {
        n->guest_hdr_len = sizeof(struct virtio_net_hdr_mrg_rxbuf);
    } else {
        n->guest_hdr_len = n->mergeable_rx_bufs ?
            sizeof(struct virtio_net_hdr_mrg_rxbuf) :
            sizeof(struct virtio_net_hdr);
    }
495

J
Jason Wang 已提交
496 497 498 499
    for (i = 0; i < n->max_queues; i++) {
        nc = qemu_get_subqueue(n->nic, i);

        if (peer_has_vnet_hdr(n) &&
500 501
            qemu_has_vnet_hdr_len(nc->peer, n->guest_hdr_len)) {
            qemu_set_vnet_hdr_len(nc->peer, n->guest_hdr_len);
J
Jason Wang 已提交
502 503
            n->host_hdr_len = n->guest_hdr_len;
        }
504 505 506
    }
}

507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524
static int virtio_net_max_tx_queue_size(VirtIONet *n)
{
    NetClientState *peer = n->nic_conf.peers.ncs[0];

    /*
     * Backends other than vhost-user don't support max queue size.
     */
    if (!peer) {
        return VIRTIO_NET_TX_QUEUE_DEFAULT_SIZE;
    }

    if (peer->info->type != NET_CLIENT_DRIVER_VHOST_USER) {
        return VIRTIO_NET_TX_QUEUE_DEFAULT_SIZE;
    }

    return VIRTQUEUE_MAX_SIZE;
}

J
Jason Wang 已提交
525 526 527 528 529 530 531 532
static int peer_attach(VirtIONet *n, int index)
{
    NetClientState *nc = qemu_get_subqueue(n->nic, index);

    if (!nc->peer) {
        return 0;
    }

533
    if (nc->peer->info->type == NET_CLIENT_DRIVER_VHOST_USER) {
534 535 536
        vhost_set_vring_enable(nc->peer, 1);
    }

537
    if (nc->peer->info->type != NET_CLIENT_DRIVER_TAP) {
J
Jason Wang 已提交
538 539 540
        return 0;
    }

541 542 543 544
    if (n->max_queues == 1) {
        return 0;
    }

J
Jason Wang 已提交
545 546 547 548 549 550 551 552 553 554 555
    return tap_enable(nc->peer);
}

static int peer_detach(VirtIONet *n, int index)
{
    NetClientState *nc = qemu_get_subqueue(n->nic, index);

    if (!nc->peer) {
        return 0;
    }

556
    if (nc->peer->info->type == NET_CLIENT_DRIVER_VHOST_USER) {
557 558 559
        vhost_set_vring_enable(nc->peer, 0);
    }

560
    if (nc->peer->info->type !=  NET_CLIENT_DRIVER_TAP) {
J
Jason Wang 已提交
561 562 563 564 565 566 567 568 569
        return 0;
    }

    return tap_disable(nc->peer);
}

static void virtio_net_set_queues(VirtIONet *n)
{
    int i;
570
    int r;
J
Jason Wang 已提交
571

572 573 574 575
    if (n->nic->peer_deleted) {
        return;
    }

J
Jason Wang 已提交
576 577
    for (i = 0; i < n->max_queues; i++) {
        if (i < n->curr_queues) {
578 579
            r = peer_attach(n, i);
            assert(!r);
J
Jason Wang 已提交
580
        } else {
581 582
            r = peer_detach(n, i);
            assert(!r);
J
Jason Wang 已提交
583 584 585 586
        }
    }
}

J
Jason Wang 已提交
587
static void virtio_net_set_multiqueue(VirtIONet *n, int multiqueue);
J
Jason Wang 已提交
588

J
Jason Wang 已提交
589 590
static uint64_t virtio_net_get_features(VirtIODevice *vdev, uint64_t features,
                                        Error **errp)
A
aliguori 已提交
591
{
592
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
593
    NetClientState *nc = qemu_get_queue(n->nic);
A
aliguori 已提交
594

595 596 597
    /* Firstly sync all virtio-net possible supported features */
    features |= n->host_features;

598
    virtio_add_feature(&features, VIRTIO_NET_F_MAC);
599

600
    if (!peer_has_vnet_hdr(n)) {
601 602 603 604
        virtio_clear_feature(&features, VIRTIO_NET_F_CSUM);
        virtio_clear_feature(&features, VIRTIO_NET_F_HOST_TSO4);
        virtio_clear_feature(&features, VIRTIO_NET_F_HOST_TSO6);
        virtio_clear_feature(&features, VIRTIO_NET_F_HOST_ECN);
605

606 607 608 609
        virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_CSUM);
        virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_TSO4);
        virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_TSO6);
        virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_ECN);
610
    }
M
Mark McLoughlin 已提交
611

612
    if (!peer_has_vnet_hdr(n) || !peer_has_ufo(n)) {
613 614
        virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_UFO);
        virtio_clear_feature(&features, VIRTIO_NET_F_HOST_UFO);
M
Mark McLoughlin 已提交
615 616
    }

617
    if (!get_vhost_net(nc->peer)) {
618 619
        return features;
    }
620 621 622 623 624 625 626 627 628
    features = vhost_net_get_features(get_vhost_net(nc->peer), features);
    vdev->backend_features = features;

    if (n->mtu_bypass_backend &&
            (n->host_features & 1ULL << VIRTIO_NET_F_MTU)) {
        features |= (1ULL << VIRTIO_NET_F_MTU);
    }

    return features;
A
aliguori 已提交
629 630
}

G
Gerd Hoffmann 已提交
631
static uint64_t virtio_net_bad_features(VirtIODevice *vdev)
632
{
G
Gerd Hoffmann 已提交
633
    uint64_t features = 0;
634 635 636

    /* Linux kernel 2.6.25.  It understood MAC (as everyone must),
     * but also these: */
637 638 639 640 641
    virtio_add_feature(&features, VIRTIO_NET_F_MAC);
    virtio_add_feature(&features, VIRTIO_NET_F_CSUM);
    virtio_add_feature(&features, VIRTIO_NET_F_HOST_TSO4);
    virtio_add_feature(&features, VIRTIO_NET_F_HOST_TSO6);
    virtio_add_feature(&features, VIRTIO_NET_F_HOST_ECN);
642

643
    return features;
644 645
}

646 647
static void virtio_net_apply_guest_offloads(VirtIONet *n)
{
648
    qemu_set_offload(qemu_get_queue(n->nic)->peer,
649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673
            !!(n->curr_guest_offloads & (1ULL << VIRTIO_NET_F_GUEST_CSUM)),
            !!(n->curr_guest_offloads & (1ULL << VIRTIO_NET_F_GUEST_TSO4)),
            !!(n->curr_guest_offloads & (1ULL << VIRTIO_NET_F_GUEST_TSO6)),
            !!(n->curr_guest_offloads & (1ULL << VIRTIO_NET_F_GUEST_ECN)),
            !!(n->curr_guest_offloads & (1ULL << VIRTIO_NET_F_GUEST_UFO)));
}

static uint64_t virtio_net_guest_offloads_by_features(uint32_t features)
{
    static const uint64_t guest_offloads_mask =
        (1ULL << VIRTIO_NET_F_GUEST_CSUM) |
        (1ULL << VIRTIO_NET_F_GUEST_TSO4) |
        (1ULL << VIRTIO_NET_F_GUEST_TSO6) |
        (1ULL << VIRTIO_NET_F_GUEST_ECN)  |
        (1ULL << VIRTIO_NET_F_GUEST_UFO);

    return guest_offloads_mask & features;
}

static inline uint64_t virtio_net_supported_guest_offloads(VirtIONet *n)
{
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
    return virtio_net_guest_offloads_by_features(vdev->guest_features);
}

G
Gerd Hoffmann 已提交
674
static void virtio_net_set_features(VirtIODevice *vdev, uint64_t features)
A
aliguori 已提交
675
{
676
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
677 678
    int i;

679 680 681 682 683
    if (n->mtu_bypass_backend &&
            !virtio_has_feature(vdev->backend_features, VIRTIO_NET_F_MTU)) {
        features &= ~(1ULL << VIRTIO_NET_F_MTU);
    }

684
    virtio_net_set_multiqueue(n,
685
                              virtio_has_feature(features, VIRTIO_NET_F_MQ));
A
aliguori 已提交
686

687
    virtio_net_set_mrg_rx_bufs(n,
688 689 690 691
                               virtio_has_feature(features,
                                                  VIRTIO_NET_F_MRG_RXBUF),
                               virtio_has_feature(features,
                                                  VIRTIO_F_VERSION_1));
692 693

    if (n->has_vnet_hdr) {
694 695 696
        n->curr_guest_offloads =
            virtio_net_guest_offloads_by_features(features);
        virtio_net_apply_guest_offloads(n);
697
    }
J
Jason Wang 已提交
698 699 700 701

    for (i = 0;  i < n->max_queues; i++) {
        NetClientState *nc = qemu_get_subqueue(n->nic, i);

702
        if (!get_vhost_net(nc->peer)) {
J
Jason Wang 已提交
703 704
            continue;
        }
705
        vhost_net_ack_features(get_vhost_net(nc->peer), features);
D
David L Stevens 已提交
706
    }
707

708
    if (virtio_has_feature(features, VIRTIO_NET_F_CTRL_VLAN)) {
709 710 711 712
        memset(n->vlans, 0, MAX_VLAN >> 3);
    } else {
        memset(n->vlans, 0xff, MAX_VLAN >> 3);
    }
A
aliguori 已提交
713 714
}

715
static int virtio_net_handle_rx_mode(VirtIONet *n, uint8_t cmd,
716
                                     struct iovec *iov, unsigned int iov_cnt)
717 718
{
    uint8_t on;
719
    size_t s;
720
    NetClientState *nc = qemu_get_queue(n->nic);
721

722 723 724
    s = iov_to_buf(iov, iov_cnt, 0, &on, sizeof(on));
    if (s != sizeof(on)) {
        return VIRTIO_NET_ERR;
725 726
    }

A
Amos Kong 已提交
727
    if (cmd == VIRTIO_NET_CTRL_RX_PROMISC) {
728
        n->promisc = on;
A
Amos Kong 已提交
729
    } else if (cmd == VIRTIO_NET_CTRL_RX_ALLMULTI) {
730
        n->allmulti = on;
A
Amos Kong 已提交
731
    } else if (cmd == VIRTIO_NET_CTRL_RX_ALLUNI) {
732
        n->alluni = on;
A
Amos Kong 已提交
733
    } else if (cmd == VIRTIO_NET_CTRL_RX_NOMULTI) {
734
        n->nomulti = on;
A
Amos Kong 已提交
735
    } else if (cmd == VIRTIO_NET_CTRL_RX_NOUNI) {
736
        n->nouni = on;
A
Amos Kong 已提交
737
    } else if (cmd == VIRTIO_NET_CTRL_RX_NOBCAST) {
738
        n->nobcast = on;
739
    } else {
740
        return VIRTIO_NET_ERR;
741
    }
742

743 744
    rxfilter_notify(nc);

745 746 747
    return VIRTIO_NET_OK;
}

748 749 750 751 752 753 754
static int virtio_net_handle_offloads(VirtIONet *n, uint8_t cmd,
                                     struct iovec *iov, unsigned int iov_cnt)
{
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
    uint64_t offloads;
    size_t s;

755
    if (!virtio_vdev_has_feature(vdev, VIRTIO_NET_F_CTRL_GUEST_OFFLOADS)) {
756 757 758 759 760 761 762 763 764 765 766
        return VIRTIO_NET_ERR;
    }

    s = iov_to_buf(iov, iov_cnt, 0, &offloads, sizeof(offloads));
    if (s != sizeof(offloads)) {
        return VIRTIO_NET_ERR;
    }

    if (cmd == VIRTIO_NET_CTRL_GUEST_OFFLOADS_SET) {
        uint64_t supported_offloads;

J
Jason Wang 已提交
767 768
        offloads = virtio_ldq_p(vdev, &offloads);

769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786
        if (!n->has_vnet_hdr) {
            return VIRTIO_NET_ERR;
        }

        supported_offloads = virtio_net_supported_guest_offloads(n);
        if (offloads & ~supported_offloads) {
            return VIRTIO_NET_ERR;
        }

        n->curr_guest_offloads = offloads;
        virtio_net_apply_guest_offloads(n);

        return VIRTIO_NET_OK;
    } else {
        return VIRTIO_NET_ERR;
    }
}

787
static int virtio_net_handle_mac(VirtIONet *n, uint8_t cmd,
788
                                 struct iovec *iov, unsigned int iov_cnt)
789
{
790
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
791
    struct virtio_net_ctrl_mac mac_data;
792
    size_t s;
793
    NetClientState *nc = qemu_get_queue(n->nic);
794

795 796 797 798 799 800
    if (cmd == VIRTIO_NET_CTRL_MAC_ADDR_SET) {
        if (iov_size(iov, iov_cnt) != sizeof(n->mac)) {
            return VIRTIO_NET_ERR;
        }
        s = iov_to_buf(iov, iov_cnt, 0, &n->mac, sizeof(n->mac));
        assert(s == sizeof(n->mac));
J
Jason Wang 已提交
801
        qemu_format_nic_info_str(qemu_get_queue(n->nic), n->mac);
802 803
        rxfilter_notify(nc);

804 805 806
        return VIRTIO_NET_OK;
    }

807
    if (cmd != VIRTIO_NET_CTRL_MAC_TABLE_SET) {
808
        return VIRTIO_NET_ERR;
809
    }
810

811 812 813 814 815
    int in_use = 0;
    int first_multi = 0;
    uint8_t uni_overflow = 0;
    uint8_t multi_overflow = 0;
    uint8_t *macs = g_malloc0(MAC_TABLE_ENTRIES * ETH_ALEN);
816

817 818
    s = iov_to_buf(iov, iov_cnt, 0, &mac_data.entries,
                   sizeof(mac_data.entries));
819
    mac_data.entries = virtio_ldl_p(vdev, &mac_data.entries);
820
    if (s != sizeof(mac_data.entries)) {
821
        goto error;
822 823
    }
    iov_discard_front(&iov, &iov_cnt, s);
824

825
    if (mac_data.entries * ETH_ALEN > iov_size(iov, iov_cnt)) {
826
        goto error;
827
    }
828 829

    if (mac_data.entries <= MAC_TABLE_ENTRIES) {
830
        s = iov_to_buf(iov, iov_cnt, 0, macs,
831 832
                       mac_data.entries * ETH_ALEN);
        if (s != mac_data.entries * ETH_ALEN) {
833
            goto error;
834
        }
835
        in_use += mac_data.entries;
836
    } else {
837
        uni_overflow = 1;
838 839
    }

840 841
    iov_discard_front(&iov, &iov_cnt, mac_data.entries * ETH_ALEN);

842
    first_multi = in_use;
843

844 845
    s = iov_to_buf(iov, iov_cnt, 0, &mac_data.entries,
                   sizeof(mac_data.entries));
846
    mac_data.entries = virtio_ldl_p(vdev, &mac_data.entries);
847
    if (s != sizeof(mac_data.entries)) {
848
        goto error;
849 850 851
    }

    iov_discard_front(&iov, &iov_cnt, s);
852

853
    if (mac_data.entries * ETH_ALEN != iov_size(iov, iov_cnt)) {
854
        goto error;
855
    }
856

857
    if (mac_data.entries <= MAC_TABLE_ENTRIES - in_use) {
858
        s = iov_to_buf(iov, iov_cnt, 0, &macs[in_use * ETH_ALEN],
859 860
                       mac_data.entries * ETH_ALEN);
        if (s != mac_data.entries * ETH_ALEN) {
861
            goto error;
862
        }
863
        in_use += mac_data.entries;
864
    } else {
865
        multi_overflow = 1;
866 867
    }

868 869 870 871 872 873
    n->mac_table.in_use = in_use;
    n->mac_table.first_multi = first_multi;
    n->mac_table.uni_overflow = uni_overflow;
    n->mac_table.multi_overflow = multi_overflow;
    memcpy(n->mac_table.macs, macs, MAC_TABLE_ENTRIES * ETH_ALEN);
    g_free(macs);
874 875
    rxfilter_notify(nc);

876
    return VIRTIO_NET_OK;
877 878

error:
879
    g_free(macs);
880
    return VIRTIO_NET_ERR;
881 882
}

883
static int virtio_net_handle_vlan_table(VirtIONet *n, uint8_t cmd,
884
                                        struct iovec *iov, unsigned int iov_cnt)
885
{
886
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
887
    uint16_t vid;
888
    size_t s;
889
    NetClientState *nc = qemu_get_queue(n->nic);
890

891
    s = iov_to_buf(iov, iov_cnt, 0, &vid, sizeof(vid));
892
    vid = virtio_lduw_p(vdev, &vid);
893
    if (s != sizeof(vid)) {
894 895 896 897 898 899 900 901 902 903 904 905 906
        return VIRTIO_NET_ERR;
    }

    if (vid >= MAX_VLAN)
        return VIRTIO_NET_ERR;

    if (cmd == VIRTIO_NET_CTRL_VLAN_ADD)
        n->vlans[vid >> 5] |= (1U << (vid & 0x1f));
    else if (cmd == VIRTIO_NET_CTRL_VLAN_DEL)
        n->vlans[vid >> 5] &= ~(1U << (vid & 0x1f));
    else
        return VIRTIO_NET_ERR;

907 908
    rxfilter_notify(nc);

909 910 911
    return VIRTIO_NET_OK;
}

J
Jason Wang 已提交
912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928
static int virtio_net_handle_announce(VirtIONet *n, uint8_t cmd,
                                      struct iovec *iov, unsigned int iov_cnt)
{
    if (cmd == VIRTIO_NET_CTRL_ANNOUNCE_ACK &&
        n->status & VIRTIO_NET_S_ANNOUNCE) {
        n->status &= ~VIRTIO_NET_S_ANNOUNCE;
        if (n->announce_counter) {
            timer_mod(n->announce_timer,
                      qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL) +
                      self_announce_delay(n->announce_counter));
        }
        return VIRTIO_NET_OK;
    } else {
        return VIRTIO_NET_ERR;
    }
}

J
Jason Wang 已提交
929
static int virtio_net_handle_mq(VirtIONet *n, uint8_t cmd,
930
                                struct iovec *iov, unsigned int iov_cnt)
J
Jason Wang 已提交
931
{
932
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
933 934 935
    struct virtio_net_ctrl_mq mq;
    size_t s;
    uint16_t queues;
J
Jason Wang 已提交
936

937 938
    s = iov_to_buf(iov, iov_cnt, 0, &mq, sizeof(mq));
    if (s != sizeof(mq)) {
J
Jason Wang 已提交
939 940 941 942 943 944 945
        return VIRTIO_NET_ERR;
    }

    if (cmd != VIRTIO_NET_CTRL_MQ_VQ_PAIRS_SET) {
        return VIRTIO_NET_ERR;
    }

946
    queues = virtio_lduw_p(vdev, &mq.virtqueue_pairs);
J
Jason Wang 已提交
947

948 949 950
    if (queues < VIRTIO_NET_CTRL_MQ_VQ_PAIRS_MIN ||
        queues > VIRTIO_NET_CTRL_MQ_VQ_PAIRS_MAX ||
        queues > n->max_queues ||
J
Jason Wang 已提交
951 952 953 954
        !n->multiqueue) {
        return VIRTIO_NET_ERR;
    }

955
    n->curr_queues = queues;
J
Jason Wang 已提交
956 957
    /* stop the backend before changing the number of queues to avoid handling a
     * disabled queue */
958
    virtio_net_set_status(vdev, vdev->status);
J
Jason Wang 已提交
959 960 961 962
    virtio_net_set_queues(n);

    return VIRTIO_NET_OK;
}
963

964 965
static void virtio_net_handle_ctrl(VirtIODevice *vdev, VirtQueue *vq)
{
966
    VirtIONet *n = VIRTIO_NET(vdev);
967 968
    struct virtio_net_ctrl_hdr ctrl;
    virtio_net_ctrl_ack status = VIRTIO_NET_ERR;
969
    VirtQueueElement *elem;
970
    size_t s;
J
Jason Wang 已提交
971
    struct iovec *iov, *iov2;
972
    unsigned int iov_cnt;
973

974 975 976 977 978 979 980
    for (;;) {
        elem = virtqueue_pop(vq, sizeof(VirtQueueElement));
        if (!elem) {
            break;
        }
        if (iov_size(elem->in_sg, elem->in_num) < sizeof(status) ||
            iov_size(elem->out_sg, elem->out_num) < sizeof(ctrl)) {
981 982 983 984
            virtio_error(vdev, "virtio-net ctrl missing headers");
            virtqueue_detach_element(vq, elem, 0);
            g_free(elem);
            break;
985 986
        }

987 988
        iov_cnt = elem->out_num;
        iov2 = iov = g_memdup(elem->out_sg, sizeof(struct iovec) * elem->out_num);
989 990 991 992
        s = iov_to_buf(iov, iov_cnt, 0, &ctrl, sizeof(ctrl));
        iov_discard_front(&iov, &iov_cnt, sizeof(ctrl));
        if (s != sizeof(ctrl)) {
            status = VIRTIO_NET_ERR;
A
Amos Kong 已提交
993
        } else if (ctrl.class == VIRTIO_NET_CTRL_RX) {
994 995 996 997 998
            status = virtio_net_handle_rx_mode(n, ctrl.cmd, iov, iov_cnt);
        } else if (ctrl.class == VIRTIO_NET_CTRL_MAC) {
            status = virtio_net_handle_mac(n, ctrl.cmd, iov, iov_cnt);
        } else if (ctrl.class == VIRTIO_NET_CTRL_VLAN) {
            status = virtio_net_handle_vlan_table(n, ctrl.cmd, iov, iov_cnt);
J
Jason Wang 已提交
999 1000
        } else if (ctrl.class == VIRTIO_NET_CTRL_ANNOUNCE) {
            status = virtio_net_handle_announce(n, ctrl.cmd, iov, iov_cnt);
J
Jason Wang 已提交
1001
        } else if (ctrl.class == VIRTIO_NET_CTRL_MQ) {
1002
            status = virtio_net_handle_mq(n, ctrl.cmd, iov, iov_cnt);
1003 1004
        } else if (ctrl.class == VIRTIO_NET_CTRL_GUEST_OFFLOADS) {
            status = virtio_net_handle_offloads(n, ctrl.cmd, iov, iov_cnt);
1005 1006
        }

1007
        s = iov_from_buf(elem->in_sg, elem->in_num, 0, &status, sizeof(status));
1008
        assert(s == sizeof(status));
1009

1010
        virtqueue_push(vq, elem, sizeof(status));
1011
        virtio_notify(vdev, vq);
J
Jason Wang 已提交
1012
        g_free(iov2);
1013
        g_free(elem);
1014 1015 1016
    }
}

A
aliguori 已提交
1017 1018 1019 1020
/* RX */

static void virtio_net_handle_rx(VirtIODevice *vdev, VirtQueue *vq)
{
1021
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
1022
    int queue_index = vq2q(virtio_get_queue_index(vq));
1023

J
Jason Wang 已提交
1024
    qemu_flush_queued_packets(qemu_get_subqueue(n->nic, queue_index));
A
aliguori 已提交
1025 1026
}

1027
static int virtio_net_can_receive(NetClientState *nc)
A
aliguori 已提交
1028
{
J
Jason Wang 已提交
1029
    VirtIONet *n = qemu_get_nic_opaque(nc);
1030
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
J
Jason Wang 已提交
1031
    VirtIONetQueue *q = virtio_net_get_subqueue(nc);
1032

1033
    if (!vdev->vm_running) {
1034 1035
        return 0;
    }
1036

J
Jason Wang 已提交
1037 1038 1039 1040
    if (nc->queue_index >= n->curr_queues) {
        return 0;
    }

1041
    if (!virtio_queue_ready(q->rx_vq) ||
1042
        !(vdev->status & VIRTIO_CONFIG_S_DRIVER_OK)) {
A
aliguori 已提交
1043
        return 0;
1044
    }
A
aliguori 已提交
1045

1046 1047 1048
    return 1;
}

1049
static int virtio_net_has_buffers(VirtIONetQueue *q, int bufsize)
1050
{
1051 1052
    VirtIONet *n = q->n;
    if (virtio_queue_empty(q->rx_vq) ||
A
aliguori 已提交
1053
        (n->mergeable_rx_bufs &&
1054 1055
         !virtqueue_avail_bytes(q->rx_vq, bufsize, 0))) {
        virtio_queue_set_notification(q->rx_vq, 1);
1056 1057 1058 1059 1060

        /* To avoid a race condition where the guest has made some buffers
         * available after the above check but before notification was
         * enabled, check for available buffers again.
         */
1061
        if (virtio_queue_empty(q->rx_vq) ||
1062
            (n->mergeable_rx_bufs &&
1063
             !virtqueue_avail_bytes(q->rx_vq, bufsize, 0))) {
1064
            return 0;
1065
        }
A
aliguori 已提交
1066 1067
    }

1068
    virtio_queue_set_notification(q->rx_vq, 0);
A
aliguori 已提交
1069 1070 1071
    return 1;
}

1072
static void virtio_net_hdr_swap(VirtIODevice *vdev, struct virtio_net_hdr *hdr)
1073
{
1074 1075 1076 1077
    virtio_tswap16s(vdev, &hdr->hdr_len);
    virtio_tswap16s(vdev, &hdr->gso_size);
    virtio_tswap16s(vdev, &hdr->csum_start);
    virtio_tswap16s(vdev, &hdr->csum_offset);
1078 1079
}

A
Anthony Liguori 已提交
1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094
/* dhclient uses AF_PACKET but doesn't pass auxdata to the kernel so
 * it never finds out that the packets don't have valid checksums.  This
 * causes dhclient to get upset.  Fedora's carried a patch for ages to
 * fix this with Xen but it hasn't appeared in an upstream release of
 * dhclient yet.
 *
 * To avoid breaking existing guests, we catch udp packets and add
 * checksums.  This is terrible but it's better than hacking the guest
 * kernels.
 *
 * N.B. if we introduce a zero-copy API, this operation is no longer free so
 * we should provide a mechanism to disable it to avoid polluting the host
 * cache.
 */
static void work_around_broken_dhclient(struct virtio_net_hdr *hdr,
M
Michael S. Tsirkin 已提交
1095
                                        uint8_t *buf, size_t size)
A
Anthony Liguori 已提交
1096 1097 1098 1099 1100 1101
{
    if ((hdr->flags & VIRTIO_NET_HDR_F_NEEDS_CSUM) && /* missing csum */
        (size > 27 && size < 1500) && /* normal sized MTU */
        (buf[12] == 0x08 && buf[13] == 0x00) && /* ethertype == IPv4 */
        (buf[23] == 17) && /* ip.protocol == UDP */
        (buf[34] == 0 && buf[35] == 67)) { /* udp.srcport == bootps */
M
Michael S. Tsirkin 已提交
1102
        net_checksum_calculate(buf, size);
A
Anthony Liguori 已提交
1103 1104 1105 1106
        hdr->flags &= ~VIRTIO_NET_HDR_F_NEEDS_CSUM;
    }
}

1107 1108
static void receive_header(VirtIONet *n, const struct iovec *iov, int iov_cnt,
                           const void *buf, size_t size)
A
aliguori 已提交
1109
{
M
Mark McLoughlin 已提交
1110
    if (n->has_vnet_hdr) {
M
Michael S. Tsirkin 已提交
1111 1112
        /* FIXME this cast is evil */
        void *wbuf = (void *)buf;
1113 1114
        work_around_broken_dhclient(wbuf, wbuf + n->host_hdr_len,
                                    size - n->host_hdr_len);
1115 1116 1117 1118

        if (n->needs_vnet_hdr_swap) {
            virtio_net_hdr_swap(VIRTIO_DEVICE(n), wbuf);
        }
1119
        iov_from_buf(iov, iov_cnt, 0, buf, sizeof(struct virtio_net_hdr));
M
Michael S. Tsirkin 已提交
1120 1121 1122 1123 1124 1125
    } else {
        struct virtio_net_hdr hdr = {
            .flags = 0,
            .gso_type = VIRTIO_NET_HDR_GSO_NONE
        };
        iov_from_buf(iov, iov_cnt, 0, &hdr, sizeof hdr);
M
Mark McLoughlin 已提交
1126
    }
A
aliguori 已提交
1127 1128
}

1129 1130 1131
static int receive_filter(VirtIONet *n, const uint8_t *buf, int size)
{
    static const uint8_t bcast[] = {0xff, 0xff, 0xff, 0xff, 0xff, 0xff};
1132
    static const uint8_t vlan[] = {0x81, 0x00};
1133
    uint8_t *ptr = (uint8_t *)buf;
1134
    int i;
1135 1136 1137 1138

    if (n->promisc)
        return 1;

1139
    ptr += n->host_hdr_len;
M
Mark McLoughlin 已提交
1140

1141
    if (!memcmp(&ptr[12], vlan, sizeof(vlan))) {
1142
        int vid = lduw_be_p(ptr + 14) & 0xfff;
1143 1144 1145 1146
        if (!(n->vlans[vid >> 5] & (1U << (vid & 0x1f))))
            return 0;
    }

1147 1148
    if (ptr[0] & 1) { // multicast
        if (!memcmp(ptr, bcast, sizeof(bcast))) {
1149 1150 1151
            return !n->nobcast;
        } else if (n->nomulti) {
            return 0;
1152
        } else if (n->allmulti || n->mac_table.multi_overflow) {
1153 1154
            return 1;
        }
1155 1156 1157 1158 1159 1160

        for (i = n->mac_table.first_multi; i < n->mac_table.in_use; i++) {
            if (!memcmp(ptr, &n->mac_table.macs[i * ETH_ALEN], ETH_ALEN)) {
                return 1;
            }
        }
1161
    } else { // unicast
1162 1163 1164
        if (n->nouni) {
            return 0;
        } else if (n->alluni || n->mac_table.uni_overflow) {
1165 1166
            return 1;
        } else if (!memcmp(ptr, n->mac, ETH_ALEN)) {
1167 1168
            return 1;
        }
1169

1170 1171 1172 1173 1174
        for (i = 0; i < n->mac_table.first_multi; i++) {
            if (!memcmp(ptr, &n->mac_table.macs[i * ETH_ALEN], ETH_ALEN)) {
                return 1;
            }
        }
1175 1176
    }

1177 1178 1179
    return 0;
}

1180 1181
static ssize_t virtio_net_receive_rcu(NetClientState *nc, const uint8_t *buf,
                                      size_t size)
A
aliguori 已提交
1182
{
J
Jason Wang 已提交
1183
    VirtIONet *n = qemu_get_nic_opaque(nc);
J
Jason Wang 已提交
1184
    VirtIONetQueue *q = virtio_net_get_subqueue(nc);
1185
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
1186 1187 1188
    struct iovec mhdr_sg[VIRTQUEUE_MAX_SIZE];
    struct virtio_net_hdr_mrg_rxbuf mhdr;
    unsigned mhdr_cnt = 0;
M
Michael S. Tsirkin 已提交
1189
    size_t offset, i, guest_offset;
A
aliguori 已提交
1190

J
Jason Wang 已提交
1191
    if (!virtio_net_can_receive(nc)) {
1192
        return -1;
J
Jason Wang 已提交
1193
    }
1194

1195
    /* hdr_len refers to the header we supply to the guest */
1196
    if (!virtio_net_has_buffers(q, size + n->guest_hdr_len - n->host_hdr_len)) {
1197
        return 0;
1198
    }
A
aliguori 已提交
1199

1200
    if (!receive_filter(n, buf, size))
1201
        return size;
1202

A
aliguori 已提交
1203 1204 1205
    offset = i = 0;

    while (offset < size) {
1206
        VirtQueueElement *elem;
A
aliguori 已提交
1207
        int len, total;
1208
        const struct iovec *sg;
A
aliguori 已提交
1209

A
Amit Shah 已提交
1210
        total = 0;
A
aliguori 已提交
1211

1212 1213
        elem = virtqueue_pop(q->rx_vq, sizeof(VirtQueueElement));
        if (!elem) {
1214 1215 1216 1217 1218 1219 1220 1221 1222 1223
            if (i) {
                virtio_error(vdev, "virtio-net unexpected empty queue: "
                             "i %zd mergeable %d offset %zd, size %zd, "
                             "guest hdr len %zd, host hdr len %zd "
                             "guest features 0x%" PRIx64,
                             i, n->mergeable_rx_bufs, offset, size,
                             n->guest_hdr_len, n->host_hdr_len,
                             vdev->guest_features);
            }
            return -1;
A
aliguori 已提交
1224 1225
        }

1226
        if (elem->in_num < 1) {
1227 1228 1229 1230 1231
            virtio_error(vdev,
                         "virtio-net receive queue contains no in buffers");
            virtqueue_detach_element(q->rx_vq, elem, 0);
            g_free(elem);
            return -1;
A
aliguori 已提交
1232 1233
        }

1234
        sg = elem->in_sg;
A
aliguori 已提交
1235
        if (i == 0) {
1236
            assert(offset == 0);
1237 1238
            if (n->mergeable_rx_bufs) {
                mhdr_cnt = iov_copy(mhdr_sg, ARRAY_SIZE(mhdr_sg),
1239
                                    sg, elem->in_num,
1240 1241 1242
                                    offsetof(typeof(mhdr), num_buffers),
                                    sizeof(mhdr.num_buffers));
            }
A
aliguori 已提交
1243

1244
            receive_header(n, sg, elem->in_num, buf, size);
1245
            offset = n->host_hdr_len;
1246
            total += n->guest_hdr_len;
M
Michael S. Tsirkin 已提交
1247 1248 1249
            guest_offset = n->guest_hdr_len;
        } else {
            guest_offset = 0;
A
aliguori 已提交
1250 1251 1252
        }

        /* copy in packet.  ugh */
1253
        len = iov_from_buf(sg, elem->in_num, guest_offset,
1254
                           buf + offset, size - offset);
A
aliguori 已提交
1255
        total += len;
1256 1257 1258 1259 1260
        offset += len;
        /* If buffers can't be merged, at this point we
         * must have consumed the complete packet.
         * Otherwise, drop it. */
        if (!n->mergeable_rx_bufs && offset < size) {
1261
            virtqueue_unpop(q->rx_vq, elem, total);
1262
            g_free(elem);
1263 1264
            return size;
        }
A
aliguori 已提交
1265 1266

        /* signal other side */
1267 1268
        virtqueue_fill(q->rx_vq, elem, total, i++);
        g_free(elem);
A
aliguori 已提交
1269 1270
    }

1271
    if (mhdr_cnt) {
1272
        virtio_stw_p(vdev, &mhdr.num_buffers, i);
1273 1274 1275
        iov_from_buf(mhdr_sg, mhdr_cnt,
                     0,
                     &mhdr.num_buffers, sizeof mhdr.num_buffers);
1276
    }
A
aliguori 已提交
1277

1278
    virtqueue_flush(q->rx_vq, i);
1279
    virtio_notify(vdev, q->rx_vq);
1280 1281

    return size;
A
aliguori 已提交
1282 1283
}

1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294
static ssize_t virtio_net_receive(NetClientState *nc, const uint8_t *buf,
                                  size_t size)
{
    ssize_t r;

    rcu_read_lock();
    r = virtio_net_receive_rcu(nc, buf, size);
    rcu_read_unlock();
    return r;
}

1295
static int32_t virtio_net_flush_tx(VirtIONetQueue *q);
1296

1297
static void virtio_net_tx_complete(NetClientState *nc, ssize_t len)
1298
{
J
Jason Wang 已提交
1299
    VirtIONet *n = qemu_get_nic_opaque(nc);
J
Jason Wang 已提交
1300
    VirtIONetQueue *q = virtio_net_get_subqueue(nc);
1301
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
1302

1303
    virtqueue_push(q->tx_vq, q->async_tx.elem, 0);
1304
    virtio_notify(vdev, q->tx_vq);
1305

1306 1307
    g_free(q->async_tx.elem);
    q->async_tx.elem = NULL;
1308

1309 1310
    virtio_queue_set_notification(q->tx_vq, 1);
    virtio_net_flush_tx(q);
1311 1312
}

A
aliguori 已提交
1313
/* TX */
1314
static int32_t virtio_net_flush_tx(VirtIONetQueue *q)
A
aliguori 已提交
1315
{
1316
    VirtIONet *n = q->n;
1317
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
1318
    VirtQueueElement *elem;
1319
    int32_t num_packets = 0;
J
Jason Wang 已提交
1320
    int queue_index = vq2q(virtio_get_queue_index(q->tx_vq));
1321
    if (!(vdev->status & VIRTIO_CONFIG_S_DRIVER_OK)) {
1322 1323
        return num_packets;
    }
A
aliguori 已提交
1324

1325
    if (q->async_tx.elem) {
1326
        virtio_queue_set_notification(q->tx_vq, 0);
1327
        return num_packets;
1328 1329
    }

1330
    for (;;) {
J
Jason Wang 已提交
1331
        ssize_t ret;
1332 1333
        unsigned int out_num;
        struct iovec sg[VIRTQUEUE_MAX_SIZE], sg2[VIRTQUEUE_MAX_SIZE + 1], *out_sg;
J
Jason Wang 已提交
1334
        struct virtio_net_hdr_mrg_rxbuf mhdr;
A
aliguori 已提交
1335

1336 1337 1338 1339 1340 1341 1342
        elem = virtqueue_pop(q->tx_vq, sizeof(VirtQueueElement));
        if (!elem) {
            break;
        }

        out_num = elem->out_num;
        out_sg = elem->out_sg;
M
Michael S. Tsirkin 已提交
1343
        if (out_num < 1) {
1344 1345 1346 1347
            virtio_error(vdev, "virtio-net header not in first element");
            virtqueue_detach_element(q->tx_vq, elem, 0);
            g_free(elem);
            return -EINVAL;
A
aliguori 已提交
1348 1349
        }

1350
        if (n->has_vnet_hdr) {
J
Jason Wang 已提交
1351 1352
            if (iov_to_buf(out_sg, out_num, 0, &mhdr, n->guest_hdr_len) <
                n->guest_hdr_len) {
1353 1354 1355 1356
                virtio_error(vdev, "virtio-net header incorrect");
                virtqueue_detach_element(q->tx_vq, elem, 0);
                g_free(elem);
                return -EINVAL;
1357
            }
1358
            if (n->needs_vnet_hdr_swap) {
J
Jason Wang 已提交
1359 1360 1361 1362 1363 1364 1365 1366 1367 1368 1369 1370
                virtio_net_hdr_swap(vdev, (void *) &mhdr);
                sg2[0].iov_base = &mhdr;
                sg2[0].iov_len = n->guest_hdr_len;
                out_num = iov_copy(&sg2[1], ARRAY_SIZE(sg2) - 1,
                                   out_sg, out_num,
                                   n->guest_hdr_len, -1);
                if (out_num == VIRTQUEUE_MAX_SIZE) {
                    goto drop;
		}
                out_num += 1;
                out_sg = sg2;
	    }
1371
        }
1372 1373 1374 1375 1376 1377 1378 1379 1380 1381 1382 1383 1384 1385 1386
        /*
         * If host wants to see the guest header as is, we can
         * pass it on unchanged. Otherwise, copy just the parts
         * that host is interested in.
         */
        assert(n->host_hdr_len <= n->guest_hdr_len);
        if (n->host_hdr_len != n->guest_hdr_len) {
            unsigned sg_num = iov_copy(sg, ARRAY_SIZE(sg),
                                       out_sg, out_num,
                                       0, n->host_hdr_len);
            sg_num += iov_copy(sg + sg_num, ARRAY_SIZE(sg) - sg_num,
                             out_sg, out_num,
                             n->guest_hdr_len, -1);
            out_num = sg_num;
            out_sg = sg;
A
aliguori 已提交
1387 1388
        }

J
Jason Wang 已提交
1389 1390
        ret = qemu_sendv_packet_async(qemu_get_subqueue(n->nic, queue_index),
                                      out_sg, out_num, virtio_net_tx_complete);
1391
        if (ret == 0) {
1392 1393
            virtio_queue_set_notification(q->tx_vq, 0);
            q->async_tx.elem = elem;
1394
            return -EBUSY;
1395 1396
        }

J
Jason Wang 已提交
1397
drop:
1398
        virtqueue_push(q->tx_vq, elem, 0);
1399
        virtio_notify(vdev, q->tx_vq);
1400
        g_free(elem);
1401 1402 1403 1404

        if (++num_packets >= n->tx_burst) {
            break;
        }
A
aliguori 已提交
1405
    }
1406
    return num_packets;
A
aliguori 已提交
1407 1408
}

1409
static void virtio_net_handle_tx_timer(VirtIODevice *vdev, VirtQueue *vq)
A
aliguori 已提交
1410
{
1411
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
1412
    VirtIONetQueue *q = &n->vqs[vq2q(virtio_get_queue_index(vq))];
A
aliguori 已提交
1413

1414 1415 1416 1417 1418
    if (unlikely((n->status & VIRTIO_NET_S_LINK_UP) == 0)) {
        virtio_net_drop_tx_queue_data(vdev, vq);
        return;
    }

1419
    /* This happens when device was stopped but VCPU wasn't. */
1420
    if (!vdev->vm_running) {
1421
        q->tx_waiting = 1;
1422 1423 1424
        return;
    }

1425
    if (q->tx_waiting) {
A
aliguori 已提交
1426
        virtio_queue_set_notification(vq, 1);
1427
        timer_del(q->tx_timer);
1428
        q->tx_waiting = 0;
1429 1430 1431
        if (virtio_net_flush_tx(q) == -EINVAL) {
            return;
        }
A
aliguori 已提交
1432
    } else {
1433 1434
        timer_mod(q->tx_timer,
                       qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + n->tx_timeout);
1435
        q->tx_waiting = 1;
A
aliguori 已提交
1436 1437 1438 1439
        virtio_queue_set_notification(vq, 0);
    }
}

1440 1441
static void virtio_net_handle_tx_bh(VirtIODevice *vdev, VirtQueue *vq)
{
1442
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
1443
    VirtIONetQueue *q = &n->vqs[vq2q(virtio_get_queue_index(vq))];
1444

1445 1446 1447 1448 1449
    if (unlikely((n->status & VIRTIO_NET_S_LINK_UP) == 0)) {
        virtio_net_drop_tx_queue_data(vdev, vq);
        return;
    }

1450
    if (unlikely(q->tx_waiting)) {
1451 1452
        return;
    }
1453
    q->tx_waiting = 1;
1454
    /* This happens when device was stopped but VCPU wasn't. */
1455
    if (!vdev->vm_running) {
1456 1457
        return;
    }
1458
    virtio_queue_set_notification(vq, 0);
1459
    qemu_bh_schedule(q->tx_bh);
1460 1461
}

A
aliguori 已提交
1462 1463
static void virtio_net_tx_timer(void *opaque)
{
1464 1465
    VirtIONetQueue *q = opaque;
    VirtIONet *n = q->n;
1466
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
1467 1468 1469 1470 1471 1472
    /* This happens when device was stopped but BH wasn't. */
    if (!vdev->vm_running) {
        /* Make sure tx waiting is set, so we'll run when restarted. */
        assert(q->tx_waiting);
        return;
    }
A
aliguori 已提交
1473

1474
    q->tx_waiting = 0;
A
aliguori 已提交
1475 1476

    /* Just in case the driver is not ready on more */
1477
    if (!(vdev->status & VIRTIO_CONFIG_S_DRIVER_OK)) {
A
aliguori 已提交
1478
        return;
1479
    }
A
aliguori 已提交
1480

1481 1482
    virtio_queue_set_notification(q->tx_vq, 1);
    virtio_net_flush_tx(q);
A
aliguori 已提交
1483 1484
}

1485 1486
static void virtio_net_tx_bh(void *opaque)
{
1487 1488
    VirtIONetQueue *q = opaque;
    VirtIONet *n = q->n;
1489
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
1490 1491
    int32_t ret;

1492 1493 1494 1495 1496 1497
    /* This happens when device was stopped but BH wasn't. */
    if (!vdev->vm_running) {
        /* Make sure tx waiting is set, so we'll run when restarted. */
        assert(q->tx_waiting);
        return;
    }
1498

1499
    q->tx_waiting = 0;
1500 1501

    /* Just in case the driver is not ready on more */
1502
    if (unlikely(!(vdev->status & VIRTIO_CONFIG_S_DRIVER_OK))) {
1503
        return;
1504
    }
1505

1506
    ret = virtio_net_flush_tx(q);
1507 1508 1509
    if (ret == -EBUSY || ret == -EINVAL) {
        return; /* Notification re-enable handled by tx_complete or device
                 * broken */
1510 1511 1512 1513 1514
    }

    /* If we flush a full burst of packets, assume there are
     * more coming and immediately reschedule */
    if (ret >= n->tx_burst) {
1515 1516
        qemu_bh_schedule(q->tx_bh);
        q->tx_waiting = 1;
1517 1518 1519 1520 1521 1522
        return;
    }

    /* If less than a full burst, re-enable notification and flush
     * anything that may have come in while we weren't looking.  If
     * we find something, assume the guest is still active and reschedule */
1523
    virtio_queue_set_notification(q->tx_vq, 1);
1524 1525 1526 1527
    ret = virtio_net_flush_tx(q);
    if (ret == -EINVAL) {
        return;
    } else if (ret > 0) {
1528 1529 1530
        virtio_queue_set_notification(q->tx_vq, 0);
        qemu_bh_schedule(q->tx_bh);
        q->tx_waiting = 1;
1531 1532 1533
    }
}

1534 1535 1536 1537
static void virtio_net_add_queue(VirtIONet *n, int index)
{
    VirtIODevice *vdev = VIRTIO_DEVICE(n);

1538 1539
    n->vqs[index].rx_vq = virtio_add_queue(vdev, n->net_conf.rx_queue_size,
                                           virtio_net_handle_rx);
1540

1541 1542
    if (n->net_conf.tx && !strcmp(n->net_conf.tx, "timer")) {
        n->vqs[index].tx_vq =
1543 1544
            virtio_add_queue(vdev, n->net_conf.tx_queue_size,
                             virtio_net_handle_tx_timer);
1545 1546 1547 1548 1549
        n->vqs[index].tx_timer = timer_new_ns(QEMU_CLOCK_VIRTUAL,
                                              virtio_net_tx_timer,
                                              &n->vqs[index]);
    } else {
        n->vqs[index].tx_vq =
1550 1551
            virtio_add_queue(vdev, n->net_conf.tx_queue_size,
                             virtio_net_handle_tx_bh);
1552 1553 1554 1555 1556 1557 1558 1559 1560 1561 1562 1563 1564 1565 1566 1567 1568 1569 1570
        n->vqs[index].tx_bh = qemu_bh_new(virtio_net_tx_bh, &n->vqs[index]);
    }

    n->vqs[index].tx_waiting = 0;
    n->vqs[index].n = n;
}

static void virtio_net_del_queue(VirtIONet *n, int index)
{
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
    VirtIONetQueue *q = &n->vqs[index];
    NetClientState *nc = qemu_get_subqueue(n->nic, index);

    qemu_purge_queued_packets(nc);

    virtio_del_queue(vdev, index * 2);
    if (q->tx_timer) {
        timer_del(q->tx_timer);
        timer_free(q->tx_timer);
1571
        q->tx_timer = NULL;
1572 1573
    } else {
        qemu_bh_delete(q->tx_bh);
1574
        q->tx_bh = NULL;
1575
    }
1576
    q->tx_waiting = 0;
1577 1578 1579 1580 1581 1582 1583 1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612 1613 1614
    virtio_del_queue(vdev, index * 2 + 1);
}

static void virtio_net_change_num_queues(VirtIONet *n, int new_max_queues)
{
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
    int old_num_queues = virtio_get_num_queues(vdev);
    int new_num_queues = new_max_queues * 2 + 1;
    int i;

    assert(old_num_queues >= 3);
    assert(old_num_queues % 2 == 1);

    if (old_num_queues == new_num_queues) {
        return;
    }

    /*
     * We always need to remove and add ctrl vq if
     * old_num_queues != new_num_queues. Remove ctrl_vq first,
     * and then we only enter one of the following too loops.
     */
    virtio_del_queue(vdev, old_num_queues - 1);

    for (i = new_num_queues - 1; i < old_num_queues - 1; i += 2) {
        /* new_num_queues < old_num_queues */
        virtio_net_del_queue(n, i / 2);
    }

    for (i = old_num_queues - 1; i < new_num_queues - 1; i += 2) {
        /* new_num_queues > old_num_queues */
        virtio_net_add_queue(n, i / 2);
    }

    /* add ctrl_vq last */
    n->ctrl_vq = virtio_add_queue(vdev, 64, virtio_net_handle_ctrl);
}

J
Jason Wang 已提交
1615
static void virtio_net_set_multiqueue(VirtIONet *n, int multiqueue)
J
Jason Wang 已提交
1616
{
1617 1618
    int max = multiqueue ? n->max_queues : 1;

J
Jason Wang 已提交
1619
    n->multiqueue = multiqueue;
1620
    virtio_net_change_num_queues(n, max);
J
Jason Wang 已提交
1621 1622 1623 1624

    virtio_net_set_queues(n);
}

1625
static int virtio_net_post_load_device(void *opaque, int version_id)
1626
{
1627 1628
    VirtIONet *n = opaque;
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
1629
    int i, link_down;
A
aliguori 已提交
1630

1631
    virtio_net_set_mrg_rx_bufs(n, n->mergeable_rx_bufs,
1632 1633
                               virtio_vdev_has_feature(vdev,
                                                       VIRTIO_F_VERSION_1));
A
aliguori 已提交
1634

1635
    /* MAC_TABLE_ENTRIES may be different from the saved image */
1636
    if (n->mac_table.in_use > MAC_TABLE_ENTRIES) {
1637
        n->mac_table.in_use = 0;
1638
    }
1639

1640
    if (!virtio_vdev_has_feature(vdev, VIRTIO_NET_F_CTRL_GUEST_OFFLOADS)) {
1641 1642 1643 1644 1645 1646 1647
        n->curr_guest_offloads = virtio_net_supported_guest_offloads(n);
    }

    if (peer_has_vnet_hdr(n)) {
        virtio_net_apply_guest_offloads(n);
    }

1648 1649
    virtio_net_set_queues(n);

1650 1651 1652 1653 1654 1655 1656
    /* Find the first multicast entry in the saved MAC filter */
    for (i = 0; i < n->mac_table.in_use; i++) {
        if (n->mac_table.macs[i * ETH_ALEN] & 1) {
            break;
        }
    }
    n->mac_table.first_multi = i;
1657 1658 1659

    /* nc.link_down can't be migrated, so infer link_down according
     * to link status bit in n->status */
1660 1661 1662 1663
    link_down = (n->status & VIRTIO_NET_S_LINK_UP) == 0;
    for (i = 0; i < n->max_queues; i++) {
        qemu_get_subqueue(n->nic, i)->link_down = link_down;
    }
1664

1665 1666 1667 1668 1669 1670
    if (virtio_vdev_has_feature(vdev, VIRTIO_NET_F_GUEST_ANNOUNCE) &&
        virtio_vdev_has_feature(vdev, VIRTIO_NET_F_CTRL_VQ)) {
        n->announce_counter = SELF_ANNOUNCE_ROUNDS;
        timer_mod(n->announce_timer, qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL));
    }

A
aliguori 已提交
1671 1672 1673
    return 0;
}

1674 1675 1676 1677 1678 1679 1680 1681 1682 1683 1684 1685 1686 1687 1688 1689 1690 1691 1692 1693 1694 1695 1696 1697 1698 1699 1700 1701 1702 1703 1704 1705 1706 1707 1708 1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720
/* tx_waiting field of a VirtIONetQueue */
static const VMStateDescription vmstate_virtio_net_queue_tx_waiting = {
    .name = "virtio-net-queue-tx_waiting",
    .fields = (VMStateField[]) {
        VMSTATE_UINT32(tx_waiting, VirtIONetQueue),
        VMSTATE_END_OF_LIST()
   },
};

static bool max_queues_gt_1(void *opaque, int version_id)
{
    return VIRTIO_NET(opaque)->max_queues > 1;
}

static bool has_ctrl_guest_offloads(void *opaque, int version_id)
{
    return virtio_vdev_has_feature(VIRTIO_DEVICE(opaque),
                                   VIRTIO_NET_F_CTRL_GUEST_OFFLOADS);
}

static bool mac_table_fits(void *opaque, int version_id)
{
    return VIRTIO_NET(opaque)->mac_table.in_use <= MAC_TABLE_ENTRIES;
}

static bool mac_table_doesnt_fit(void *opaque, int version_id)
{
    return !mac_table_fits(opaque, version_id);
}

/* This temporary type is shared by all the WITH_TMP methods
 * although only some fields are used by each.
 */
struct VirtIONetMigTmp {
    VirtIONet      *parent;
    VirtIONetQueue *vqs_1;
    uint16_t        curr_queues_1;
    uint8_t         has_ufo;
    uint32_t        has_vnet_hdr;
};

/* The 2nd and subsequent tx_waiting flags are loaded later than
 * the 1st entry in the queues and only if there's more than one
 * entry.  We use the tmp mechanism to calculate a temporary
 * pointer and count and also validate the count.
 */

1721
static int virtio_net_tx_waiting_pre_save(void *opaque)
1722 1723 1724 1725 1726 1727 1728 1729
{
    struct VirtIONetMigTmp *tmp = opaque;

    tmp->vqs_1 = tmp->parent->vqs + 1;
    tmp->curr_queues_1 = tmp->parent->curr_queues - 1;
    if (tmp->parent->curr_queues == 0) {
        tmp->curr_queues_1 = 0;
    }
1730 1731

    return 0;
1732 1733 1734 1735 1736 1737 1738 1739 1740 1741 1742 1743 1744 1745 1746 1747 1748 1749 1750 1751 1752 1753 1754 1755 1756 1757 1758 1759 1760 1761 1762 1763 1764 1765 1766 1767 1768 1769 1770 1771 1772 1773 1774 1775 1776 1777 1778
}

static int virtio_net_tx_waiting_pre_load(void *opaque)
{
    struct VirtIONetMigTmp *tmp = opaque;

    /* Reuse the pointer setup from save */
    virtio_net_tx_waiting_pre_save(opaque);

    if (tmp->parent->curr_queues > tmp->parent->max_queues) {
        error_report("virtio-net: curr_queues %x > max_queues %x",
            tmp->parent->curr_queues, tmp->parent->max_queues);

        return -EINVAL;
    }

    return 0; /* all good */
}

static const VMStateDescription vmstate_virtio_net_tx_waiting = {
    .name      = "virtio-net-tx_waiting",
    .pre_load  = virtio_net_tx_waiting_pre_load,
    .pre_save  = virtio_net_tx_waiting_pre_save,
    .fields    = (VMStateField[]) {
        VMSTATE_STRUCT_VARRAY_POINTER_UINT16(vqs_1, struct VirtIONetMigTmp,
                                     curr_queues_1,
                                     vmstate_virtio_net_queue_tx_waiting,
                                     struct VirtIONetQueue),
        VMSTATE_END_OF_LIST()
    },
};

/* the 'has_ufo' flag is just tested; if the incoming stream has the
 * flag set we need to check that we have it
 */
static int virtio_net_ufo_post_load(void *opaque, int version_id)
{
    struct VirtIONetMigTmp *tmp = opaque;

    if (tmp->has_ufo && !peer_has_ufo(tmp->parent)) {
        error_report("virtio-net: saved image requires TUN_F_UFO support");
        return -EINVAL;
    }

    return 0;
}

1779
static int virtio_net_ufo_pre_save(void *opaque)
1780 1781 1782 1783
{
    struct VirtIONetMigTmp *tmp = opaque;

    tmp->has_ufo = tmp->parent->has_ufo;
1784 1785

    return 0;
1786 1787 1788 1789 1790 1791 1792 1793 1794 1795 1796 1797 1798 1799 1800 1801 1802 1803 1804 1805 1806 1807 1808 1809 1810 1811 1812
}

static const VMStateDescription vmstate_virtio_net_has_ufo = {
    .name      = "virtio-net-ufo",
    .post_load = virtio_net_ufo_post_load,
    .pre_save  = virtio_net_ufo_pre_save,
    .fields    = (VMStateField[]) {
        VMSTATE_UINT8(has_ufo, struct VirtIONetMigTmp),
        VMSTATE_END_OF_LIST()
    },
};

/* the 'has_vnet_hdr' flag is just tested; if the incoming stream has the
 * flag set we need to check that we have it
 */
static int virtio_net_vnet_post_load(void *opaque, int version_id)
{
    struct VirtIONetMigTmp *tmp = opaque;

    if (tmp->has_vnet_hdr && !peer_has_vnet_hdr(tmp->parent)) {
        error_report("virtio-net: saved image requires vnet_hdr=on");
        return -EINVAL;
    }

    return 0;
}

1813
static int virtio_net_vnet_pre_save(void *opaque)
1814 1815 1816 1817
{
    struct VirtIONetMigTmp *tmp = opaque;

    tmp->has_vnet_hdr = tmp->parent->has_vnet_hdr;
1818 1819

    return 0;
1820 1821 1822 1823 1824 1825 1826 1827 1828 1829 1830 1831 1832 1833 1834 1835 1836 1837 1838 1839 1840 1841 1842 1843 1844 1845 1846 1847 1848 1849 1850 1851 1852 1853 1854 1855 1856 1857 1858 1859 1860 1861 1862 1863 1864 1865 1866 1867 1868 1869 1870 1871 1872 1873 1874 1875 1876 1877 1878 1879 1880 1881 1882 1883
}

static const VMStateDescription vmstate_virtio_net_has_vnet = {
    .name      = "virtio-net-vnet",
    .post_load = virtio_net_vnet_post_load,
    .pre_save  = virtio_net_vnet_pre_save,
    .fields    = (VMStateField[]) {
        VMSTATE_UINT32(has_vnet_hdr, struct VirtIONetMigTmp),
        VMSTATE_END_OF_LIST()
    },
};

static const VMStateDescription vmstate_virtio_net_device = {
    .name = "virtio-net-device",
    .version_id = VIRTIO_NET_VM_VERSION,
    .minimum_version_id = VIRTIO_NET_VM_VERSION,
    .post_load = virtio_net_post_load_device,
    .fields = (VMStateField[]) {
        VMSTATE_UINT8_ARRAY(mac, VirtIONet, ETH_ALEN),
        VMSTATE_STRUCT_POINTER(vqs, VirtIONet,
                               vmstate_virtio_net_queue_tx_waiting,
                               VirtIONetQueue),
        VMSTATE_UINT32(mergeable_rx_bufs, VirtIONet),
        VMSTATE_UINT16(status, VirtIONet),
        VMSTATE_UINT8(promisc, VirtIONet),
        VMSTATE_UINT8(allmulti, VirtIONet),
        VMSTATE_UINT32(mac_table.in_use, VirtIONet),

        /* Guarded pair: If it fits we load it, else we throw it away
         * - can happen if source has a larger MAC table.; post-load
         *  sets flags in this case.
         */
        VMSTATE_VBUFFER_MULTIPLY(mac_table.macs, VirtIONet,
                                0, mac_table_fits, mac_table.in_use,
                                 ETH_ALEN),
        VMSTATE_UNUSED_VARRAY_UINT32(VirtIONet, mac_table_doesnt_fit, 0,
                                     mac_table.in_use, ETH_ALEN),

        /* Note: This is an array of uint32's that's always been saved as a
         * buffer; hold onto your endiannesses; it's actually used as a bitmap
         * but based on the uint.
         */
        VMSTATE_BUFFER_POINTER_UNSAFE(vlans, VirtIONet, 0, MAX_VLAN >> 3),
        VMSTATE_WITH_TMP(VirtIONet, struct VirtIONetMigTmp,
                         vmstate_virtio_net_has_vnet),
        VMSTATE_UINT8(mac_table.multi_overflow, VirtIONet),
        VMSTATE_UINT8(mac_table.uni_overflow, VirtIONet),
        VMSTATE_UINT8(alluni, VirtIONet),
        VMSTATE_UINT8(nomulti, VirtIONet),
        VMSTATE_UINT8(nouni, VirtIONet),
        VMSTATE_UINT8(nobcast, VirtIONet),
        VMSTATE_WITH_TMP(VirtIONet, struct VirtIONetMigTmp,
                         vmstate_virtio_net_has_ufo),
        VMSTATE_SINGLE_TEST(max_queues, VirtIONet, max_queues_gt_1, 0,
                            vmstate_info_uint16_equal, uint16_t),
        VMSTATE_UINT16_TEST(curr_queues, VirtIONet, max_queues_gt_1),
        VMSTATE_WITH_TMP(VirtIONet, struct VirtIONetMigTmp,
                         vmstate_virtio_net_tx_waiting),
        VMSTATE_UINT64_TEST(curr_guest_offloads, VirtIONet,
                            has_ctrl_guest_offloads),
        VMSTATE_END_OF_LIST()
   },
};

M
Mark McLoughlin 已提交
1884
static NetClientInfo net_virtio_info = {
1885
    .type = NET_CLIENT_DRIVER_NIC,
M
Mark McLoughlin 已提交
1886 1887 1888 1889
    .size = sizeof(NICState),
    .can_receive = virtio_net_can_receive,
    .receive = virtio_net_receive,
    .link_status_changed = virtio_net_set_link_status,
1890
    .query_rx_filter = virtio_net_query_rxfilter,
M
Mark McLoughlin 已提交
1891 1892
};

1893 1894
static bool virtio_net_guest_notifier_pending(VirtIODevice *vdev, int idx)
{
1895
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
1896
    NetClientState *nc = qemu_get_subqueue(n->nic, vq2q(idx));
1897
    assert(n->vhost_started);
1898
    return vhost_net_virtqueue_pending(get_vhost_net(nc->peer), idx);
1899 1900 1901 1902 1903
}

static void virtio_net_guest_notifier_mask(VirtIODevice *vdev, int idx,
                                           bool mask)
{
1904
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
1905
    NetClientState *nc = qemu_get_subqueue(n->nic, vq2q(idx));
1906
    assert(n->vhost_started);
1907
    vhost_net_virtqueue_mask(get_vhost_net(nc->peer),
1908 1909 1910
                             vdev, idx, mask);
}

G
Gerd Hoffmann 已提交
1911
static void virtio_net_set_config_size(VirtIONet *n, uint64_t host_features)
A
aliguori 已提交
1912
{
1913
    int i, config_size = 0;
1914
    virtio_add_feature(&host_features, VIRTIO_NET_F_MAC);
1915

1916 1917 1918 1919 1920
    for (i = 0; feature_sizes[i].flags != 0; i++) {
        if (host_features & feature_sizes[i].flags) {
            config_size = MAX(feature_sizes[i].end, config_size);
        }
    }
1921 1922 1923
    n->config_size = config_size;
}

1924 1925 1926 1927 1928 1929 1930 1931
void virtio_net_set_netclient_name(VirtIONet *n, const char *name,
                                   const char *type)
{
    /*
     * The name can be NULL, the netclient name will be type.x.
     */
    assert(type != NULL);

1932 1933
    g_free(n->netclient_name);
    g_free(n->netclient_type);
1934
    n->netclient_name = g_strdup(name);
1935 1936 1937
    n->netclient_type = g_strdup(type);
}

1938
static void virtio_net_device_realize(DeviceState *dev, Error **errp)
1939
{
1940
    VirtIODevice *vdev = VIRTIO_DEVICE(dev);
1941
    VirtIONet *n = VIRTIO_NET(dev);
1942
    NetClientState *nc;
1943
    int i;
1944

1945
    if (n->net_conf.mtu) {
1946
        n->host_features |= (1ULL << VIRTIO_NET_F_MTU);
1947 1948
    }

1949 1950 1951 1952 1953 1954 1955 1956 1957 1958 1959 1960 1961 1962 1963 1964 1965 1966 1967
    if (n->net_conf.duplex_str) {
        if (strncmp(n->net_conf.duplex_str, "half", 5) == 0) {
            n->net_conf.duplex = DUPLEX_HALF;
        } else if (strncmp(n->net_conf.duplex_str, "full", 5) == 0) {
            n->net_conf.duplex = DUPLEX_FULL;
        } else {
            error_setg(errp, "'duplex' must be 'half' or 'full'");
        }
        n->host_features |= (1ULL << VIRTIO_NET_F_SPEED_DUPLEX);
    } else {
        n->net_conf.duplex = DUPLEX_UNKNOWN;
    }

    if (n->net_conf.speed < SPEED_UNKNOWN) {
        error_setg(errp, "'speed' must be between 0 and INT_MAX");
    } else if (n->net_conf.speed >= 0) {
        n->host_features |= (1ULL << VIRTIO_NET_F_SPEED_DUPLEX);
    }

1968
    virtio_net_set_config_size(n, n->host_features);
1969
    virtio_init(vdev, "virtio-net", VIRTIO_ID_NET, n->config_size);
A
aliguori 已提交
1970

1971 1972 1973 1974 1975 1976 1977
    /*
     * We set a lower limit on RX queue size to what it always was.
     * Guests that want a smaller ring can always resize it without
     * help from us (using virtio 1 and up).
     */
    if (n->net_conf.rx_queue_size < VIRTIO_NET_RX_QUEUE_MIN_SIZE ||
        n->net_conf.rx_queue_size > VIRTQUEUE_MAX_SIZE ||
1978
        !is_power_of_2(n->net_conf.rx_queue_size)) {
1979 1980 1981 1982 1983 1984 1985 1986
        error_setg(errp, "Invalid rx_queue_size (= %" PRIu16 "), "
                   "must be a power of 2 between %d and %d.",
                   n->net_conf.rx_queue_size, VIRTIO_NET_RX_QUEUE_MIN_SIZE,
                   VIRTQUEUE_MAX_SIZE);
        virtio_cleanup(vdev);
        return;
    }

1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 1997
    if (n->net_conf.tx_queue_size < VIRTIO_NET_TX_QUEUE_MIN_SIZE ||
        n->net_conf.tx_queue_size > VIRTQUEUE_MAX_SIZE ||
        !is_power_of_2(n->net_conf.tx_queue_size)) {
        error_setg(errp, "Invalid tx_queue_size (= %" PRIu16 "), "
                   "must be a power of 2 between %d and %d",
                   n->net_conf.tx_queue_size, VIRTIO_NET_TX_QUEUE_MIN_SIZE,
                   VIRTQUEUE_MAX_SIZE);
        virtio_cleanup(vdev);
        return;
    }

1998
    n->max_queues = MAX(n->nic_conf.peers.queues, 1);
1999
    if (n->max_queues * 2 + 1 > VIRTIO_QUEUE_MAX) {
2000
        error_setg(errp, "Invalid number of queues (= %" PRIu32 "), "
S
Stefan Weil 已提交
2001
                   "must be a positive integer less than %d.",
2002
                   n->max_queues, (VIRTIO_QUEUE_MAX - 1) / 2);
2003 2004 2005
        virtio_cleanup(vdev);
        return;
    }
2006
    n->vqs = g_malloc0(sizeof(VirtIONetQueue) * n->max_queues);
J
Jason Wang 已提交
2007
    n->curr_queues = 1;
2008
    n->tx_timeout = n->net_conf.txtimer;
2009

2010 2011
    if (n->net_conf.tx && strcmp(n->net_conf.tx, "timer")
                       && strcmp(n->net_conf.tx, "bh")) {
2012 2013
        error_report("virtio-net: "
                     "Unknown option tx=%s, valid options: \"timer\" \"bh\"",
2014
                     n->net_conf.tx);
2015
        error_report("Defaulting to \"bh\"");
2016 2017
    }

2018 2019
    n->net_conf.tx_queue_size = MIN(virtio_net_max_tx_queue_size(n),
                                    n->net_conf.tx_queue_size);
2020

2021
    for (i = 0; i < n->max_queues; i++) {
2022
        virtio_net_add_queue(n, i);
2023
    }
2024

2025
    n->ctrl_vq = virtio_add_queue(vdev, 64, virtio_net_handle_ctrl);
2026 2027
    qemu_macaddr_default_if_unset(&n->nic_conf.macaddr);
    memcpy(&n->mac[0], &n->nic_conf.macaddr, sizeof(n->mac));
2028
    n->status = VIRTIO_NET_S_LINK_UP;
J
Jason Wang 已提交
2029 2030
    n->announce_timer = timer_new_ms(QEMU_CLOCK_VIRTUAL,
                                     virtio_net_announce_timer, n);
A
aliguori 已提交
2031

2032 2033 2034 2035 2036 2037 2038 2039
    if (n->netclient_type) {
        /*
         * Happen when virtio_net_set_netclient_name has been called.
         */
        n->nic = qemu_new_nic(&net_virtio_info, &n->nic_conf,
                              n->netclient_type, n->netclient_name, n);
    } else {
        n->nic = qemu_new_nic(&net_virtio_info, &n->nic_conf,
2040
                              object_get_typename(OBJECT(dev)), dev->id, n);
2041 2042
    }

2043 2044
    peer_test_vnet_hdr(n);
    if (peer_has_vnet_hdr(n)) {
J
Jason Wang 已提交
2045
        for (i = 0; i < n->max_queues; i++) {
2046
            qemu_using_vnet_hdr(qemu_get_subqueue(n->nic, i)->peer, true);
J
Jason Wang 已提交
2047
        }
2048 2049 2050 2051
        n->host_hdr_len = sizeof(struct virtio_net_hdr);
    } else {
        n->host_hdr_len = 0;
    }
M
Mark McLoughlin 已提交
2052

2053
    qemu_format_nic_info_str(qemu_get_queue(n->nic), n->nic_conf.macaddr.a);
2054

J
Jason Wang 已提交
2055
    n->vqs[0].tx_waiting = 0;
2056
    n->tx_burst = n->net_conf.txburst;
2057
    virtio_net_set_mrg_rx_bufs(n, 0, 0);
2058
    n->promisc = 1; /* for compatibility */
A
aliguori 已提交
2059

2060
    n->mac_table.macs = g_malloc0(MAC_TABLE_ENTRIES * ETH_ALEN);
2061

2062
    n->vlans = g_malloc0(MAX_VLAN >> 3);
2063

2064 2065 2066
    nc = qemu_get_queue(n->nic);
    nc->rxfilter_notify_enabled = 1;

2067
    n->qdev = dev;
2068 2069
}

2070
static void virtio_net_device_unrealize(DeviceState *dev, Error **errp)
2071
{
2072 2073
    VirtIODevice *vdev = VIRTIO_DEVICE(dev);
    VirtIONet *n = VIRTIO_NET(dev);
2074
    int i, max_queues;
2075 2076 2077 2078

    /* This will stop vhost backend if appropriate. */
    virtio_net_set_status(vdev, 0);

2079 2080 2081 2082
    g_free(n->netclient_name);
    n->netclient_name = NULL;
    g_free(n->netclient_type);
    n->netclient_type = NULL;
2083

2084 2085 2086
    g_free(n->mac_table.macs);
    g_free(n->vlans);

2087 2088 2089
    max_queues = n->multiqueue ? n->max_queues : 1;
    for (i = 0; i < max_queues; i++) {
        virtio_net_del_queue(n, i);
2090 2091
    }

J
Jason Wang 已提交
2092 2093
    timer_del(n->announce_timer);
    timer_free(n->announce_timer);
2094 2095
    g_free(n->vqs);
    qemu_del_nic(n->nic);
2096
    virtio_cleanup(vdev);
2097 2098 2099 2100 2101 2102 2103 2104 2105 2106 2107
}

static void virtio_net_instance_init(Object *obj)
{
    VirtIONet *n = VIRTIO_NET(obj);

    /*
     * The default config_size is sizeof(struct virtio_net_config).
     * Can be overriden with virtio_net_set_config_size.
     */
    n->config_size = sizeof(struct virtio_net_config);
2108 2109 2110
    device_add_bootindex_property(obj, &n->nic_conf.bootindex,
                                  "bootindex", "/ethernet-phy@0",
                                  DEVICE(n), NULL);
2111 2112
}

2113
static int virtio_net_pre_save(void *opaque)
2114 2115 2116 2117 2118 2119
{
    VirtIONet *n = opaque;

    /* At this point, backend must be stopped, otherwise
     * it might keep writing to memory. */
    assert(!n->vhost_started);
2120 2121

    return 0;
2122 2123 2124 2125 2126 2127 2128 2129 2130 2131 2132 2133
}

static const VMStateDescription vmstate_virtio_net = {
    .name = "virtio-net",
    .minimum_version_id = VIRTIO_NET_VM_VERSION,
    .version_id = VIRTIO_NET_VM_VERSION,
    .fields = (VMStateField[]) {
        VMSTATE_VIRTIO_DEVICE,
        VMSTATE_END_OF_LIST()
    },
    .pre_save = virtio_net_pre_save,
};
2134

2135
static Property virtio_net_properties[] = {
2136 2137 2138
    DEFINE_PROP_BIT64("csum", VirtIONet, host_features,
                    VIRTIO_NET_F_CSUM, true),
    DEFINE_PROP_BIT64("guest_csum", VirtIONet, host_features,
2139
                    VIRTIO_NET_F_GUEST_CSUM, true),
2140 2141
    DEFINE_PROP_BIT64("gso", VirtIONet, host_features, VIRTIO_NET_F_GSO, true),
    DEFINE_PROP_BIT64("guest_tso4", VirtIONet, host_features,
2142
                    VIRTIO_NET_F_GUEST_TSO4, true),
2143
    DEFINE_PROP_BIT64("guest_tso6", VirtIONet, host_features,
2144
                    VIRTIO_NET_F_GUEST_TSO6, true),
2145
    DEFINE_PROP_BIT64("guest_ecn", VirtIONet, host_features,
2146
                    VIRTIO_NET_F_GUEST_ECN, true),
2147
    DEFINE_PROP_BIT64("guest_ufo", VirtIONet, host_features,
2148
                    VIRTIO_NET_F_GUEST_UFO, true),
2149
    DEFINE_PROP_BIT64("guest_announce", VirtIONet, host_features,
2150
                    VIRTIO_NET_F_GUEST_ANNOUNCE, true),
2151
    DEFINE_PROP_BIT64("host_tso4", VirtIONet, host_features,
2152
                    VIRTIO_NET_F_HOST_TSO4, true),
2153
    DEFINE_PROP_BIT64("host_tso6", VirtIONet, host_features,
2154
                    VIRTIO_NET_F_HOST_TSO6, true),
2155
    DEFINE_PROP_BIT64("host_ecn", VirtIONet, host_features,
2156
                    VIRTIO_NET_F_HOST_ECN, true),
2157
    DEFINE_PROP_BIT64("host_ufo", VirtIONet, host_features,
2158
                    VIRTIO_NET_F_HOST_UFO, true),
2159
    DEFINE_PROP_BIT64("mrg_rxbuf", VirtIONet, host_features,
2160
                    VIRTIO_NET_F_MRG_RXBUF, true),
2161
    DEFINE_PROP_BIT64("status", VirtIONet, host_features,
2162
                    VIRTIO_NET_F_STATUS, true),
2163
    DEFINE_PROP_BIT64("ctrl_vq", VirtIONet, host_features,
2164
                    VIRTIO_NET_F_CTRL_VQ, true),
2165
    DEFINE_PROP_BIT64("ctrl_rx", VirtIONet, host_features,
2166
                    VIRTIO_NET_F_CTRL_RX, true),
2167
    DEFINE_PROP_BIT64("ctrl_vlan", VirtIONet, host_features,
2168
                    VIRTIO_NET_F_CTRL_VLAN, true),
2169
    DEFINE_PROP_BIT64("ctrl_rx_extra", VirtIONet, host_features,
2170
                    VIRTIO_NET_F_CTRL_RX_EXTRA, true),
2171
    DEFINE_PROP_BIT64("ctrl_mac_addr", VirtIONet, host_features,
2172
                    VIRTIO_NET_F_CTRL_MAC_ADDR, true),
2173
    DEFINE_PROP_BIT64("ctrl_guest_offloads", VirtIONet, host_features,
2174
                    VIRTIO_NET_F_CTRL_GUEST_OFFLOADS, true),
2175
    DEFINE_PROP_BIT64("mq", VirtIONet, host_features, VIRTIO_NET_F_MQ, false),
2176 2177
    DEFINE_NIC_PROPERTIES(VirtIONet, nic_conf),
    DEFINE_PROP_UINT32("x-txtimer", VirtIONet, net_conf.txtimer,
2178
                       TX_TIMER_INTERVAL),
2179 2180
    DEFINE_PROP_INT32("x-txburst", VirtIONet, net_conf.txburst, TX_BURST),
    DEFINE_PROP_STRING("tx", VirtIONet, net_conf.tx),
2181 2182
    DEFINE_PROP_UINT16("rx_queue_size", VirtIONet, net_conf.rx_queue_size,
                       VIRTIO_NET_RX_QUEUE_DEFAULT_SIZE),
2183 2184
    DEFINE_PROP_UINT16("tx_queue_size", VirtIONet, net_conf.tx_queue_size,
                       VIRTIO_NET_TX_QUEUE_DEFAULT_SIZE),
2185
    DEFINE_PROP_UINT16("host_mtu", VirtIONet, net_conf.mtu, 0),
2186 2187
    DEFINE_PROP_BOOL("x-mtu-bypass-backend", VirtIONet, mtu_bypass_backend,
                     true),
2188 2189
    DEFINE_PROP_INT32("speed", VirtIONet, net_conf.speed, SPEED_UNKNOWN),
    DEFINE_PROP_STRING("duplex", VirtIONet, net_conf.duplex_str),
2190 2191 2192 2193 2194 2195 2196
    DEFINE_PROP_END_OF_LIST(),
};

static void virtio_net_class_init(ObjectClass *klass, void *data)
{
    DeviceClass *dc = DEVICE_CLASS(klass);
    VirtioDeviceClass *vdc = VIRTIO_DEVICE_CLASS(klass);
2197

2198
    dc->props = virtio_net_properties;
2199
    dc->vmsd = &vmstate_virtio_net;
2200
    set_bit(DEVICE_CATEGORY_NETWORK, dc->categories);
2201
    vdc->realize = virtio_net_device_realize;
2202
    vdc->unrealize = virtio_net_device_unrealize;
2203 2204 2205 2206 2207 2208 2209 2210 2211
    vdc->get_config = virtio_net_get_config;
    vdc->set_config = virtio_net_set_config;
    vdc->get_features = virtio_net_get_features;
    vdc->set_features = virtio_net_set_features;
    vdc->bad_features = virtio_net_bad_features;
    vdc->reset = virtio_net_reset;
    vdc->set_status = virtio_net_set_status;
    vdc->guest_notifier_mask = virtio_net_guest_notifier_mask;
    vdc->guest_notifier_pending = virtio_net_guest_notifier_pending;
2212
    vdc->legacy_features |= (0x1 << VIRTIO_NET_F_GSO);
2213
    vdc->vmsd = &vmstate_virtio_net_device;
2214 2215 2216 2217 2218 2219 2220 2221 2222 2223 2224 2225 2226 2227 2228 2229
}

static const TypeInfo virtio_net_info = {
    .name = TYPE_VIRTIO_NET,
    .parent = TYPE_VIRTIO_DEVICE,
    .instance_size = sizeof(VirtIONet),
    .instance_init = virtio_net_instance_init,
    .class_init = virtio_net_class_init,
};

static void virtio_register_types(void)
{
    type_register_static(&virtio_net_info);
}

type_init(virtio_register_types)