virtio-net.c 68.3 KB
Newer Older
A
aliguori 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13
/*
 * Virtio Network Device
 *
 * Copyright IBM, Corp. 2007
 *
 * Authors:
 *  Anthony Liguori   <aliguori@us.ibm.com>
 *
 * This work is licensed under the terms of the GNU GPL, version 2.  See
 * the COPYING file in the top-level directory.
 *
 */

P
Peter Maydell 已提交
14
#include "qemu/osdep.h"
15
#include "qemu/iov.h"
P
Paolo Bonzini 已提交
16
#include "hw/virtio/virtio.h"
P
Paolo Bonzini 已提交
17
#include "net/net.h"
18
#include "net/checksum.h"
19
#include "net/tap.h"
20 21
#include "qemu/error-report.h"
#include "qemu/timer.h"
P
Paolo Bonzini 已提交
22 23
#include "hw/virtio/virtio-net.h"
#include "net/vhost_net.h"
24
#include "hw/virtio/virtio-bus.h"
25
#include "qapi/error.h"
26
#include "qapi/qapi-events-net.h"
27
#include "hw/virtio/virtio-access.h"
28
#include "migration/misc.h"
29
#include "standard-headers/linux/ethtool.h"
A
aliguori 已提交
30

31
#define VIRTIO_NET_VM_VERSION    11
32

33
#define MAC_TABLE_ENTRIES    64
34
#define MAX_VLAN    (1 << 12)   /* Per 802.1Q definition */
35

36 37
/* previously fixed value */
#define VIRTIO_NET_RX_QUEUE_DEFAULT_SIZE 256
38 39
#define VIRTIO_NET_TX_QUEUE_DEFAULT_SIZE 256

40 41
/* for now, only allow larger queues; with virtio-1, guest can downsize */
#define VIRTIO_NET_RX_QUEUE_MIN_SIZE VIRTIO_NET_RX_QUEUE_DEFAULT_SIZE
42
#define VIRTIO_NET_TX_QUEUE_MIN_SIZE VIRTIO_NET_TX_QUEUE_DEFAULT_SIZE
43

44 45 46 47 48
/*
 * Calculate the number of bytes up to and including the given 'field' of
 * 'container'.
 */
#define endof(container, field) \
49
    (offsetof(container, field) + sizeof_field(container, field))
50 51

typedef struct VirtIOFeature {
52
    uint64_t flags;
53 54 55 56
    size_t end;
} VirtIOFeature;

static VirtIOFeature feature_sizes[] = {
57
    {.flags = 1ULL << VIRTIO_NET_F_MAC,
58
     .end = endof(struct virtio_net_config, mac)},
59
    {.flags = 1ULL << VIRTIO_NET_F_STATUS,
60
     .end = endof(struct virtio_net_config, status)},
61
    {.flags = 1ULL << VIRTIO_NET_F_MQ,
62
     .end = endof(struct virtio_net_config, max_virtqueue_pairs)},
63
    {.flags = 1ULL << VIRTIO_NET_F_MTU,
64
     .end = endof(struct virtio_net_config, mtu)},
65 66
    {.flags = 1ULL << VIRTIO_NET_F_SPEED_DUPLEX,
     .end = endof(struct virtio_net_config, duplex)},
67 68 69
    {}
};

J
Jason Wang 已提交
70
static VirtIONetQueue *virtio_net_get_subqueue(NetClientState *nc)
71 72 73
{
    VirtIONet *n = qemu_get_nic_opaque(nc);

J
Jason Wang 已提交
74
    return &n->vqs[nc->queue_index];
75
}
J
Jason Wang 已提交
76 77 78 79 80 81

static int vq2q(int queue_index)
{
    return queue_index / 2;
}

A
aliguori 已提交
82 83 84 85
/* TODO
 * - we could suppress RX interrupt if we were so inclined.
 */

86
static void virtio_net_get_config(VirtIODevice *vdev, uint8_t *config)
A
aliguori 已提交
87
{
88
    VirtIONet *n = VIRTIO_NET(vdev);
A
aliguori 已提交
89 90
    struct virtio_net_config netcfg;

91 92
    virtio_stw_p(vdev, &netcfg.status, n->status);
    virtio_stw_p(vdev, &netcfg.max_virtqueue_pairs, n->max_queues);
93
    virtio_stw_p(vdev, &netcfg.mtu, n->net_conf.mtu);
94
    memcpy(netcfg.mac, n->mac, ETH_ALEN);
95 96
    virtio_stl_p(vdev, &netcfg.speed, n->net_conf.speed);
    netcfg.duplex = n->net_conf.duplex;
97
    memcpy(config, &netcfg, n->config_size);
A
aliguori 已提交
98 99
}

100 101
static void virtio_net_set_config(VirtIODevice *vdev, const uint8_t *config)
{
102
    VirtIONet *n = VIRTIO_NET(vdev);
103
    struct virtio_net_config netcfg = {};
104

105
    memcpy(&netcfg, config, n->config_size);
106

107 108
    if (!virtio_vdev_has_feature(vdev, VIRTIO_NET_F_CTRL_MAC_ADDR) &&
        !virtio_vdev_has_feature(vdev, VIRTIO_F_VERSION_1) &&
109
        memcmp(netcfg.mac, n->mac, ETH_ALEN)) {
110
        memcpy(n->mac, netcfg.mac, ETH_ALEN);
J
Jason Wang 已提交
111
        qemu_format_nic_info_str(qemu_get_queue(n->nic), n->mac);
112 113 114
    }
}

115 116
static bool virtio_net_started(VirtIONet *n, uint8_t status)
{
117
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
118
    return (status & VIRTIO_CONFIG_S_DRIVER_OK) &&
119
        (n->status & VIRTIO_NET_S_LINK_UP) && vdev->vm_running;
120 121
}

J
Jason Wang 已提交
122 123 124 125 126 127 128 129 130 131
static void virtio_net_announce_timer(void *opaque)
{
    VirtIONet *n = opaque;
    VirtIODevice *vdev = VIRTIO_DEVICE(n);

    n->announce_counter--;
    n->status |= VIRTIO_NET_S_ANNOUNCE;
    virtio_notify_config(vdev);
}

132
static void virtio_net_vhost_status(VirtIONet *n, uint8_t status)
133
{
134
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
J
Jason Wang 已提交
135
    NetClientState *nc = qemu_get_queue(n->nic);
J
Jason Wang 已提交
136
    int queues = n->multiqueue ? n->max_queues : 1;
J
Jason Wang 已提交
137

138
    if (!get_vhost_net(nc->peer)) {
139 140
        return;
    }
J
Jason Wang 已提交
141

142 143
    if ((virtio_net_started(n, status) && !nc->peer->link_down) ==
        !!n->vhost_started) {
144 145 146
        return;
    }
    if (!n->vhost_started) {
147 148
        int r, i;

149 150 151 152 153 154 155
        if (n->needs_vnet_hdr_swap) {
            error_report("backend does not support %s vnet headers; "
                         "falling back on userspace virtio",
                         virtio_is_big_endian(vdev) ? "BE" : "LE");
            return;
        }

156 157 158 159 160 161 162 163 164 165 166
        /* Any packets outstanding? Purge them to avoid touching rings
         * when vhost is running.
         */
        for (i = 0;  i < queues; i++) {
            NetClientState *qnc = qemu_get_subqueue(n->nic, i);

            /* Purge both directions: TX and RX. */
            qemu_net_queue_purge(qnc->peer->incoming_queue, qnc);
            qemu_net_queue_purge(qnc->incoming_queue, qnc->peer);
        }

167 168 169 170 171 172 173 174 175 176
        if (virtio_has_feature(vdev->guest_features, VIRTIO_NET_F_MTU)) {
            r = vhost_net_set_mtu(get_vhost_net(nc->peer), n->net_conf.mtu);
            if (r < 0) {
                error_report("%uBytes MTU not supported by the backend",
                             n->net_conf.mtu);

                return;
            }
        }

177
        n->vhost_started = 1;
178
        r = vhost_net_start(vdev, n->nic->ncs, queues);
179
        if (r < 0) {
180 181
            error_report("unable to start vhost net: %d: "
                         "falling back on userspace virtio", -r);
182
            n->vhost_started = 0;
183 184
        }
    } else {
185
        vhost_net_stop(vdev, n->nic->ncs, queues);
186 187 188 189
        n->vhost_started = 0;
    }
}

190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242
static int virtio_net_set_vnet_endian_one(VirtIODevice *vdev,
                                          NetClientState *peer,
                                          bool enable)
{
    if (virtio_is_big_endian(vdev)) {
        return qemu_set_vnet_be(peer, enable);
    } else {
        return qemu_set_vnet_le(peer, enable);
    }
}

static bool virtio_net_set_vnet_endian(VirtIODevice *vdev, NetClientState *ncs,
                                       int queues, bool enable)
{
    int i;

    for (i = 0; i < queues; i++) {
        if (virtio_net_set_vnet_endian_one(vdev, ncs[i].peer, enable) < 0 &&
            enable) {
            while (--i >= 0) {
                virtio_net_set_vnet_endian_one(vdev, ncs[i].peer, false);
            }

            return true;
        }
    }

    return false;
}

static void virtio_net_vnet_endian_status(VirtIONet *n, uint8_t status)
{
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
    int queues = n->multiqueue ? n->max_queues : 1;

    if (virtio_net_started(n, status)) {
        /* Before using the device, we tell the network backend about the
         * endianness to use when parsing vnet headers. If the backend
         * can't do it, we fallback onto fixing the headers in the core
         * virtio-net code.
         */
        n->needs_vnet_hdr_swap = virtio_net_set_vnet_endian(vdev, n->nic->ncs,
                                                            queues, true);
    } else if (virtio_net_started(n, vdev->status)) {
        /* After using the device, we need to reset the network backend to
         * the default (guest native endianness), otherwise the guest may
         * lose network connectivity if it is rebooted into a different
         * endianness.
         */
        virtio_net_set_vnet_endian(vdev, n->nic->ncs, queues, false);
    }
}

243 244 245 246 247 248 249 250
static void virtio_net_drop_tx_queue_data(VirtIODevice *vdev, VirtQueue *vq)
{
    unsigned int dropped = virtqueue_drop_all(vq);
    if (dropped) {
        virtio_notify(vdev, vq);
    }
}

251 252
static void virtio_net_set_status(struct VirtIODevice *vdev, uint8_t status)
{
253
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
254 255 256
    VirtIONetQueue *q;
    int i;
    uint8_t queue_status;
257

258
    virtio_net_vnet_endian_status(n, status);
259 260
    virtio_net_vhost_status(n, status);

J
Jason Wang 已提交
261
    for (i = 0; i < n->max_queues; i++) {
262 263
        NetClientState *ncs = qemu_get_subqueue(n->nic, i);
        bool queue_started;
J
Jason Wang 已提交
264
        q = &n->vqs[i];
265

J
Jason Wang 已提交
266 267
        if ((!n->multiqueue && i != 0) || i >= n->curr_queues) {
            queue_status = 0;
268
        } else {
J
Jason Wang 已提交
269
            queue_status = status;
270
        }
271 272 273 274 275 276
        queue_started =
            virtio_net_started(n, queue_status) && !n->vhost_started;

        if (queue_started) {
            qemu_flush_queued_packets(ncs);
        }
J
Jason Wang 已提交
277 278 279 280 281

        if (!q->tx_waiting) {
            continue;
        }

282
        if (queue_started) {
J
Jason Wang 已提交
283
            if (q->tx_timer) {
284 285
                timer_mod(q->tx_timer,
                               qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + n->tx_timeout);
J
Jason Wang 已提交
286 287 288
            } else {
                qemu_bh_schedule(q->tx_bh);
            }
289
        } else {
J
Jason Wang 已提交
290
            if (q->tx_timer) {
291
                timer_del(q->tx_timer);
J
Jason Wang 已提交
292 293 294
            } else {
                qemu_bh_cancel(q->tx_bh);
            }
295
            if ((n->status & VIRTIO_NET_S_LINK_UP) == 0 &&
296 297
                (queue_status & VIRTIO_CONFIG_S_DRIVER_OK) &&
                vdev->vm_running) {
298 299 300 301 302 303
                /* if tx is waiting we are likely have some packets in tx queue
                 * and disabled notification */
                q->tx_waiting = 0;
                virtio_queue_set_notification(q->tx_vq, 1);
                virtio_net_drop_tx_queue_data(vdev, q->tx_vq);
            }
304 305 306 307
        }
    }
}

308
static void virtio_net_set_link_status(NetClientState *nc)
309
{
J
Jason Wang 已提交
310
    VirtIONet *n = qemu_get_nic_opaque(nc);
311
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
312 313
    uint16_t old_status = n->status;

M
Mark McLoughlin 已提交
314
    if (nc->link_down)
315 316 317 318 319
        n->status &= ~VIRTIO_NET_S_LINK_UP;
    else
        n->status |= VIRTIO_NET_S_LINK_UP;

    if (n->status != old_status)
320
        virtio_notify_config(vdev);
321

322
    virtio_net_set_status(vdev, vdev->status);
323 324
}

325 326 327 328 329
static void rxfilter_notify(NetClientState *nc)
{
    VirtIONet *n = qemu_get_nic_opaque(nc);

    if (nc->rxfilter_notify_enabled) {
330
        gchar *path = object_get_canonical_path(OBJECT(n->qdev));
331 332
        qapi_event_send_nic_rx_filter_changed(!!n->netclient_name,
                                              n->netclient_name, path, &error_abort);
333
        g_free(path);
334 335 336 337 338 339

        /* disable event notification to avoid events flooding */
        nc->rxfilter_notify_enabled = 0;
    }
}

340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359
static intList *get_vlan_table(VirtIONet *n)
{
    intList *list, *entry;
    int i, j;

    list = NULL;
    for (i = 0; i < MAX_VLAN >> 5; i++) {
        for (j = 0; n->vlans[i] && j <= 0x1f; j++) {
            if (n->vlans[i] & (1U << j)) {
                entry = g_malloc0(sizeof(*entry));
                entry->value = (i << 5) + j;
                entry->next = list;
                list = entry;
            }
        }
    }

    return list;
}

360 361 362
static RxFilterInfo *virtio_net_query_rxfilter(NetClientState *nc)
{
    VirtIONet *n = qemu_get_nic_opaque(nc);
363
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
364 365
    RxFilterInfo *info;
    strList *str_list, *entry;
366
    int i;
367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391

    info = g_malloc0(sizeof(*info));
    info->name = g_strdup(nc->name);
    info->promiscuous = n->promisc;

    if (n->nouni) {
        info->unicast = RX_STATE_NONE;
    } else if (n->alluni) {
        info->unicast = RX_STATE_ALL;
    } else {
        info->unicast = RX_STATE_NORMAL;
    }

    if (n->nomulti) {
        info->multicast = RX_STATE_NONE;
    } else if (n->allmulti) {
        info->multicast = RX_STATE_ALL;
    } else {
        info->multicast = RX_STATE_NORMAL;
    }

    info->broadcast_allowed = n->nobcast;
    info->multicast_overflow = n->mac_table.multi_overflow;
    info->unicast_overflow = n->mac_table.uni_overflow;

392
    info->main_mac = qemu_mac_strdup_printf(n->mac);
393 394 395 396

    str_list = NULL;
    for (i = 0; i < n->mac_table.first_multi; i++) {
        entry = g_malloc0(sizeof(*entry));
397
        entry->value = qemu_mac_strdup_printf(n->mac_table.macs + i * ETH_ALEN);
398 399 400 401 402 403 404 405
        entry->next = str_list;
        str_list = entry;
    }
    info->unicast_table = str_list;

    str_list = NULL;
    for (i = n->mac_table.first_multi; i < n->mac_table.in_use; i++) {
        entry = g_malloc0(sizeof(*entry));
406
        entry->value = qemu_mac_strdup_printf(n->mac_table.macs + i * ETH_ALEN);
407 408 409 410
        entry->next = str_list;
        str_list = entry;
    }
    info->multicast_table = str_list;
411
    info->vlan_table = get_vlan_table(n);
412

413
    if (!virtio_vdev_has_feature(vdev, VIRTIO_NET_F_CTRL_VLAN)) {
414 415 416 417 418
        info->vlan = RX_STATE_ALL;
    } else if (!info->vlan_table) {
        info->vlan = RX_STATE_NONE;
    } else {
        info->vlan = RX_STATE_NORMAL;
419 420 421 422 423 424 425 426
    }

    /* enable event notification after query */
    nc->rxfilter_notify_enabled = 1;

    return info;
}

427 428
static void virtio_net_reset(VirtIODevice *vdev)
{
429
    VirtIONet *n = VIRTIO_NET(vdev);
430
    int i;
431 432 433 434

    /* Reset back to compatibility mode */
    n->promisc = 1;
    n->allmulti = 0;
435 436 437 438
    n->alluni = 0;
    n->nomulti = 0;
    n->nouni = 0;
    n->nobcast = 0;
J
Jason Wang 已提交
439 440
    /* multiqueue is disabled by default */
    n->curr_queues = 1;
J
Jason Wang 已提交
441 442 443
    timer_del(n->announce_timer);
    n->announce_counter = 0;
    n->status &= ~VIRTIO_NET_S_ANNOUNCE;
444

445
    /* Flush any MAC and VLAN filter table state */
446
    n->mac_table.in_use = 0;
447
    n->mac_table.first_multi = 0;
448 449
    n->mac_table.multi_overflow = 0;
    n->mac_table.uni_overflow = 0;
450
    memset(n->mac_table.macs, 0, MAC_TABLE_ENTRIES * ETH_ALEN);
451
    memcpy(&n->mac[0], &n->nic->conf->macaddr, sizeof(n->mac));
452
    qemu_format_nic_info_str(qemu_get_queue(n->nic), n->mac);
453
    memset(n->vlans, 0, MAX_VLAN >> 3);
454 455 456 457 458 459 460 461 462 463

    /* Flush any async TX */
    for (i = 0;  i < n->max_queues; i++) {
        NetClientState *nc = qemu_get_subqueue(n->nic, i);

        if (nc->peer) {
            qemu_flush_or_purge_queued_packets(nc->peer, true);
            assert(!virtio_net_get_subqueue(nc)->async_tx.elem);
        }
    }
464 465
}

466
static void peer_test_vnet_hdr(VirtIONet *n)
M
Mark McLoughlin 已提交
467
{
J
Jason Wang 已提交
468 469
    NetClientState *nc = qemu_get_queue(n->nic);
    if (!nc->peer) {
470
        return;
J
Jason Wang 已提交
471
    }
M
Mark McLoughlin 已提交
472

473
    n->has_vnet_hdr = qemu_has_vnet_hdr(nc->peer);
474
}
M
Mark McLoughlin 已提交
475

476 477
static int peer_has_vnet_hdr(VirtIONet *n)
{
M
Mark McLoughlin 已提交
478 479 480
    return n->has_vnet_hdr;
}

481 482 483 484 485
static int peer_has_ufo(VirtIONet *n)
{
    if (!peer_has_vnet_hdr(n))
        return 0;

486
    n->has_ufo = qemu_has_ufo(qemu_get_queue(n->nic)->peer);
487 488 489 490

    return n->has_ufo;
}

491 492
static void virtio_net_set_mrg_rx_bufs(VirtIONet *n, int mergeable_rx_bufs,
                                       int version_1)
493
{
J
Jason Wang 已提交
494 495 496
    int i;
    NetClientState *nc;

497 498
    n->mergeable_rx_bufs = mergeable_rx_bufs;

499 500 501 502 503 504 505
    if (version_1) {
        n->guest_hdr_len = sizeof(struct virtio_net_hdr_mrg_rxbuf);
    } else {
        n->guest_hdr_len = n->mergeable_rx_bufs ?
            sizeof(struct virtio_net_hdr_mrg_rxbuf) :
            sizeof(struct virtio_net_hdr);
    }
506

J
Jason Wang 已提交
507 508 509 510
    for (i = 0; i < n->max_queues; i++) {
        nc = qemu_get_subqueue(n->nic, i);

        if (peer_has_vnet_hdr(n) &&
511 512
            qemu_has_vnet_hdr_len(nc->peer, n->guest_hdr_len)) {
            qemu_set_vnet_hdr_len(nc->peer, n->guest_hdr_len);
J
Jason Wang 已提交
513 514
            n->host_hdr_len = n->guest_hdr_len;
        }
515 516 517
    }
}

518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535
static int virtio_net_max_tx_queue_size(VirtIONet *n)
{
    NetClientState *peer = n->nic_conf.peers.ncs[0];

    /*
     * Backends other than vhost-user don't support max queue size.
     */
    if (!peer) {
        return VIRTIO_NET_TX_QUEUE_DEFAULT_SIZE;
    }

    if (peer->info->type != NET_CLIENT_DRIVER_VHOST_USER) {
        return VIRTIO_NET_TX_QUEUE_DEFAULT_SIZE;
    }

    return VIRTQUEUE_MAX_SIZE;
}

J
Jason Wang 已提交
536 537 538 539 540 541 542 543
static int peer_attach(VirtIONet *n, int index)
{
    NetClientState *nc = qemu_get_subqueue(n->nic, index);

    if (!nc->peer) {
        return 0;
    }

544
    if (nc->peer->info->type == NET_CLIENT_DRIVER_VHOST_USER) {
545 546 547
        vhost_set_vring_enable(nc->peer, 1);
    }

548
    if (nc->peer->info->type != NET_CLIENT_DRIVER_TAP) {
J
Jason Wang 已提交
549 550 551
        return 0;
    }

552 553 554 555
    if (n->max_queues == 1) {
        return 0;
    }

J
Jason Wang 已提交
556 557 558 559 560 561 562 563 564 565 566
    return tap_enable(nc->peer);
}

static int peer_detach(VirtIONet *n, int index)
{
    NetClientState *nc = qemu_get_subqueue(n->nic, index);

    if (!nc->peer) {
        return 0;
    }

567
    if (nc->peer->info->type == NET_CLIENT_DRIVER_VHOST_USER) {
568 569 570
        vhost_set_vring_enable(nc->peer, 0);
    }

571
    if (nc->peer->info->type !=  NET_CLIENT_DRIVER_TAP) {
J
Jason Wang 已提交
572 573 574 575 576 577 578 579 580
        return 0;
    }

    return tap_disable(nc->peer);
}

static void virtio_net_set_queues(VirtIONet *n)
{
    int i;
581
    int r;
J
Jason Wang 已提交
582

583 584 585 586
    if (n->nic->peer_deleted) {
        return;
    }

J
Jason Wang 已提交
587 588
    for (i = 0; i < n->max_queues; i++) {
        if (i < n->curr_queues) {
589 590
            r = peer_attach(n, i);
            assert(!r);
J
Jason Wang 已提交
591
        } else {
592 593
            r = peer_detach(n, i);
            assert(!r);
J
Jason Wang 已提交
594 595 596 597
        }
    }
}

J
Jason Wang 已提交
598
static void virtio_net_set_multiqueue(VirtIONet *n, int multiqueue);
J
Jason Wang 已提交
599

J
Jason Wang 已提交
600 601
static uint64_t virtio_net_get_features(VirtIODevice *vdev, uint64_t features,
                                        Error **errp)
A
aliguori 已提交
602
{
603
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
604
    NetClientState *nc = qemu_get_queue(n->nic);
A
aliguori 已提交
605

606 607 608
    /* Firstly sync all virtio-net possible supported features */
    features |= n->host_features;

609
    virtio_add_feature(&features, VIRTIO_NET_F_MAC);
610

611
    if (!peer_has_vnet_hdr(n)) {
612 613 614 615
        virtio_clear_feature(&features, VIRTIO_NET_F_CSUM);
        virtio_clear_feature(&features, VIRTIO_NET_F_HOST_TSO4);
        virtio_clear_feature(&features, VIRTIO_NET_F_HOST_TSO6);
        virtio_clear_feature(&features, VIRTIO_NET_F_HOST_ECN);
616

617 618 619 620
        virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_CSUM);
        virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_TSO4);
        virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_TSO6);
        virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_ECN);
621
    }
M
Mark McLoughlin 已提交
622

623
    if (!peer_has_vnet_hdr(n) || !peer_has_ufo(n)) {
624 625
        virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_UFO);
        virtio_clear_feature(&features, VIRTIO_NET_F_HOST_UFO);
M
Mark McLoughlin 已提交
626 627
    }

628
    if (!get_vhost_net(nc->peer)) {
629 630
        return features;
    }
631 632 633 634 635 636 637 638 639
    features = vhost_net_get_features(get_vhost_net(nc->peer), features);
    vdev->backend_features = features;

    if (n->mtu_bypass_backend &&
            (n->host_features & 1ULL << VIRTIO_NET_F_MTU)) {
        features |= (1ULL << VIRTIO_NET_F_MTU);
    }

    return features;
A
aliguori 已提交
640 641
}

G
Gerd Hoffmann 已提交
642
static uint64_t virtio_net_bad_features(VirtIODevice *vdev)
643
{
G
Gerd Hoffmann 已提交
644
    uint64_t features = 0;
645 646 647

    /* Linux kernel 2.6.25.  It understood MAC (as everyone must),
     * but also these: */
648 649 650 651 652
    virtio_add_feature(&features, VIRTIO_NET_F_MAC);
    virtio_add_feature(&features, VIRTIO_NET_F_CSUM);
    virtio_add_feature(&features, VIRTIO_NET_F_HOST_TSO4);
    virtio_add_feature(&features, VIRTIO_NET_F_HOST_TSO6);
    virtio_add_feature(&features, VIRTIO_NET_F_HOST_ECN);
653

654
    return features;
655 656
}

657 658
static void virtio_net_apply_guest_offloads(VirtIONet *n)
{
659
    qemu_set_offload(qemu_get_queue(n->nic)->peer,
660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684
            !!(n->curr_guest_offloads & (1ULL << VIRTIO_NET_F_GUEST_CSUM)),
            !!(n->curr_guest_offloads & (1ULL << VIRTIO_NET_F_GUEST_TSO4)),
            !!(n->curr_guest_offloads & (1ULL << VIRTIO_NET_F_GUEST_TSO6)),
            !!(n->curr_guest_offloads & (1ULL << VIRTIO_NET_F_GUEST_ECN)),
            !!(n->curr_guest_offloads & (1ULL << VIRTIO_NET_F_GUEST_UFO)));
}

static uint64_t virtio_net_guest_offloads_by_features(uint32_t features)
{
    static const uint64_t guest_offloads_mask =
        (1ULL << VIRTIO_NET_F_GUEST_CSUM) |
        (1ULL << VIRTIO_NET_F_GUEST_TSO4) |
        (1ULL << VIRTIO_NET_F_GUEST_TSO6) |
        (1ULL << VIRTIO_NET_F_GUEST_ECN)  |
        (1ULL << VIRTIO_NET_F_GUEST_UFO);

    return guest_offloads_mask & features;
}

static inline uint64_t virtio_net_supported_guest_offloads(VirtIONet *n)
{
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
    return virtio_net_guest_offloads_by_features(vdev->guest_features);
}

G
Gerd Hoffmann 已提交
685
static void virtio_net_set_features(VirtIODevice *vdev, uint64_t features)
A
aliguori 已提交
686
{
687
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
688 689
    int i;

690 691 692 693 694
    if (n->mtu_bypass_backend &&
            !virtio_has_feature(vdev->backend_features, VIRTIO_NET_F_MTU)) {
        features &= ~(1ULL << VIRTIO_NET_F_MTU);
    }

695
    virtio_net_set_multiqueue(n,
696
                              virtio_has_feature(features, VIRTIO_NET_F_MQ));
A
aliguori 已提交
697

698
    virtio_net_set_mrg_rx_bufs(n,
699 700 701 702
                               virtio_has_feature(features,
                                                  VIRTIO_NET_F_MRG_RXBUF),
                               virtio_has_feature(features,
                                                  VIRTIO_F_VERSION_1));
703 704

    if (n->has_vnet_hdr) {
705 706 707
        n->curr_guest_offloads =
            virtio_net_guest_offloads_by_features(features);
        virtio_net_apply_guest_offloads(n);
708
    }
J
Jason Wang 已提交
709 710 711 712

    for (i = 0;  i < n->max_queues; i++) {
        NetClientState *nc = qemu_get_subqueue(n->nic, i);

713
        if (!get_vhost_net(nc->peer)) {
J
Jason Wang 已提交
714 715
            continue;
        }
716
        vhost_net_ack_features(get_vhost_net(nc->peer), features);
D
David L Stevens 已提交
717
    }
718

719
    if (virtio_has_feature(features, VIRTIO_NET_F_CTRL_VLAN)) {
720 721 722 723
        memset(n->vlans, 0, MAX_VLAN >> 3);
    } else {
        memset(n->vlans, 0xff, MAX_VLAN >> 3);
    }
A
aliguori 已提交
724 725
}

726
static int virtio_net_handle_rx_mode(VirtIONet *n, uint8_t cmd,
727
                                     struct iovec *iov, unsigned int iov_cnt)
728 729
{
    uint8_t on;
730
    size_t s;
731
    NetClientState *nc = qemu_get_queue(n->nic);
732

733 734 735
    s = iov_to_buf(iov, iov_cnt, 0, &on, sizeof(on));
    if (s != sizeof(on)) {
        return VIRTIO_NET_ERR;
736 737
    }

A
Amos Kong 已提交
738
    if (cmd == VIRTIO_NET_CTRL_RX_PROMISC) {
739
        n->promisc = on;
A
Amos Kong 已提交
740
    } else if (cmd == VIRTIO_NET_CTRL_RX_ALLMULTI) {
741
        n->allmulti = on;
A
Amos Kong 已提交
742
    } else if (cmd == VIRTIO_NET_CTRL_RX_ALLUNI) {
743
        n->alluni = on;
A
Amos Kong 已提交
744
    } else if (cmd == VIRTIO_NET_CTRL_RX_NOMULTI) {
745
        n->nomulti = on;
A
Amos Kong 已提交
746
    } else if (cmd == VIRTIO_NET_CTRL_RX_NOUNI) {
747
        n->nouni = on;
A
Amos Kong 已提交
748
    } else if (cmd == VIRTIO_NET_CTRL_RX_NOBCAST) {
749
        n->nobcast = on;
750
    } else {
751
        return VIRTIO_NET_ERR;
752
    }
753

754 755
    rxfilter_notify(nc);

756 757 758
    return VIRTIO_NET_OK;
}

759 760 761 762 763 764 765
static int virtio_net_handle_offloads(VirtIONet *n, uint8_t cmd,
                                     struct iovec *iov, unsigned int iov_cnt)
{
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
    uint64_t offloads;
    size_t s;

766
    if (!virtio_vdev_has_feature(vdev, VIRTIO_NET_F_CTRL_GUEST_OFFLOADS)) {
767 768 769 770 771 772 773 774 775 776 777
        return VIRTIO_NET_ERR;
    }

    s = iov_to_buf(iov, iov_cnt, 0, &offloads, sizeof(offloads));
    if (s != sizeof(offloads)) {
        return VIRTIO_NET_ERR;
    }

    if (cmd == VIRTIO_NET_CTRL_GUEST_OFFLOADS_SET) {
        uint64_t supported_offloads;

J
Jason Wang 已提交
778 779
        offloads = virtio_ldq_p(vdev, &offloads);

780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797
        if (!n->has_vnet_hdr) {
            return VIRTIO_NET_ERR;
        }

        supported_offloads = virtio_net_supported_guest_offloads(n);
        if (offloads & ~supported_offloads) {
            return VIRTIO_NET_ERR;
        }

        n->curr_guest_offloads = offloads;
        virtio_net_apply_guest_offloads(n);

        return VIRTIO_NET_OK;
    } else {
        return VIRTIO_NET_ERR;
    }
}

798
static int virtio_net_handle_mac(VirtIONet *n, uint8_t cmd,
799
                                 struct iovec *iov, unsigned int iov_cnt)
800
{
801
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
802
    struct virtio_net_ctrl_mac mac_data;
803
    size_t s;
804
    NetClientState *nc = qemu_get_queue(n->nic);
805

806 807 808 809 810 811
    if (cmd == VIRTIO_NET_CTRL_MAC_ADDR_SET) {
        if (iov_size(iov, iov_cnt) != sizeof(n->mac)) {
            return VIRTIO_NET_ERR;
        }
        s = iov_to_buf(iov, iov_cnt, 0, &n->mac, sizeof(n->mac));
        assert(s == sizeof(n->mac));
J
Jason Wang 已提交
812
        qemu_format_nic_info_str(qemu_get_queue(n->nic), n->mac);
813 814
        rxfilter_notify(nc);

815 816 817
        return VIRTIO_NET_OK;
    }

818
    if (cmd != VIRTIO_NET_CTRL_MAC_TABLE_SET) {
819
        return VIRTIO_NET_ERR;
820
    }
821

822 823 824 825 826
    int in_use = 0;
    int first_multi = 0;
    uint8_t uni_overflow = 0;
    uint8_t multi_overflow = 0;
    uint8_t *macs = g_malloc0(MAC_TABLE_ENTRIES * ETH_ALEN);
827

828 829
    s = iov_to_buf(iov, iov_cnt, 0, &mac_data.entries,
                   sizeof(mac_data.entries));
830
    mac_data.entries = virtio_ldl_p(vdev, &mac_data.entries);
831
    if (s != sizeof(mac_data.entries)) {
832
        goto error;
833 834
    }
    iov_discard_front(&iov, &iov_cnt, s);
835

836
    if (mac_data.entries * ETH_ALEN > iov_size(iov, iov_cnt)) {
837
        goto error;
838
    }
839 840

    if (mac_data.entries <= MAC_TABLE_ENTRIES) {
841
        s = iov_to_buf(iov, iov_cnt, 0, macs,
842 843
                       mac_data.entries * ETH_ALEN);
        if (s != mac_data.entries * ETH_ALEN) {
844
            goto error;
845
        }
846
        in_use += mac_data.entries;
847
    } else {
848
        uni_overflow = 1;
849 850
    }

851 852
    iov_discard_front(&iov, &iov_cnt, mac_data.entries * ETH_ALEN);

853
    first_multi = in_use;
854

855 856
    s = iov_to_buf(iov, iov_cnt, 0, &mac_data.entries,
                   sizeof(mac_data.entries));
857
    mac_data.entries = virtio_ldl_p(vdev, &mac_data.entries);
858
    if (s != sizeof(mac_data.entries)) {
859
        goto error;
860 861 862
    }

    iov_discard_front(&iov, &iov_cnt, s);
863

864
    if (mac_data.entries * ETH_ALEN != iov_size(iov, iov_cnt)) {
865
        goto error;
866
    }
867

868
    if (mac_data.entries <= MAC_TABLE_ENTRIES - in_use) {
869
        s = iov_to_buf(iov, iov_cnt, 0, &macs[in_use * ETH_ALEN],
870 871
                       mac_data.entries * ETH_ALEN);
        if (s != mac_data.entries * ETH_ALEN) {
872
            goto error;
873
        }
874
        in_use += mac_data.entries;
875
    } else {
876
        multi_overflow = 1;
877 878
    }

879 880 881 882 883 884
    n->mac_table.in_use = in_use;
    n->mac_table.first_multi = first_multi;
    n->mac_table.uni_overflow = uni_overflow;
    n->mac_table.multi_overflow = multi_overflow;
    memcpy(n->mac_table.macs, macs, MAC_TABLE_ENTRIES * ETH_ALEN);
    g_free(macs);
885 886
    rxfilter_notify(nc);

887
    return VIRTIO_NET_OK;
888 889

error:
890
    g_free(macs);
891
    return VIRTIO_NET_ERR;
892 893
}

894
static int virtio_net_handle_vlan_table(VirtIONet *n, uint8_t cmd,
895
                                        struct iovec *iov, unsigned int iov_cnt)
896
{
897
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
898
    uint16_t vid;
899
    size_t s;
900
    NetClientState *nc = qemu_get_queue(n->nic);
901

902
    s = iov_to_buf(iov, iov_cnt, 0, &vid, sizeof(vid));
903
    vid = virtio_lduw_p(vdev, &vid);
904
    if (s != sizeof(vid)) {
905 906 907 908 909 910 911 912 913 914 915 916 917
        return VIRTIO_NET_ERR;
    }

    if (vid >= MAX_VLAN)
        return VIRTIO_NET_ERR;

    if (cmd == VIRTIO_NET_CTRL_VLAN_ADD)
        n->vlans[vid >> 5] |= (1U << (vid & 0x1f));
    else if (cmd == VIRTIO_NET_CTRL_VLAN_DEL)
        n->vlans[vid >> 5] &= ~(1U << (vid & 0x1f));
    else
        return VIRTIO_NET_ERR;

918 919
    rxfilter_notify(nc);

920 921 922
    return VIRTIO_NET_OK;
}

J
Jason Wang 已提交
923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939
static int virtio_net_handle_announce(VirtIONet *n, uint8_t cmd,
                                      struct iovec *iov, unsigned int iov_cnt)
{
    if (cmd == VIRTIO_NET_CTRL_ANNOUNCE_ACK &&
        n->status & VIRTIO_NET_S_ANNOUNCE) {
        n->status &= ~VIRTIO_NET_S_ANNOUNCE;
        if (n->announce_counter) {
            timer_mod(n->announce_timer,
                      qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL) +
                      self_announce_delay(n->announce_counter));
        }
        return VIRTIO_NET_OK;
    } else {
        return VIRTIO_NET_ERR;
    }
}

J
Jason Wang 已提交
940
static int virtio_net_handle_mq(VirtIONet *n, uint8_t cmd,
941
                                struct iovec *iov, unsigned int iov_cnt)
J
Jason Wang 已提交
942
{
943
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
944 945 946
    struct virtio_net_ctrl_mq mq;
    size_t s;
    uint16_t queues;
J
Jason Wang 已提交
947

948 949
    s = iov_to_buf(iov, iov_cnt, 0, &mq, sizeof(mq));
    if (s != sizeof(mq)) {
J
Jason Wang 已提交
950 951 952 953 954 955 956
        return VIRTIO_NET_ERR;
    }

    if (cmd != VIRTIO_NET_CTRL_MQ_VQ_PAIRS_SET) {
        return VIRTIO_NET_ERR;
    }

957
    queues = virtio_lduw_p(vdev, &mq.virtqueue_pairs);
J
Jason Wang 已提交
958

959 960 961
    if (queues < VIRTIO_NET_CTRL_MQ_VQ_PAIRS_MIN ||
        queues > VIRTIO_NET_CTRL_MQ_VQ_PAIRS_MAX ||
        queues > n->max_queues ||
J
Jason Wang 已提交
962 963 964 965
        !n->multiqueue) {
        return VIRTIO_NET_ERR;
    }

966
    n->curr_queues = queues;
J
Jason Wang 已提交
967 968
    /* stop the backend before changing the number of queues to avoid handling a
     * disabled queue */
969
    virtio_net_set_status(vdev, vdev->status);
J
Jason Wang 已提交
970 971 972 973
    virtio_net_set_queues(n);

    return VIRTIO_NET_OK;
}
974

975 976
static void virtio_net_handle_ctrl(VirtIODevice *vdev, VirtQueue *vq)
{
977
    VirtIONet *n = VIRTIO_NET(vdev);
978 979
    struct virtio_net_ctrl_hdr ctrl;
    virtio_net_ctrl_ack status = VIRTIO_NET_ERR;
980
    VirtQueueElement *elem;
981
    size_t s;
J
Jason Wang 已提交
982
    struct iovec *iov, *iov2;
983
    unsigned int iov_cnt;
984

985 986 987 988 989 990 991
    for (;;) {
        elem = virtqueue_pop(vq, sizeof(VirtQueueElement));
        if (!elem) {
            break;
        }
        if (iov_size(elem->in_sg, elem->in_num) < sizeof(status) ||
            iov_size(elem->out_sg, elem->out_num) < sizeof(ctrl)) {
992 993 994 995
            virtio_error(vdev, "virtio-net ctrl missing headers");
            virtqueue_detach_element(vq, elem, 0);
            g_free(elem);
            break;
996 997
        }

998 999
        iov_cnt = elem->out_num;
        iov2 = iov = g_memdup(elem->out_sg, sizeof(struct iovec) * elem->out_num);
1000 1001 1002 1003
        s = iov_to_buf(iov, iov_cnt, 0, &ctrl, sizeof(ctrl));
        iov_discard_front(&iov, &iov_cnt, sizeof(ctrl));
        if (s != sizeof(ctrl)) {
            status = VIRTIO_NET_ERR;
A
Amos Kong 已提交
1004
        } else if (ctrl.class == VIRTIO_NET_CTRL_RX) {
1005 1006 1007 1008 1009
            status = virtio_net_handle_rx_mode(n, ctrl.cmd, iov, iov_cnt);
        } else if (ctrl.class == VIRTIO_NET_CTRL_MAC) {
            status = virtio_net_handle_mac(n, ctrl.cmd, iov, iov_cnt);
        } else if (ctrl.class == VIRTIO_NET_CTRL_VLAN) {
            status = virtio_net_handle_vlan_table(n, ctrl.cmd, iov, iov_cnt);
J
Jason Wang 已提交
1010 1011
        } else if (ctrl.class == VIRTIO_NET_CTRL_ANNOUNCE) {
            status = virtio_net_handle_announce(n, ctrl.cmd, iov, iov_cnt);
J
Jason Wang 已提交
1012
        } else if (ctrl.class == VIRTIO_NET_CTRL_MQ) {
1013
            status = virtio_net_handle_mq(n, ctrl.cmd, iov, iov_cnt);
1014 1015
        } else if (ctrl.class == VIRTIO_NET_CTRL_GUEST_OFFLOADS) {
            status = virtio_net_handle_offloads(n, ctrl.cmd, iov, iov_cnt);
1016 1017
        }

1018
        s = iov_from_buf(elem->in_sg, elem->in_num, 0, &status, sizeof(status));
1019
        assert(s == sizeof(status));
1020

1021
        virtqueue_push(vq, elem, sizeof(status));
1022
        virtio_notify(vdev, vq);
J
Jason Wang 已提交
1023
        g_free(iov2);
1024
        g_free(elem);
1025 1026 1027
    }
}

A
aliguori 已提交
1028 1029 1030 1031
/* RX */

static void virtio_net_handle_rx(VirtIODevice *vdev, VirtQueue *vq)
{
1032
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
1033
    int queue_index = vq2q(virtio_get_queue_index(vq));
1034

J
Jason Wang 已提交
1035
    qemu_flush_queued_packets(qemu_get_subqueue(n->nic, queue_index));
A
aliguori 已提交
1036 1037
}

1038
static int virtio_net_can_receive(NetClientState *nc)
A
aliguori 已提交
1039
{
J
Jason Wang 已提交
1040
    VirtIONet *n = qemu_get_nic_opaque(nc);
1041
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
J
Jason Wang 已提交
1042
    VirtIONetQueue *q = virtio_net_get_subqueue(nc);
1043

1044
    if (!vdev->vm_running) {
1045 1046
        return 0;
    }
1047

J
Jason Wang 已提交
1048 1049 1050 1051
    if (nc->queue_index >= n->curr_queues) {
        return 0;
    }

1052
    if (!virtio_queue_ready(q->rx_vq) ||
1053
        !(vdev->status & VIRTIO_CONFIG_S_DRIVER_OK)) {
A
aliguori 已提交
1054
        return 0;
1055
    }
A
aliguori 已提交
1056

1057 1058 1059
    return 1;
}

1060
static int virtio_net_has_buffers(VirtIONetQueue *q, int bufsize)
1061
{
1062 1063
    VirtIONet *n = q->n;
    if (virtio_queue_empty(q->rx_vq) ||
A
aliguori 已提交
1064
        (n->mergeable_rx_bufs &&
1065 1066
         !virtqueue_avail_bytes(q->rx_vq, bufsize, 0))) {
        virtio_queue_set_notification(q->rx_vq, 1);
1067 1068 1069 1070 1071

        /* To avoid a race condition where the guest has made some buffers
         * available after the above check but before notification was
         * enabled, check for available buffers again.
         */
1072
        if (virtio_queue_empty(q->rx_vq) ||
1073
            (n->mergeable_rx_bufs &&
1074
             !virtqueue_avail_bytes(q->rx_vq, bufsize, 0))) {
1075
            return 0;
1076
        }
A
aliguori 已提交
1077 1078
    }

1079
    virtio_queue_set_notification(q->rx_vq, 0);
A
aliguori 已提交
1080 1081 1082
    return 1;
}

1083
static void virtio_net_hdr_swap(VirtIODevice *vdev, struct virtio_net_hdr *hdr)
1084
{
1085 1086 1087 1088
    virtio_tswap16s(vdev, &hdr->hdr_len);
    virtio_tswap16s(vdev, &hdr->gso_size);
    virtio_tswap16s(vdev, &hdr->csum_start);
    virtio_tswap16s(vdev, &hdr->csum_offset);
1089 1090
}

A
Anthony Liguori 已提交
1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105
/* dhclient uses AF_PACKET but doesn't pass auxdata to the kernel so
 * it never finds out that the packets don't have valid checksums.  This
 * causes dhclient to get upset.  Fedora's carried a patch for ages to
 * fix this with Xen but it hasn't appeared in an upstream release of
 * dhclient yet.
 *
 * To avoid breaking existing guests, we catch udp packets and add
 * checksums.  This is terrible but it's better than hacking the guest
 * kernels.
 *
 * N.B. if we introduce a zero-copy API, this operation is no longer free so
 * we should provide a mechanism to disable it to avoid polluting the host
 * cache.
 */
static void work_around_broken_dhclient(struct virtio_net_hdr *hdr,
M
Michael S. Tsirkin 已提交
1106
                                        uint8_t *buf, size_t size)
A
Anthony Liguori 已提交
1107 1108 1109 1110 1111 1112
{
    if ((hdr->flags & VIRTIO_NET_HDR_F_NEEDS_CSUM) && /* missing csum */
        (size > 27 && size < 1500) && /* normal sized MTU */
        (buf[12] == 0x08 && buf[13] == 0x00) && /* ethertype == IPv4 */
        (buf[23] == 17) && /* ip.protocol == UDP */
        (buf[34] == 0 && buf[35] == 67)) { /* udp.srcport == bootps */
M
Michael S. Tsirkin 已提交
1113
        net_checksum_calculate(buf, size);
A
Anthony Liguori 已提交
1114 1115 1116 1117
        hdr->flags &= ~VIRTIO_NET_HDR_F_NEEDS_CSUM;
    }
}

1118 1119
static void receive_header(VirtIONet *n, const struct iovec *iov, int iov_cnt,
                           const void *buf, size_t size)
A
aliguori 已提交
1120
{
M
Mark McLoughlin 已提交
1121
    if (n->has_vnet_hdr) {
M
Michael S. Tsirkin 已提交
1122 1123
        /* FIXME this cast is evil */
        void *wbuf = (void *)buf;
1124 1125
        work_around_broken_dhclient(wbuf, wbuf + n->host_hdr_len,
                                    size - n->host_hdr_len);
1126 1127 1128 1129

        if (n->needs_vnet_hdr_swap) {
            virtio_net_hdr_swap(VIRTIO_DEVICE(n), wbuf);
        }
1130
        iov_from_buf(iov, iov_cnt, 0, buf, sizeof(struct virtio_net_hdr));
M
Michael S. Tsirkin 已提交
1131 1132 1133 1134 1135 1136
    } else {
        struct virtio_net_hdr hdr = {
            .flags = 0,
            .gso_type = VIRTIO_NET_HDR_GSO_NONE
        };
        iov_from_buf(iov, iov_cnt, 0, &hdr, sizeof hdr);
M
Mark McLoughlin 已提交
1137
    }
A
aliguori 已提交
1138 1139
}

1140 1141 1142
static int receive_filter(VirtIONet *n, const uint8_t *buf, int size)
{
    static const uint8_t bcast[] = {0xff, 0xff, 0xff, 0xff, 0xff, 0xff};
1143
    static const uint8_t vlan[] = {0x81, 0x00};
1144
    uint8_t *ptr = (uint8_t *)buf;
1145
    int i;
1146 1147 1148 1149

    if (n->promisc)
        return 1;

1150
    ptr += n->host_hdr_len;
M
Mark McLoughlin 已提交
1151

1152
    if (!memcmp(&ptr[12], vlan, sizeof(vlan))) {
1153
        int vid = lduw_be_p(ptr + 14) & 0xfff;
1154 1155 1156 1157
        if (!(n->vlans[vid >> 5] & (1U << (vid & 0x1f))))
            return 0;
    }

1158 1159
    if (ptr[0] & 1) { // multicast
        if (!memcmp(ptr, bcast, sizeof(bcast))) {
1160 1161 1162
            return !n->nobcast;
        } else if (n->nomulti) {
            return 0;
1163
        } else if (n->allmulti || n->mac_table.multi_overflow) {
1164 1165
            return 1;
        }
1166 1167 1168 1169 1170 1171

        for (i = n->mac_table.first_multi; i < n->mac_table.in_use; i++) {
            if (!memcmp(ptr, &n->mac_table.macs[i * ETH_ALEN], ETH_ALEN)) {
                return 1;
            }
        }
1172
    } else { // unicast
1173 1174 1175
        if (n->nouni) {
            return 0;
        } else if (n->alluni || n->mac_table.uni_overflow) {
1176 1177
            return 1;
        } else if (!memcmp(ptr, n->mac, ETH_ALEN)) {
1178 1179
            return 1;
        }
1180

1181 1182 1183 1184 1185
        for (i = 0; i < n->mac_table.first_multi; i++) {
            if (!memcmp(ptr, &n->mac_table.macs[i * ETH_ALEN], ETH_ALEN)) {
                return 1;
            }
        }
1186 1187
    }

1188 1189 1190
    return 0;
}

1191 1192
static ssize_t virtio_net_receive_rcu(NetClientState *nc, const uint8_t *buf,
                                      size_t size)
A
aliguori 已提交
1193
{
J
Jason Wang 已提交
1194
    VirtIONet *n = qemu_get_nic_opaque(nc);
J
Jason Wang 已提交
1195
    VirtIONetQueue *q = virtio_net_get_subqueue(nc);
1196
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
1197 1198 1199
    struct iovec mhdr_sg[VIRTQUEUE_MAX_SIZE];
    struct virtio_net_hdr_mrg_rxbuf mhdr;
    unsigned mhdr_cnt = 0;
M
Michael S. Tsirkin 已提交
1200
    size_t offset, i, guest_offset;
A
aliguori 已提交
1201

J
Jason Wang 已提交
1202
    if (!virtio_net_can_receive(nc)) {
1203
        return -1;
J
Jason Wang 已提交
1204
    }
1205

1206
    /* hdr_len refers to the header we supply to the guest */
1207
    if (!virtio_net_has_buffers(q, size + n->guest_hdr_len - n->host_hdr_len)) {
1208
        return 0;
1209
    }
A
aliguori 已提交
1210

1211
    if (!receive_filter(n, buf, size))
1212
        return size;
1213

A
aliguori 已提交
1214 1215 1216
    offset = i = 0;

    while (offset < size) {
1217
        VirtQueueElement *elem;
A
aliguori 已提交
1218
        int len, total;
1219
        const struct iovec *sg;
A
aliguori 已提交
1220

A
Amit Shah 已提交
1221
        total = 0;
A
aliguori 已提交
1222

1223 1224
        elem = virtqueue_pop(q->rx_vq, sizeof(VirtQueueElement));
        if (!elem) {
1225 1226 1227 1228 1229 1230 1231 1232 1233 1234
            if (i) {
                virtio_error(vdev, "virtio-net unexpected empty queue: "
                             "i %zd mergeable %d offset %zd, size %zd, "
                             "guest hdr len %zd, host hdr len %zd "
                             "guest features 0x%" PRIx64,
                             i, n->mergeable_rx_bufs, offset, size,
                             n->guest_hdr_len, n->host_hdr_len,
                             vdev->guest_features);
            }
            return -1;
A
aliguori 已提交
1235 1236
        }

1237
        if (elem->in_num < 1) {
1238 1239 1240 1241 1242
            virtio_error(vdev,
                         "virtio-net receive queue contains no in buffers");
            virtqueue_detach_element(q->rx_vq, elem, 0);
            g_free(elem);
            return -1;
A
aliguori 已提交
1243 1244
        }

1245
        sg = elem->in_sg;
A
aliguori 已提交
1246
        if (i == 0) {
1247
            assert(offset == 0);
1248 1249
            if (n->mergeable_rx_bufs) {
                mhdr_cnt = iov_copy(mhdr_sg, ARRAY_SIZE(mhdr_sg),
1250
                                    sg, elem->in_num,
1251 1252 1253
                                    offsetof(typeof(mhdr), num_buffers),
                                    sizeof(mhdr.num_buffers));
            }
A
aliguori 已提交
1254

1255
            receive_header(n, sg, elem->in_num, buf, size);
1256
            offset = n->host_hdr_len;
1257
            total += n->guest_hdr_len;
M
Michael S. Tsirkin 已提交
1258 1259 1260
            guest_offset = n->guest_hdr_len;
        } else {
            guest_offset = 0;
A
aliguori 已提交
1261 1262 1263
        }

        /* copy in packet.  ugh */
1264
        len = iov_from_buf(sg, elem->in_num, guest_offset,
1265
                           buf + offset, size - offset);
A
aliguori 已提交
1266
        total += len;
1267 1268 1269 1270 1271
        offset += len;
        /* If buffers can't be merged, at this point we
         * must have consumed the complete packet.
         * Otherwise, drop it. */
        if (!n->mergeable_rx_bufs && offset < size) {
1272
            virtqueue_unpop(q->rx_vq, elem, total);
1273
            g_free(elem);
1274 1275
            return size;
        }
A
aliguori 已提交
1276 1277

        /* signal other side */
1278 1279
        virtqueue_fill(q->rx_vq, elem, total, i++);
        g_free(elem);
A
aliguori 已提交
1280 1281
    }

1282
    if (mhdr_cnt) {
1283
        virtio_stw_p(vdev, &mhdr.num_buffers, i);
1284 1285 1286
        iov_from_buf(mhdr_sg, mhdr_cnt,
                     0,
                     &mhdr.num_buffers, sizeof mhdr.num_buffers);
1287
    }
A
aliguori 已提交
1288

1289
    virtqueue_flush(q->rx_vq, i);
1290
    virtio_notify(vdev, q->rx_vq);
1291 1292

    return size;
A
aliguori 已提交
1293 1294
}

1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305
static ssize_t virtio_net_receive(NetClientState *nc, const uint8_t *buf,
                                  size_t size)
{
    ssize_t r;

    rcu_read_lock();
    r = virtio_net_receive_rcu(nc, buf, size);
    rcu_read_unlock();
    return r;
}

1306
static int32_t virtio_net_flush_tx(VirtIONetQueue *q);
1307

1308
static void virtio_net_tx_complete(NetClientState *nc, ssize_t len)
1309
{
J
Jason Wang 已提交
1310
    VirtIONet *n = qemu_get_nic_opaque(nc);
J
Jason Wang 已提交
1311
    VirtIONetQueue *q = virtio_net_get_subqueue(nc);
1312
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
1313

1314
    virtqueue_push(q->tx_vq, q->async_tx.elem, 0);
1315
    virtio_notify(vdev, q->tx_vq);
1316

1317 1318
    g_free(q->async_tx.elem);
    q->async_tx.elem = NULL;
1319

1320 1321
    virtio_queue_set_notification(q->tx_vq, 1);
    virtio_net_flush_tx(q);
1322 1323
}

A
aliguori 已提交
1324
/* TX */
1325
static int32_t virtio_net_flush_tx(VirtIONetQueue *q)
A
aliguori 已提交
1326
{
1327
    VirtIONet *n = q->n;
1328
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
1329
    VirtQueueElement *elem;
1330
    int32_t num_packets = 0;
J
Jason Wang 已提交
1331
    int queue_index = vq2q(virtio_get_queue_index(q->tx_vq));
1332
    if (!(vdev->status & VIRTIO_CONFIG_S_DRIVER_OK)) {
1333 1334
        return num_packets;
    }
A
aliguori 已提交
1335

1336
    if (q->async_tx.elem) {
1337
        virtio_queue_set_notification(q->tx_vq, 0);
1338
        return num_packets;
1339 1340
    }

1341
    for (;;) {
J
Jason Wang 已提交
1342
        ssize_t ret;
1343 1344
        unsigned int out_num;
        struct iovec sg[VIRTQUEUE_MAX_SIZE], sg2[VIRTQUEUE_MAX_SIZE + 1], *out_sg;
J
Jason Wang 已提交
1345
        struct virtio_net_hdr_mrg_rxbuf mhdr;
A
aliguori 已提交
1346

1347 1348 1349 1350 1351 1352 1353
        elem = virtqueue_pop(q->tx_vq, sizeof(VirtQueueElement));
        if (!elem) {
            break;
        }

        out_num = elem->out_num;
        out_sg = elem->out_sg;
M
Michael S. Tsirkin 已提交
1354
        if (out_num < 1) {
1355 1356 1357 1358
            virtio_error(vdev, "virtio-net header not in first element");
            virtqueue_detach_element(q->tx_vq, elem, 0);
            g_free(elem);
            return -EINVAL;
A
aliguori 已提交
1359 1360
        }

1361
        if (n->has_vnet_hdr) {
J
Jason Wang 已提交
1362 1363
            if (iov_to_buf(out_sg, out_num, 0, &mhdr, n->guest_hdr_len) <
                n->guest_hdr_len) {
1364 1365 1366 1367
                virtio_error(vdev, "virtio-net header incorrect");
                virtqueue_detach_element(q->tx_vq, elem, 0);
                g_free(elem);
                return -EINVAL;
1368
            }
1369
            if (n->needs_vnet_hdr_swap) {
J
Jason Wang 已提交
1370 1371 1372 1373 1374 1375 1376 1377 1378 1379 1380 1381
                virtio_net_hdr_swap(vdev, (void *) &mhdr);
                sg2[0].iov_base = &mhdr;
                sg2[0].iov_len = n->guest_hdr_len;
                out_num = iov_copy(&sg2[1], ARRAY_SIZE(sg2) - 1,
                                   out_sg, out_num,
                                   n->guest_hdr_len, -1);
                if (out_num == VIRTQUEUE_MAX_SIZE) {
                    goto drop;
		}
                out_num += 1;
                out_sg = sg2;
	    }
1382
        }
1383 1384 1385 1386 1387 1388 1389 1390 1391 1392 1393 1394 1395 1396 1397
        /*
         * If host wants to see the guest header as is, we can
         * pass it on unchanged. Otherwise, copy just the parts
         * that host is interested in.
         */
        assert(n->host_hdr_len <= n->guest_hdr_len);
        if (n->host_hdr_len != n->guest_hdr_len) {
            unsigned sg_num = iov_copy(sg, ARRAY_SIZE(sg),
                                       out_sg, out_num,
                                       0, n->host_hdr_len);
            sg_num += iov_copy(sg + sg_num, ARRAY_SIZE(sg) - sg_num,
                             out_sg, out_num,
                             n->guest_hdr_len, -1);
            out_num = sg_num;
            out_sg = sg;
A
aliguori 已提交
1398 1399
        }

J
Jason Wang 已提交
1400 1401
        ret = qemu_sendv_packet_async(qemu_get_subqueue(n->nic, queue_index),
                                      out_sg, out_num, virtio_net_tx_complete);
1402
        if (ret == 0) {
1403 1404
            virtio_queue_set_notification(q->tx_vq, 0);
            q->async_tx.elem = elem;
1405
            return -EBUSY;
1406 1407
        }

J
Jason Wang 已提交
1408
drop:
1409
        virtqueue_push(q->tx_vq, elem, 0);
1410
        virtio_notify(vdev, q->tx_vq);
1411
        g_free(elem);
1412 1413 1414 1415

        if (++num_packets >= n->tx_burst) {
            break;
        }
A
aliguori 已提交
1416
    }
1417
    return num_packets;
A
aliguori 已提交
1418 1419
}

1420
static void virtio_net_handle_tx_timer(VirtIODevice *vdev, VirtQueue *vq)
A
aliguori 已提交
1421
{
1422
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
1423
    VirtIONetQueue *q = &n->vqs[vq2q(virtio_get_queue_index(vq))];
A
aliguori 已提交
1424

1425 1426 1427 1428 1429
    if (unlikely((n->status & VIRTIO_NET_S_LINK_UP) == 0)) {
        virtio_net_drop_tx_queue_data(vdev, vq);
        return;
    }

1430
    /* This happens when device was stopped but VCPU wasn't. */
1431
    if (!vdev->vm_running) {
1432
        q->tx_waiting = 1;
1433 1434 1435
        return;
    }

1436
    if (q->tx_waiting) {
A
aliguori 已提交
1437
        virtio_queue_set_notification(vq, 1);
1438
        timer_del(q->tx_timer);
1439
        q->tx_waiting = 0;
1440 1441 1442
        if (virtio_net_flush_tx(q) == -EINVAL) {
            return;
        }
A
aliguori 已提交
1443
    } else {
1444 1445
        timer_mod(q->tx_timer,
                       qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + n->tx_timeout);
1446
        q->tx_waiting = 1;
A
aliguori 已提交
1447 1448 1449 1450
        virtio_queue_set_notification(vq, 0);
    }
}

1451 1452
static void virtio_net_handle_tx_bh(VirtIODevice *vdev, VirtQueue *vq)
{
1453
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
1454
    VirtIONetQueue *q = &n->vqs[vq2q(virtio_get_queue_index(vq))];
1455

1456 1457 1458 1459 1460
    if (unlikely((n->status & VIRTIO_NET_S_LINK_UP) == 0)) {
        virtio_net_drop_tx_queue_data(vdev, vq);
        return;
    }

1461
    if (unlikely(q->tx_waiting)) {
1462 1463
        return;
    }
1464
    q->tx_waiting = 1;
1465
    /* This happens when device was stopped but VCPU wasn't. */
1466
    if (!vdev->vm_running) {
1467 1468
        return;
    }
1469
    virtio_queue_set_notification(vq, 0);
1470
    qemu_bh_schedule(q->tx_bh);
1471 1472
}

A
aliguori 已提交
1473 1474
static void virtio_net_tx_timer(void *opaque)
{
1475 1476
    VirtIONetQueue *q = opaque;
    VirtIONet *n = q->n;
1477
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
1478 1479 1480 1481 1482 1483
    /* This happens when device was stopped but BH wasn't. */
    if (!vdev->vm_running) {
        /* Make sure tx waiting is set, so we'll run when restarted. */
        assert(q->tx_waiting);
        return;
    }
A
aliguori 已提交
1484

1485
    q->tx_waiting = 0;
A
aliguori 已提交
1486 1487

    /* Just in case the driver is not ready on more */
1488
    if (!(vdev->status & VIRTIO_CONFIG_S_DRIVER_OK)) {
A
aliguori 已提交
1489
        return;
1490
    }
A
aliguori 已提交
1491

1492 1493
    virtio_queue_set_notification(q->tx_vq, 1);
    virtio_net_flush_tx(q);
A
aliguori 已提交
1494 1495
}

1496 1497
static void virtio_net_tx_bh(void *opaque)
{
1498 1499
    VirtIONetQueue *q = opaque;
    VirtIONet *n = q->n;
1500
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
1501 1502
    int32_t ret;

1503 1504 1505 1506 1507 1508
    /* This happens when device was stopped but BH wasn't. */
    if (!vdev->vm_running) {
        /* Make sure tx waiting is set, so we'll run when restarted. */
        assert(q->tx_waiting);
        return;
    }
1509

1510
    q->tx_waiting = 0;
1511 1512

    /* Just in case the driver is not ready on more */
1513
    if (unlikely(!(vdev->status & VIRTIO_CONFIG_S_DRIVER_OK))) {
1514
        return;
1515
    }
1516

1517
    ret = virtio_net_flush_tx(q);
1518 1519 1520
    if (ret == -EBUSY || ret == -EINVAL) {
        return; /* Notification re-enable handled by tx_complete or device
                 * broken */
1521 1522 1523 1524 1525
    }

    /* If we flush a full burst of packets, assume there are
     * more coming and immediately reschedule */
    if (ret >= n->tx_burst) {
1526 1527
        qemu_bh_schedule(q->tx_bh);
        q->tx_waiting = 1;
1528 1529 1530 1531 1532 1533
        return;
    }

    /* If less than a full burst, re-enable notification and flush
     * anything that may have come in while we weren't looking.  If
     * we find something, assume the guest is still active and reschedule */
1534
    virtio_queue_set_notification(q->tx_vq, 1);
1535 1536 1537 1538
    ret = virtio_net_flush_tx(q);
    if (ret == -EINVAL) {
        return;
    } else if (ret > 0) {
1539 1540 1541
        virtio_queue_set_notification(q->tx_vq, 0);
        qemu_bh_schedule(q->tx_bh);
        q->tx_waiting = 1;
1542 1543 1544
    }
}

1545 1546 1547 1548
static void virtio_net_add_queue(VirtIONet *n, int index)
{
    VirtIODevice *vdev = VIRTIO_DEVICE(n);

1549 1550
    n->vqs[index].rx_vq = virtio_add_queue(vdev, n->net_conf.rx_queue_size,
                                           virtio_net_handle_rx);
1551

1552 1553
    if (n->net_conf.tx && !strcmp(n->net_conf.tx, "timer")) {
        n->vqs[index].tx_vq =
1554 1555
            virtio_add_queue(vdev, n->net_conf.tx_queue_size,
                             virtio_net_handle_tx_timer);
1556 1557 1558 1559 1560
        n->vqs[index].tx_timer = timer_new_ns(QEMU_CLOCK_VIRTUAL,
                                              virtio_net_tx_timer,
                                              &n->vqs[index]);
    } else {
        n->vqs[index].tx_vq =
1561 1562
            virtio_add_queue(vdev, n->net_conf.tx_queue_size,
                             virtio_net_handle_tx_bh);
1563 1564 1565 1566 1567 1568 1569 1570 1571 1572 1573 1574 1575 1576 1577 1578 1579 1580 1581
        n->vqs[index].tx_bh = qemu_bh_new(virtio_net_tx_bh, &n->vqs[index]);
    }

    n->vqs[index].tx_waiting = 0;
    n->vqs[index].n = n;
}

static void virtio_net_del_queue(VirtIONet *n, int index)
{
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
    VirtIONetQueue *q = &n->vqs[index];
    NetClientState *nc = qemu_get_subqueue(n->nic, index);

    qemu_purge_queued_packets(nc);

    virtio_del_queue(vdev, index * 2);
    if (q->tx_timer) {
        timer_del(q->tx_timer);
        timer_free(q->tx_timer);
1582
        q->tx_timer = NULL;
1583 1584
    } else {
        qemu_bh_delete(q->tx_bh);
1585
        q->tx_bh = NULL;
1586
    }
1587
    q->tx_waiting = 0;
1588 1589 1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612 1613 1614 1615 1616 1617 1618 1619 1620 1621 1622 1623 1624 1625
    virtio_del_queue(vdev, index * 2 + 1);
}

static void virtio_net_change_num_queues(VirtIONet *n, int new_max_queues)
{
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
    int old_num_queues = virtio_get_num_queues(vdev);
    int new_num_queues = new_max_queues * 2 + 1;
    int i;

    assert(old_num_queues >= 3);
    assert(old_num_queues % 2 == 1);

    if (old_num_queues == new_num_queues) {
        return;
    }

    /*
     * We always need to remove and add ctrl vq if
     * old_num_queues != new_num_queues. Remove ctrl_vq first,
     * and then we only enter one of the following too loops.
     */
    virtio_del_queue(vdev, old_num_queues - 1);

    for (i = new_num_queues - 1; i < old_num_queues - 1; i += 2) {
        /* new_num_queues < old_num_queues */
        virtio_net_del_queue(n, i / 2);
    }

    for (i = old_num_queues - 1; i < new_num_queues - 1; i += 2) {
        /* new_num_queues > old_num_queues */
        virtio_net_add_queue(n, i / 2);
    }

    /* add ctrl_vq last */
    n->ctrl_vq = virtio_add_queue(vdev, 64, virtio_net_handle_ctrl);
}

J
Jason Wang 已提交
1626
static void virtio_net_set_multiqueue(VirtIONet *n, int multiqueue)
J
Jason Wang 已提交
1627
{
1628 1629
    int max = multiqueue ? n->max_queues : 1;

J
Jason Wang 已提交
1630
    n->multiqueue = multiqueue;
1631
    virtio_net_change_num_queues(n, max);
J
Jason Wang 已提交
1632 1633 1634 1635

    virtio_net_set_queues(n);
}

1636
static int virtio_net_post_load_device(void *opaque, int version_id)
1637
{
1638 1639
    VirtIONet *n = opaque;
    VirtIODevice *vdev = VIRTIO_DEVICE(n);
1640
    int i, link_down;
A
aliguori 已提交
1641

1642
    virtio_net_set_mrg_rx_bufs(n, n->mergeable_rx_bufs,
1643 1644
                               virtio_vdev_has_feature(vdev,
                                                       VIRTIO_F_VERSION_1));
A
aliguori 已提交
1645

1646
    /* MAC_TABLE_ENTRIES may be different from the saved image */
1647
    if (n->mac_table.in_use > MAC_TABLE_ENTRIES) {
1648
        n->mac_table.in_use = 0;
1649
    }
1650

1651
    if (!virtio_vdev_has_feature(vdev, VIRTIO_NET_F_CTRL_GUEST_OFFLOADS)) {
1652 1653 1654 1655 1656 1657 1658
        n->curr_guest_offloads = virtio_net_supported_guest_offloads(n);
    }

    if (peer_has_vnet_hdr(n)) {
        virtio_net_apply_guest_offloads(n);
    }

1659 1660
    virtio_net_set_queues(n);

1661 1662 1663 1664 1665 1666 1667
    /* Find the first multicast entry in the saved MAC filter */
    for (i = 0; i < n->mac_table.in_use; i++) {
        if (n->mac_table.macs[i * ETH_ALEN] & 1) {
            break;
        }
    }
    n->mac_table.first_multi = i;
1668 1669 1670

    /* nc.link_down can't be migrated, so infer link_down according
     * to link status bit in n->status */
1671 1672 1673 1674
    link_down = (n->status & VIRTIO_NET_S_LINK_UP) == 0;
    for (i = 0; i < n->max_queues; i++) {
        qemu_get_subqueue(n->nic, i)->link_down = link_down;
    }
1675

1676 1677 1678 1679 1680 1681
    if (virtio_vdev_has_feature(vdev, VIRTIO_NET_F_GUEST_ANNOUNCE) &&
        virtio_vdev_has_feature(vdev, VIRTIO_NET_F_CTRL_VQ)) {
        n->announce_counter = SELF_ANNOUNCE_ROUNDS;
        timer_mod(n->announce_timer, qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL));
    }

A
aliguori 已提交
1682 1683 1684
    return 0;
}

1685 1686 1687 1688 1689 1690 1691 1692 1693 1694 1695 1696 1697 1698 1699 1700 1701 1702 1703 1704 1705 1706 1707 1708 1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720 1721 1722 1723 1724 1725 1726 1727 1728 1729 1730 1731
/* tx_waiting field of a VirtIONetQueue */
static const VMStateDescription vmstate_virtio_net_queue_tx_waiting = {
    .name = "virtio-net-queue-tx_waiting",
    .fields = (VMStateField[]) {
        VMSTATE_UINT32(tx_waiting, VirtIONetQueue),
        VMSTATE_END_OF_LIST()
   },
};

static bool max_queues_gt_1(void *opaque, int version_id)
{
    return VIRTIO_NET(opaque)->max_queues > 1;
}

static bool has_ctrl_guest_offloads(void *opaque, int version_id)
{
    return virtio_vdev_has_feature(VIRTIO_DEVICE(opaque),
                                   VIRTIO_NET_F_CTRL_GUEST_OFFLOADS);
}

static bool mac_table_fits(void *opaque, int version_id)
{
    return VIRTIO_NET(opaque)->mac_table.in_use <= MAC_TABLE_ENTRIES;
}

static bool mac_table_doesnt_fit(void *opaque, int version_id)
{
    return !mac_table_fits(opaque, version_id);
}

/* This temporary type is shared by all the WITH_TMP methods
 * although only some fields are used by each.
 */
struct VirtIONetMigTmp {
    VirtIONet      *parent;
    VirtIONetQueue *vqs_1;
    uint16_t        curr_queues_1;
    uint8_t         has_ufo;
    uint32_t        has_vnet_hdr;
};

/* The 2nd and subsequent tx_waiting flags are loaded later than
 * the 1st entry in the queues and only if there's more than one
 * entry.  We use the tmp mechanism to calculate a temporary
 * pointer and count and also validate the count.
 */

1732
static int virtio_net_tx_waiting_pre_save(void *opaque)
1733 1734 1735 1736 1737 1738 1739 1740
{
    struct VirtIONetMigTmp *tmp = opaque;

    tmp->vqs_1 = tmp->parent->vqs + 1;
    tmp->curr_queues_1 = tmp->parent->curr_queues - 1;
    if (tmp->parent->curr_queues == 0) {
        tmp->curr_queues_1 = 0;
    }
1741 1742

    return 0;
1743 1744 1745 1746 1747 1748 1749 1750 1751 1752 1753 1754 1755 1756 1757 1758 1759 1760 1761 1762 1763 1764 1765 1766 1767 1768 1769 1770 1771 1772 1773 1774 1775 1776 1777 1778 1779 1780 1781 1782 1783 1784 1785 1786 1787 1788 1789
}

static int virtio_net_tx_waiting_pre_load(void *opaque)
{
    struct VirtIONetMigTmp *tmp = opaque;

    /* Reuse the pointer setup from save */
    virtio_net_tx_waiting_pre_save(opaque);

    if (tmp->parent->curr_queues > tmp->parent->max_queues) {
        error_report("virtio-net: curr_queues %x > max_queues %x",
            tmp->parent->curr_queues, tmp->parent->max_queues);

        return -EINVAL;
    }

    return 0; /* all good */
}

static const VMStateDescription vmstate_virtio_net_tx_waiting = {
    .name      = "virtio-net-tx_waiting",
    .pre_load  = virtio_net_tx_waiting_pre_load,
    .pre_save  = virtio_net_tx_waiting_pre_save,
    .fields    = (VMStateField[]) {
        VMSTATE_STRUCT_VARRAY_POINTER_UINT16(vqs_1, struct VirtIONetMigTmp,
                                     curr_queues_1,
                                     vmstate_virtio_net_queue_tx_waiting,
                                     struct VirtIONetQueue),
        VMSTATE_END_OF_LIST()
    },
};

/* the 'has_ufo' flag is just tested; if the incoming stream has the
 * flag set we need to check that we have it
 */
static int virtio_net_ufo_post_load(void *opaque, int version_id)
{
    struct VirtIONetMigTmp *tmp = opaque;

    if (tmp->has_ufo && !peer_has_ufo(tmp->parent)) {
        error_report("virtio-net: saved image requires TUN_F_UFO support");
        return -EINVAL;
    }

    return 0;
}

1790
static int virtio_net_ufo_pre_save(void *opaque)
1791 1792 1793 1794
{
    struct VirtIONetMigTmp *tmp = opaque;

    tmp->has_ufo = tmp->parent->has_ufo;
1795 1796

    return 0;
1797 1798 1799 1800 1801 1802 1803 1804 1805 1806 1807 1808 1809 1810 1811 1812 1813 1814 1815 1816 1817 1818 1819 1820 1821 1822 1823
}

static const VMStateDescription vmstate_virtio_net_has_ufo = {
    .name      = "virtio-net-ufo",
    .post_load = virtio_net_ufo_post_load,
    .pre_save  = virtio_net_ufo_pre_save,
    .fields    = (VMStateField[]) {
        VMSTATE_UINT8(has_ufo, struct VirtIONetMigTmp),
        VMSTATE_END_OF_LIST()
    },
};

/* the 'has_vnet_hdr' flag is just tested; if the incoming stream has the
 * flag set we need to check that we have it
 */
static int virtio_net_vnet_post_load(void *opaque, int version_id)
{
    struct VirtIONetMigTmp *tmp = opaque;

    if (tmp->has_vnet_hdr && !peer_has_vnet_hdr(tmp->parent)) {
        error_report("virtio-net: saved image requires vnet_hdr=on");
        return -EINVAL;
    }

    return 0;
}

1824
static int virtio_net_vnet_pre_save(void *opaque)
1825 1826 1827 1828
{
    struct VirtIONetMigTmp *tmp = opaque;

    tmp->has_vnet_hdr = tmp->parent->has_vnet_hdr;
1829 1830

    return 0;
1831 1832 1833 1834 1835 1836 1837 1838 1839 1840 1841 1842 1843 1844 1845 1846 1847 1848 1849 1850 1851 1852 1853 1854 1855 1856 1857 1858 1859 1860 1861 1862 1863 1864 1865 1866 1867 1868 1869 1870 1871 1872 1873 1874 1875 1876 1877 1878 1879 1880 1881 1882 1883 1884 1885 1886 1887 1888 1889 1890 1891 1892 1893 1894
}

static const VMStateDescription vmstate_virtio_net_has_vnet = {
    .name      = "virtio-net-vnet",
    .post_load = virtio_net_vnet_post_load,
    .pre_save  = virtio_net_vnet_pre_save,
    .fields    = (VMStateField[]) {
        VMSTATE_UINT32(has_vnet_hdr, struct VirtIONetMigTmp),
        VMSTATE_END_OF_LIST()
    },
};

static const VMStateDescription vmstate_virtio_net_device = {
    .name = "virtio-net-device",
    .version_id = VIRTIO_NET_VM_VERSION,
    .minimum_version_id = VIRTIO_NET_VM_VERSION,
    .post_load = virtio_net_post_load_device,
    .fields = (VMStateField[]) {
        VMSTATE_UINT8_ARRAY(mac, VirtIONet, ETH_ALEN),
        VMSTATE_STRUCT_POINTER(vqs, VirtIONet,
                               vmstate_virtio_net_queue_tx_waiting,
                               VirtIONetQueue),
        VMSTATE_UINT32(mergeable_rx_bufs, VirtIONet),
        VMSTATE_UINT16(status, VirtIONet),
        VMSTATE_UINT8(promisc, VirtIONet),
        VMSTATE_UINT8(allmulti, VirtIONet),
        VMSTATE_UINT32(mac_table.in_use, VirtIONet),

        /* Guarded pair: If it fits we load it, else we throw it away
         * - can happen if source has a larger MAC table.; post-load
         *  sets flags in this case.
         */
        VMSTATE_VBUFFER_MULTIPLY(mac_table.macs, VirtIONet,
                                0, mac_table_fits, mac_table.in_use,
                                 ETH_ALEN),
        VMSTATE_UNUSED_VARRAY_UINT32(VirtIONet, mac_table_doesnt_fit, 0,
                                     mac_table.in_use, ETH_ALEN),

        /* Note: This is an array of uint32's that's always been saved as a
         * buffer; hold onto your endiannesses; it's actually used as a bitmap
         * but based on the uint.
         */
        VMSTATE_BUFFER_POINTER_UNSAFE(vlans, VirtIONet, 0, MAX_VLAN >> 3),
        VMSTATE_WITH_TMP(VirtIONet, struct VirtIONetMigTmp,
                         vmstate_virtio_net_has_vnet),
        VMSTATE_UINT8(mac_table.multi_overflow, VirtIONet),
        VMSTATE_UINT8(mac_table.uni_overflow, VirtIONet),
        VMSTATE_UINT8(alluni, VirtIONet),
        VMSTATE_UINT8(nomulti, VirtIONet),
        VMSTATE_UINT8(nouni, VirtIONet),
        VMSTATE_UINT8(nobcast, VirtIONet),
        VMSTATE_WITH_TMP(VirtIONet, struct VirtIONetMigTmp,
                         vmstate_virtio_net_has_ufo),
        VMSTATE_SINGLE_TEST(max_queues, VirtIONet, max_queues_gt_1, 0,
                            vmstate_info_uint16_equal, uint16_t),
        VMSTATE_UINT16_TEST(curr_queues, VirtIONet, max_queues_gt_1),
        VMSTATE_WITH_TMP(VirtIONet, struct VirtIONetMigTmp,
                         vmstate_virtio_net_tx_waiting),
        VMSTATE_UINT64_TEST(curr_guest_offloads, VirtIONet,
                            has_ctrl_guest_offloads),
        VMSTATE_END_OF_LIST()
   },
};

M
Mark McLoughlin 已提交
1895
static NetClientInfo net_virtio_info = {
1896
    .type = NET_CLIENT_DRIVER_NIC,
M
Mark McLoughlin 已提交
1897 1898 1899 1900
    .size = sizeof(NICState),
    .can_receive = virtio_net_can_receive,
    .receive = virtio_net_receive,
    .link_status_changed = virtio_net_set_link_status,
1901
    .query_rx_filter = virtio_net_query_rxfilter,
M
Mark McLoughlin 已提交
1902 1903
};

1904 1905
static bool virtio_net_guest_notifier_pending(VirtIODevice *vdev, int idx)
{
1906
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
1907
    NetClientState *nc = qemu_get_subqueue(n->nic, vq2q(idx));
1908
    assert(n->vhost_started);
1909
    return vhost_net_virtqueue_pending(get_vhost_net(nc->peer), idx);
1910 1911 1912 1913 1914
}

static void virtio_net_guest_notifier_mask(VirtIODevice *vdev, int idx,
                                           bool mask)
{
1915
    VirtIONet *n = VIRTIO_NET(vdev);
J
Jason Wang 已提交
1916
    NetClientState *nc = qemu_get_subqueue(n->nic, vq2q(idx));
1917
    assert(n->vhost_started);
1918
    vhost_net_virtqueue_mask(get_vhost_net(nc->peer),
1919 1920 1921
                             vdev, idx, mask);
}

G
Gerd Hoffmann 已提交
1922
static void virtio_net_set_config_size(VirtIONet *n, uint64_t host_features)
A
aliguori 已提交
1923
{
1924
    int i, config_size = 0;
1925
    virtio_add_feature(&host_features, VIRTIO_NET_F_MAC);
1926

1927 1928 1929 1930 1931
    for (i = 0; feature_sizes[i].flags != 0; i++) {
        if (host_features & feature_sizes[i].flags) {
            config_size = MAX(feature_sizes[i].end, config_size);
        }
    }
1932 1933 1934
    n->config_size = config_size;
}

1935 1936 1937 1938 1939 1940 1941 1942
void virtio_net_set_netclient_name(VirtIONet *n, const char *name,
                                   const char *type)
{
    /*
     * The name can be NULL, the netclient name will be type.x.
     */
    assert(type != NULL);

1943 1944
    g_free(n->netclient_name);
    g_free(n->netclient_type);
1945
    n->netclient_name = g_strdup(name);
1946 1947 1948
    n->netclient_type = g_strdup(type);
}

1949
static void virtio_net_device_realize(DeviceState *dev, Error **errp)
1950
{
1951
    VirtIODevice *vdev = VIRTIO_DEVICE(dev);
1952
    VirtIONet *n = VIRTIO_NET(dev);
1953
    NetClientState *nc;
1954
    int i;
1955

1956
    if (n->net_conf.mtu) {
1957
        n->host_features |= (1ULL << VIRTIO_NET_F_MTU);
1958 1959
    }

1960 1961 1962 1963 1964 1965 1966 1967 1968 1969 1970 1971 1972 1973 1974 1975 1976 1977 1978
    if (n->net_conf.duplex_str) {
        if (strncmp(n->net_conf.duplex_str, "half", 5) == 0) {
            n->net_conf.duplex = DUPLEX_HALF;
        } else if (strncmp(n->net_conf.duplex_str, "full", 5) == 0) {
            n->net_conf.duplex = DUPLEX_FULL;
        } else {
            error_setg(errp, "'duplex' must be 'half' or 'full'");
        }
        n->host_features |= (1ULL << VIRTIO_NET_F_SPEED_DUPLEX);
    } else {
        n->net_conf.duplex = DUPLEX_UNKNOWN;
    }

    if (n->net_conf.speed < SPEED_UNKNOWN) {
        error_setg(errp, "'speed' must be between 0 and INT_MAX");
    } else if (n->net_conf.speed >= 0) {
        n->host_features |= (1ULL << VIRTIO_NET_F_SPEED_DUPLEX);
    }

1979
    virtio_net_set_config_size(n, n->host_features);
1980
    virtio_init(vdev, "virtio-net", VIRTIO_ID_NET, n->config_size);
A
aliguori 已提交
1981

1982 1983 1984 1985 1986 1987 1988
    /*
     * We set a lower limit on RX queue size to what it always was.
     * Guests that want a smaller ring can always resize it without
     * help from us (using virtio 1 and up).
     */
    if (n->net_conf.rx_queue_size < VIRTIO_NET_RX_QUEUE_MIN_SIZE ||
        n->net_conf.rx_queue_size > VIRTQUEUE_MAX_SIZE ||
1989
        !is_power_of_2(n->net_conf.rx_queue_size)) {
1990 1991 1992 1993 1994 1995 1996 1997
        error_setg(errp, "Invalid rx_queue_size (= %" PRIu16 "), "
                   "must be a power of 2 between %d and %d.",
                   n->net_conf.rx_queue_size, VIRTIO_NET_RX_QUEUE_MIN_SIZE,
                   VIRTQUEUE_MAX_SIZE);
        virtio_cleanup(vdev);
        return;
    }

1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008
    if (n->net_conf.tx_queue_size < VIRTIO_NET_TX_QUEUE_MIN_SIZE ||
        n->net_conf.tx_queue_size > VIRTQUEUE_MAX_SIZE ||
        !is_power_of_2(n->net_conf.tx_queue_size)) {
        error_setg(errp, "Invalid tx_queue_size (= %" PRIu16 "), "
                   "must be a power of 2 between %d and %d",
                   n->net_conf.tx_queue_size, VIRTIO_NET_TX_QUEUE_MIN_SIZE,
                   VIRTQUEUE_MAX_SIZE);
        virtio_cleanup(vdev);
        return;
    }

2009
    n->max_queues = MAX(n->nic_conf.peers.queues, 1);
2010
    if (n->max_queues * 2 + 1 > VIRTIO_QUEUE_MAX) {
2011
        error_setg(errp, "Invalid number of queues (= %" PRIu32 "), "
S
Stefan Weil 已提交
2012
                   "must be a positive integer less than %d.",
2013
                   n->max_queues, (VIRTIO_QUEUE_MAX - 1) / 2);
2014 2015 2016
        virtio_cleanup(vdev);
        return;
    }
2017
    n->vqs = g_malloc0(sizeof(VirtIONetQueue) * n->max_queues);
J
Jason Wang 已提交
2018
    n->curr_queues = 1;
2019
    n->tx_timeout = n->net_conf.txtimer;
2020

2021 2022
    if (n->net_conf.tx && strcmp(n->net_conf.tx, "timer")
                       && strcmp(n->net_conf.tx, "bh")) {
2023 2024
        error_report("virtio-net: "
                     "Unknown option tx=%s, valid options: \"timer\" \"bh\"",
2025
                     n->net_conf.tx);
2026
        error_report("Defaulting to \"bh\"");
2027 2028
    }

2029 2030
    n->net_conf.tx_queue_size = MIN(virtio_net_max_tx_queue_size(n),
                                    n->net_conf.tx_queue_size);
2031

2032
    for (i = 0; i < n->max_queues; i++) {
2033
        virtio_net_add_queue(n, i);
2034
    }
2035

2036
    n->ctrl_vq = virtio_add_queue(vdev, 64, virtio_net_handle_ctrl);
2037 2038
    qemu_macaddr_default_if_unset(&n->nic_conf.macaddr);
    memcpy(&n->mac[0], &n->nic_conf.macaddr, sizeof(n->mac));
2039
    n->status = VIRTIO_NET_S_LINK_UP;
J
Jason Wang 已提交
2040 2041
    n->announce_timer = timer_new_ms(QEMU_CLOCK_VIRTUAL,
                                     virtio_net_announce_timer, n);
A
aliguori 已提交
2042

2043 2044 2045 2046 2047 2048 2049 2050
    if (n->netclient_type) {
        /*
         * Happen when virtio_net_set_netclient_name has been called.
         */
        n->nic = qemu_new_nic(&net_virtio_info, &n->nic_conf,
                              n->netclient_type, n->netclient_name, n);
    } else {
        n->nic = qemu_new_nic(&net_virtio_info, &n->nic_conf,
2051
                              object_get_typename(OBJECT(dev)), dev->id, n);
2052 2053
    }

2054 2055
    peer_test_vnet_hdr(n);
    if (peer_has_vnet_hdr(n)) {
J
Jason Wang 已提交
2056
        for (i = 0; i < n->max_queues; i++) {
2057
            qemu_using_vnet_hdr(qemu_get_subqueue(n->nic, i)->peer, true);
J
Jason Wang 已提交
2058
        }
2059 2060 2061 2062
        n->host_hdr_len = sizeof(struct virtio_net_hdr);
    } else {
        n->host_hdr_len = 0;
    }
M
Mark McLoughlin 已提交
2063

2064
    qemu_format_nic_info_str(qemu_get_queue(n->nic), n->nic_conf.macaddr.a);
2065

J
Jason Wang 已提交
2066
    n->vqs[0].tx_waiting = 0;
2067
    n->tx_burst = n->net_conf.txburst;
2068
    virtio_net_set_mrg_rx_bufs(n, 0, 0);
2069
    n->promisc = 1; /* for compatibility */
A
aliguori 已提交
2070

2071
    n->mac_table.macs = g_malloc0(MAC_TABLE_ENTRIES * ETH_ALEN);
2072

2073
    n->vlans = g_malloc0(MAX_VLAN >> 3);
2074

2075 2076 2077
    nc = qemu_get_queue(n->nic);
    nc->rxfilter_notify_enabled = 1;

2078
    n->qdev = dev;
2079 2080
}

2081
static void virtio_net_device_unrealize(DeviceState *dev, Error **errp)
2082
{
2083 2084
    VirtIODevice *vdev = VIRTIO_DEVICE(dev);
    VirtIONet *n = VIRTIO_NET(dev);
2085
    int i, max_queues;
2086 2087 2088 2089

    /* This will stop vhost backend if appropriate. */
    virtio_net_set_status(vdev, 0);

2090 2091 2092 2093
    g_free(n->netclient_name);
    n->netclient_name = NULL;
    g_free(n->netclient_type);
    n->netclient_type = NULL;
2094

2095 2096 2097
    g_free(n->mac_table.macs);
    g_free(n->vlans);

2098 2099 2100
    max_queues = n->multiqueue ? n->max_queues : 1;
    for (i = 0; i < max_queues; i++) {
        virtio_net_del_queue(n, i);
2101 2102
    }

J
Jason Wang 已提交
2103 2104
    timer_del(n->announce_timer);
    timer_free(n->announce_timer);
2105 2106
    g_free(n->vqs);
    qemu_del_nic(n->nic);
2107
    virtio_cleanup(vdev);
2108 2109 2110 2111 2112 2113 2114 2115 2116 2117 2118
}

static void virtio_net_instance_init(Object *obj)
{
    VirtIONet *n = VIRTIO_NET(obj);

    /*
     * The default config_size is sizeof(struct virtio_net_config).
     * Can be overriden with virtio_net_set_config_size.
     */
    n->config_size = sizeof(struct virtio_net_config);
2119 2120 2121
    device_add_bootindex_property(obj, &n->nic_conf.bootindex,
                                  "bootindex", "/ethernet-phy@0",
                                  DEVICE(n), NULL);
2122 2123
}

2124
static int virtio_net_pre_save(void *opaque)
2125 2126 2127 2128 2129 2130
{
    VirtIONet *n = opaque;

    /* At this point, backend must be stopped, otherwise
     * it might keep writing to memory. */
    assert(!n->vhost_started);
2131 2132

    return 0;
2133 2134 2135 2136 2137 2138 2139 2140 2141 2142 2143 2144
}

static const VMStateDescription vmstate_virtio_net = {
    .name = "virtio-net",
    .minimum_version_id = VIRTIO_NET_VM_VERSION,
    .version_id = VIRTIO_NET_VM_VERSION,
    .fields = (VMStateField[]) {
        VMSTATE_VIRTIO_DEVICE,
        VMSTATE_END_OF_LIST()
    },
    .pre_save = virtio_net_pre_save,
};
2145

2146
static Property virtio_net_properties[] = {
2147 2148 2149
    DEFINE_PROP_BIT64("csum", VirtIONet, host_features,
                    VIRTIO_NET_F_CSUM, true),
    DEFINE_PROP_BIT64("guest_csum", VirtIONet, host_features,
2150
                    VIRTIO_NET_F_GUEST_CSUM, true),
2151 2152
    DEFINE_PROP_BIT64("gso", VirtIONet, host_features, VIRTIO_NET_F_GSO, true),
    DEFINE_PROP_BIT64("guest_tso4", VirtIONet, host_features,
2153
                    VIRTIO_NET_F_GUEST_TSO4, true),
2154
    DEFINE_PROP_BIT64("guest_tso6", VirtIONet, host_features,
2155
                    VIRTIO_NET_F_GUEST_TSO6, true),
2156
    DEFINE_PROP_BIT64("guest_ecn", VirtIONet, host_features,
2157
                    VIRTIO_NET_F_GUEST_ECN, true),
2158
    DEFINE_PROP_BIT64("guest_ufo", VirtIONet, host_features,
2159
                    VIRTIO_NET_F_GUEST_UFO, true),
2160
    DEFINE_PROP_BIT64("guest_announce", VirtIONet, host_features,
2161
                    VIRTIO_NET_F_GUEST_ANNOUNCE, true),
2162
    DEFINE_PROP_BIT64("host_tso4", VirtIONet, host_features,
2163
                    VIRTIO_NET_F_HOST_TSO4, true),
2164
    DEFINE_PROP_BIT64("host_tso6", VirtIONet, host_features,
2165
                    VIRTIO_NET_F_HOST_TSO6, true),
2166
    DEFINE_PROP_BIT64("host_ecn", VirtIONet, host_features,
2167
                    VIRTIO_NET_F_HOST_ECN, true),
2168
    DEFINE_PROP_BIT64("host_ufo", VirtIONet, host_features,
2169
                    VIRTIO_NET_F_HOST_UFO, true),
2170
    DEFINE_PROP_BIT64("mrg_rxbuf", VirtIONet, host_features,
2171
                    VIRTIO_NET_F_MRG_RXBUF, true),
2172
    DEFINE_PROP_BIT64("status", VirtIONet, host_features,
2173
                    VIRTIO_NET_F_STATUS, true),
2174
    DEFINE_PROP_BIT64("ctrl_vq", VirtIONet, host_features,
2175
                    VIRTIO_NET_F_CTRL_VQ, true),
2176
    DEFINE_PROP_BIT64("ctrl_rx", VirtIONet, host_features,
2177
                    VIRTIO_NET_F_CTRL_RX, true),
2178
    DEFINE_PROP_BIT64("ctrl_vlan", VirtIONet, host_features,
2179
                    VIRTIO_NET_F_CTRL_VLAN, true),
2180
    DEFINE_PROP_BIT64("ctrl_rx_extra", VirtIONet, host_features,
2181
                    VIRTIO_NET_F_CTRL_RX_EXTRA, true),
2182
    DEFINE_PROP_BIT64("ctrl_mac_addr", VirtIONet, host_features,
2183
                    VIRTIO_NET_F_CTRL_MAC_ADDR, true),
2184
    DEFINE_PROP_BIT64("ctrl_guest_offloads", VirtIONet, host_features,
2185
                    VIRTIO_NET_F_CTRL_GUEST_OFFLOADS, true),
2186
    DEFINE_PROP_BIT64("mq", VirtIONet, host_features, VIRTIO_NET_F_MQ, false),
2187 2188
    DEFINE_NIC_PROPERTIES(VirtIONet, nic_conf),
    DEFINE_PROP_UINT32("x-txtimer", VirtIONet, net_conf.txtimer,
2189
                       TX_TIMER_INTERVAL),
2190 2191
    DEFINE_PROP_INT32("x-txburst", VirtIONet, net_conf.txburst, TX_BURST),
    DEFINE_PROP_STRING("tx", VirtIONet, net_conf.tx),
2192 2193
    DEFINE_PROP_UINT16("rx_queue_size", VirtIONet, net_conf.rx_queue_size,
                       VIRTIO_NET_RX_QUEUE_DEFAULT_SIZE),
2194 2195
    DEFINE_PROP_UINT16("tx_queue_size", VirtIONet, net_conf.tx_queue_size,
                       VIRTIO_NET_TX_QUEUE_DEFAULT_SIZE),
2196
    DEFINE_PROP_UINT16("host_mtu", VirtIONet, net_conf.mtu, 0),
2197 2198
    DEFINE_PROP_BOOL("x-mtu-bypass-backend", VirtIONet, mtu_bypass_backend,
                     true),
2199 2200
    DEFINE_PROP_INT32("speed", VirtIONet, net_conf.speed, SPEED_UNKNOWN),
    DEFINE_PROP_STRING("duplex", VirtIONet, net_conf.duplex_str),
2201 2202 2203 2204 2205 2206 2207
    DEFINE_PROP_END_OF_LIST(),
};

static void virtio_net_class_init(ObjectClass *klass, void *data)
{
    DeviceClass *dc = DEVICE_CLASS(klass);
    VirtioDeviceClass *vdc = VIRTIO_DEVICE_CLASS(klass);
2208

2209
    dc->props = virtio_net_properties;
2210
    dc->vmsd = &vmstate_virtio_net;
2211
    set_bit(DEVICE_CATEGORY_NETWORK, dc->categories);
2212
    vdc->realize = virtio_net_device_realize;
2213
    vdc->unrealize = virtio_net_device_unrealize;
2214 2215 2216 2217 2218 2219 2220 2221 2222
    vdc->get_config = virtio_net_get_config;
    vdc->set_config = virtio_net_set_config;
    vdc->get_features = virtio_net_get_features;
    vdc->set_features = virtio_net_set_features;
    vdc->bad_features = virtio_net_bad_features;
    vdc->reset = virtio_net_reset;
    vdc->set_status = virtio_net_set_status;
    vdc->guest_notifier_mask = virtio_net_guest_notifier_mask;
    vdc->guest_notifier_pending = virtio_net_guest_notifier_pending;
2223
    vdc->legacy_features |= (0x1 << VIRTIO_NET_F_GSO);
2224
    vdc->vmsd = &vmstate_virtio_net_device;
2225 2226 2227 2228 2229 2230 2231 2232 2233 2234 2235 2236 2237 2238 2239 2240
}

static const TypeInfo virtio_net_info = {
    .name = TYPE_VIRTIO_NET,
    .parent = TYPE_VIRTIO_DEVICE,
    .instance_size = sizeof(VirtIONet),
    .instance_init = virtio_net_instance_init,
    .class_init = virtio_net_class_init,
};

static void virtio_register_types(void)
{
    type_register_static(&virtio_net_info);
}

type_init(virtio_register_types)