On Tue, Apr 16, 2024 at 1:43 PM Yuri Benditovich <yuri.benditov...@daynix.com> wrote: > > On Tue, Apr 16, 2024 at 7:00 AM Jason Wang <jasow...@redhat.com> wrote: > > > > On Mon, Apr 15, 2024 at 10:05 PM Yuri Benditovich > > <yuri.benditov...@daynix.com> wrote: > > > > > > On Wed, Apr 3, 2024 at 2:11 PM Akihiko Odaki <akihiko.od...@daynix.com> > > > wrote: > > > > > > > > vhost requires eBPF for RSS. When eBPF is not available, virtio-net > > > > implicitly disables RSS even if the user explicitly requests it. Return > > > > an error instead of implicitly disabling RSS if RSS is requested but not > > > > available. > > > > > > > > Signed-off-by: Akihiko Odaki <akihiko.od...@daynix.com> > > > > --- > > > > hw/net/virtio-net.c | 97 > > > > ++++++++++++++++++++++++++--------------------------- > > > > 1 file changed, 48 insertions(+), 49 deletions(-) > > > > > > > > diff --git a/hw/net/virtio-net.c b/hw/net/virtio-net.c > > > > index 61b49e335dea..3d53eba88cfc 100644 > > > > --- a/hw/net/virtio-net.c > > > > +++ b/hw/net/virtio-net.c > > > > @@ -793,9 +793,6 @@ static uint64_t > > > > virtio_net_get_features(VirtIODevice *vdev, uint64_t features, > > > > return features; > > > > } > > > > > > > > - if (!ebpf_rss_is_loaded(&n->ebpf_rss)) { > > > > - virtio_clear_feature(&features, VIRTIO_NET_F_RSS); > > > > - } > > > > features = vhost_net_get_features(get_vhost_net(nc->peer), > > > > features); > > > > vdev->backend_features = features; > > > > > > > > @@ -3591,6 +3588,50 @@ static bool > > > > failover_hide_primary_device(DeviceListener *listener, > > > > return qatomic_read(&n->failover_primary_hidden); > > > > } > > > > > > > > +static void virtio_net_device_unrealize(DeviceState *dev) > > > > +{ > > > > + VirtIODevice *vdev = VIRTIO_DEVICE(dev); > > > > + VirtIONet *n = VIRTIO_NET(dev); > > > > + int i, max_queue_pairs; > > > > + > > > > + if (virtio_has_feature(n->host_features, VIRTIO_NET_F_RSS)) { > > > > + virtio_net_unload_ebpf(n); > > > > + } > > > > + > > > > + /* This will stop vhost backend if appropriate. */ > > > > + virtio_net_set_status(vdev, 0); > > > > + > > > > + g_free(n->netclient_name); > > > > + n->netclient_name = NULL; > > > > + g_free(n->netclient_type); > > > > + n->netclient_type = NULL; > > > > + > > > > + g_free(n->mac_table.macs); > > > > + g_free(n->vlans); > > > > + > > > > + if (n->failover) { > > > > + qobject_unref(n->primary_opts); > > > > + device_listener_unregister(&n->primary_listener); > > > > + migration_remove_notifier(&n->migration_state); > > > > + } else { > > > > + assert(n->primary_opts == NULL); > > > > + } > > > > + > > > > + max_queue_pairs = n->multiqueue ? n->max_queue_pairs : 1; > > > > + for (i = 0; i < max_queue_pairs; i++) { > > > > + virtio_net_del_queue(n, i); > > > > + } > > > > + /* delete also control vq */ > > > > + virtio_del_queue(vdev, max_queue_pairs * 2); > > > > + qemu_announce_timer_del(&n->announce_timer, false); > > > > + g_free(n->vqs); > > > > + qemu_del_nic(n->nic); > > > > + virtio_net_rsc_cleanup(n); > > > > + g_free(n->rss_data.indirections_table); > > > > + net_rx_pkt_uninit(n->rx_pkt); > > > > + virtio_cleanup(vdev); > > > > +} > > > > + > > > > static void virtio_net_device_realize(DeviceState *dev, Error **errp) > > > > { > > > > VirtIODevice *vdev = VIRTIO_DEVICE(dev); > > > > @@ -3760,53 +3801,11 @@ static void > > > > virtio_net_device_realize(DeviceState *dev, Error **errp) > > > > > > > > net_rx_pkt_init(&n->rx_pkt); > > > > > > > > - if (virtio_has_feature(n->host_features, VIRTIO_NET_F_RSS)) { > > > > - virtio_net_load_ebpf(n); > > > > - } > > > > -} > > > > - > > > > -static void virtio_net_device_unrealize(DeviceState *dev) > > > > -{ > > > > - VirtIODevice *vdev = VIRTIO_DEVICE(dev); > > > > - VirtIONet *n = VIRTIO_NET(dev); > > > > - int i, max_queue_pairs; > > > > - > > > > - if (virtio_has_feature(n->host_features, VIRTIO_NET_F_RSS)) { > > > > - virtio_net_unload_ebpf(n); > > > > + if (virtio_has_feature(n->host_features, VIRTIO_NET_F_RSS) && > > > > + !virtio_net_load_ebpf(n) && get_vhost_net(nc->peer)) { > > > > + virtio_net_device_unrealize(dev); > > > > + error_setg(errp, "Can't load eBPF RSS for vhost"); > > > > } > > > > > > As I already mentioned, I think this is an extremely bad idea to > > > fail to run qemu due to such a reason as .absence of one feature. > > > What I suggest is: > > > 1. Redefine rss as tri-state (off|auto|on) > > > 2. Fail to run only if rss is on and not available via ebpf > > > 3. On auto - silently drop it > > > > "Auto" might be promatic for migration compatibility which is hard to > > be used by management layers like libvirt. The reason is that there's > > no way for libvirt to know if it is supported by device or not. > > In terms of migration every feature that somehow depends on the kernel > is problematic, not only RSS.
True, but if we can avoid more, it would still be better. > Last time we added the USO feature - is > it different? I may miss something but we never define tristate for USO? DEFINE_PROP_BIT64("guest_uso4", VirtIONet, host_features, VIRTIO_NET_F_GUEST_USO4, true), DEFINE_PROP_BIT64("guest_uso6", VirtIONet, host_features, VIRTIO_NET_F_GUEST_USO6, true), DEFINE_PROP_BIT64("host_uso", VirtIONet, host_features, VIRTIO_NET_F_HOST_USO, true), ? > And in terms of migration "rss=on" is problematic the same way as "rss=auto". Failing early when launching Qemu is better than failing silently as a guest after a migration. > Can you please show one scenario of migration where they will behave > differently? If you mean the problem of "auto", here's one: Assuming auto is used in both src and dst. On source, rss is enabled but not destination. RSS failed to work after migration. > And in terms of regular experience there is a big advantage. Similarly, silent clearing a feature is also not good: if (!peer_has_vnet_hdr(n)) { virtio_clear_feature(&features, VIRTIO_NET_F_CSUM); virtio_clear_feature(&features, VIRTIO_NET_F_HOST_TSO4); virtio_clear_feature(&features, VIRTIO_NET_F_HOST_TSO6); virtio_clear_feature(&features, VIRTIO_NET_F_HOST_ECN); virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_CSUM); virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_TSO4); virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_TSO6); virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_ECN); virtio_clear_feature(&features, VIRTIO_NET_F_HOST_USO); virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_USO4); virtio_clear_feature(&features, VIRTIO_NET_F_GUEST_USO6); virtio_clear_feature(&features, VIRTIO_NET_F_HASH_REPORT); } The reason we never see complaints is probably because vhost/TAP are the only backend that supports migration where vnet support there has been more than a decade. Thanks > > > > > > Thanks > > > > > 4. The same with 'hash' option - it is not compatible with vhost (at > > > least at the moment) > > > 5. Reformat the patch as it is hard to review it due to replacing > > > entire procedures, i.e. one patch with replacing without changes, > > > another one - with real changes. > > > If this is hard to review only for me - please ignore that. > > > > > > > - > > > > - /* This will stop vhost backend if appropriate. */ > > > > - virtio_net_set_status(vdev, 0); > > > > - > > > > - g_free(n->netclient_name); > > > > - n->netclient_name = NULL; > > > > - g_free(n->netclient_type); > > > > - n->netclient_type = NULL; > > > > - > > > > - g_free(n->mac_table.macs); > > > > - g_free(n->vlans); > > > > - > > > > - if (n->failover) { > > > > - qobject_unref(n->primary_opts); > > > > - device_listener_unregister(&n->primary_listener); > > > > - migration_remove_notifier(&n->migration_state); > > > > - } else { > > > > - assert(n->primary_opts == NULL); > > > > - } > > > > - > > > > - max_queue_pairs = n->multiqueue ? n->max_queue_pairs : 1; > > > > - for (i = 0; i < max_queue_pairs; i++) { > > > > - virtio_net_del_queue(n, i); > > > > - } > > > > - /* delete also control vq */ > > > > - virtio_del_queue(vdev, max_queue_pairs * 2); > > > > - qemu_announce_timer_del(&n->announce_timer, false); > > > > - g_free(n->vqs); > > > > - qemu_del_nic(n->nic); > > > > - virtio_net_rsc_cleanup(n); > > > > - g_free(n->rss_data.indirections_table); > > > > - net_rx_pkt_uninit(n->rx_pkt); > > > > - virtio_cleanup(vdev); > > > > } > > > > > > > > static void virtio_net_reset(VirtIODevice *vdev) > > > > > > > > -- > > > > 2.44.0 > > > > > > > > > >