|
From: | Xueming(Steven) Li |
Subject: | Re: [PATCH v6 1/2] vhost-user: remove VirtQ notifier restore |
Date: | Mon, 7 Feb 2022 13:49:09 +0000 |
User-agent: | Evolution 3.42.2 |
v7 sent with more description, please check.
Thanks,
Xueming
On Fri, 2022-02-04 at 07:25 -0500, Michael S. Tsirkin wrote:
I dropped this for now as I'm a bit lost with this patchset.Cc Raphael maybe he'll understand it better.On Wed, Jan 12, 2022 at 03:05:15PM +0000, Xueming(Steven) Li wrote:On Wed, 2021-11-03 at 16:30 -0400, Michael S. Tsirkin wrote:On Wed, Nov 03, 2021 at 02:48:41PM +0000, Xueming(Steven) Li wrote:On Tue, 2021-11-02 at 02:49 -0400, Michael S. Tsirkin wrote:On Tue, Nov 02, 2021 at 06:08:39AM +0000, Xueming(Steven) Li wrote:On Mon, 2021-11-01 at 17:06 -0400, Michael S. Tsirkin wrote:On Mon, Nov 01, 2021 at 04:38:12PM +0800, Xueming Li wrote:When vhost-user vdpa client suspend, backend may close all resources,VQ notifier mmap address become invalid, restore MR which containsthe invalid address is wrong. vdpa client will set VQ notifier afterreconnect.This patch removes VQ notifier restore and related flags to avoid reusinginvalid address.Fixes: 44866521bd6e ("vhost-user: support registering external host notifiers")Cc: qemu-stable@nongnu.orgCc: Yuwei Zhang <zhangyuwei.9149@bytedance.com>Signed-off-by: Xueming Li <xuemingl@nvidia.com>---hw/virtio/vhost-user.c | 19 +------------------include/hw/virtio/vhost-user.h | 1 -2 files changed, 1 insertion(+), 19 deletions(-)diff --git a/hw/virtio/vhost-user.c b/hw/virtio/vhost-user.cindex bf6e50223c..c671719e9b 100644--- a/hw/virtio/vhost-user.c+++ b/hw/virtio/vhost-user.c@@ -1143,19 +1143,6 @@ static int vhost_user_set_vring_num(struct vhost_dev *dev,return vhost_set_vring(dev, VHOST_USER_SET_VRING_NUM, ring);}-static void vhost_user_host_notifier_restore(struct vhost_dev *dev,- int queue_idx)-{- struct vhost_user *u = dev->opaque;- VhostUserHostNotifier *n = &u->user->notifier[queue_idx];- VirtIODevice *vdev = dev->vdev;-- if (n->addr && !n->set) {- virtio_queue_set_host_notifier_mr(vdev, queue_idx, &n->mr, true);- n->set = true;- }-}-static void vhost_user_host_notifier_remove(struct vhost_dev *dev,int queue_idx){@@ -1163,17 +1150,14 @@ static void vhost_user_host_notifier_remove(struct vhost_dev *dev,VhostUserHostNotifier *n = &u->user->notifier[queue_idx];VirtIODevice *vdev = dev->vdev;- if (n->addr && n->set) {+ if (n->addr) {virtio_queue_set_host_notifier_mr(vdev, queue_idx, &n->mr, false);- n->set = false;}}So on vq stop we still remove the notifier...static int vhost_user_set_vring_base(struct vhost_dev *dev,struct vhost_vring_state *ring){- vhost_user_host_notifier_restore(dev, ring->index);-return vhost_set_vring(dev, VHOST_USER_SET_VRING_BASE, ring);}but on vq start we do not reinstate it? Does this not mean thatnotifiers won't work after stop then start?Yes, backend initially work w/o host notifier, request VM driver toinstall notifier if needed - call this function through slave socket.I think it's cleaner if qemu handles this itself like it did before, itknows vm is stopped without getting called.If vhost play as server, there are 2 scenario that remove the notifier:1. VM suspend: backend still there, it's okay to keep mmap address.2. vhost backend stopped or process killed: resources from backendshould be released. That's why patch 2/2 munmap in notifier removefunction. Then the restore function get nothing to restore, maybe Ishouldn't reverse patch order.I can't say I understand what you mean here. Do you plan to changethe patchset in some way?When you do, pls include a cover letter with a changelog andCc all people you include on patches on the cover letter too.Here is the detail of the problem I encountered, my vhost backend isDPDK vdpa user space application. Notifier address is set when vdpa askqemu to mmap a FD and an offset from vdpa, see functionvhost_user_slave_handle_vring_host_notifier(). If the vdpa applicationrestart of get killed for some reason,vhost_user_host_notifier_remove() is called and notifier MR isuninstalled, the notifier address that retrieved from mmap isreferencing to an invalid FD, not released. This will cause HWresources on kernel side still referenced, most important, when vdpaconnection restored, this invalid notifier will be be restored asnotifier MR.To resolve it, have to remove the notifer restore mechanism, vDPAapplication will issue client socket request again to install notifierto VM, so no concern that the notifier will be lost after resume.Since vdpa might be killed, no chance to notify qemu to removenotifier. An alternative solution is to detect sock disconnection andunmmap notifier, but it looks more complex to me. How do you think?@@ -1538,7 +1522,6 @@ static int vhost_user_slave_handle_vring_host_notifier(struct vhost_dev *dev,}n->addr = addr;- n->set = true;return 0;}diff --git a/include/hw/virtio/vhost-user.h b/include/hw/virtio/vhost-user.hindex a9abca3288..f6012b2078 100644--- a/include/hw/virtio/vhost-user.h+++ b/include/hw/virtio/vhost-user.h@@ -14,7 +14,6 @@typedef struct VhostUserHostNotifier {MemoryRegion mr;void *addr;- bool set;} VhostUserHostNotifier;typedef struct VhostUserState {--2.33.0
[Prev in Thread] | Current Thread | [Next in Thread] |