* [PATCH V2 0/4] vDPA/ifcvf: implement shared IRQ feature
@ 2022-01-25 9:17 Zhu Lingshan
2022-01-25 9:17 ` [PATCH V2 1/4] vDPA/ifcvf: implement IO read/write helpers in the header file Zhu Lingshan
` (4 more replies)
0 siblings, 5 replies; 13+ messages in thread
From: Zhu Lingshan @ 2022-01-25 9:17 UTC (permalink / raw)
To: mst, jasowang; +Cc: netdev, virtualization, Zhu Lingshan
It has been observed that on some platforms/devices, there may
not be enough MSI vectors for virtqueues and the config change.
Under such circumstances, the interrupt sources of a device
have to share vectors/IRQs.
This series implemented a shared IRQ feature for ifcvf.
Please help review.
Changes from V1:
(1) Enable config interrupt when only one vector is allocated(Michael)
(2) Clean vectors/IRQs if failed to request config interrupt
since config interrupt is a must(Michael)
(3) Keep local vdpa_ops, disable irq_bypass by setting IRQ = -EINVAL
for shared IRQ case(Michael)
(4) Improvements on error messages(Michael)
(5) Squash functions implementation patches to the callers(Michael)
Zhu Lingshan (4):
vDPA/ifcvf: implement IO read/write helpers in the header file
vDPA/ifcvf: implement device MSIX vector allocator
vhost_vdpa: don't setup irq offloading when irq_num < 0
vDPA/ifcvf: implement shared IRQ feature
drivers/vdpa/ifcvf/ifcvf_base.c | 67 +++------
drivers/vdpa/ifcvf/ifcvf_base.h | 60 +++++++-
drivers/vdpa/ifcvf/ifcvf_main.c | 254 ++++++++++++++++++++++++++++----
drivers/vhost/vdpa.c | 3 +
4 files changed, 305 insertions(+), 79 deletions(-)
--
2.27.0
^ permalink raw reply [flat|nested] 13+ messages in thread* [PATCH V2 1/4] vDPA/ifcvf: implement IO read/write helpers in the header file 2022-01-25 9:17 [PATCH V2 0/4] vDPA/ifcvf: implement shared IRQ feature Zhu Lingshan @ 2022-01-25 9:17 ` Zhu Lingshan 2022-01-25 9:17 ` [PATCH V2 2/4] vDPA/ifcvf: implement device MSIX vector allocator Zhu Lingshan ` (3 subsequent siblings) 4 siblings, 0 replies; 13+ messages in thread From: Zhu Lingshan @ 2022-01-25 9:17 UTC (permalink / raw) To: mst, jasowang; +Cc: netdev, virtualization, Zhu Lingshan re-implement IO read/write helpers in the header file, so that they can be utilized among modules. Signed-off-by: Zhu Lingshan <lingshan.zhu@intel.com> --- drivers/vdpa/ifcvf/ifcvf_base.c | 36 -------------------------------- drivers/vdpa/ifcvf/ifcvf_base.h | 37 +++++++++++++++++++++++++++++++++ 2 files changed, 37 insertions(+), 36 deletions(-) diff --git a/drivers/vdpa/ifcvf/ifcvf_base.c b/drivers/vdpa/ifcvf/ifcvf_base.c index 7d41dfe48ade..397692ae671c 100644 --- a/drivers/vdpa/ifcvf/ifcvf_base.c +++ b/drivers/vdpa/ifcvf/ifcvf_base.c @@ -10,42 +10,6 @@ #include "ifcvf_base.h" -static inline u8 ifc_ioread8(u8 __iomem *addr) -{ - return ioread8(addr); -} -static inline u16 ifc_ioread16 (__le16 __iomem *addr) -{ - return ioread16(addr); -} - -static inline u32 ifc_ioread32(__le32 __iomem *addr) -{ - return ioread32(addr); -} - -static inline void ifc_iowrite8(u8 value, u8 __iomem *addr) -{ - iowrite8(value, addr); -} - -static inline void ifc_iowrite16(u16 value, __le16 __iomem *addr) -{ - iowrite16(value, addr); -} - -static inline void ifc_iowrite32(u32 value, __le32 __iomem *addr) -{ - iowrite32(value, addr); -} - -static void ifc_iowrite64_twopart(u64 val, - __le32 __iomem *lo, __le32 __iomem *hi) -{ - ifc_iowrite32((u32)val, lo); - ifc_iowrite32(val >> 32, hi); -} - struct ifcvf_adapter *vf_to_adapter(struct ifcvf_hw *hw) { return container_of(hw, struct ifcvf_adapter, vf); diff --git a/drivers/vdpa/ifcvf/ifcvf_base.h b/drivers/vdpa/ifcvf/ifcvf_base.h index c486873f370a..949b4fb9d554 100644 --- a/drivers/vdpa/ifcvf/ifcvf_base.h +++ b/drivers/vdpa/ifcvf/ifcvf_base.h @@ -42,6 +42,43 @@ #define ifcvf_private_to_vf(adapter) \ (&((struct ifcvf_adapter *)adapter)->vf) +static inline u8 ifc_ioread8(u8 __iomem *addr) +{ + return ioread8(addr); +} + +static inline u16 ifc_ioread16(__le16 __iomem *addr) +{ + return ioread16(addr); +} + +static inline u32 ifc_ioread32(__le32 __iomem *addr) +{ + return ioread32(addr); +} + +static inline void ifc_iowrite8(u8 value, u8 __iomem *addr) +{ + iowrite8(value, addr); +} + +static inline void ifc_iowrite16(u16 value, __le16 __iomem *addr) +{ + iowrite16(value, addr); +} + +static inline void ifc_iowrite32(u32 value, __le32 __iomem *addr) +{ + iowrite32(value, addr); +} + +static inline void ifc_iowrite64_twopart(u64 val, + __le32 __iomem *lo, __le32 __iomem *hi) +{ + ifc_iowrite32((u32)val, lo); + ifc_iowrite32(val >> 32, hi); +} + struct vring_info { u64 desc; u64 avail; -- 2.27.0 ^ permalink raw reply related [flat|nested] 13+ messages in thread
* [PATCH V2 2/4] vDPA/ifcvf: implement device MSIX vector allocator 2022-01-25 9:17 [PATCH V2 0/4] vDPA/ifcvf: implement shared IRQ feature Zhu Lingshan 2022-01-25 9:17 ` [PATCH V2 1/4] vDPA/ifcvf: implement IO read/write helpers in the header file Zhu Lingshan @ 2022-01-25 9:17 ` Zhu Lingshan 2022-01-25 19:36 ` Michael S. Tsirkin 2022-01-25 9:17 ` [PATCH V2 3/4] vhost_vdpa: don't setup irq offloading when irq_num < 0 Zhu Lingshan ` (2 subsequent siblings) 4 siblings, 1 reply; 13+ messages in thread From: Zhu Lingshan @ 2022-01-25 9:17 UTC (permalink / raw) To: mst, jasowang; +Cc: netdev, virtualization, Zhu Lingshan This commit implements a MSIX vector allocation helper for vqs and config interrupts. Signed-off-by: Zhu Lingshan <lingshan.zhu@intel.com> --- drivers/vdpa/ifcvf/ifcvf_main.c | 30 ++++++++++++++++++++++++++++-- 1 file changed, 28 insertions(+), 2 deletions(-) diff --git a/drivers/vdpa/ifcvf/ifcvf_main.c b/drivers/vdpa/ifcvf/ifcvf_main.c index d1a6b5ab543c..7e2af2d2aaf5 100644 --- a/drivers/vdpa/ifcvf/ifcvf_main.c +++ b/drivers/vdpa/ifcvf/ifcvf_main.c @@ -58,14 +58,40 @@ static void ifcvf_free_irq(struct ifcvf_adapter *adapter, int queues) ifcvf_free_irq_vectors(pdev); } +static int ifcvf_alloc_vectors(struct ifcvf_adapter *adapter) +{ + struct pci_dev *pdev = adapter->pdev; + struct ifcvf_hw *vf = &adapter->vf; + u16 max_intr, ret; + + /* all queues and config interrupt */ + max_intr = vf->nr_vring + 1; + ret = pci_alloc_irq_vectors(pdev, 1, max_intr, PCI_IRQ_MSIX | PCI_IRQ_AFFINITY); + + if (ret < 0) { + IFCVF_ERR(pdev, "Failed to alloc IRQ vectors\n"); + return ret; + } + + if (ret < max_intr) + IFCVF_INFO(pdev, + "Requested %u vectors, however only %u allocated, lower performance\n", + max_intr, ret); + + return ret; +} + static int ifcvf_request_irq(struct ifcvf_adapter *adapter) { struct pci_dev *pdev = adapter->pdev; struct ifcvf_hw *vf = &adapter->vf; int vector, i, ret, irq; - u16 max_intr; + u16 nvectors, max_intr; + + nvectors = ifcvf_alloc_vectors(adapter); + if (!(nvectors > 0)) + return nvectors; - /* all queues and config interrupt */ max_intr = vf->nr_vring + 1; ret = pci_alloc_irq_vectors(pdev, max_intr, -- 2.27.0 ^ permalink raw reply related [flat|nested] 13+ messages in thread
* Re: [PATCH V2 2/4] vDPA/ifcvf: implement device MSIX vector allocator 2022-01-25 9:17 ` [PATCH V2 2/4] vDPA/ifcvf: implement device MSIX vector allocator Zhu Lingshan @ 2022-01-25 19:36 ` Michael S. Tsirkin 2022-01-26 12:18 ` Zhu, Lingshan 0 siblings, 1 reply; 13+ messages in thread From: Michael S. Tsirkin @ 2022-01-25 19:36 UTC (permalink / raw) To: Zhu Lingshan; +Cc: jasowang, netdev, virtualization On Tue, Jan 25, 2022 at 05:17:42PM +0800, Zhu Lingshan wrote: > This commit implements a MSIX vector allocation helper > for vqs and config interrupts. > > Signed-off-by: Zhu Lingshan <lingshan.zhu@intel.com> > --- > drivers/vdpa/ifcvf/ifcvf_main.c | 30 ++++++++++++++++++++++++++++-- > 1 file changed, 28 insertions(+), 2 deletions(-) > > diff --git a/drivers/vdpa/ifcvf/ifcvf_main.c b/drivers/vdpa/ifcvf/ifcvf_main.c > index d1a6b5ab543c..7e2af2d2aaf5 100644 > --- a/drivers/vdpa/ifcvf/ifcvf_main.c > +++ b/drivers/vdpa/ifcvf/ifcvf_main.c > @@ -58,14 +58,40 @@ static void ifcvf_free_irq(struct ifcvf_adapter *adapter, int queues) > ifcvf_free_irq_vectors(pdev); > } > > +static int ifcvf_alloc_vectors(struct ifcvf_adapter *adapter) So the helper returns an int... > +{ > + struct pci_dev *pdev = adapter->pdev; > + struct ifcvf_hw *vf = &adapter->vf; > + u16 max_intr, ret; > + > + /* all queues and config interrupt */ > + max_intr = vf->nr_vring + 1; > + ret = pci_alloc_irq_vectors(pdev, 1, max_intr, PCI_IRQ_MSIX | PCI_IRQ_AFFINITY); > + > + if (ret < 0) { > + IFCVF_ERR(pdev, "Failed to alloc IRQ vectors\n"); > + return ret; which is negative on error... > + } > + > + if (ret < max_intr) > + IFCVF_INFO(pdev, > + "Requested %u vectors, however only %u allocated, lower performance\n", > + max_intr, ret); > + > + return ret; > +} > + > static int ifcvf_request_irq(struct ifcvf_adapter *adapter) > { > struct pci_dev *pdev = adapter->pdev; > struct ifcvf_hw *vf = &adapter->vf; > int vector, i, ret, irq; > - u16 max_intr; > + u16 nvectors, max_intr; > + > + nvectors = ifcvf_alloc_vectors(adapter); which you proceed to stash into an unsigned int ... > + if (!(nvectors > 0)) and then compare to zero ... > + return nvectors; > correct error handling is unlikely as a result. > - /* all queues and config interrupt */ > max_intr = vf->nr_vring + 1; > > ret = pci_alloc_irq_vectors(pdev, max_intr, As long as you are introducing a helper, document it's return type and behaviour and then use correctly. > -- > 2.27.0 ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: [PATCH V2 2/4] vDPA/ifcvf: implement device MSIX vector allocator 2022-01-25 19:36 ` Michael S. Tsirkin @ 2022-01-26 12:18 ` Zhu, Lingshan 0 siblings, 0 replies; 13+ messages in thread From: Zhu, Lingshan @ 2022-01-26 12:18 UTC (permalink / raw) To: Michael S. Tsirkin; +Cc: jasowang, netdev, virtualization On 1/26/2022 3:36 AM, Michael S. Tsirkin wrote: > On Tue, Jan 25, 2022 at 05:17:42PM +0800, Zhu Lingshan wrote: >> This commit implements a MSIX vector allocation helper >> for vqs and config interrupts. >> >> Signed-off-by: Zhu Lingshan <lingshan.zhu@intel.com> >> --- >> drivers/vdpa/ifcvf/ifcvf_main.c | 30 ++++++++++++++++++++++++++++-- >> 1 file changed, 28 insertions(+), 2 deletions(-) >> >> diff --git a/drivers/vdpa/ifcvf/ifcvf_main.c b/drivers/vdpa/ifcvf/ifcvf_main.c >> index d1a6b5ab543c..7e2af2d2aaf5 100644 >> --- a/drivers/vdpa/ifcvf/ifcvf_main.c >> +++ b/drivers/vdpa/ifcvf/ifcvf_main.c >> @@ -58,14 +58,40 @@ static void ifcvf_free_irq(struct ifcvf_adapter *adapter, int queues) >> ifcvf_free_irq_vectors(pdev); >> } >> >> +static int ifcvf_alloc_vectors(struct ifcvf_adapter *adapter) > > So the helper returns an int... can fix in V3, thanks for your help! > >> +{ >> + struct pci_dev *pdev = adapter->pdev; >> + struct ifcvf_hw *vf = &adapter->vf; >> + u16 max_intr, ret; >> + >> + /* all queues and config interrupt */ >> + max_intr = vf->nr_vring + 1; >> + ret = pci_alloc_irq_vectors(pdev, 1, max_intr, PCI_IRQ_MSIX | PCI_IRQ_AFFINITY); >> + >> + if (ret < 0) { >> + IFCVF_ERR(pdev, "Failed to alloc IRQ vectors\n"); >> + return ret; > > which is negative on error... > >> + } >> + >> + if (ret < max_intr) >> + IFCVF_INFO(pdev, >> + "Requested %u vectors, however only %u allocated, lower performance\n", >> + max_intr, ret); >> + >> + return ret; >> +} >> + >> static int ifcvf_request_irq(struct ifcvf_adapter *adapter) >> { >> struct pci_dev *pdev = adapter->pdev; >> struct ifcvf_hw *vf = &adapter->vf; >> int vector, i, ret, irq; >> - u16 max_intr; >> + u16 nvectors, max_intr; >> + >> + nvectors = ifcvf_alloc_vectors(adapter); > which you proceed to stash into an unsigned int ... > >> + if (!(nvectors > 0)) > and then compare to zero ... > >> + return nvectors; >> > correct error handling is unlikely as a result. > > >> - /* all queues and config interrupt */ >> max_intr = vf->nr_vring + 1; >> >> ret = pci_alloc_irq_vectors(pdev, max_intr, > > As long as you are introducing a helper, document it's return > type and behaviour and then use correctly. > >> -- >> 2.27.0 ^ permalink raw reply [flat|nested] 13+ messages in thread
* [PATCH V2 3/4] vhost_vdpa: don't setup irq offloading when irq_num < 0 2022-01-25 9:17 [PATCH V2 0/4] vDPA/ifcvf: implement shared IRQ feature Zhu Lingshan 2022-01-25 9:17 ` [PATCH V2 1/4] vDPA/ifcvf: implement IO read/write helpers in the header file Zhu Lingshan 2022-01-25 9:17 ` [PATCH V2 2/4] vDPA/ifcvf: implement device MSIX vector allocator Zhu Lingshan @ 2022-01-25 9:17 ` Zhu Lingshan 2022-01-25 19:17 ` kernel test robot 2022-01-25 19:30 ` Michael S. Tsirkin 2022-01-25 9:17 ` [PATCH V2 4/4] vDPA/ifcvf: implement shared IRQ feature Zhu Lingshan 2022-01-25 19:32 ` [PATCH V2 0/4] " Michael S. Tsirkin 4 siblings, 2 replies; 13+ messages in thread From: Zhu Lingshan @ 2022-01-25 9:17 UTC (permalink / raw) To: mst, jasowang; +Cc: netdev, virtualization, Zhu Lingshan When irq number is negative(e.g., -EINVAL), the virtqueue may be disabled or the virtqueues are sharing a device irq. In such case, we should not setup irq offloading for a virtqueue. Signed-off-by: Zhu Lingshan <lingshan.zhu@intel.com> --- drivers/vhost/vdpa.c | 3 +++ 1 file changed, 3 insertions(+) diff --git a/drivers/vhost/vdpa.c b/drivers/vhost/vdpa.c index 851539807bc9..909891d518e8 100644 --- a/drivers/vhost/vdpa.c +++ b/drivers/vhost/vdpa.c @@ -96,6 +96,9 @@ static void vhost_vdpa_setup_vq_irq(struct vhost_vdpa *v, u16 qid) if (!ops->get_vq_irq) return; + if (irq < 0) + return; + irq = ops->get_vq_irq(vdpa, qid); irq_bypass_unregister_producer(&vq->call_ctx.producer); if (!vq->call_ctx.ctx || irq < 0) -- 2.27.0 ^ permalink raw reply related [flat|nested] 13+ messages in thread
* Re: [PATCH V2 3/4] vhost_vdpa: don't setup irq offloading when irq_num < 0 2022-01-25 9:17 ` [PATCH V2 3/4] vhost_vdpa: don't setup irq offloading when irq_num < 0 Zhu Lingshan @ 2022-01-25 19:17 ` kernel test robot 2022-01-25 19:30 ` Michael S. Tsirkin 1 sibling, 0 replies; 13+ messages in thread From: kernel test robot @ 2022-01-25 19:17 UTC (permalink / raw) To: Zhu Lingshan, mst, jasowang Cc: llvm, kbuild-all, netdev, virtualization, Zhu Lingshan Hi Zhu, Thank you for the patch! Perhaps something to improve: [auto build test WARNING on mst-vhost/linux-next] [also build test WARNING on horms-ipvs/master linus/master v5.17-rc1 next-20220125] [If your patch is applied to the wrong git tree, kindly drop us a note. And when submitting patch, we suggest to use '--base' as documented in https://git-scm.com/docs/git-format-patch] url: https://github.com/0day-ci/linux/commits/Zhu-Lingshan/vDPA-ifcvf-implement-shared-IRQ-feature/20220125-174020 base: https://git.kernel.org/pub/scm/linux/kernel/git/mst/vhost.git linux-next config: arm-randconfig-c002-20220124 (https://download.01.org/0day-ci/archive/20220126/202201260245.1yTB6YwE-lkp@intel.com/config) compiler: clang version 14.0.0 (https://github.com/llvm/llvm-project 997e128e2a78f5a5434fc75997441ae1ee76f8a4) reproduce (this is a W=1 build): wget https://raw.githubusercontent.com/intel/lkp-tests/master/sbin/make.cross -O ~/bin/make.cross chmod +x ~/bin/make.cross # install arm cross compiling tool for clang build # apt-get install binutils-arm-linux-gnueabi # https://github.com/0day-ci/linux/commit/9242eae873643db8562d24857da7d05a2950ecfe git remote add linux-review https://github.com/0day-ci/linux git fetch --no-tags linux-review Zhu-Lingshan/vDPA-ifcvf-implement-shared-IRQ-feature/20220125-174020 git checkout 9242eae873643db8562d24857da7d05a2950ecfe # save the config file to linux build tree mkdir build_dir COMPILER_INSTALL_PATH=$HOME/0day COMPILER=clang make.cross W=1 O=build_dir ARCH=arm SHELL=/bin/bash drivers/vhost/ If you fix the issue, kindly add following tag as appropriate Reported-by: kernel test robot <lkp@intel.com> All warnings (new ones prefixed by >>): >> drivers/vhost/vdpa.c:99:6: warning: variable 'irq' is uninitialized when used here [-Wuninitialized] if (irq < 0) ^~~ drivers/vhost/vdpa.c:94:14: note: initialize the variable 'irq' to silence this warning int ret, irq; ^ = 0 1 warning generated. vim +/irq +99 drivers/vhost/vdpa.c 88 89 static void vhost_vdpa_setup_vq_irq(struct vhost_vdpa *v, u16 qid) 90 { 91 struct vhost_virtqueue *vq = &v->vqs[qid]; 92 const struct vdpa_config_ops *ops = v->vdpa->config; 93 struct vdpa_device *vdpa = v->vdpa; 94 int ret, irq; 95 96 if (!ops->get_vq_irq) 97 return; 98 > 99 if (irq < 0) 100 return; 101 102 irq = ops->get_vq_irq(vdpa, qid); 103 irq_bypass_unregister_producer(&vq->call_ctx.producer); 104 if (!vq->call_ctx.ctx || irq < 0) 105 return; 106 107 vq->call_ctx.producer.token = vq->call_ctx.ctx; 108 vq->call_ctx.producer.irq = irq; 109 ret = irq_bypass_register_producer(&vq->call_ctx.producer); 110 if (unlikely(ret)) 111 dev_info(&v->dev, "vq %u, irq bypass producer (token %p) registration fails, ret = %d\n", 112 qid, vq->call_ctx.producer.token, ret); 113 } 114 --- 0-DAY CI Kernel Test Service, Intel Corporation https://lists.01.org/hyperkitty/list/kbuild-all@lists.01.org ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: [PATCH V2 3/4] vhost_vdpa: don't setup irq offloading when irq_num < 0 2022-01-25 9:17 ` [PATCH V2 3/4] vhost_vdpa: don't setup irq offloading when irq_num < 0 Zhu Lingshan 2022-01-25 19:17 ` kernel test robot @ 2022-01-25 19:30 ` Michael S. Tsirkin 2022-01-26 12:18 ` Zhu, Lingshan 1 sibling, 1 reply; 13+ messages in thread From: Michael S. Tsirkin @ 2022-01-25 19:30 UTC (permalink / raw) To: Zhu Lingshan; +Cc: jasowang, netdev, virtualization On Tue, Jan 25, 2022 at 05:17:43PM +0800, Zhu Lingshan wrote: > When irq number is negative(e.g., -EINVAL), the virtqueue > may be disabled or the virtqueues are sharing a device irq. > In such case, we should not setup irq offloading for a virtqueue. > > Signed-off-by: Zhu Lingshan <lingshan.zhu@intel.com> > --- > drivers/vhost/vdpa.c | 3 +++ > 1 file changed, 3 insertions(+) > > diff --git a/drivers/vhost/vdpa.c b/drivers/vhost/vdpa.c > index 851539807bc9..909891d518e8 100644 > --- a/drivers/vhost/vdpa.c > +++ b/drivers/vhost/vdpa.c > @@ -96,6 +96,9 @@ static void vhost_vdpa_setup_vq_irq(struct vhost_vdpa *v, u16 qid) > if (!ops->get_vq_irq) > return; > > + if (irq < 0) > + return; > + > irq = ops->get_vq_irq(vdpa, qid); So it's used before it's initialized. Ugh. How was this patchset tested? > irq_bypass_unregister_producer(&vq->call_ctx.producer); > if (!vq->call_ctx.ctx || irq < 0) > -- > 2.27.0 ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: [PATCH V2 3/4] vhost_vdpa: don't setup irq offloading when irq_num < 0 2022-01-25 19:30 ` Michael S. Tsirkin @ 2022-01-26 12:18 ` Zhu, Lingshan 0 siblings, 0 replies; 13+ messages in thread From: Zhu, Lingshan @ 2022-01-26 12:18 UTC (permalink / raw) To: Michael S. Tsirkin; +Cc: jasowang, netdev, virtualization On 1/26/2022 3:30 AM, Michael S. Tsirkin wrote: > On Tue, Jan 25, 2022 at 05:17:43PM +0800, Zhu Lingshan wrote: >> When irq number is negative(e.g., -EINVAL), the virtqueue >> may be disabled or the virtqueues are sharing a device irq. >> In such case, we should not setup irq offloading for a virtqueue. >> >> Signed-off-by: Zhu Lingshan <lingshan.zhu@intel.com> >> --- >> drivers/vhost/vdpa.c | 3 +++ >> 1 file changed, 3 insertions(+) >> >> diff --git a/drivers/vhost/vdpa.c b/drivers/vhost/vdpa.c >> index 851539807bc9..909891d518e8 100644 >> --- a/drivers/vhost/vdpa.c >> +++ b/drivers/vhost/vdpa.c >> @@ -96,6 +96,9 @@ static void vhost_vdpa_setup_vq_irq(struct vhost_vdpa *v, u16 qid) >> if (!ops->get_vq_irq) >> return; >> >> + if (irq < 0) >> + return; >> + >> irq = ops->get_vq_irq(vdpa, qid); > So it's used before it's initialized. Ugh. > How was this patchset tested? Sorry, my bad, it is not rebased properly, V3 can fix this for sure. Thanks > >> irq_bypass_unregister_producer(&vq->call_ctx.producer); >> if (!vq->call_ctx.ctx || irq < 0) >> -- >> 2.27.0 ^ permalink raw reply [flat|nested] 13+ messages in thread
* [PATCH V2 4/4] vDPA/ifcvf: implement shared IRQ feature 2022-01-25 9:17 [PATCH V2 0/4] vDPA/ifcvf: implement shared IRQ feature Zhu Lingshan ` (2 preceding siblings ...) 2022-01-25 9:17 ` [PATCH V2 3/4] vhost_vdpa: don't setup irq offloading when irq_num < 0 Zhu Lingshan @ 2022-01-25 9:17 ` Zhu Lingshan 2022-01-25 19:42 ` Michael S. Tsirkin 2022-01-25 19:32 ` [PATCH V2 0/4] " Michael S. Tsirkin 4 siblings, 1 reply; 13+ messages in thread From: Zhu Lingshan @ 2022-01-25 9:17 UTC (permalink / raw) To: mst, jasowang; +Cc: netdev, virtualization, Zhu Lingshan On some platforms/devices, there may not be enough MSI vector slots allocated for virtqueues and config changes. In such a case, the interrupt sources(virtqueues, config changes) must share an IRQ/vector, to avoid initialization failures, keep the device functional. This commit handles three cases: (1) number of the allocated vectors == the number of virtqueues + 1 (config changes), every virtqueue and the config interrupt has a separated vector/IRQ, the best and the most likely case. (2) number of the allocated vectors is less than the best case, but greater than 1. In this case, all virtqueues share a vector/IRQ, the config interrupt has a separated vector/IRQ (3) only one vector is allocated, in this case, the virtqueues and the config interrupt share a vector/IRQ. The worst and most unlikely case. Otherwise, it needs to fail. This commit introduces some helper functions: ifcvf_set_vq_vector() and ifcvf_set_config_vector() sets virtqueue vector and config vector in the device config space, so that the device can send interrupt DMA. This commit adds some fields in struct ifcvf_hw and re-placed the existed fields to be aligned with the cacheline. Signed-off-by: Zhu Lingshan <lingshan.zhu@intel.com> --- drivers/vdpa/ifcvf/ifcvf_base.c | 47 +++++-- drivers/vdpa/ifcvf/ifcvf_base.h | 23 ++- drivers/vdpa/ifcvf/ifcvf_main.c | 240 +++++++++++++++++++++++++++----- 3 files changed, 253 insertions(+), 57 deletions(-) diff --git a/drivers/vdpa/ifcvf/ifcvf_base.c b/drivers/vdpa/ifcvf/ifcvf_base.c index 397692ae671c..18dcb63ab1e3 100644 --- a/drivers/vdpa/ifcvf/ifcvf_base.c +++ b/drivers/vdpa/ifcvf/ifcvf_base.c @@ -15,6 +15,36 @@ struct ifcvf_adapter *vf_to_adapter(struct ifcvf_hw *hw) return container_of(hw, struct ifcvf_adapter, vf); } +int ifcvf_set_vq_vector(struct ifcvf_hw *hw, u16 qid, int vector) +{ + struct virtio_pci_common_cfg __iomem *cfg = hw->common_cfg; + struct ifcvf_adapter *ifcvf = vf_to_adapter(hw); + + ifc_iowrite16(qid, &cfg->queue_select); + ifc_iowrite16(vector, &cfg->queue_msix_vector); + if (ifc_ioread16(&cfg->queue_msix_vector) == VIRTIO_MSI_NO_VECTOR) { + IFCVF_ERR(ifcvf->pdev, "No msix vector for queue %u\n", qid); + return -EINVAL; + } + + return 0; +} + +int ifcvf_set_config_vector(struct ifcvf_hw *hw, int vector) +{ + struct virtio_pci_common_cfg __iomem *cfg = hw->common_cfg; + struct ifcvf_adapter *ifcvf = vf_to_adapter(hw); + + cfg = hw->common_cfg; + ifc_iowrite16(vector, &cfg->msix_config); + if (ifc_ioread16(&cfg->msix_config) == VIRTIO_MSI_NO_VECTOR) { + IFCVF_ERR(ifcvf->pdev, "No msix vector for device config\n"); + return -EINVAL; + } + + return 0; +} + static void __iomem *get_cap_addr(struct ifcvf_hw *hw, struct virtio_pci_cap *cap) { @@ -140,6 +170,8 @@ int ifcvf_init_hw(struct ifcvf_hw *hw, struct pci_dev *pdev) hw->common_cfg, hw->notify_base, hw->isr, hw->dev_cfg, hw->notify_off_multiplier); + hw->vqs_shared_irq = -EINVAL; + return 0; } @@ -321,12 +353,6 @@ static int ifcvf_hw_enable(struct ifcvf_hw *hw) ifcvf = vf_to_adapter(hw); cfg = hw->common_cfg; - ifc_iowrite16(IFCVF_MSI_CONFIG_OFF, &cfg->msix_config); - - if (ifc_ioread16(&cfg->msix_config) == VIRTIO_MSI_NO_VECTOR) { - IFCVF_ERR(ifcvf->pdev, "No msix vector for device config\n"); - return -EINVAL; - } for (i = 0; i < hw->nr_vring; i++) { if (!hw->vring[i].ready) @@ -340,15 +366,6 @@ static int ifcvf_hw_enable(struct ifcvf_hw *hw) ifc_iowrite64_twopart(hw->vring[i].used, &cfg->queue_used_lo, &cfg->queue_used_hi); ifc_iowrite16(hw->vring[i].size, &cfg->queue_size); - ifc_iowrite16(i + IFCVF_MSI_QUEUE_OFF, &cfg->queue_msix_vector); - - if (ifc_ioread16(&cfg->queue_msix_vector) == - VIRTIO_MSI_NO_VECTOR) { - IFCVF_ERR(ifcvf->pdev, - "No msix vector for queue %u\n", i); - return -EINVAL; - } - ifcvf_set_vq_state(hw, i, hw->vring[i].last_avail_idx); ifc_iowrite16(1, &cfg->queue_enable); } diff --git a/drivers/vdpa/ifcvf/ifcvf_base.h b/drivers/vdpa/ifcvf/ifcvf_base.h index 949b4fb9d554..d2a2a526f0fc 100644 --- a/drivers/vdpa/ifcvf/ifcvf_base.h +++ b/drivers/vdpa/ifcvf/ifcvf_base.h @@ -27,8 +27,6 @@ #define IFCVF_QUEUE_ALIGNMENT PAGE_SIZE #define IFCVF_QUEUE_MAX 32768 -#define IFCVF_MSI_CONFIG_OFF 0 -#define IFCVF_MSI_QUEUE_OFF 1 #define IFCVF_PCI_MAX_RESOURCE 6 #define IFCVF_LM_CFG_SIZE 0x40 @@ -42,6 +40,13 @@ #define ifcvf_private_to_vf(adapter) \ (&((struct ifcvf_adapter *)adapter)->vf) +/* all vqs and config interrupt has its own vector */ +#define MSIX_VECTOR_PER_VQ_AND_CONFIG 1 +/* all vqs share a vector, and config interrupt has a separate vector */ +#define MSIX_VECTOR_SHARED_VQ_AND_CONFIG 2 +/* all vqs and config interrupt share a vector */ +#define MSIX_VECTOR_DEV_SHARED 3 + static inline u8 ifc_ioread8(u8 __iomem *addr) { return ioread8(addr); @@ -97,25 +102,27 @@ struct ifcvf_hw { u8 __iomem *isr; /* Live migration */ u8 __iomem *lm_cfg; - u16 nr_vring; /* Notification bar number */ u8 notify_bar; + u8 msix_vector_status; + /* virtio-net or virtio-blk device config size */ + u32 config_size; /* Notificaiton bar address */ void __iomem *notify_base; phys_addr_t notify_base_pa; u32 notify_off_multiplier; + u32 dev_type; u64 req_features; u64 hw_features; - u32 dev_type; struct virtio_pci_common_cfg __iomem *common_cfg; void __iomem *dev_cfg; struct vring_info vring[IFCVF_MAX_QUEUES]; void __iomem * const *base; char config_msix_name[256]; struct vdpa_callback config_cb; - unsigned int config_irq; - /* virtio-net or virtio-blk device config size */ - u32 config_size; + u32 config_irq; + u32 vqs_shared_irq; + u16 nr_vring; }; struct ifcvf_adapter { @@ -160,4 +167,6 @@ int ifcvf_set_vq_state(struct ifcvf_hw *hw, u16 qid, u16 num); struct ifcvf_adapter *vf_to_adapter(struct ifcvf_hw *hw); int ifcvf_probed_virtio_net(struct ifcvf_hw *hw); u32 ifcvf_get_config_size(struct ifcvf_hw *hw); +int ifcvf_set_vq_vector(struct ifcvf_hw *hw, u16 qid, int vector); +int ifcvf_set_config_vector(struct ifcvf_hw *hw, int vector); #endif /* _IFCVF_H_ */ diff --git a/drivers/vdpa/ifcvf/ifcvf_main.c b/drivers/vdpa/ifcvf/ifcvf_main.c index 7e2af2d2aaf5..c7070ff01776 100644 --- a/drivers/vdpa/ifcvf/ifcvf_main.c +++ b/drivers/vdpa/ifcvf/ifcvf_main.c @@ -17,6 +17,7 @@ #define DRIVER_AUTHOR "Intel Corporation" #define IFCVF_DRIVER_NAME "ifcvf" +/* handles config interrupt */ static irqreturn_t ifcvf_config_changed(int irq, void *arg) { struct ifcvf_hw *vf = arg; @@ -27,6 +28,7 @@ static irqreturn_t ifcvf_config_changed(int irq, void *arg) return IRQ_HANDLED; } +/* handles vqs interrupt */ static irqreturn_t ifcvf_intr_handler(int irq, void *arg) { struct vring_info *vring = arg; @@ -37,24 +39,77 @@ static irqreturn_t ifcvf_intr_handler(int irq, void *arg) return IRQ_HANDLED; } +/* handls vqs shared interrupt */ +static irqreturn_t ifcvf_vq_shared_intr_handler(int irq, void *arg) +{ + struct ifcvf_hw *vf = arg; + struct vring_info *vring; + int i; + + for (i = 0; i < vf->nr_vring; i++) { + vring = &vf->vring[i]; + if (vring->cb.callback) + vf->vring->cb.callback(vring->cb.private); + } + + return IRQ_HANDLED; +} + +/* handles a shared interrupt for vqs and config */ +static irqreturn_t ifcvf_dev_shared_intr_handler(int irq, void *arg) +{ + struct ifcvf_hw *vf = arg; + u8 isr; + + isr = ifc_ioread8(vf->isr); + if (isr & VIRTIO_PCI_ISR_CONFIG) + ifcvf_config_changed(irq, arg); + + return ifcvf_vq_shared_intr_handler(irq, arg); +} + static void ifcvf_free_irq_vectors(void *data) { pci_free_irq_vectors(data); } -static void ifcvf_free_irq(struct ifcvf_adapter *adapter, int queues) +static void ifcvf_free_vq_irq(struct ifcvf_adapter *adapter, int queues) { struct pci_dev *pdev = adapter->pdev; struct ifcvf_hw *vf = &adapter->vf; int i; + if (vf->msix_vector_status == MSIX_VECTOR_PER_VQ_AND_CONFIG) + for (i = 0; i < queues; i++) { + devm_free_irq(&pdev->dev, vf->vring[i].irq, &vf->vring[i]); + vf->vring[i].irq = -EINVAL; + } + else { + devm_free_irq(&pdev->dev, vf->vqs_shared_irq, vf); + } vf->vqs_shared_irq = -EINVAL; +} - for (i = 0; i < queues; i++) { - devm_free_irq(&pdev->dev, vf->vring[i].irq, &vf->vring[i]); - vf->vring[i].irq = -EINVAL; +static void ifcvf_free_config_irq(struct ifcvf_adapter *adapter) +{ + struct pci_dev *pdev = adapter->pdev; + struct ifcvf_hw *vf = &adapter->vf; + + /* If the irq is shared by all vqs and the config interrupt, + * it is already freed in ifcvf_free_vq_irq, so here only + * need to free config irq when msix_vector_status != MSIX_VECTOR_DEV_SHARED + */ + if (vf->msix_vector_status != MSIX_VECTOR_DEV_SHARED) { + devm_free_irq(&pdev->dev, vf->config_irq, vf); + vf->config_irq = -EINVAL; } +} + +static void ifcvf_free_irq(struct ifcvf_adapter *adapter, int queues) +{ + struct pci_dev *pdev = adapter->pdev; - devm_free_irq(&pdev->dev, vf->config_irq, vf); + ifcvf_free_vq_irq(adapter, queues); + ifcvf_free_config_irq(adapter); ifcvf_free_irq_vectors(pdev); } @@ -81,58 +136,170 @@ static int ifcvf_alloc_vectors(struct ifcvf_adapter *adapter) return ret; } -static int ifcvf_request_irq(struct ifcvf_adapter *adapter) +static int ifcvf_request_per_vq_irq(struct ifcvf_adapter *adapter) { struct pci_dev *pdev = adapter->pdev; struct ifcvf_hw *vf = &adapter->vf; - int vector, i, ret, irq; - u16 nvectors, max_intr; + int i, vector, ret, irq; - nvectors = ifcvf_alloc_vectors(adapter); - if (!(nvectors > 0)) - return nvectors; + for (i = 0; i < vf->nr_vring; i++) { + snprintf(vf->vring[i].msix_name, 256, "ifcvf[%s]-%d\n", pci_name(pdev), i); + vector = i; + irq = pci_irq_vector(pdev, vector); + ret = devm_request_irq(&pdev->dev, irq, + ifcvf_intr_handler, 0, + vf->vring[i].msix_name, + &vf->vring[i]); + if (ret) { + IFCVF_ERR(pdev, "Failed to request irq for vq %d\n", i); + ifcvf_free_vq_irq(adapter, i); + } else { + vf->vring[i].irq = irq; + ifcvf_set_vq_vector(vf, i, vector); + } + } - max_intr = vf->nr_vring + 1; + vf->vqs_shared_irq = -EINVAL; + + return 0; +} + +static int ifcvf_request_shared_vq_irq(struct ifcvf_adapter *adapter) +{ + struct pci_dev *pdev = adapter->pdev; + struct ifcvf_hw *vf = &adapter->vf; + int i, vector, ret, irq; + + vector = 0; + irq = pci_irq_vector(pdev, vector); + ret = devm_request_irq(&pdev->dev, irq, + ifcvf_vq_shared_intr_handler, 0, + "ifcvf_vq_shared_irq", + vf); + if (ret) { + IFCVF_ERR(pdev, "Failed to request shared irq for vf\n"); + + return ret; + } + + vf->vqs_shared_irq = irq; + for (i = 0; i < vf->nr_vring; i++) { + vf->vring[i].irq = -EINVAL; + ifcvf_set_vq_vector(vf, i, vector); + } + + return 0; + +} + +static int ifcvf_request_dev_shared_irq(struct ifcvf_adapter *adapter) +{ + struct pci_dev *pdev = adapter->pdev; + struct ifcvf_hw *vf = &adapter->vf; + int i, vector, ret, irq; + + vector = 0; + irq = pci_irq_vector(pdev, vector); + ret = devm_request_irq(&pdev->dev, irq, + ifcvf_dev_shared_intr_handler, 0, + "ifcvf_dev_shared_irq", + vf); + if (ret) { + IFCVF_ERR(pdev, "Failed to request shared irq for vf\n"); - ret = pci_alloc_irq_vectors(pdev, max_intr, - max_intr, PCI_IRQ_MSIX); - if (ret < 0) { - IFCVF_ERR(pdev, "Failed to alloc IRQ vectors\n"); return ret; } + vf->vqs_shared_irq = irq; + for (i = 0; i < vf->nr_vring; i++) { + vf->vring[i].irq = -EINVAL; + ifcvf_set_vq_vector(vf, i, vector); + } + + vf->config_irq = irq; + ifcvf_set_config_vector(vf, vector); + + return 0; + +} + +static int ifcvf_request_vq_irq(struct ifcvf_adapter *adapter) +{ + struct ifcvf_hw *vf = &adapter->vf; + int ret; + + if (vf->msix_vector_status == MSIX_VECTOR_PER_VQ_AND_CONFIG) + ret = ifcvf_request_per_vq_irq(adapter); + else + ret = ifcvf_request_shared_vq_irq(adapter); + + return ret; +} + +static int ifcvf_request_config_irq(struct ifcvf_adapter *adapter) +{ + struct pci_dev *pdev = adapter->pdev; + struct ifcvf_hw *vf = &adapter->vf; + int config_vector, ret; + + if (vf->msix_vector_status == MSIX_VECTOR_DEV_SHARED) + return 0; + + if (vf->msix_vector_status == MSIX_VECTOR_PER_VQ_AND_CONFIG) + /* vector 0 ~ vf->nr_vring for vqs, num vf->nr_vring vector for config interrupt */ + config_vector = vf->nr_vring; + + if (vf->msix_vector_status == MSIX_VECTOR_SHARED_VQ_AND_CONFIG) + /* vector 0 for vqs and 1 for config interrupt */ + config_vector = 1; + snprintf(vf->config_msix_name, 256, "ifcvf[%s]-config\n", pci_name(pdev)); - vector = 0; - vf->config_irq = pci_irq_vector(pdev, vector); + vf->config_irq = pci_irq_vector(pdev, config_vector); ret = devm_request_irq(&pdev->dev, vf->config_irq, ifcvf_config_changed, 0, vf->config_msix_name, vf); if (ret) { IFCVF_ERR(pdev, "Failed to request config irq\n"); + ifcvf_free_vq_irq(adapter, vf->nr_vring); return ret; } + ifcvf_set_config_vector(vf, config_vector); - for (i = 0; i < vf->nr_vring; i++) { - snprintf(vf->vring[i].msix_name, 256, "ifcvf[%s]-%d\n", - pci_name(pdev), i); - vector = i + IFCVF_MSI_QUEUE_OFF; - irq = pci_irq_vector(pdev, vector); - ret = devm_request_irq(&pdev->dev, irq, - ifcvf_intr_handler, 0, - vf->vring[i].msix_name, - &vf->vring[i]); - if (ret) { - IFCVF_ERR(pdev, - "Failed to request irq for vq %d\n", i); - ifcvf_free_irq(adapter, i); + return 0; +} - return ret; - } +static int ifcvf_request_irq(struct ifcvf_adapter *adapter) +{ + struct ifcvf_hw *vf = &adapter->vf; + u16 nvectors, max_intr; + int ret; - vf->vring[i].irq = irq; + nvectors = ifcvf_alloc_vectors(adapter); + if (!(nvectors > 0)) + return nvectors; + + vf->msix_vector_status = MSIX_VECTOR_PER_VQ_AND_CONFIG; + max_intr = vf->nr_vring + 1; + if (nvectors < max_intr) + vf->msix_vector_status = MSIX_VECTOR_SHARED_VQ_AND_CONFIG; + + if (nvectors == 1) { + vf->msix_vector_status = MSIX_VECTOR_DEV_SHARED; + ret = ifcvf_request_dev_shared_irq(adapter); + + return ret; } + ret = ifcvf_request_vq_irq(adapter); + if (ret) + return ret; + + ret = ifcvf_request_config_irq(adapter); + + if (ret) + return ret; + return 0; } @@ -436,7 +603,10 @@ static int ifcvf_vdpa_get_vq_irq(struct vdpa_device *vdpa_dev, { struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev); - return vf->vring[qid].irq; + if (vf->vqs_shared_irq < 0) + return vf->vring[qid].irq; + else + return -EINVAL; } static struct vdpa_notification_area ifcvf_get_vq_notification(struct vdpa_device *vdpa_dev, -- 2.27.0 ^ permalink raw reply related [flat|nested] 13+ messages in thread
* Re: [PATCH V2 4/4] vDPA/ifcvf: implement shared IRQ feature 2022-01-25 9:17 ` [PATCH V2 4/4] vDPA/ifcvf: implement shared IRQ feature Zhu Lingshan @ 2022-01-25 19:42 ` Michael S. Tsirkin 2022-01-26 12:22 ` Zhu, Lingshan 0 siblings, 1 reply; 13+ messages in thread From: Michael S. Tsirkin @ 2022-01-25 19:42 UTC (permalink / raw) To: Zhu Lingshan; +Cc: jasowang, netdev, virtualization On Tue, Jan 25, 2022 at 05:17:44PM +0800, Zhu Lingshan wrote: > On some platforms/devices, there may not be enough MSI vector > slots allocated for virtqueues and config changes. In such a case, > the interrupt sources(virtqueues, config changes) must share > an IRQ/vector, to avoid initialization failures, keep > the device functional. > > This commit handles three cases: > (1) number of the allocated vectors == the number of virtqueues + 1 > (config changes), every virtqueue and the config interrupt has > a separated vector/IRQ, the best and the most likely case. > (2) number of the allocated vectors is less than the best case, but > greater than 1. In this case, all virtqueues share a vector/IRQ, > the config interrupt has a separated vector/IRQ > (3) only one vector is allocated, in this case, the virtqueues and > the config interrupt share a vector/IRQ. The worst and most > unlikely case. > > Otherwise, it needs to fail. > > This commit introduces some helper functions: > ifcvf_set_vq_vector() and ifcvf_set_config_vector() sets virtqueue > vector and config vector in the device config space, so that > the device can send interrupt DMA. > > This commit adds some fields in struct ifcvf_hw and re-placed > the existed fields to be aligned with the cacheline. > > Signed-off-by: Zhu Lingshan <lingshan.zhu@intel.com> > --- > drivers/vdpa/ifcvf/ifcvf_base.c | 47 +++++-- > drivers/vdpa/ifcvf/ifcvf_base.h | 23 ++- > drivers/vdpa/ifcvf/ifcvf_main.c | 240 +++++++++++++++++++++++++++----- > 3 files changed, 253 insertions(+), 57 deletions(-) > > diff --git a/drivers/vdpa/ifcvf/ifcvf_base.c b/drivers/vdpa/ifcvf/ifcvf_base.c > index 397692ae671c..18dcb63ab1e3 100644 > --- a/drivers/vdpa/ifcvf/ifcvf_base.c > +++ b/drivers/vdpa/ifcvf/ifcvf_base.c > @@ -15,6 +15,36 @@ struct ifcvf_adapter *vf_to_adapter(struct ifcvf_hw *hw) > return container_of(hw, struct ifcvf_adapter, vf); > } > > +int ifcvf_set_vq_vector(struct ifcvf_hw *hw, u16 qid, int vector) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = hw->common_cfg; > + struct ifcvf_adapter *ifcvf = vf_to_adapter(hw); > + > + ifc_iowrite16(qid, &cfg->queue_select); > + ifc_iowrite16(vector, &cfg->queue_msix_vector); > + if (ifc_ioread16(&cfg->queue_msix_vector) == VIRTIO_MSI_NO_VECTOR) { > + IFCVF_ERR(ifcvf->pdev, "No msix vector for queue %u\n", qid); > + return -EINVAL; > + } > + > + return 0; > +} > + > +int ifcvf_set_config_vector(struct ifcvf_hw *hw, int vector) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = hw->common_cfg; > + struct ifcvf_adapter *ifcvf = vf_to_adapter(hw); > + > + cfg = hw->common_cfg; > + ifc_iowrite16(vector, &cfg->msix_config); > + if (ifc_ioread16(&cfg->msix_config) == VIRTIO_MSI_NO_VECTOR) { > + IFCVF_ERR(ifcvf->pdev, "No msix vector for device config\n"); > + return -EINVAL; > + } > + > + return 0; > +} > + > static void __iomem *get_cap_addr(struct ifcvf_hw *hw, > struct virtio_pci_cap *cap) > { > @@ -140,6 +170,8 @@ int ifcvf_init_hw(struct ifcvf_hw *hw, struct pci_dev *pdev) > hw->common_cfg, hw->notify_base, hw->isr, > hw->dev_cfg, hw->notify_off_multiplier); > > + hw->vqs_shared_irq = -EINVAL; > + This stuffing of errno values into u32 is not a good idea. > return 0; > } > > @@ -321,12 +353,6 @@ static int ifcvf_hw_enable(struct ifcvf_hw *hw) > > ifcvf = vf_to_adapter(hw); > cfg = hw->common_cfg; > - ifc_iowrite16(IFCVF_MSI_CONFIG_OFF, &cfg->msix_config); > - > - if (ifc_ioread16(&cfg->msix_config) == VIRTIO_MSI_NO_VECTOR) { > - IFCVF_ERR(ifcvf->pdev, "No msix vector for device config\n"); > - return -EINVAL; > - } > > for (i = 0; i < hw->nr_vring; i++) { > if (!hw->vring[i].ready) > @@ -340,15 +366,6 @@ static int ifcvf_hw_enable(struct ifcvf_hw *hw) > ifc_iowrite64_twopart(hw->vring[i].used, &cfg->queue_used_lo, > &cfg->queue_used_hi); > ifc_iowrite16(hw->vring[i].size, &cfg->queue_size); > - ifc_iowrite16(i + IFCVF_MSI_QUEUE_OFF, &cfg->queue_msix_vector); > - > - if (ifc_ioread16(&cfg->queue_msix_vector) == > - VIRTIO_MSI_NO_VECTOR) { > - IFCVF_ERR(ifcvf->pdev, > - "No msix vector for queue %u\n", i); > - return -EINVAL; > - } > - > ifcvf_set_vq_state(hw, i, hw->vring[i].last_avail_idx); > ifc_iowrite16(1, &cfg->queue_enable); > } > diff --git a/drivers/vdpa/ifcvf/ifcvf_base.h b/drivers/vdpa/ifcvf/ifcvf_base.h > index 949b4fb9d554..d2a2a526f0fc 100644 > --- a/drivers/vdpa/ifcvf/ifcvf_base.h > +++ b/drivers/vdpa/ifcvf/ifcvf_base.h > @@ -27,8 +27,6 @@ > > #define IFCVF_QUEUE_ALIGNMENT PAGE_SIZE > #define IFCVF_QUEUE_MAX 32768 > -#define IFCVF_MSI_CONFIG_OFF 0 > -#define IFCVF_MSI_QUEUE_OFF 1 > #define IFCVF_PCI_MAX_RESOURCE 6 > > #define IFCVF_LM_CFG_SIZE 0x40 > @@ -42,6 +40,13 @@ > #define ifcvf_private_to_vf(adapter) \ > (&((struct ifcvf_adapter *)adapter)->vf) > > +/* all vqs and config interrupt has its own vector */ > +#define MSIX_VECTOR_PER_VQ_AND_CONFIG 1 > +/* all vqs share a vector, and config interrupt has a separate vector */ > +#define MSIX_VECTOR_SHARED_VQ_AND_CONFIG 2 > +/* all vqs and config interrupt share a vector */ > +#define MSIX_VECTOR_DEV_SHARED 3 > + > static inline u8 ifc_ioread8(u8 __iomem *addr) > { > return ioread8(addr); > @@ -97,25 +102,27 @@ struct ifcvf_hw { > u8 __iomem *isr; > /* Live migration */ > u8 __iomem *lm_cfg; > - u16 nr_vring; > /* Notification bar number */ > u8 notify_bar; > + u8 msix_vector_status; > + /* virtio-net or virtio-blk device config size */ > + u32 config_size; > /* Notificaiton bar address */ > void __iomem *notify_base; > phys_addr_t notify_base_pa; > u32 notify_off_multiplier; > + u32 dev_type; > u64 req_features; > u64 hw_features; > - u32 dev_type; > struct virtio_pci_common_cfg __iomem *common_cfg; > void __iomem *dev_cfg; > struct vring_info vring[IFCVF_MAX_QUEUES]; > void __iomem * const *base; > char config_msix_name[256]; > struct vdpa_callback config_cb; > - unsigned int config_irq; > - /* virtio-net or virtio-blk device config size */ > - u32 config_size; > + u32 config_irq; > + u32 vqs_shared_irq; > + u16 nr_vring; > }; > > struct ifcvf_adapter { > @@ -160,4 +167,6 @@ int ifcvf_set_vq_state(struct ifcvf_hw *hw, u16 qid, u16 num); > struct ifcvf_adapter *vf_to_adapter(struct ifcvf_hw *hw); > int ifcvf_probed_virtio_net(struct ifcvf_hw *hw); > u32 ifcvf_get_config_size(struct ifcvf_hw *hw); > +int ifcvf_set_vq_vector(struct ifcvf_hw *hw, u16 qid, int vector); > +int ifcvf_set_config_vector(struct ifcvf_hw *hw, int vector); > #endif /* _IFCVF_H_ */ > diff --git a/drivers/vdpa/ifcvf/ifcvf_main.c b/drivers/vdpa/ifcvf/ifcvf_main.c > index 7e2af2d2aaf5..c7070ff01776 100644 > --- a/drivers/vdpa/ifcvf/ifcvf_main.c > +++ b/drivers/vdpa/ifcvf/ifcvf_main.c > @@ -17,6 +17,7 @@ > #define DRIVER_AUTHOR "Intel Corporation" > #define IFCVF_DRIVER_NAME "ifcvf" > > +/* handles config interrupt */ > static irqreturn_t ifcvf_config_changed(int irq, void *arg) > { > struct ifcvf_hw *vf = arg; > @@ -27,6 +28,7 @@ static irqreturn_t ifcvf_config_changed(int irq, void *arg) > return IRQ_HANDLED; > } > > +/* handles vqs interrupt */ > static irqreturn_t ifcvf_intr_handler(int irq, void *arg) > { > struct vring_info *vring = arg; > @@ -37,24 +39,77 @@ static irqreturn_t ifcvf_intr_handler(int irq, void *arg) > return IRQ_HANDLED; > } > > +/* handls vqs shared interrupt */ > +static irqreturn_t ifcvf_vq_shared_intr_handler(int irq, void *arg) > +{ > + struct ifcvf_hw *vf = arg; > + struct vring_info *vring; > + int i; > + > + for (i = 0; i < vf->nr_vring; i++) { > + vring = &vf->vring[i]; > + if (vring->cb.callback) > + vf->vring->cb.callback(vring->cb.private); > + } > + > + return IRQ_HANDLED; > +} > + > +/* handles a shared interrupt for vqs and config */ > +static irqreturn_t ifcvf_dev_shared_intr_handler(int irq, void *arg) > +{ > + struct ifcvf_hw *vf = arg; > + u8 isr; > + > + isr = ifc_ioread8(vf->isr); > + if (isr & VIRTIO_PCI_ISR_CONFIG) > + ifcvf_config_changed(irq, arg); > + > + return ifcvf_vq_shared_intr_handler(irq, arg); > +} > + > static void ifcvf_free_irq_vectors(void *data) > { > pci_free_irq_vectors(data); > } > > -static void ifcvf_free_irq(struct ifcvf_adapter *adapter, int queues) > +static void ifcvf_free_vq_irq(struct ifcvf_adapter *adapter, int queues) > { > struct pci_dev *pdev = adapter->pdev; > struct ifcvf_hw *vf = &adapter->vf; > int i; > > + if (vf->msix_vector_status == MSIX_VECTOR_PER_VQ_AND_CONFIG) > + for (i = 0; i < queues; i++) { > + devm_free_irq(&pdev->dev, vf->vring[i].irq, &vf->vring[i]); > + vf->vring[i].irq = -EINVAL; > + } > + else { > + devm_free_irq(&pdev->dev, vf->vqs_shared_irq, vf); > + } vf->vqs_shared_irq = -EINVAL; If you use {} in the else branch, use it on the then branch, too. > +} > > - for (i = 0; i < queues; i++) { > - devm_free_irq(&pdev->dev, vf->vring[i].irq, &vf->vring[i]); > - vf->vring[i].irq = -EINVAL; > +static void ifcvf_free_config_irq(struct ifcvf_adapter *adapter) > +{ > + struct pci_dev *pdev = adapter->pdev; > + struct ifcvf_hw *vf = &adapter->vf; > + > + /* If the irq is shared by all vqs and the config interrupt, > + * it is already freed in ifcvf_free_vq_irq, so here only > + * need to free config irq when msix_vector_status != MSIX_VECTOR_DEV_SHARED > + */ > + if (vf->msix_vector_status != MSIX_VECTOR_DEV_SHARED) { > + devm_free_irq(&pdev->dev, vf->config_irq, vf); > + vf->config_irq = -EINVAL; > } > +} > + > +static void ifcvf_free_irq(struct ifcvf_adapter *adapter, int queues) > +{ > + struct pci_dev *pdev = adapter->pdev; > > - devm_free_irq(&pdev->dev, vf->config_irq, vf); > + ifcvf_free_vq_irq(adapter, queues); > + ifcvf_free_config_irq(adapter); > ifcvf_free_irq_vectors(pdev); > } > > @@ -81,58 +136,170 @@ static int ifcvf_alloc_vectors(struct ifcvf_adapter *adapter) > return ret; > } > > -static int ifcvf_request_irq(struct ifcvf_adapter *adapter) > +static int ifcvf_request_per_vq_irq(struct ifcvf_adapter *adapter) > { > struct pci_dev *pdev = adapter->pdev; > struct ifcvf_hw *vf = &adapter->vf; > - int vector, i, ret, irq; > - u16 nvectors, max_intr; > + int i, vector, ret, irq; > > - nvectors = ifcvf_alloc_vectors(adapter); > - if (!(nvectors > 0)) > - return nvectors; > + for (i = 0; i < vf->nr_vring; i++) { > + snprintf(vf->vring[i].msix_name, 256, "ifcvf[%s]-%d\n", pci_name(pdev), i); > + vector = i; > + irq = pci_irq_vector(pdev, vector); > + ret = devm_request_irq(&pdev->dev, irq, > + ifcvf_intr_handler, 0, > + vf->vring[i].msix_name, > + &vf->vring[i]); > + if (ret) { > + IFCVF_ERR(pdev, "Failed to request irq for vq %d\n", i); > + ifcvf_free_vq_irq(adapter, i); > + } else { > + vf->vring[i].irq = irq; > + ifcvf_set_vq_vector(vf, i, vector); > + } > + } > > - max_intr = vf->nr_vring + 1; > + vf->vqs_shared_irq = -EINVAL; > + > + return 0; > +} > + > +static int ifcvf_request_shared_vq_irq(struct ifcvf_adapter *adapter) > +{ > + struct pci_dev *pdev = adapter->pdev; > + struct ifcvf_hw *vf = &adapter->vf; > + int i, vector, ret, irq; > + > + vector = 0; > + irq = pci_irq_vector(pdev, vector); > + ret = devm_request_irq(&pdev->dev, irq, > + ifcvf_vq_shared_intr_handler, 0, > + "ifcvf_vq_shared_irq", > + vf); > + if (ret) { > + IFCVF_ERR(pdev, "Failed to request shared irq for vf\n"); > + > + return ret; > + } > + > + vf->vqs_shared_irq = irq; > + for (i = 0; i < vf->nr_vring; i++) { > + vf->vring[i].irq = -EINVAL; > + ifcvf_set_vq_vector(vf, i, vector); > + } > + > + return 0; > + > +} > + > +static int ifcvf_request_dev_shared_irq(struct ifcvf_adapter *adapter) > +{ > + struct pci_dev *pdev = adapter->pdev; > + struct ifcvf_hw *vf = &adapter->vf; > + int i, vector, ret, irq; > + > + vector = 0; > + irq = pci_irq_vector(pdev, vector); > + ret = devm_request_irq(&pdev->dev, irq, > + ifcvf_dev_shared_intr_handler, 0, > + "ifcvf_dev_shared_irq", > + vf); > + if (ret) { > + IFCVF_ERR(pdev, "Failed to request shared irq for vf\n"); > > - ret = pci_alloc_irq_vectors(pdev, max_intr, > - max_intr, PCI_IRQ_MSIX); > - if (ret < 0) { > - IFCVF_ERR(pdev, "Failed to alloc IRQ vectors\n"); > return ret; > } > > + vf->vqs_shared_irq = irq; > + for (i = 0; i < vf->nr_vring; i++) { > + vf->vring[i].irq = -EINVAL; > + ifcvf_set_vq_vector(vf, i, vector); > + } > + > + vf->config_irq = irq; > + ifcvf_set_config_vector(vf, vector); > + > + return 0; > + > +} > + > +static int ifcvf_request_vq_irq(struct ifcvf_adapter *adapter) > +{ > + struct ifcvf_hw *vf = &adapter->vf; > + int ret; > + > + if (vf->msix_vector_status == MSIX_VECTOR_PER_VQ_AND_CONFIG) > + ret = ifcvf_request_per_vq_irq(adapter); > + else > + ret = ifcvf_request_shared_vq_irq(adapter); > + > + return ret; > +} > + > +static int ifcvf_request_config_irq(struct ifcvf_adapter *adapter) > +{ > + struct pci_dev *pdev = adapter->pdev; > + struct ifcvf_hw *vf = &adapter->vf; > + int config_vector, ret; > + > + if (vf->msix_vector_status == MSIX_VECTOR_DEV_SHARED) > + return 0; > + > + if (vf->msix_vector_status == MSIX_VECTOR_PER_VQ_AND_CONFIG) > + /* vector 0 ~ vf->nr_vring for vqs, num vf->nr_vring vector for config interrupt */ > + config_vector = vf->nr_vring; > + > + if (vf->msix_vector_status == MSIX_VECTOR_SHARED_VQ_AND_CONFIG) > + /* vector 0 for vqs and 1 for config interrupt */ > + config_vector = 1; > + > snprintf(vf->config_msix_name, 256, "ifcvf[%s]-config\n", > pci_name(pdev)); > - vector = 0; > - vf->config_irq = pci_irq_vector(pdev, vector); > + vf->config_irq = pci_irq_vector(pdev, config_vector); > ret = devm_request_irq(&pdev->dev, vf->config_irq, > ifcvf_config_changed, 0, > vf->config_msix_name, vf); > if (ret) { > IFCVF_ERR(pdev, "Failed to request config irq\n"); > + ifcvf_free_vq_irq(adapter, vf->nr_vring); > return ret; > } > + ifcvf_set_config_vector(vf, config_vector); > > - for (i = 0; i < vf->nr_vring; i++) { > - snprintf(vf->vring[i].msix_name, 256, "ifcvf[%s]-%d\n", > - pci_name(pdev), i); > - vector = i + IFCVF_MSI_QUEUE_OFF; > - irq = pci_irq_vector(pdev, vector); > - ret = devm_request_irq(&pdev->dev, irq, > - ifcvf_intr_handler, 0, > - vf->vring[i].msix_name, > - &vf->vring[i]); > - if (ret) { > - IFCVF_ERR(pdev, > - "Failed to request irq for vq %d\n", i); > - ifcvf_free_irq(adapter, i); > + return 0; > +} > > - return ret; > - } > +static int ifcvf_request_irq(struct ifcvf_adapter *adapter) This function was not moved, was it? Try --minimal or --histogram maybe. Or just --anchored=ifcvf_request_irq > +{ > + struct ifcvf_hw *vf = &adapter->vf; > + u16 nvectors, max_intr; > + int ret; > > - vf->vring[i].irq = irq; > + nvectors = ifcvf_alloc_vectors(adapter); > + if (!(nvectors > 0)) > + return nvectors; > + > + vf->msix_vector_status = MSIX_VECTOR_PER_VQ_AND_CONFIG; > + max_intr = vf->nr_vring + 1; > + if (nvectors < max_intr) > + vf->msix_vector_status = MSIX_VECTOR_SHARED_VQ_AND_CONFIG; > + > + if (nvectors == 1) { > + vf->msix_vector_status = MSIX_VECTOR_DEV_SHARED; > + ret = ifcvf_request_dev_shared_irq(adapter); > + > + return ret; > } > > + ret = ifcvf_request_vq_irq(adapter); > + if (ret) > + return ret; > + > + ret = ifcvf_request_config_irq(adapter); > + > + if (ret) > + return ret; > + > return 0; > } > > @@ -436,7 +603,10 @@ static int ifcvf_vdpa_get_vq_irq(struct vdpa_device *vdpa_dev, > { > struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev); > > - return vf->vring[qid].irq; > + if (vf->vqs_shared_irq < 0) > + return vf->vring[qid].irq; > + else > + return -EINVAL; > } > > static struct vdpa_notification_area ifcvf_get_vq_notification(struct vdpa_device *vdpa_dev, > -- > 2.27.0 ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: [PATCH V2 4/4] vDPA/ifcvf: implement shared IRQ feature 2022-01-25 19:42 ` Michael S. Tsirkin @ 2022-01-26 12:22 ` Zhu, Lingshan 0 siblings, 0 replies; 13+ messages in thread From: Zhu, Lingshan @ 2022-01-26 12:22 UTC (permalink / raw) To: Michael S. Tsirkin; +Cc: jasowang, netdev, virtualization On 1/26/2022 3:42 AM, Michael S. Tsirkin wrote: > On Tue, Jan 25, 2022 at 05:17:44PM +0800, Zhu Lingshan wrote: >> On some platforms/devices, there may not be enough MSI vector >> slots allocated for virtqueues and config changes. In such a case, >> the interrupt sources(virtqueues, config changes) must share >> an IRQ/vector, to avoid initialization failures, keep >> the device functional. >> >> This commit handles three cases: >> (1) number of the allocated vectors == the number of virtqueues + 1 >> (config changes), every virtqueue and the config interrupt has >> a separated vector/IRQ, the best and the most likely case. >> (2) number of the allocated vectors is less than the best case, but >> greater than 1. In this case, all virtqueues share a vector/IRQ, >> the config interrupt has a separated vector/IRQ >> (3) only one vector is allocated, in this case, the virtqueues and >> the config interrupt share a vector/IRQ. The worst and most >> unlikely case. >> >> Otherwise, it needs to fail. >> >> This commit introduces some helper functions: >> ifcvf_set_vq_vector() and ifcvf_set_config_vector() sets virtqueue >> vector and config vector in the device config space, so that >> the device can send interrupt DMA. >> >> This commit adds some fields in struct ifcvf_hw and re-placed >> the existed fields to be aligned with the cacheline. >> >> Signed-off-by: Zhu Lingshan <lingshan.zhu@intel.com> >> --- >> drivers/vdpa/ifcvf/ifcvf_base.c | 47 +++++-- >> drivers/vdpa/ifcvf/ifcvf_base.h | 23 ++- >> drivers/vdpa/ifcvf/ifcvf_main.c | 240 +++++++++++++++++++++++++++----- >> 3 files changed, 253 insertions(+), 57 deletions(-) >> >> diff --git a/drivers/vdpa/ifcvf/ifcvf_base.c b/drivers/vdpa/ifcvf/ifcvf_base.c >> index 397692ae671c..18dcb63ab1e3 100644 >> --- a/drivers/vdpa/ifcvf/ifcvf_base.c >> +++ b/drivers/vdpa/ifcvf/ifcvf_base.c >> @@ -15,6 +15,36 @@ struct ifcvf_adapter *vf_to_adapter(struct ifcvf_hw *hw) >> return container_of(hw, struct ifcvf_adapter, vf); >> } >> >> +int ifcvf_set_vq_vector(struct ifcvf_hw *hw, u16 qid, int vector) >> +{ >> + struct virtio_pci_common_cfg __iomem *cfg = hw->common_cfg; >> + struct ifcvf_adapter *ifcvf = vf_to_adapter(hw); >> + >> + ifc_iowrite16(qid, &cfg->queue_select); >> + ifc_iowrite16(vector, &cfg->queue_msix_vector); >> + if (ifc_ioread16(&cfg->queue_msix_vector) == VIRTIO_MSI_NO_VECTOR) { >> + IFCVF_ERR(ifcvf->pdev, "No msix vector for queue %u\n", qid); >> + return -EINVAL; >> + } >> + >> + return 0; >> +} >> + >> +int ifcvf_set_config_vector(struct ifcvf_hw *hw, int vector) >> +{ >> + struct virtio_pci_common_cfg __iomem *cfg = hw->common_cfg; >> + struct ifcvf_adapter *ifcvf = vf_to_adapter(hw); >> + >> + cfg = hw->common_cfg; >> + ifc_iowrite16(vector, &cfg->msix_config); >> + if (ifc_ioread16(&cfg->msix_config) == VIRTIO_MSI_NO_VECTOR) { >> + IFCVF_ERR(ifcvf->pdev, "No msix vector for device config\n"); >> + return -EINVAL; >> + } >> + >> + return 0; >> +} >> + >> static void __iomem *get_cap_addr(struct ifcvf_hw *hw, >> struct virtio_pci_cap *cap) >> { >> @@ -140,6 +170,8 @@ int ifcvf_init_hw(struct ifcvf_hw *hw, struct pci_dev *pdev) >> hw->common_cfg, hw->notify_base, hw->isr, >> hw->dev_cfg, hw->notify_off_multiplier); >> >> + hw->vqs_shared_irq = -EINVAL; >> + > This stuffing of errno values into u32 is not a good idea. will fix this in V3 > > >> return 0; >> } >> >> @@ -321,12 +353,6 @@ static int ifcvf_hw_enable(struct ifcvf_hw *hw) >> >> ifcvf = vf_to_adapter(hw); >> cfg = hw->common_cfg; >> - ifc_iowrite16(IFCVF_MSI_CONFIG_OFF, &cfg->msix_config); >> - >> - if (ifc_ioread16(&cfg->msix_config) == VIRTIO_MSI_NO_VECTOR) { >> - IFCVF_ERR(ifcvf->pdev, "No msix vector for device config\n"); >> - return -EINVAL; >> - } >> >> for (i = 0; i < hw->nr_vring; i++) { >> if (!hw->vring[i].ready) >> @@ -340,15 +366,6 @@ static int ifcvf_hw_enable(struct ifcvf_hw *hw) >> ifc_iowrite64_twopart(hw->vring[i].used, &cfg->queue_used_lo, >> &cfg->queue_used_hi); >> ifc_iowrite16(hw->vring[i].size, &cfg->queue_size); >> - ifc_iowrite16(i + IFCVF_MSI_QUEUE_OFF, &cfg->queue_msix_vector); >> - >> - if (ifc_ioread16(&cfg->queue_msix_vector) == >> - VIRTIO_MSI_NO_VECTOR) { >> - IFCVF_ERR(ifcvf->pdev, >> - "No msix vector for queue %u\n", i); >> - return -EINVAL; >> - } >> - >> ifcvf_set_vq_state(hw, i, hw->vring[i].last_avail_idx); >> ifc_iowrite16(1, &cfg->queue_enable); >> } >> diff --git a/drivers/vdpa/ifcvf/ifcvf_base.h b/drivers/vdpa/ifcvf/ifcvf_base.h >> index 949b4fb9d554..d2a2a526f0fc 100644 >> --- a/drivers/vdpa/ifcvf/ifcvf_base.h >> +++ b/drivers/vdpa/ifcvf/ifcvf_base.h >> @@ -27,8 +27,6 @@ >> >> #define IFCVF_QUEUE_ALIGNMENT PAGE_SIZE >> #define IFCVF_QUEUE_MAX 32768 >> -#define IFCVF_MSI_CONFIG_OFF 0 >> -#define IFCVF_MSI_QUEUE_OFF 1 >> #define IFCVF_PCI_MAX_RESOURCE 6 >> >> #define IFCVF_LM_CFG_SIZE 0x40 >> @@ -42,6 +40,13 @@ >> #define ifcvf_private_to_vf(adapter) \ >> (&((struct ifcvf_adapter *)adapter)->vf) >> >> +/* all vqs and config interrupt has its own vector */ >> +#define MSIX_VECTOR_PER_VQ_AND_CONFIG 1 >> +/* all vqs share a vector, and config interrupt has a separate vector */ >> +#define MSIX_VECTOR_SHARED_VQ_AND_CONFIG 2 >> +/* all vqs and config interrupt share a vector */ >> +#define MSIX_VECTOR_DEV_SHARED 3 >> + >> static inline u8 ifc_ioread8(u8 __iomem *addr) >> { >> return ioread8(addr); >> @@ -97,25 +102,27 @@ struct ifcvf_hw { >> u8 __iomem *isr; >> /* Live migration */ >> u8 __iomem *lm_cfg; >> - u16 nr_vring; >> /* Notification bar number */ >> u8 notify_bar; >> + u8 msix_vector_status; >> + /* virtio-net or virtio-blk device config size */ >> + u32 config_size; >> /* Notificaiton bar address */ >> void __iomem *notify_base; >> phys_addr_t notify_base_pa; >> u32 notify_off_multiplier; >> + u32 dev_type; >> u64 req_features; >> u64 hw_features; >> - u32 dev_type; >> struct virtio_pci_common_cfg __iomem *common_cfg; >> void __iomem *dev_cfg; >> struct vring_info vring[IFCVF_MAX_QUEUES]; >> void __iomem * const *base; >> char config_msix_name[256]; >> struct vdpa_callback config_cb; >> - unsigned int config_irq; >> - /* virtio-net or virtio-blk device config size */ >> - u32 config_size; >> + u32 config_irq; >> + u32 vqs_shared_irq; >> + u16 nr_vring; >> }; >> >> struct ifcvf_adapter { >> @@ -160,4 +167,6 @@ int ifcvf_set_vq_state(struct ifcvf_hw *hw, u16 qid, u16 num); >> struct ifcvf_adapter *vf_to_adapter(struct ifcvf_hw *hw); >> int ifcvf_probed_virtio_net(struct ifcvf_hw *hw); >> u32 ifcvf_get_config_size(struct ifcvf_hw *hw); >> +int ifcvf_set_vq_vector(struct ifcvf_hw *hw, u16 qid, int vector); >> +int ifcvf_set_config_vector(struct ifcvf_hw *hw, int vector); >> #endif /* _IFCVF_H_ */ >> diff --git a/drivers/vdpa/ifcvf/ifcvf_main.c b/drivers/vdpa/ifcvf/ifcvf_main.c >> index 7e2af2d2aaf5..c7070ff01776 100644 >> --- a/drivers/vdpa/ifcvf/ifcvf_main.c >> +++ b/drivers/vdpa/ifcvf/ifcvf_main.c >> @@ -17,6 +17,7 @@ >> #define DRIVER_AUTHOR "Intel Corporation" >> #define IFCVF_DRIVER_NAME "ifcvf" >> >> +/* handles config interrupt */ >> static irqreturn_t ifcvf_config_changed(int irq, void *arg) >> { >> struct ifcvf_hw *vf = arg; >> @@ -27,6 +28,7 @@ static irqreturn_t ifcvf_config_changed(int irq, void *arg) >> return IRQ_HANDLED; >> } >> >> +/* handles vqs interrupt */ >> static irqreturn_t ifcvf_intr_handler(int irq, void *arg) >> { >> struct vring_info *vring = arg; >> @@ -37,24 +39,77 @@ static irqreturn_t ifcvf_intr_handler(int irq, void *arg) >> return IRQ_HANDLED; >> } >> >> +/* handls vqs shared interrupt */ >> +static irqreturn_t ifcvf_vq_shared_intr_handler(int irq, void *arg) >> +{ >> + struct ifcvf_hw *vf = arg; >> + struct vring_info *vring; >> + int i; >> + >> + for (i = 0; i < vf->nr_vring; i++) { >> + vring = &vf->vring[i]; >> + if (vring->cb.callback) >> + vf->vring->cb.callback(vring->cb.private); >> + } >> + >> + return IRQ_HANDLED; >> +} >> + >> +/* handles a shared interrupt for vqs and config */ >> +static irqreturn_t ifcvf_dev_shared_intr_handler(int irq, void *arg) >> +{ >> + struct ifcvf_hw *vf = arg; >> + u8 isr; >> + >> + isr = ifc_ioread8(vf->isr); >> + if (isr & VIRTIO_PCI_ISR_CONFIG) >> + ifcvf_config_changed(irq, arg); >> + >> + return ifcvf_vq_shared_intr_handler(irq, arg); >> +} >> + >> static void ifcvf_free_irq_vectors(void *data) >> { >> pci_free_irq_vectors(data); >> } >> >> -static void ifcvf_free_irq(struct ifcvf_adapter *adapter, int queues) >> +static void ifcvf_free_vq_irq(struct ifcvf_adapter *adapter, int queues) >> { >> struct pci_dev *pdev = adapter->pdev; >> struct ifcvf_hw *vf = &adapter->vf; >> int i; >> >> + if (vf->msix_vector_status == MSIX_VECTOR_PER_VQ_AND_CONFIG) >> + for (i = 0; i < queues; i++) { >> + devm_free_irq(&pdev->dev, vf->vring[i].irq, &vf->vring[i]); >> + vf->vring[i].irq = -EINVAL; >> + } >> + else { >> + devm_free_irq(&pdev->dev, vf->vqs_shared_irq, vf); >> + } vf->vqs_shared_irq = -EINVAL; > If you use {} in the else branch, use it on the then branch, too. sure > > >> +} >> >> - for (i = 0; i < queues; i++) { >> - devm_free_irq(&pdev->dev, vf->vring[i].irq, &vf->vring[i]); >> - vf->vring[i].irq = -EINVAL; >> +static void ifcvf_free_config_irq(struct ifcvf_adapter *adapter) >> +{ >> + struct pci_dev *pdev = adapter->pdev; >> + struct ifcvf_hw *vf = &adapter->vf; >> + >> + /* If the irq is shared by all vqs and the config interrupt, >> + * it is already freed in ifcvf_free_vq_irq, so here only >> + * need to free config irq when msix_vector_status != MSIX_VECTOR_DEV_SHARED >> + */ >> + if (vf->msix_vector_status != MSIX_VECTOR_DEV_SHARED) { >> + devm_free_irq(&pdev->dev, vf->config_irq, vf); >> + vf->config_irq = -EINVAL; >> } >> +} >> + >> +static void ifcvf_free_irq(struct ifcvf_adapter *adapter, int queues) >> +{ >> + struct pci_dev *pdev = adapter->pdev; >> >> - devm_free_irq(&pdev->dev, vf->config_irq, vf); >> + ifcvf_free_vq_irq(adapter, queues); >> + ifcvf_free_config_irq(adapter); >> ifcvf_free_irq_vectors(pdev); >> } >> >> @@ -81,58 +136,170 @@ static int ifcvf_alloc_vectors(struct ifcvf_adapter *adapter) >> return ret; >> } >> >> -static int ifcvf_request_irq(struct ifcvf_adapter *adapter) >> +static int ifcvf_request_per_vq_irq(struct ifcvf_adapter *adapter) >> { >> struct pci_dev *pdev = adapter->pdev; >> struct ifcvf_hw *vf = &adapter->vf; >> - int vector, i, ret, irq; >> - u16 nvectors, max_intr; >> + int i, vector, ret, irq; >> >> - nvectors = ifcvf_alloc_vectors(adapter); >> - if (!(nvectors > 0)) >> - return nvectors; >> + for (i = 0; i < vf->nr_vring; i++) { >> + snprintf(vf->vring[i].msix_name, 256, "ifcvf[%s]-%d\n", pci_name(pdev), i); >> + vector = i; >> + irq = pci_irq_vector(pdev, vector); >> + ret = devm_request_irq(&pdev->dev, irq, >> + ifcvf_intr_handler, 0, >> + vf->vring[i].msix_name, >> + &vf->vring[i]); >> + if (ret) { >> + IFCVF_ERR(pdev, "Failed to request irq for vq %d\n", i); >> + ifcvf_free_vq_irq(adapter, i); >> + } else { >> + vf->vring[i].irq = irq; >> + ifcvf_set_vq_vector(vf, i, vector); >> + } >> + } >> >> - max_intr = vf->nr_vring + 1; >> + vf->vqs_shared_irq = -EINVAL; >> + >> + return 0; >> +} >> + >> +static int ifcvf_request_shared_vq_irq(struct ifcvf_adapter *adapter) >> +{ >> + struct pci_dev *pdev = adapter->pdev; >> + struct ifcvf_hw *vf = &adapter->vf; >> + int i, vector, ret, irq; >> + >> + vector = 0; >> + irq = pci_irq_vector(pdev, vector); >> + ret = devm_request_irq(&pdev->dev, irq, >> + ifcvf_vq_shared_intr_handler, 0, >> + "ifcvf_vq_shared_irq", >> + vf); >> + if (ret) { >> + IFCVF_ERR(pdev, "Failed to request shared irq for vf\n"); >> + >> + return ret; >> + } >> + >> + vf->vqs_shared_irq = irq; >> + for (i = 0; i < vf->nr_vring; i++) { >> + vf->vring[i].irq = -EINVAL; >> + ifcvf_set_vq_vector(vf, i, vector); >> + } >> + >> + return 0; >> + >> +} >> + >> +static int ifcvf_request_dev_shared_irq(struct ifcvf_adapter *adapter) >> +{ >> + struct pci_dev *pdev = adapter->pdev; >> + struct ifcvf_hw *vf = &adapter->vf; >> + int i, vector, ret, irq; >> + >> + vector = 0; >> + irq = pci_irq_vector(pdev, vector); >> + ret = devm_request_irq(&pdev->dev, irq, >> + ifcvf_dev_shared_intr_handler, 0, >> + "ifcvf_dev_shared_irq", >> + vf); >> + if (ret) { >> + IFCVF_ERR(pdev, "Failed to request shared irq for vf\n"); >> >> - ret = pci_alloc_irq_vectors(pdev, max_intr, >> - max_intr, PCI_IRQ_MSIX); >> - if (ret < 0) { >> - IFCVF_ERR(pdev, "Failed to alloc IRQ vectors\n"); >> return ret; >> } >> >> + vf->vqs_shared_irq = irq; >> + for (i = 0; i < vf->nr_vring; i++) { >> + vf->vring[i].irq = -EINVAL; >> + ifcvf_set_vq_vector(vf, i, vector); >> + } >> + >> + vf->config_irq = irq; >> + ifcvf_set_config_vector(vf, vector); >> + >> + return 0; >> + >> +} >> + >> +static int ifcvf_request_vq_irq(struct ifcvf_adapter *adapter) >> +{ >> + struct ifcvf_hw *vf = &adapter->vf; >> + int ret; >> + >> + if (vf->msix_vector_status == MSIX_VECTOR_PER_VQ_AND_CONFIG) >> + ret = ifcvf_request_per_vq_irq(adapter); >> + else >> + ret = ifcvf_request_shared_vq_irq(adapter); >> + >> + return ret; >> +} >> + >> +static int ifcvf_request_config_irq(struct ifcvf_adapter *adapter) >> +{ >> + struct pci_dev *pdev = adapter->pdev; >> + struct ifcvf_hw *vf = &adapter->vf; >> + int config_vector, ret; >> + >> + if (vf->msix_vector_status == MSIX_VECTOR_DEV_SHARED) >> + return 0; >> + >> + if (vf->msix_vector_status == MSIX_VECTOR_PER_VQ_AND_CONFIG) >> + /* vector 0 ~ vf->nr_vring for vqs, num vf->nr_vring vector for config interrupt */ >> + config_vector = vf->nr_vring; >> + >> + if (vf->msix_vector_status == MSIX_VECTOR_SHARED_VQ_AND_CONFIG) >> + /* vector 0 for vqs and 1 for config interrupt */ >> + config_vector = 1; >> + >> snprintf(vf->config_msix_name, 256, "ifcvf[%s]-config\n", >> pci_name(pdev)); >> - vector = 0; >> - vf->config_irq = pci_irq_vector(pdev, vector); >> + vf->config_irq = pci_irq_vector(pdev, config_vector); >> ret = devm_request_irq(&pdev->dev, vf->config_irq, >> ifcvf_config_changed, 0, >> vf->config_msix_name, vf); >> if (ret) { >> IFCVF_ERR(pdev, "Failed to request config irq\n"); >> + ifcvf_free_vq_irq(adapter, vf->nr_vring); >> return ret; >> } >> + ifcvf_set_config_vector(vf, config_vector); >> >> - for (i = 0; i < vf->nr_vring; i++) { >> - snprintf(vf->vring[i].msix_name, 256, "ifcvf[%s]-%d\n", >> - pci_name(pdev), i); >> - vector = i + IFCVF_MSI_QUEUE_OFF; >> - irq = pci_irq_vector(pdev, vector); >> - ret = devm_request_irq(&pdev->dev, irq, >> - ifcvf_intr_handler, 0, >> - vf->vring[i].msix_name, >> - &vf->vring[i]); >> - if (ret) { >> - IFCVF_ERR(pdev, >> - "Failed to request irq for vq %d\n", i); >> - ifcvf_free_irq(adapter, i); >> + return 0; >> +} >> >> - return ret; >> - } >> +static int ifcvf_request_irq(struct ifcvf_adapter *adapter) > This function was not moved, was it? > > Try --minimal or --histogram maybe. Or just > --anchored=ifcvf_request_irq Yes, it is not moved, just changed. I tried these parameters, there are some improvements though seems not significant. It seems git still finds ifcvf_request_per_vq_irq() uses some lines of original ifcvf_request_irq(). Thanks for your review and help Zhu Lingshan > > >> +{ >> + struct ifcvf_hw *vf = &adapter->vf; >> + u16 nvectors, max_intr; >> + int ret; >> >> - vf->vring[i].irq = irq; >> + nvectors = ifcvf_alloc_vectors(adapter); >> + if (!(nvectors > 0)) >> + return nvectors; >> + > >> + vf->msix_vector_status = MSIX_VECTOR_PER_VQ_AND_CONFIG; >> + max_intr = vf->nr_vring + 1; >> + if (nvectors < max_intr) >> + vf->msix_vector_status = MSIX_VECTOR_SHARED_VQ_AND_CONFIG; >> + >> + if (nvectors == 1) { >> + vf->msix_vector_status = MSIX_VECTOR_DEV_SHARED; >> + ret = ifcvf_request_dev_shared_irq(adapter); >> + >> + return ret; >> } >> >> + ret = ifcvf_request_vq_irq(adapter); >> + if (ret) >> + return ret; >> + >> + ret = ifcvf_request_config_irq(adapter); >> + >> + if (ret) >> + return ret; >> + >> return 0; >> } >> >> @@ -436,7 +603,10 @@ static int ifcvf_vdpa_get_vq_irq(struct vdpa_device *vdpa_dev, >> { >> struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev); >> >> - return vf->vring[qid].irq; >> + if (vf->vqs_shared_irq < 0) >> + return vf->vring[qid].irq; >> + else >> + return -EINVAL; >> } >> >> static struct vdpa_notification_area ifcvf_get_vq_notification(struct vdpa_device *vdpa_dev, >> -- >> 2.27.0 ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: [PATCH V2 0/4] vDPA/ifcvf: implement shared IRQ feature 2022-01-25 9:17 [PATCH V2 0/4] vDPA/ifcvf: implement shared IRQ feature Zhu Lingshan ` (3 preceding siblings ...) 2022-01-25 9:17 ` [PATCH V2 4/4] vDPA/ifcvf: implement shared IRQ feature Zhu Lingshan @ 2022-01-25 19:32 ` Michael S. Tsirkin 4 siblings, 0 replies; 13+ messages in thread From: Michael S. Tsirkin @ 2022-01-25 19:32 UTC (permalink / raw) To: Zhu Lingshan; +Cc: jasowang, netdev, virtualization On Tue, Jan 25, 2022 at 05:17:40PM +0800, Zhu Lingshan wrote: > It has been observed that on some platforms/devices, there may > not be enough MSI vectors for virtqueues and the config change. > Under such circumstances, the interrupt sources of a device > have to share vectors/IRQs. > > This series implemented a shared IRQ feature for ifcvf. Which configurations did you test with this, and what were the results? Given patch 3 is broken ... > Please help review. > > Changes from V1: > (1) Enable config interrupt when only one vector is allocated(Michael) > (2) Clean vectors/IRQs if failed to request config interrupt > since config interrupt is a must(Michael) > (3) Keep local vdpa_ops, disable irq_bypass by setting IRQ = -EINVAL > for shared IRQ case(Michael) > (4) Improvements on error messages(Michael) > (5) Squash functions implementation patches to the callers(Michael) > > Zhu Lingshan (4): > vDPA/ifcvf: implement IO read/write helpers in the header file > vDPA/ifcvf: implement device MSIX vector allocator > vhost_vdpa: don't setup irq offloading when irq_num < 0 > vDPA/ifcvf: implement shared IRQ feature > > drivers/vdpa/ifcvf/ifcvf_base.c | 67 +++------ > drivers/vdpa/ifcvf/ifcvf_base.h | 60 +++++++- > drivers/vdpa/ifcvf/ifcvf_main.c | 254 ++++++++++++++++++++++++++++---- > drivers/vhost/vdpa.c | 3 + > 4 files changed, 305 insertions(+), 79 deletions(-) > > -- > 2.27.0 ^ permalink raw reply [flat|nested] 13+ messages in thread
end of thread, other threads:[~2022-01-26 12:22 UTC | newest] Thread overview: 13+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2022-01-25 9:17 [PATCH V2 0/4] vDPA/ifcvf: implement shared IRQ feature Zhu Lingshan 2022-01-25 9:17 ` [PATCH V2 1/4] vDPA/ifcvf: implement IO read/write helpers in the header file Zhu Lingshan 2022-01-25 9:17 ` [PATCH V2 2/4] vDPA/ifcvf: implement device MSIX vector allocator Zhu Lingshan 2022-01-25 19:36 ` Michael S. Tsirkin 2022-01-26 12:18 ` Zhu, Lingshan 2022-01-25 9:17 ` [PATCH V2 3/4] vhost_vdpa: don't setup irq offloading when irq_num < 0 Zhu Lingshan 2022-01-25 19:17 ` kernel test robot 2022-01-25 19:30 ` Michael S. Tsirkin 2022-01-26 12:18 ` Zhu, Lingshan 2022-01-25 9:17 ` [PATCH V2 4/4] vDPA/ifcvf: implement shared IRQ feature Zhu Lingshan 2022-01-25 19:42 ` Michael S. Tsirkin 2022-01-26 12:22 ` Zhu, Lingshan 2022-01-25 19:32 ` [PATCH V2 0/4] " Michael S. Tsirkin
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).