virtualization.lists.linux-foundation.org archive mirror
 help / color / mirror / Atom feed
From: Jason Wang <jasowang@redhat.com>
To: Zhu Lingshan <lingshan.zhu@intel.com>, mst@redhat.com
Cc: netdev@vger.kernel.org, kvm@vger.kernel.org,
	virtualization@lists.linux-foundation.org
Subject: Re: [PATCH 1/2] vDPA/ifcvf: detect and use the onboard number of queues directly
Date: Thu, 19 Aug 2021 12:09:19 +0800	[thread overview]
Message-ID: <78c25559-0423-7ef6-9cdd-d2b81df111ca@redhat.com> (raw)
In-Reply-To: <20210818095714.3220-2-lingshan.zhu@intel.com>


在 2021/8/18 下午5:57, Zhu Lingshan 写道:
> To enable this multi-queue feature for ifcvf, this commit
> intends to detect and use the onboard number of queues
> directly than IFCVF_MAX_QUEUE_PAIRS = 1 (removed)
>
> Signed-off-by: Zhu Lingshan <lingshan.zhu@intel.com>
> ---
>   drivers/vdpa/ifcvf/ifcvf_base.c |  8 +++++---
>   drivers/vdpa/ifcvf/ifcvf_base.h | 10 ++++------
>   drivers/vdpa/ifcvf/ifcvf_main.c | 21 ++++++++++++---------
>   3 files changed, 21 insertions(+), 18 deletions(-)
>
> diff --git a/drivers/vdpa/ifcvf/ifcvf_base.c b/drivers/vdpa/ifcvf/ifcvf_base.c
> index 6e197fe0fcf9..2808f1ba9f7b 100644
> --- a/drivers/vdpa/ifcvf/ifcvf_base.c
> +++ b/drivers/vdpa/ifcvf/ifcvf_base.c
> @@ -158,7 +158,9 @@ int ifcvf_init_hw(struct ifcvf_hw *hw, struct pci_dev *pdev)
>   		return -EIO;
>   	}
>   
> -	for (i = 0; i < IFCVF_MAX_QUEUE_PAIRS * 2; i++) {
> +	hw->nr_vring = ifc_ioread16(&hw->common_cfg->num_queues);
> +
> +	for (i = 0; i < hw->nr_vring; i++) {
>   		ifc_iowrite16(i, &hw->common_cfg->queue_select);
>   		notify_off = ifc_ioread16(&hw->common_cfg->queue_notify_off);
>   		hw->vring[i].notify_addr = hw->notify_base +
> @@ -304,7 +306,7 @@ u16 ifcvf_get_vq_state(struct ifcvf_hw *hw, u16 qid)
>   	u32 q_pair_id;
>   
>   	ifcvf_lm = (struct ifcvf_lm_cfg __iomem *)hw->lm_cfg;
> -	q_pair_id = qid / (IFCVF_MAX_QUEUE_PAIRS * 2);
> +	q_pair_id = qid / hw->nr_vring;
>   	avail_idx_addr = &ifcvf_lm->vring_lm_cfg[q_pair_id].idx_addr[qid % 2];
>   	last_avail_idx = ifc_ioread16(avail_idx_addr);
>   
> @@ -318,7 +320,7 @@ int ifcvf_set_vq_state(struct ifcvf_hw *hw, u16 qid, u16 num)
>   	u32 q_pair_id;
>   
>   	ifcvf_lm = (struct ifcvf_lm_cfg __iomem *)hw->lm_cfg;
> -	q_pair_id = qid / (IFCVF_MAX_QUEUE_PAIRS * 2);
> +	q_pair_id = qid / hw->nr_vring;
>   	avail_idx_addr = &ifcvf_lm->vring_lm_cfg[q_pair_id].idx_addr[qid % 2];
>   	hw->vring[qid].last_avail_idx = num;
>   	ifc_iowrite16(num, avail_idx_addr);
> diff --git a/drivers/vdpa/ifcvf/ifcvf_base.h b/drivers/vdpa/ifcvf/ifcvf_base.h
> index 1601e87870da..97d9019a3ec0 100644
> --- a/drivers/vdpa/ifcvf/ifcvf_base.h
> +++ b/drivers/vdpa/ifcvf/ifcvf_base.h
> @@ -31,8 +31,8 @@
>   		 (1ULL << VIRTIO_F_ACCESS_PLATFORM)		| \
>   		 (1ULL << VIRTIO_NET_F_MRG_RXBUF))
>   
> -/* Only one queue pair for now. */
> -#define IFCVF_MAX_QUEUE_PAIRS	1
> +/* Max 8 data queue pairs(16 queues) and one control vq for now. */
> +#define IFCVF_MAX_QUEUES	17


While at it, I wonder if we can get rid of this.

Other than this,

Acked-by: Jason Wang <jasowang@redhat.com>


>   
>   #define IFCVF_QUEUE_ALIGNMENT	PAGE_SIZE
>   #define IFCVF_QUEUE_MAX		32768
> @@ -51,8 +51,6 @@
>   #define ifcvf_private_to_vf(adapter) \
>   	(&((struct ifcvf_adapter *)adapter)->vf)
>   
> -#define IFCVF_MAX_INTR (IFCVF_MAX_QUEUE_PAIRS * 2 + 1)
> -
>   struct vring_info {
>   	u64 desc;
>   	u64 avail;
> @@ -83,7 +81,7 @@ struct ifcvf_hw {
>   	u32 dev_type;
>   	struct virtio_pci_common_cfg __iomem *common_cfg;
>   	void __iomem *net_cfg;
> -	struct vring_info vring[IFCVF_MAX_QUEUE_PAIRS * 2];
> +	struct vring_info vring[IFCVF_MAX_QUEUES];
>   	void __iomem * const *base;
>   	char config_msix_name[256];
>   	struct vdpa_callback config_cb;
> @@ -103,7 +101,7 @@ struct ifcvf_vring_lm_cfg {
>   
>   struct ifcvf_lm_cfg {
>   	u8 reserved[IFCVF_LM_RING_STATE_OFFSET];
> -	struct ifcvf_vring_lm_cfg vring_lm_cfg[IFCVF_MAX_QUEUE_PAIRS];
> +	struct ifcvf_vring_lm_cfg vring_lm_cfg[IFCVF_MAX_QUEUES];
>   };
>   
>   struct ifcvf_vdpa_mgmt_dev {
> diff --git a/drivers/vdpa/ifcvf/ifcvf_main.c b/drivers/vdpa/ifcvf/ifcvf_main.c
> index 4b623253f460..e34c2ec2b69b 100644
> --- a/drivers/vdpa/ifcvf/ifcvf_main.c
> +++ b/drivers/vdpa/ifcvf/ifcvf_main.c
> @@ -63,9 +63,13 @@ static int ifcvf_request_irq(struct ifcvf_adapter *adapter)
>   	struct pci_dev *pdev = adapter->pdev;
>   	struct ifcvf_hw *vf = &adapter->vf;
>   	int vector, i, ret, irq;
> +	u16 max_intr;
>   
> -	ret = pci_alloc_irq_vectors(pdev, IFCVF_MAX_INTR,
> -				    IFCVF_MAX_INTR, PCI_IRQ_MSIX);
> +	/* all queues and config interrupt  */
> +	max_intr = vf->nr_vring + 1;
> +
> +	ret = pci_alloc_irq_vectors(pdev, max_intr,
> +				    max_intr, PCI_IRQ_MSIX);
>   	if (ret < 0) {
>   		IFCVF_ERR(pdev, "Failed to alloc IRQ vectors\n");
>   		return ret;
> @@ -83,7 +87,7 @@ static int ifcvf_request_irq(struct ifcvf_adapter *adapter)
>   		return ret;
>   	}
>   
> -	for (i = 0; i < IFCVF_MAX_QUEUE_PAIRS * 2; i++) {
> +	for (i = 0; i < vf->nr_vring; i++) {
>   		snprintf(vf->vring[i].msix_name, 256, "ifcvf[%s]-%d\n",
>   			 pci_name(pdev), i);
>   		vector = i + IFCVF_MSI_QUEUE_OFF;
> @@ -112,7 +116,6 @@ static int ifcvf_start_datapath(void *private)
>   	u8 status;
>   	int ret;
>   
> -	vf->nr_vring = IFCVF_MAX_QUEUE_PAIRS * 2;
>   	ret = ifcvf_start_hw(vf);
>   	if (ret < 0) {
>   		status = ifcvf_get_status(vf);
> @@ -128,7 +131,7 @@ static int ifcvf_stop_datapath(void *private)
>   	struct ifcvf_hw *vf = ifcvf_private_to_vf(private);
>   	int i;
>   
> -	for (i = 0; i < IFCVF_MAX_QUEUE_PAIRS * 2; i++)
> +	for (i = 0; i < vf->nr_vring; i++)
>   		vf->vring[i].cb.callback = NULL;
>   
>   	ifcvf_stop_hw(vf);
> @@ -141,7 +144,7 @@ static void ifcvf_reset_vring(struct ifcvf_adapter *adapter)
>   	struct ifcvf_hw *vf = ifcvf_private_to_vf(adapter);
>   	int i;
>   
> -	for (i = 0; i < IFCVF_MAX_QUEUE_PAIRS * 2; i++) {
> +	for (i = 0; i < vf->nr_vring; i++) {
>   		vf->vring[i].last_avail_idx = 0;
>   		vf->vring[i].desc = 0;
>   		vf->vring[i].avail = 0;
> @@ -227,7 +230,7 @@ static void ifcvf_vdpa_set_status(struct vdpa_device *vdpa_dev, u8 status)
>   	if ((status_old & VIRTIO_CONFIG_S_DRIVER_OK) &&
>   	    !(status & VIRTIO_CONFIG_S_DRIVER_OK)) {
>   		ifcvf_stop_datapath(adapter);
> -		ifcvf_free_irq(adapter, IFCVF_MAX_QUEUE_PAIRS * 2);
> +		ifcvf_free_irq(adapter, vf->nr_vring);
>   	}
>   
>   	if (status == 0) {
> @@ -526,13 +529,13 @@ static int ifcvf_vdpa_dev_add(struct vdpa_mgmt_dev *mdev, const char *name)
>   		goto err;
>   	}
>   
> -	for (i = 0; i < IFCVF_MAX_QUEUE_PAIRS * 2; i++)
> +	for (i = 0; i < vf->nr_vring; i++)
>   		vf->vring[i].irq = -EINVAL;
>   
>   	vf->hw_features = ifcvf_get_hw_features(vf);
>   
>   	adapter->vdpa.mdev = &ifcvf_mgmt_dev->mdev;
> -	ret = _vdpa_register_device(&adapter->vdpa, IFCVF_MAX_QUEUE_PAIRS * 2);
> +	ret = _vdpa_register_device(&adapter->vdpa, vf->nr_vring);
>   	if (ret) {
>   		IFCVF_ERR(pdev, "Failed to register to vDPA bus");
>   		goto err;

_______________________________________________
Virtualization mailing list
Virtualization@lists.linux-foundation.org
https://lists.linuxfoundation.org/mailman/listinfo/virtualization

       reply	other threads:[~2021-08-19  4:09 UTC|newest]

Thread overview: 4+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <20210818095714.3220-1-lingshan.zhu@intel.com>
     [not found] ` <20210818095714.3220-2-lingshan.zhu@intel.com>
2021-08-19  4:09   ` Jason Wang [this message]
     [not found] ` <20210818095714.3220-3-lingshan.zhu@intel.com>
2021-08-19  4:09   ` [PATCH 2/2] vDPA/ifcvf: enable multiqueue and control vq Jason Wang
2021-08-19  4:11 ` [PATCH 0/2] " Jason Wang
     [not found]   ` <9e6f6cb0-eaed-9d83-c297-3a89f5cc9efd@intel.com>
2021-08-19  7:15     ` Jason Wang

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=78c25559-0423-7ef6-9cdd-d2b81df111ca@redhat.com \
    --to=jasowang@redhat.com \
    --cc=kvm@vger.kernel.org \
    --cc=lingshan.zhu@intel.com \
    --cc=mst@redhat.com \
    --cc=netdev@vger.kernel.org \
    --cc=virtualization@lists.linux-foundation.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).