* (unknown)
2013-01-06 6:36 [PATCH] tcm_vhost: Use llist for cmd completion list Asias He
@ 2013-01-08 6:52 ` Michael S. Tsirkin
2013-01-12 16:06 ` [PATCH] tcm_vhost: Use llist for cmd completion list Paolo Bonzini
` (2 subsequent siblings)
3 siblings, 0 replies; 5+ messages in thread
From: Michael S. Tsirkin @ 2013-01-08 6:52 UTC (permalink / raw)
To: Asias He; +Cc: kvm, virtualization
On Sun, Jan 06, 2013 at 02:36:13PM +0800, Asias He wrote:
> This drops the cmd completion list spin lock and makes the cmd
> completion queue lock-less.
>
> Signed-off-by: Asias He <asias@redhat.com>
Nicholas, any feedback?
> ---
> drivers/vhost/tcm_vhost.c | 46 +++++++++++++---------------------------------
> drivers/vhost/tcm_vhost.h | 2 +-
> 2 files changed, 14 insertions(+), 34 deletions(-)
>
> diff --git a/drivers/vhost/tcm_vhost.c b/drivers/vhost/tcm_vhost.c
> index b20df5c..3720604 100644
> --- a/drivers/vhost/tcm_vhost.c
> +++ b/drivers/vhost/tcm_vhost.c
> @@ -47,6 +47,7 @@
> #include <linux/vhost.h>
> #include <linux/virtio_net.h> /* TODO vhost.h currently depends on this */
> #include <linux/virtio_scsi.h>
> +#include <linux/llist.h>
>
> #include "vhost.c"
> #include "vhost.h"
> @@ -64,8 +65,7 @@ struct vhost_scsi {
> struct vhost_virtqueue vqs[3];
>
> struct vhost_work vs_completion_work; /* cmd completion work item */
> - struct list_head vs_completion_list; /* cmd completion queue */
> - spinlock_t vs_completion_lock; /* protects s_completion_list */
> + struct llist_head vs_completion_list; /* cmd completion queue */
> };
>
> /* Local pointer to allocated TCM configfs fabric module */
> @@ -301,9 +301,7 @@ static void vhost_scsi_complete_cmd(struct tcm_vhost_cmd *tv_cmd)
> {
> struct vhost_scsi *vs = tv_cmd->tvc_vhost;
>
> - spin_lock_bh(&vs->vs_completion_lock);
> - list_add_tail(&tv_cmd->tvc_completion_list, &vs->vs_completion_list);
> - spin_unlock_bh(&vs->vs_completion_lock);
> + llist_add(&tv_cmd->tvc_completion_list, &vs->vs_completion_list);
>
> vhost_work_queue(&vs->dev, &vs->vs_completion_work);
> }
> @@ -347,27 +345,6 @@ static void vhost_scsi_free_cmd(struct tcm_vhost_cmd *tv_cmd)
> kfree(tv_cmd);
> }
>
> -/* Dequeue a command from the completion list */
> -static struct tcm_vhost_cmd *vhost_scsi_get_cmd_from_completion(
> - struct vhost_scsi *vs)
> -{
> - struct tcm_vhost_cmd *tv_cmd = NULL;
> -
> - spin_lock_bh(&vs->vs_completion_lock);
> - if (list_empty(&vs->vs_completion_list)) {
> - spin_unlock_bh(&vs->vs_completion_lock);
> - return NULL;
> - }
> -
> - list_for_each_entry(tv_cmd, &vs->vs_completion_list,
> - tvc_completion_list) {
> - list_del(&tv_cmd->tvc_completion_list);
> - break;
> - }
> - spin_unlock_bh(&vs->vs_completion_lock);
> - return tv_cmd;
> -}
> -
> /* Fill in status and signal that we are done processing this command
> *
> * This is scheduled in the vhost work queue so we are called with the owner
> @@ -377,12 +354,18 @@ static void vhost_scsi_complete_cmd_work(struct vhost_work *work)
> {
> struct vhost_scsi *vs = container_of(work, struct vhost_scsi,
> vs_completion_work);
> + struct virtio_scsi_cmd_resp v_rsp;
> struct tcm_vhost_cmd *tv_cmd;
> + struct llist_node *llnode;
> + struct se_cmd *se_cmd;
> + int ret;
>
> - while ((tv_cmd = vhost_scsi_get_cmd_from_completion(vs))) {
> - struct virtio_scsi_cmd_resp v_rsp;
> - struct se_cmd *se_cmd = &tv_cmd->tvc_se_cmd;
> - int ret;
> + llnode = llist_del_all(&vs->vs_completion_list);
> + while (llnode) {
> + tv_cmd = llist_entry(llnode, struct tcm_vhost_cmd,
> + tvc_completion_list);
> + llnode = llist_next(llnode);
> + se_cmd = &tv_cmd->tvc_se_cmd;
>
> pr_debug("%s tv_cmd %p resid %u status %#02x\n", __func__,
> tv_cmd, se_cmd->residual_count, se_cmd->scsi_status);
> @@ -426,7 +409,6 @@ static struct tcm_vhost_cmd *vhost_scsi_allocate_cmd(
> pr_err("Unable to allocate struct tcm_vhost_cmd\n");
> return ERR_PTR(-ENOMEM);
> }
> - INIT_LIST_HEAD(&tv_cmd->tvc_completion_list);
> tv_cmd->tvc_tag = v_req->tag;
> tv_cmd->tvc_task_attr = v_req->task_attr;
> tv_cmd->tvc_exp_data_len = exp_data_len;
> @@ -859,8 +841,6 @@ static int vhost_scsi_open(struct inode *inode, struct file *f)
> return -ENOMEM;
>
> vhost_work_init(&s->vs_completion_work, vhost_scsi_complete_cmd_work);
> - INIT_LIST_HEAD(&s->vs_completion_list);
> - spin_lock_init(&s->vs_completion_lock);
>
> s->vqs[VHOST_SCSI_VQ_CTL].handle_kick = vhost_scsi_ctl_handle_kick;
> s->vqs[VHOST_SCSI_VQ_EVT].handle_kick = vhost_scsi_evt_handle_kick;
> diff --git a/drivers/vhost/tcm_vhost.h b/drivers/vhost/tcm_vhost.h
> index 7e87c63..47ee80b 100644
> --- a/drivers/vhost/tcm_vhost.h
> +++ b/drivers/vhost/tcm_vhost.h
> @@ -34,7 +34,7 @@ struct tcm_vhost_cmd {
> /* Sense buffer that will be mapped into outgoing status */
> unsigned char tvc_sense_buf[TRANSPORT_SENSE_BUFFER];
> /* Completed commands list, serviced from vhost worker thread */
> - struct list_head tvc_completion_list;
> + struct llist_node tvc_completion_list;
> };
>
> struct tcm_vhost_nexus {
> --
> 1.7.11.7
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [PATCH] tcm_vhost: Use llist for cmd completion list
2013-01-06 6:36 [PATCH] tcm_vhost: Use llist for cmd completion list Asias He
2013-01-08 6:52 ` (unknown) Michael S. Tsirkin
@ 2013-01-12 16:06 ` Paolo Bonzini
2013-01-17 20:35 ` Nicholas A. Bellinger
[not found] ` <1358454953.18551.9.camel@haakon2.linux-iscsi.org>
3 siblings, 0 replies; 5+ messages in thread
From: Paolo Bonzini @ 2013-01-12 16:06 UTC (permalink / raw)
To: Asias He; +Cc: Michael S. Tsirkin, kvm, virtualization
Il 06/01/2013 07:36, Asias He ha scritto:
> This drops the cmd completion list spin lock and makes the cmd
> completion queue lock-less.
>
> Signed-off-by: Asias He <asias@redhat.com>
> ---
> drivers/vhost/tcm_vhost.c | 46 +++++++++++++---------------------------------
> drivers/vhost/tcm_vhost.h | 2 +-
> 2 files changed, 14 insertions(+), 34 deletions(-)
>
> diff --git a/drivers/vhost/tcm_vhost.c b/drivers/vhost/tcm_vhost.c
> index b20df5c..3720604 100644
> --- a/drivers/vhost/tcm_vhost.c
> +++ b/drivers/vhost/tcm_vhost.c
> @@ -47,6 +47,7 @@
> #include <linux/vhost.h>
> #include <linux/virtio_net.h> /* TODO vhost.h currently depends on this */
> #include <linux/virtio_scsi.h>
> +#include <linux/llist.h>
>
> #include "vhost.c"
> #include "vhost.h"
> @@ -64,8 +65,7 @@ struct vhost_scsi {
> struct vhost_virtqueue vqs[3];
>
> struct vhost_work vs_completion_work; /* cmd completion work item */
> - struct list_head vs_completion_list; /* cmd completion queue */
> - spinlock_t vs_completion_lock; /* protects s_completion_list */
> + struct llist_head vs_completion_list; /* cmd completion queue */
> };
>
> /* Local pointer to allocated TCM configfs fabric module */
> @@ -301,9 +301,7 @@ static void vhost_scsi_complete_cmd(struct tcm_vhost_cmd *tv_cmd)
> {
> struct vhost_scsi *vs = tv_cmd->tvc_vhost;
>
> - spin_lock_bh(&vs->vs_completion_lock);
> - list_add_tail(&tv_cmd->tvc_completion_list, &vs->vs_completion_list);
> - spin_unlock_bh(&vs->vs_completion_lock);
> + llist_add(&tv_cmd->tvc_completion_list, &vs->vs_completion_list);
>
> vhost_work_queue(&vs->dev, &vs->vs_completion_work);
> }
> @@ -347,27 +345,6 @@ static void vhost_scsi_free_cmd(struct tcm_vhost_cmd *tv_cmd)
> kfree(tv_cmd);
> }
>
> -/* Dequeue a command from the completion list */
> -static struct tcm_vhost_cmd *vhost_scsi_get_cmd_from_completion(
> - struct vhost_scsi *vs)
> -{
> - struct tcm_vhost_cmd *tv_cmd = NULL;
> -
> - spin_lock_bh(&vs->vs_completion_lock);
> - if (list_empty(&vs->vs_completion_list)) {
> - spin_unlock_bh(&vs->vs_completion_lock);
> - return NULL;
> - }
> -
> - list_for_each_entry(tv_cmd, &vs->vs_completion_list,
> - tvc_completion_list) {
> - list_del(&tv_cmd->tvc_completion_list);
> - break;
> - }
> - spin_unlock_bh(&vs->vs_completion_lock);
> - return tv_cmd;
> -}
> -
> /* Fill in status and signal that we are done processing this command
> *
> * This is scheduled in the vhost work queue so we are called with the owner
> @@ -377,12 +354,18 @@ static void vhost_scsi_complete_cmd_work(struct vhost_work *work)
> {
> struct vhost_scsi *vs = container_of(work, struct vhost_scsi,
> vs_completion_work);
> + struct virtio_scsi_cmd_resp v_rsp;
> struct tcm_vhost_cmd *tv_cmd;
> + struct llist_node *llnode;
> + struct se_cmd *se_cmd;
> + int ret;
>
> - while ((tv_cmd = vhost_scsi_get_cmd_from_completion(vs))) {
> - struct virtio_scsi_cmd_resp v_rsp;
> - struct se_cmd *se_cmd = &tv_cmd->tvc_se_cmd;
> - int ret;
> + llnode = llist_del_all(&vs->vs_completion_list);
> + while (llnode) {
> + tv_cmd = llist_entry(llnode, struct tcm_vhost_cmd,
> + tvc_completion_list);
> + llnode = llist_next(llnode);
> + se_cmd = &tv_cmd->tvc_se_cmd;
>
> pr_debug("%s tv_cmd %p resid %u status %#02x\n", __func__,
> tv_cmd, se_cmd->residual_count, se_cmd->scsi_status);
> @@ -426,7 +409,6 @@ static struct tcm_vhost_cmd *vhost_scsi_allocate_cmd(
> pr_err("Unable to allocate struct tcm_vhost_cmd\n");
> return ERR_PTR(-ENOMEM);
> }
> - INIT_LIST_HEAD(&tv_cmd->tvc_completion_list);
> tv_cmd->tvc_tag = v_req->tag;
> tv_cmd->tvc_task_attr = v_req->task_attr;
> tv_cmd->tvc_exp_data_len = exp_data_len;
> @@ -859,8 +841,6 @@ static int vhost_scsi_open(struct inode *inode, struct file *f)
> return -ENOMEM;
>
> vhost_work_init(&s->vs_completion_work, vhost_scsi_complete_cmd_work);
> - INIT_LIST_HEAD(&s->vs_completion_list);
> - spin_lock_init(&s->vs_completion_lock);
>
> s->vqs[VHOST_SCSI_VQ_CTL].handle_kick = vhost_scsi_ctl_handle_kick;
> s->vqs[VHOST_SCSI_VQ_EVT].handle_kick = vhost_scsi_evt_handle_kick;
> diff --git a/drivers/vhost/tcm_vhost.h b/drivers/vhost/tcm_vhost.h
> index 7e87c63..47ee80b 100644
> --- a/drivers/vhost/tcm_vhost.h
> +++ b/drivers/vhost/tcm_vhost.h
> @@ -34,7 +34,7 @@ struct tcm_vhost_cmd {
> /* Sense buffer that will be mapped into outgoing status */
> unsigned char tvc_sense_buf[TRANSPORT_SENSE_BUFFER];
> /* Completed commands list, serviced from vhost worker thread */
> - struct list_head tvc_completion_list;
> + struct llist_node tvc_completion_list;
> };
>
> struct tcm_vhost_nexus {
>
Reviewed-by: Paolo Bonzini <pbonzini@redhat.com>
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [PATCH] tcm_vhost: Use llist for cmd completion list
2013-01-06 6:36 [PATCH] tcm_vhost: Use llist for cmd completion list Asias He
2013-01-08 6:52 ` (unknown) Michael S. Tsirkin
2013-01-12 16:06 ` [PATCH] tcm_vhost: Use llist for cmd completion list Paolo Bonzini
@ 2013-01-17 20:35 ` Nicholas A. Bellinger
[not found] ` <1358454953.18551.9.camel@haakon2.linux-iscsi.org>
3 siblings, 0 replies; 5+ messages in thread
From: Nicholas A. Bellinger @ 2013-01-17 20:35 UTC (permalink / raw)
To: Asias He; +Cc: target-devel, virtualization, kvm, Michael S. Tsirkin
Hi Asias!
On Sun, 2013-01-06 at 14:36 +0800, Asias He wrote:
> This drops the cmd completion list spin lock and makes the cmd
> completion queue lock-less.
>
> Signed-off-by: Asias He <asias@redhat.com>
> ---
Apologies for the long delay to get back to this patch.
After some initial testing, I'm seeing about about a ~5K IOPs
performance increase to single RAMDISK_MCP (~110K to ~115K) on the heavy
mixed 75/25 4k randrw fio workload.
That said, I think it's fine to as a for-3.9 item.
Acked-by: Nicholas Bellinger <nab@linux-iscsi.org>
MST, do you want to take this via your vhost tree, or shall I merge via
target-pending/for-next..?
Thanks,
--nab
> drivers/vhost/tcm_vhost.c | 46 +++++++++++++---------------------------------
> drivers/vhost/tcm_vhost.h | 2 +-
> 2 files changed, 14 insertions(+), 34 deletions(-)
>
> diff --git a/drivers/vhost/tcm_vhost.c b/drivers/vhost/tcm_vhost.c
> index b20df5c..3720604 100644
> --- a/drivers/vhost/tcm_vhost.c
> +++ b/drivers/vhost/tcm_vhost.c
> @@ -47,6 +47,7 @@
> #include <linux/vhost.h>
> #include <linux/virtio_net.h> /* TODO vhost.h currently depends on this */
> #include <linux/virtio_scsi.h>
> +#include <linux/llist.h>
>
> #include "vhost.c"
> #include "vhost.h"
> @@ -64,8 +65,7 @@ struct vhost_scsi {
> struct vhost_virtqueue vqs[3];
>
> struct vhost_work vs_completion_work; /* cmd completion work item */
> - struct list_head vs_completion_list; /* cmd completion queue */
> - spinlock_t vs_completion_lock; /* protects s_completion_list */
> + struct llist_head vs_completion_list; /* cmd completion queue */
> };
>
> /* Local pointer to allocated TCM configfs fabric module */
> @@ -301,9 +301,7 @@ static void vhost_scsi_complete_cmd(struct tcm_vhost_cmd *tv_cmd)
> {
> struct vhost_scsi *vs = tv_cmd->tvc_vhost;
>
> - spin_lock_bh(&vs->vs_completion_lock);
> - list_add_tail(&tv_cmd->tvc_completion_list, &vs->vs_completion_list);
> - spin_unlock_bh(&vs->vs_completion_lock);
> + llist_add(&tv_cmd->tvc_completion_list, &vs->vs_completion_list);
>
> vhost_work_queue(&vs->dev, &vs->vs_completion_work);
> }
> @@ -347,27 +345,6 @@ static void vhost_scsi_free_cmd(struct tcm_vhost_cmd *tv_cmd)
> kfree(tv_cmd);
> }
>
> -/* Dequeue a command from the completion list */
> -static struct tcm_vhost_cmd *vhost_scsi_get_cmd_from_completion(
> - struct vhost_scsi *vs)
> -{
> - struct tcm_vhost_cmd *tv_cmd = NULL;
> -
> - spin_lock_bh(&vs->vs_completion_lock);
> - if (list_empty(&vs->vs_completion_list)) {
> - spin_unlock_bh(&vs->vs_completion_lock);
> - return NULL;
> - }
> -
> - list_for_each_entry(tv_cmd, &vs->vs_completion_list,
> - tvc_completion_list) {
> - list_del(&tv_cmd->tvc_completion_list);
> - break;
> - }
> - spin_unlock_bh(&vs->vs_completion_lock);
> - return tv_cmd;
> -}
> -
> /* Fill in status and signal that we are done processing this command
> *
> * This is scheduled in the vhost work queue so we are called with the owner
> @@ -377,12 +354,18 @@ static void vhost_scsi_complete_cmd_work(struct vhost_work *work)
> {
> struct vhost_scsi *vs = container_of(work, struct vhost_scsi,
> vs_completion_work);
> + struct virtio_scsi_cmd_resp v_rsp;
> struct tcm_vhost_cmd *tv_cmd;
> + struct llist_node *llnode;
> + struct se_cmd *se_cmd;
> + int ret;
>
> - while ((tv_cmd = vhost_scsi_get_cmd_from_completion(vs))) {
> - struct virtio_scsi_cmd_resp v_rsp;
> - struct se_cmd *se_cmd = &tv_cmd->tvc_se_cmd;
> - int ret;
> + llnode = llist_del_all(&vs->vs_completion_list);
> + while (llnode) {
> + tv_cmd = llist_entry(llnode, struct tcm_vhost_cmd,
> + tvc_completion_list);
> + llnode = llist_next(llnode);
> + se_cmd = &tv_cmd->tvc_se_cmd;
>
> pr_debug("%s tv_cmd %p resid %u status %#02x\n", __func__,
> tv_cmd, se_cmd->residual_count, se_cmd->scsi_status);
> @@ -426,7 +409,6 @@ static struct tcm_vhost_cmd *vhost_scsi_allocate_cmd(
> pr_err("Unable to allocate struct tcm_vhost_cmd\n");
> return ERR_PTR(-ENOMEM);
> }
> - INIT_LIST_HEAD(&tv_cmd->tvc_completion_list);
> tv_cmd->tvc_tag = v_req->tag;
> tv_cmd->tvc_task_attr = v_req->task_attr;
> tv_cmd->tvc_exp_data_len = exp_data_len;
> @@ -859,8 +841,6 @@ static int vhost_scsi_open(struct inode *inode, struct file *f)
> return -ENOMEM;
>
> vhost_work_init(&s->vs_completion_work, vhost_scsi_complete_cmd_work);
> - INIT_LIST_HEAD(&s->vs_completion_list);
> - spin_lock_init(&s->vs_completion_lock);
>
> s->vqs[VHOST_SCSI_VQ_CTL].handle_kick = vhost_scsi_ctl_handle_kick;
> s->vqs[VHOST_SCSI_VQ_EVT].handle_kick = vhost_scsi_evt_handle_kick;
> diff --git a/drivers/vhost/tcm_vhost.h b/drivers/vhost/tcm_vhost.h
> index 7e87c63..47ee80b 100644
> --- a/drivers/vhost/tcm_vhost.h
> +++ b/drivers/vhost/tcm_vhost.h
> @@ -34,7 +34,7 @@ struct tcm_vhost_cmd {
> /* Sense buffer that will be mapped into outgoing status */
> unsigned char tvc_sense_buf[TRANSPORT_SENSE_BUFFER];
> /* Completed commands list, serviced from vhost worker thread */
> - struct list_head tvc_completion_list;
> + struct llist_node tvc_completion_list;
> };
>
> struct tcm_vhost_nexus {
^ permalink raw reply [flat|nested] 5+ messages in thread
[parent not found: <1358454953.18551.9.camel@haakon2.linux-iscsi.org>]
* Re: [PATCH] tcm_vhost: Use llist for cmd completion list
[not found] ` <1358454953.18551.9.camel@haakon2.linux-iscsi.org>
@ 2013-01-18 0:42 ` Asias He
0 siblings, 0 replies; 5+ messages in thread
From: Asias He @ 2013-01-18 0:42 UTC (permalink / raw)
To: Nicholas A. Bellinger
Cc: target-devel, virtualization, kvm, Michael S. Tsirkin
On 01/18/2013 04:35 AM, Nicholas A. Bellinger wrote:
> Hi Asias!
>
> On Sun, 2013-01-06 at 14:36 +0800, Asias He wrote:
>> This drops the cmd completion list spin lock and makes the cmd
>> completion queue lock-less.
>>
>> Signed-off-by: Asias He <asias@redhat.com>
>> ---
>
> Apologies for the long delay to get back to this patch.
No problem.
>
> After some initial testing, I'm seeing about about a ~5K IOPs
> performance increase to single RAMDISK_MCP (~110K to ~115K) on the heavy
> mixed 75/25 4k randrw fio workload.
> That said, I think it's fine to as a for-3.9 item.
Okay.
> Acked-by: Nicholas Bellinger <nab@linux-iscsi.org>
>
> MST, do you want to take this via your vhost tree, or shall I merge via
> target-pending/for-next..?
>
> Thanks,
>
> --nab
>
>> drivers/vhost/tcm_vhost.c | 46 +++++++++++++---------------------------------
>> drivers/vhost/tcm_vhost.h | 2 +-
>> 2 files changed, 14 insertions(+), 34 deletions(-)
>>
>> diff --git a/drivers/vhost/tcm_vhost.c b/drivers/vhost/tcm_vhost.c
>> index b20df5c..3720604 100644
>> --- a/drivers/vhost/tcm_vhost.c
>> +++ b/drivers/vhost/tcm_vhost.c
>> @@ -47,6 +47,7 @@
>> #include <linux/vhost.h>
>> #include <linux/virtio_net.h> /* TODO vhost.h currently depends on this */
>> #include <linux/virtio_scsi.h>
>> +#include <linux/llist.h>
>>
>> #include "vhost.c"
>> #include "vhost.h"
>> @@ -64,8 +65,7 @@ struct vhost_scsi {
>> struct vhost_virtqueue vqs[3];
>>
>> struct vhost_work vs_completion_work; /* cmd completion work item */
>> - struct list_head vs_completion_list; /* cmd completion queue */
>> - spinlock_t vs_completion_lock; /* protects s_completion_list */
>> + struct llist_head vs_completion_list; /* cmd completion queue */
>> };
>>
>> /* Local pointer to allocated TCM configfs fabric module */
>> @@ -301,9 +301,7 @@ static void vhost_scsi_complete_cmd(struct tcm_vhost_cmd *tv_cmd)
>> {
>> struct vhost_scsi *vs = tv_cmd->tvc_vhost;
>>
>> - spin_lock_bh(&vs->vs_completion_lock);
>> - list_add_tail(&tv_cmd->tvc_completion_list, &vs->vs_completion_list);
>> - spin_unlock_bh(&vs->vs_completion_lock);
>> + llist_add(&tv_cmd->tvc_completion_list, &vs->vs_completion_list);
>>
>> vhost_work_queue(&vs->dev, &vs->vs_completion_work);
>> }
>> @@ -347,27 +345,6 @@ static void vhost_scsi_free_cmd(struct tcm_vhost_cmd *tv_cmd)
>> kfree(tv_cmd);
>> }
>>
>> -/* Dequeue a command from the completion list */
>> -static struct tcm_vhost_cmd *vhost_scsi_get_cmd_from_completion(
>> - struct vhost_scsi *vs)
>> -{
>> - struct tcm_vhost_cmd *tv_cmd = NULL;
>> -
>> - spin_lock_bh(&vs->vs_completion_lock);
>> - if (list_empty(&vs->vs_completion_list)) {
>> - spin_unlock_bh(&vs->vs_completion_lock);
>> - return NULL;
>> - }
>> -
>> - list_for_each_entry(tv_cmd, &vs->vs_completion_list,
>> - tvc_completion_list) {
>> - list_del(&tv_cmd->tvc_completion_list);
>> - break;
>> - }
>> - spin_unlock_bh(&vs->vs_completion_lock);
>> - return tv_cmd;
>> -}
>> -
>> /* Fill in status and signal that we are done processing this command
>> *
>> * This is scheduled in the vhost work queue so we are called with the owner
>> @@ -377,12 +354,18 @@ static void vhost_scsi_complete_cmd_work(struct vhost_work *work)
>> {
>> struct vhost_scsi *vs = container_of(work, struct vhost_scsi,
>> vs_completion_work);
>> + struct virtio_scsi_cmd_resp v_rsp;
>> struct tcm_vhost_cmd *tv_cmd;
>> + struct llist_node *llnode;
>> + struct se_cmd *se_cmd;
>> + int ret;
>>
>> - while ((tv_cmd = vhost_scsi_get_cmd_from_completion(vs))) {
>> - struct virtio_scsi_cmd_resp v_rsp;
>> - struct se_cmd *se_cmd = &tv_cmd->tvc_se_cmd;
>> - int ret;
>> + llnode = llist_del_all(&vs->vs_completion_list);
>> + while (llnode) {
>> + tv_cmd = llist_entry(llnode, struct tcm_vhost_cmd,
>> + tvc_completion_list);
>> + llnode = llist_next(llnode);
>> + se_cmd = &tv_cmd->tvc_se_cmd;
>>
>> pr_debug("%s tv_cmd %p resid %u status %#02x\n", __func__,
>> tv_cmd, se_cmd->residual_count, se_cmd->scsi_status);
>> @@ -426,7 +409,6 @@ static struct tcm_vhost_cmd *vhost_scsi_allocate_cmd(
>> pr_err("Unable to allocate struct tcm_vhost_cmd\n");
>> return ERR_PTR(-ENOMEM);
>> }
>> - INIT_LIST_HEAD(&tv_cmd->tvc_completion_list);
>> tv_cmd->tvc_tag = v_req->tag;
>> tv_cmd->tvc_task_attr = v_req->task_attr;
>> tv_cmd->tvc_exp_data_len = exp_data_len;
>> @@ -859,8 +841,6 @@ static int vhost_scsi_open(struct inode *inode, struct file *f)
>> return -ENOMEM;
>>
>> vhost_work_init(&s->vs_completion_work, vhost_scsi_complete_cmd_work);
>> - INIT_LIST_HEAD(&s->vs_completion_list);
>> - spin_lock_init(&s->vs_completion_lock);
>>
>> s->vqs[VHOST_SCSI_VQ_CTL].handle_kick = vhost_scsi_ctl_handle_kick;
>> s->vqs[VHOST_SCSI_VQ_EVT].handle_kick = vhost_scsi_evt_handle_kick;
>> diff --git a/drivers/vhost/tcm_vhost.h b/drivers/vhost/tcm_vhost.h
>> index 7e87c63..47ee80b 100644
>> --- a/drivers/vhost/tcm_vhost.h
>> +++ b/drivers/vhost/tcm_vhost.h
>> @@ -34,7 +34,7 @@ struct tcm_vhost_cmd {
>> /* Sense buffer that will be mapped into outgoing status */
>> unsigned char tvc_sense_buf[TRANSPORT_SENSE_BUFFER];
>> /* Completed commands list, serviced from vhost worker thread */
>> - struct list_head tvc_completion_list;
>> + struct llist_node tvc_completion_list;
>> };
>>
>> struct tcm_vhost_nexus {
>
>
--
Asias
^ permalink raw reply [flat|nested] 5+ messages in thread