* [PATCH 1/7] blk-mq: initialize resid_len [not found] ` <1397464212-4454-1-git-send-email-hch@lst.de> @ 2014-04-14 8:30 ` Christoph Hellwig 2014-04-14 8:30 ` [PATCH 2/7] blk-mq: do not initialize req->special Christoph Hellwig ` (6 subsequent siblings) 7 siblings, 0 replies; 39+ messages in thread From: Christoph Hellwig @ 2014-04-14 8:30 UTC (permalink / raw) To: Jens Axboe; +Cc: Matias Bjorling, linux-kernel, linux-scsi Signed-off-by: Christoph Hellwig <hch@lst.de> --- block/blk-mq.c | 2 ++ 1 file changed, 2 insertions(+) diff --git a/block/blk-mq.c b/block/blk-mq.c index 1d2a9bd..c1f4736 100644 --- a/block/blk-mq.c +++ b/block/blk-mq.c @@ -350,6 +350,8 @@ static void blk_mq_start_request(struct request *rq, bool last) trace_block_rq_issue(q, rq); + rq->resid_len = blk_rq_bytes(rq); + /* * Just mark start time and set the started bit. Due to memory * ordering, we know we'll see the correct deadline as long as -- 1.7.10.4 ^ permalink raw reply related [flat|nested] 39+ messages in thread
* [PATCH 2/7] blk-mq: do not initialize req->special [not found] ` <1397464212-4454-1-git-send-email-hch@lst.de> 2014-04-14 8:30 ` [PATCH 1/7] blk-mq: initialize resid_len Christoph Hellwig @ 2014-04-14 8:30 ` Christoph Hellwig 2014-04-14 8:30 ` [PATCH 3/7] blk-mq: make ->flush_rq fully transparent to drivers Christoph Hellwig ` (5 subsequent siblings) 7 siblings, 0 replies; 39+ messages in thread From: Christoph Hellwig @ 2014-04-14 8:30 UTC (permalink / raw) To: Jens Axboe; +Cc: Matias Bjorling, linux-kernel, linux-scsi Drivers can reach their private data easily using the blk_mq_rq_to_pdu helper and don't need req->special. By not initializing it code can be simplified nicely, and we also shave off a few more instructions from the I/O path. Signed-off-by: Christoph Hellwig <hch@lst.de> --- block/blk-flush.c | 10 ++-------- block/blk-mq.c | 15 ++------------- block/blk-mq.h | 1 - drivers/block/null_blk.c | 4 ++-- drivers/block/virtio_blk.c | 6 +++--- 5 files changed, 9 insertions(+), 27 deletions(-) diff --git a/block/blk-flush.c b/block/blk-flush.c index 43e6b47..9a0c427 100644 --- a/block/blk-flush.c +++ b/block/blk-flush.c @@ -306,22 +306,16 @@ static bool blk_kick_flush(struct request_queue *q) */ q->flush_pending_idx ^= 1; + blk_rq_init(q, q->flush_rq); if (q->mq_ops) { - struct blk_mq_ctx *ctx = first_rq->mq_ctx; - struct blk_mq_hw_ctx *hctx = q->mq_ops->map_queue(q, ctx->cpu); - - blk_mq_rq_init(hctx, q->flush_rq); - q->flush_rq->mq_ctx = ctx; - /* * Reuse the tag value from the fist waiting request, * with blk-mq the tag is generated during request * allocation and drivers can rely on it being inside * the range they asked for. */ + q->flush_rq->mq_ctx = first_rq->mq_ctx; q->flush_rq->tag = first_rq->tag; - } else { - blk_rq_init(q, q->flush_rq); } q->flush_rq->cmd_type = REQ_TYPE_FS; diff --git a/block/blk-mq.c b/block/blk-mq.c index c1f4736..747feb9 100644 --- a/block/blk-mq.c +++ b/block/blk-mq.c @@ -248,24 +248,13 @@ struct request *blk_mq_alloc_reserved_request(struct request_queue *q, int rw, } EXPORT_SYMBOL(blk_mq_alloc_reserved_request); -/* - * Re-init and set pdu, if we have it - */ -void blk_mq_rq_init(struct blk_mq_hw_ctx *hctx, struct request *rq) -{ - blk_rq_init(hctx->queue, rq); - - if (hctx->cmd_size) - rq->special = blk_mq_rq_to_pdu(rq); -} - static void __blk_mq_free_request(struct blk_mq_hw_ctx *hctx, struct blk_mq_ctx *ctx, struct request *rq) { const int tag = rq->tag; struct request_queue *q = rq->q; - blk_mq_rq_init(hctx, rq); + blk_rq_init(hctx->queue, rq); blk_mq_put_tag(hctx->tags, tag); blk_mq_queue_exit(q); @@ -1143,7 +1132,7 @@ static int blk_mq_init_rq_map(struct blk_mq_hw_ctx *hctx, left -= to_do * rq_size; for (j = 0; j < to_do; j++) { hctx->rqs[i] = p; - blk_mq_rq_init(hctx, hctx->rqs[i]); + blk_rq_init(hctx->queue, hctx->rqs[i]); p += rq_size; i++; } diff --git a/block/blk-mq.h b/block/blk-mq.h index ebbe6ba..238379a 100644 --- a/block/blk-mq.h +++ b/block/blk-mq.h @@ -27,7 +27,6 @@ void blk_mq_run_hw_queue(struct blk_mq_hw_ctx *hctx, bool async); void blk_mq_init_flush(struct request_queue *q); void blk_mq_drain_queue(struct request_queue *q); void blk_mq_free_queue(struct request_queue *q); -void blk_mq_rq_init(struct blk_mq_hw_ctx *hctx, struct request *rq); /* * CPU hotplug helpers diff --git a/drivers/block/null_blk.c b/drivers/block/null_blk.c index 091b9ea..71df69d 100644 --- a/drivers/block/null_blk.c +++ b/drivers/block/null_blk.c @@ -226,7 +226,7 @@ static void null_cmd_end_timer(struct nullb_cmd *cmd) static void null_softirq_done_fn(struct request *rq) { - end_cmd(rq->special); + end_cmd(blk_mq_rq_to_pdu(rq)); } static inline void null_handle_cmd(struct nullb_cmd *cmd) @@ -311,7 +311,7 @@ static void null_request_fn(struct request_queue *q) static int null_queue_rq(struct blk_mq_hw_ctx *hctx, struct request *rq) { - struct nullb_cmd *cmd = rq->special; + struct nullb_cmd *cmd = blk_mq_rq_to_pdu(rq); cmd->rq = rq; cmd->nq = hctx->driver_data; diff --git a/drivers/block/virtio_blk.c b/drivers/block/virtio_blk.c index 6d8a87f..c7d02bc 100644 --- a/drivers/block/virtio_blk.c +++ b/drivers/block/virtio_blk.c @@ -112,7 +112,7 @@ static int __virtblk_add_req(struct virtqueue *vq, static inline void virtblk_request_done(struct request *req) { - struct virtblk_req *vbr = req->special; + struct virtblk_req *vbr = blk_mq_rq_to_pdu(req); int error = virtblk_result(vbr); if (req->cmd_type == REQ_TYPE_BLOCK_PC) { @@ -154,7 +154,7 @@ static void virtblk_done(struct virtqueue *vq) static int virtio_queue_rq(struct blk_mq_hw_ctx *hctx, struct request *req) { struct virtio_blk *vblk = hctx->queue->queuedata; - struct virtblk_req *vbr = req->special; + struct virtblk_req *vbr = blk_mq_rq_to_pdu(req); unsigned long flags; unsigned int num; const bool last = (req->cmd_flags & REQ_END) != 0; @@ -501,7 +501,7 @@ static int virtblk_init_vbr(void *data, struct blk_mq_hw_ctx *hctx, struct request *rq, unsigned int nr) { struct virtio_blk *vblk = data; - struct virtblk_req *vbr = rq->special; + struct virtblk_req *vbr = blk_mq_rq_to_pdu(rq); sg_init_table(vbr->sg, vblk->sg_elems); return 0; -- 1.7.10.4 ^ permalink raw reply related [flat|nested] 39+ messages in thread
* [PATCH 3/7] blk-mq: make ->flush_rq fully transparent to drivers [not found] ` <1397464212-4454-1-git-send-email-hch@lst.de> 2014-04-14 8:30 ` [PATCH 1/7] blk-mq: initialize resid_len Christoph Hellwig 2014-04-14 8:30 ` [PATCH 2/7] blk-mq: do not initialize req->special Christoph Hellwig @ 2014-04-14 8:30 ` Christoph Hellwig 2014-04-14 8:30 ` [PATCH 4/7] blk-mq: add ->init_request and ->exit_request methods Christoph Hellwig ` (4 subsequent siblings) 7 siblings, 0 replies; 39+ messages in thread From: Christoph Hellwig @ 2014-04-14 8:30 UTC (permalink / raw) To: Jens Axboe; +Cc: Matias Bjorling, linux-kernel, linux-scsi Drivers shouldn't have to care about the block layer setting aside a request to implement the flush state machine. We already override the mq context and tag to make it more transparent, but so far haven't deal with the driver private data in the request. Make sure to override this as well, and while we're at it add a proper helper sitting in blk-mq.c that implements the full impersonation. Signed-off-by: Christoph Hellwig <hch@lst.de> --- block/blk-flush.c | 12 ++---------- block/blk-mq.c | 20 ++++++++++++++++++++ block/blk-mq.h | 2 ++ 3 files changed, 24 insertions(+), 10 deletions(-) diff --git a/block/blk-flush.c b/block/blk-flush.c index 9a0c427..b218f61 100644 --- a/block/blk-flush.c +++ b/block/blk-flush.c @@ -307,16 +307,8 @@ static bool blk_kick_flush(struct request_queue *q) q->flush_pending_idx ^= 1; blk_rq_init(q, q->flush_rq); - if (q->mq_ops) { - /* - * Reuse the tag value from the fist waiting request, - * with blk-mq the tag is generated during request - * allocation and drivers can rely on it being inside - * the range they asked for. - */ - q->flush_rq->mq_ctx = first_rq->mq_ctx; - q->flush_rq->tag = first_rq->tag; - } + if (q->mq_ops) + blk_mq_clone_flush_request(q->flush_rq, first_rq); q->flush_rq->cmd_type = REQ_TYPE_FS; q->flush_rq->cmd_flags = WRITE_FLUSH | REQ_FLUSH_SEQ; diff --git a/block/blk-mq.c b/block/blk-mq.c index 747feb9..67859e9 100644 --- a/block/blk-mq.c +++ b/block/blk-mq.c @@ -272,6 +272,26 @@ void blk_mq_free_request(struct request *rq) __blk_mq_free_request(hctx, ctx, rq); } +/* + * Clone all relevant state from a request that has been put on hold in + * the flush state machine into the preallocated flush request that hangs + * off the request queue. + * + * For a driver the flush request should be invisible, that's why we are + * impersonating the original request here. + */ +void blk_mq_clone_flush_request(struct request *flush_rq, + struct request *orig_rq) +{ + struct blk_mq_hw_ctx *hctx = + orig_rq->q->mq_ops->map_queue(orig_rq->q, orig_rq->mq_ctx->cpu); + + flush_rq->mq_ctx = orig_rq->mq_ctx; + flush_rq->tag = orig_rq->tag; + memcpy(blk_mq_rq_to_pdu(flush_rq), blk_mq_rq_to_pdu(orig_rq), + hctx->cmd_size); +} + bool blk_mq_end_io_partial(struct request *rq, int error, unsigned int nr_bytes) { if (blk_update_request(rq, error, blk_rq_bytes(rq))) diff --git a/block/blk-mq.h b/block/blk-mq.h index 238379a..7964dad 100644 --- a/block/blk-mq.h +++ b/block/blk-mq.h @@ -27,6 +27,8 @@ void blk_mq_run_hw_queue(struct blk_mq_hw_ctx *hctx, bool async); void blk_mq_init_flush(struct request_queue *q); void blk_mq_drain_queue(struct request_queue *q); void blk_mq_free_queue(struct request_queue *q); +void blk_mq_clone_flush_request(struct request *flush_rq, + struct request *orig_rq); /* * CPU hotplug helpers -- 1.7.10.4 ^ permalink raw reply related [flat|nested] 39+ messages in thread
* [PATCH 4/7] blk-mq: add ->init_request and ->exit_request methods [not found] ` <1397464212-4454-1-git-send-email-hch@lst.de> ` (2 preceding siblings ...) 2014-04-14 8:30 ` [PATCH 3/7] blk-mq: make ->flush_rq fully transparent to drivers Christoph Hellwig @ 2014-04-14 8:30 ` Christoph Hellwig 2014-04-14 8:30 ` [PATCH 5/7] blk-mq: initialize request on allocation Christoph Hellwig ` (3 subsequent siblings) 7 siblings, 0 replies; 39+ messages in thread From: Christoph Hellwig @ 2014-04-14 8:30 UTC (permalink / raw) To: Jens Axboe; +Cc: Matias Bjorling, linux-kernel, linux-scsi The current blk_mq_init_commands/blk_mq_free_commands interface has a two problems: 1) Because only the constructor is passed to blk_mq_init_commands there is no easy way to clean up when a comman initialization failed. The current code simply leaks the allocations done in the constructor. 2) There is no good place to call blk_mq_free_commands: before blk_cleanup_queue there is no guarantee that all outstanding commands have completed, so we can't free them yet. After blk_cleanup_queue the queue has usually been freed. This can be worked around by grabbing an unconditional reference before calling blk_cleanup_queue and dropping it after blk_mq_free_commands is done, although that's not exatly pretty and driver writers are guaranteed to get it wrong sooner or later. Both issues are easily fixed by making the request constructor and destructor normal blk_mq_ops methods. Signed-off-by: Christoph Hellwig <hch@lst.de> --- block/blk-mq.c | 105 ++++++++++++++------------------------------ drivers/block/virtio_blk.c | 23 +++++----- include/linux/blk-mq.h | 14 +++++- 3 files changed, 55 insertions(+), 87 deletions(-) diff --git a/block/blk-mq.c b/block/blk-mq.c index 67859e9..ef7acc0 100644 --- a/block/blk-mq.c +++ b/block/blk-mq.c @@ -1009,74 +1009,20 @@ static void blk_mq_hctx_notify(void *data, unsigned long action, blk_mq_run_hw_queue(hctx, true); } -static int blk_mq_init_hw_commands(struct blk_mq_hw_ctx *hctx, - int (*init)(void *, struct blk_mq_hw_ctx *, - struct request *, unsigned int), - void *data) +static void blk_mq_free_rq_map(struct blk_mq_hw_ctx *hctx, void *driver_data) { - unsigned int i; - int ret = 0; - - for (i = 0; i < hctx->queue_depth; i++) { - struct request *rq = hctx->rqs[i]; - - ret = init(data, hctx, rq, i); - if (ret) - break; - } - - return ret; -} - -int blk_mq_init_commands(struct request_queue *q, - int (*init)(void *, struct blk_mq_hw_ctx *, - struct request *, unsigned int), - void *data) -{ - struct blk_mq_hw_ctx *hctx; - unsigned int i; - int ret = 0; - - queue_for_each_hw_ctx(q, hctx, i) { - ret = blk_mq_init_hw_commands(hctx, init, data); - if (ret) - break; - } - - return ret; -} -EXPORT_SYMBOL(blk_mq_init_commands); - -static void blk_mq_free_hw_commands(struct blk_mq_hw_ctx *hctx, - void (*free)(void *, struct blk_mq_hw_ctx *, - struct request *, unsigned int), - void *data) -{ - unsigned int i; + struct page *page; - for (i = 0; i < hctx->queue_depth; i++) { - struct request *rq = hctx->rqs[i]; + if (hctx->rqs && hctx->queue->mq_ops->exit_request) { + int i; - free(data, hctx, rq, i); + for (i = 0; i < hctx->queue_depth; i++) { + if (!hctx->rqs[i]) + continue; + hctx->queue->mq_ops->exit_request(driver_data, hctx, + hctx->rqs[i], i); + } } -} - -void blk_mq_free_commands(struct request_queue *q, - void (*free)(void *, struct blk_mq_hw_ctx *, - struct request *, unsigned int), - void *data) -{ - struct blk_mq_hw_ctx *hctx; - unsigned int i; - - queue_for_each_hw_ctx(q, hctx, i) - blk_mq_free_hw_commands(hctx, free, data); -} -EXPORT_SYMBOL(blk_mq_free_commands); - -static void blk_mq_free_rq_map(struct blk_mq_hw_ctx *hctx) -{ - struct page *page; while (!list_empty(&hctx->page_list)) { page = list_first_entry(&hctx->page_list, struct page, lru); @@ -1101,10 +1047,12 @@ static size_t order_to_size(unsigned int order) } static int blk_mq_init_rq_map(struct blk_mq_hw_ctx *hctx, - unsigned int reserved_tags, int node) + struct blk_mq_reg *reg, void *driver_data, int node) { + unsigned int reserved_tags = reg->reserved_tags; unsigned int i, j, entries_per_page, max_order = 4; size_t rq_size, left; + int error; INIT_LIST_HEAD(&hctx->page_list); @@ -1153,14 +1101,23 @@ static int blk_mq_init_rq_map(struct blk_mq_hw_ctx *hctx, for (j = 0; j < to_do; j++) { hctx->rqs[i] = p; blk_rq_init(hctx->queue, hctx->rqs[i]); + if (reg->ops->init_request) { + error = reg->ops->init_request(driver_data, + hctx, hctx->rqs[i], i); + if (error) + goto err_rq_map; + } + p += rq_size; i++; } } - if (i < (reserved_tags + BLK_MQ_TAG_MIN)) + if (i < (reserved_tags + BLK_MQ_TAG_MIN)) { + error = -ENOMEM; goto err_rq_map; - else if (i != hctx->queue_depth) { + } + if (i != hctx->queue_depth) { hctx->queue_depth = i; pr_warn("%s: queue depth set to %u because of low memory\n", __func__, i); @@ -1168,12 +1125,14 @@ static int blk_mq_init_rq_map(struct blk_mq_hw_ctx *hctx, hctx->tags = blk_mq_init_tags(hctx->queue_depth, reserved_tags, node); if (!hctx->tags) { -err_rq_map: - blk_mq_free_rq_map(hctx); - return -ENOMEM; + error = -ENOMEM; + goto err_rq_map; } return 0; +err_rq_map: + blk_mq_free_rq_map(hctx, driver_data); + return error; } static int blk_mq_init_hw_queues(struct request_queue *q, @@ -1206,7 +1165,7 @@ static int blk_mq_init_hw_queues(struct request_queue *q, blk_mq_hctx_notify, hctx); blk_mq_register_cpu_notifier(&hctx->cpu_notifier); - if (blk_mq_init_rq_map(hctx, reg->reserved_tags, node)) + if (blk_mq_init_rq_map(hctx, reg, driver_data, node)) break; /* @@ -1246,7 +1205,7 @@ static int blk_mq_init_hw_queues(struct request_queue *q, reg->ops->exit_hctx(hctx, j); blk_mq_unregister_cpu_notifier(&hctx->cpu_notifier); - blk_mq_free_rq_map(hctx); + blk_mq_free_rq_map(hctx, driver_data); kfree(hctx->ctxs); } @@ -1423,7 +1382,7 @@ void blk_mq_free_queue(struct request_queue *q) queue_for_each_hw_ctx(q, hctx, i) { kfree(hctx->ctx_map); kfree(hctx->ctxs); - blk_mq_free_rq_map(hctx); + blk_mq_free_rq_map(hctx, q->queuedata); blk_mq_unregister_cpu_notifier(&hctx->cpu_notifier); if (q->mq_ops->exit_hctx) q->mq_ops->exit_hctx(hctx, i); diff --git a/drivers/block/virtio_blk.c b/drivers/block/virtio_blk.c index c7d02bc..d06206a 100644 --- a/drivers/block/virtio_blk.c +++ b/drivers/block/virtio_blk.c @@ -480,11 +480,22 @@ static const struct device_attribute dev_attr_cache_type_rw = __ATTR(cache_type, S_IRUGO|S_IWUSR, virtblk_cache_type_show, virtblk_cache_type_store); +static int virtblk_init_request(void *data, struct blk_mq_hw_ctx *hctx, + struct request *rq, unsigned int nr) +{ + struct virtio_blk *vblk = data; + struct virtblk_req *vbr = blk_mq_rq_to_pdu(rq); + + sg_init_table(vbr->sg, vblk->sg_elems); + return 0; +} + static struct blk_mq_ops virtio_mq_ops = { .queue_rq = virtio_queue_rq, .map_queue = blk_mq_map_queue, .alloc_hctx = blk_mq_alloc_single_hw_queue, .free_hctx = blk_mq_free_single_hw_queue, + .init_request = virtblk_init_request, .complete = virtblk_request_done, }; @@ -497,16 +508,6 @@ static struct blk_mq_reg virtio_mq_reg = { }; module_param_named(queue_depth, virtio_mq_reg.queue_depth, uint, 0444); -static int virtblk_init_vbr(void *data, struct blk_mq_hw_ctx *hctx, - struct request *rq, unsigned int nr) -{ - struct virtio_blk *vblk = data; - struct virtblk_req *vbr = blk_mq_rq_to_pdu(rq); - - sg_init_table(vbr->sg, vblk->sg_elems); - return 0; -} - static int virtblk_probe(struct virtio_device *vdev) { struct virtio_blk *vblk; @@ -577,8 +578,6 @@ static int virtblk_probe(struct virtio_device *vdev) goto out_put_disk; } - blk_mq_init_commands(q, virtblk_init_vbr, vblk); - q->queuedata = vblk; virtblk_name_format("vd", index, vblk->disk->disk_name, DISK_NAME_LEN); diff --git a/include/linux/blk-mq.h b/include/linux/blk-mq.h index 0120451..897ca1a 100644 --- a/include/linux/blk-mq.h +++ b/include/linux/blk-mq.h @@ -66,6 +66,10 @@ typedef struct blk_mq_hw_ctx *(alloc_hctx_fn)(struct blk_mq_reg *,unsigned int); typedef void (free_hctx_fn)(struct blk_mq_hw_ctx *, unsigned int); typedef int (init_hctx_fn)(struct blk_mq_hw_ctx *, void *, unsigned int); typedef void (exit_hctx_fn)(struct blk_mq_hw_ctx *, unsigned int); +typedef int (init_request_fn)(void *, struct blk_mq_hw_ctx *, + struct request *, unsigned int); +typedef void (exit_request_fn)(void *, struct blk_mq_hw_ctx *, + struct request *, unsigned int); struct blk_mq_ops { /* @@ -98,6 +102,14 @@ struct blk_mq_ops { */ init_hctx_fn *init_hctx; exit_hctx_fn *exit_hctx; + + /* + * Called for every command allocated by the block layer to allow + * the driver to set up driver specific data. + * Ditto for exit/teardown. + */ + init_request_fn *init_request; + exit_request_fn *exit_request; }; enum { @@ -117,8 +129,6 @@ enum { struct request_queue *blk_mq_init_queue(struct blk_mq_reg *, void *); int blk_mq_register_disk(struct gendisk *); void blk_mq_unregister_disk(struct gendisk *); -int blk_mq_init_commands(struct request_queue *, int (*init)(void *data, struct blk_mq_hw_ctx *, struct request *, unsigned int), void *data); -void blk_mq_free_commands(struct request_queue *, void (*free)(void *data, struct blk_mq_hw_ctx *, struct request *, unsigned int), void *data); void blk_mq_flush_plug_list(struct blk_plug *plug, bool from_schedule); -- 1.7.10.4 ^ permalink raw reply related [flat|nested] 39+ messages in thread
* [PATCH 5/7] blk-mq: initialize request on allocation [not found] ` <1397464212-4454-1-git-send-email-hch@lst.de> ` (3 preceding siblings ...) 2014-04-14 8:30 ` [PATCH 4/7] blk-mq: add ->init_request and ->exit_request methods Christoph Hellwig @ 2014-04-14 8:30 ` Christoph Hellwig 2014-04-17 14:54 ` Ming Lei 2014-04-14 8:30 ` [PATCH 6/7] blk-mq: split out tag initialization, support shared tags Christoph Hellwig ` (2 subsequent siblings) 7 siblings, 1 reply; 39+ messages in thread From: Christoph Hellwig @ 2014-04-14 8:30 UTC (permalink / raw) To: Jens Axboe; +Cc: Matias Bjorling, linux-kernel, linux-scsi If we want to share tag and request allocation between queues we cannot initialize the request at init/free time, but need to initialize it at allocation time as it might get used for different queues over its lifetime. Signed-off-by: Christoph Hellwig <hch@lst.de> --- block/blk-mq.c | 4 +--- 1 file changed, 1 insertion(+), 3 deletions(-) diff --git a/block/blk-mq.c b/block/blk-mq.c index ef7acc0..1cb7618 100644 --- a/block/blk-mq.c +++ b/block/blk-mq.c @@ -82,6 +82,7 @@ static struct request *__blk_mq_alloc_request(struct blk_mq_hw_ctx *hctx, tag = blk_mq_get_tag(hctx->tags, gfp, reserved); if (tag != BLK_MQ_TAG_FAIL) { rq = hctx->rqs[tag]; + blk_rq_init(hctx->queue, rq); rq->tag = tag; return rq; @@ -254,9 +255,7 @@ static void __blk_mq_free_request(struct blk_mq_hw_ctx *hctx, const int tag = rq->tag; struct request_queue *q = rq->q; - blk_rq_init(hctx->queue, rq); blk_mq_put_tag(hctx->tags, tag); - blk_mq_queue_exit(q); } @@ -1100,7 +1099,6 @@ static int blk_mq_init_rq_map(struct blk_mq_hw_ctx *hctx, left -= to_do * rq_size; for (j = 0; j < to_do; j++) { hctx->rqs[i] = p; - blk_rq_init(hctx->queue, hctx->rqs[i]); if (reg->ops->init_request) { error = reg->ops->init_request(driver_data, hctx, hctx->rqs[i], i); -- 1.7.10.4 ^ permalink raw reply related [flat|nested] 39+ messages in thread
* Re: [PATCH 5/7] blk-mq: initialize request on allocation 2014-04-14 8:30 ` [PATCH 5/7] blk-mq: initialize request on allocation Christoph Hellwig @ 2014-04-17 14:54 ` Ming Lei 2014-04-17 14:57 ` Christoph Hellwig 0 siblings, 1 reply; 39+ messages in thread From: Ming Lei @ 2014-04-17 14:54 UTC (permalink / raw) To: Christoph Hellwig Cc: Jens Axboe, Matias Bjorling, Linux Kernel Mailing List, Linux SCSI List On Mon, Apr 14, 2014 at 4:30 PM, Christoph Hellwig <hch@lst.de> wrote: > If we want to share tag and request allocation between queues we cannot > initialize the request at init/free time, but need to initialize it > at allocation time as it might get used for different queues over its > lifetime. Could you explain the use pattern? Looks you mean there are still users of the tag/req even after it is freed, that looks a bit weird since the tag/req can still be reallocated in another path after it is freed. > > Signed-off-by: Christoph Hellwig <hch@lst.de> > --- > block/blk-mq.c | 4 +--- > 1 file changed, 1 insertion(+), 3 deletions(-) > > diff --git a/block/blk-mq.c b/block/blk-mq.c > index ef7acc0..1cb7618 100644 > --- a/block/blk-mq.c > +++ b/block/blk-mq.c > @@ -82,6 +82,7 @@ static struct request *__blk_mq_alloc_request(struct blk_mq_hw_ctx *hctx, > tag = blk_mq_get_tag(hctx->tags, gfp, reserved); > if (tag != BLK_MQ_TAG_FAIL) { > rq = hctx->rqs[tag]; > + blk_rq_init(hctx->queue, rq); > rq->tag = tag; > > return rq; > @@ -254,9 +255,7 @@ static void __blk_mq_free_request(struct blk_mq_hw_ctx *hctx, > const int tag = rq->tag; > struct request_queue *q = rq->q; > > - blk_rq_init(hctx->queue, rq); > blk_mq_put_tag(hctx->tags, tag); > - > blk_mq_queue_exit(q); > } > > @@ -1100,7 +1099,6 @@ static int blk_mq_init_rq_map(struct blk_mq_hw_ctx *hctx, > left -= to_do * rq_size; > for (j = 0; j < to_do; j++) { > hctx->rqs[i] = p; > - blk_rq_init(hctx->queue, hctx->rqs[i]); > if (reg->ops->init_request) { > error = reg->ops->init_request(driver_data, > hctx, hctx->rqs[i], i); > -- > 1.7.10.4 > > -- > To unsubscribe from this list: send the line "unsubscribe linux-kernel" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html > Please read the FAQ at http://www.tux.org/lkml/ Thanks, -- Ming Lei ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: [PATCH 5/7] blk-mq: initialize request on allocation 2014-04-17 14:54 ` Ming Lei @ 2014-04-17 14:57 ` Christoph Hellwig 2014-04-17 15:07 ` Ming Lei 0 siblings, 1 reply; 39+ messages in thread From: Christoph Hellwig @ 2014-04-17 14:57 UTC (permalink / raw) To: Ming Lei Cc: Christoph Hellwig, Jens Axboe, Matias Bjorling, Linux Kernel Mailing List, Linux SCSI List On Thu, Apr 17, 2014 at 10:54:23PM +0800, Ming Lei wrote: > On Mon, Apr 14, 2014 at 4:30 PM, Christoph Hellwig <hch@lst.de> wrote: > > If we want to share tag and request allocation between queues we cannot > > initialize the request at init/free time, but need to initialize it > > at allocation time as it might get used for different queues over its > > lifetime. > > Could you explain the use pattern? Looks you mean there are > still users of the tag/req even after it is freed, that looks a bit > weird since the tag/req can still be reallocated in another path > after it is freed. No difference in use pattern. But blk_rq_init initializes the rq->q field, and a request might get reused for a different queue. ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: [PATCH 5/7] blk-mq: initialize request on allocation 2014-04-17 14:57 ` Christoph Hellwig @ 2014-04-17 15:07 ` Ming Lei 0 siblings, 0 replies; 39+ messages in thread From: Ming Lei @ 2014-04-17 15:07 UTC (permalink / raw) To: Christoph Hellwig Cc: Jens Axboe, Matias Bjorling, Linux Kernel Mailing List, Linux SCSI List On Thu, Apr 17, 2014 at 10:57 PM, Christoph Hellwig <hch@lst.de> wrote: > On Thu, Apr 17, 2014 at 10:54:23PM +0800, Ming Lei wrote: >> On Mon, Apr 14, 2014 at 4:30 PM, Christoph Hellwig <hch@lst.de> wrote: >> > If we want to share tag and request allocation between queues we cannot >> > initialize the request at init/free time, but need to initialize it >> > at allocation time as it might get used for different queues over its >> > lifetime. >> >> Could you explain the use pattern? Looks you mean there are >> still users of the tag/req even after it is freed, that looks a bit >> weird since the tag/req can still be reallocated in another path >> after it is freed. > > No difference in use pattern. But blk_rq_init initializes the rq->q field, > and a request might get reused for a different queue. If so, it may be better to only initialize the rq->q in allocation because Jens said 'It's done that way because of presumed cache hotness on completion'. Thanks, -- Ming Lei ^ permalink raw reply [flat|nested] 39+ messages in thread
* [PATCH 6/7] blk-mq: split out tag initialization, support shared tags [not found] ` <1397464212-4454-1-git-send-email-hch@lst.de> ` (4 preceding siblings ...) 2014-04-14 8:30 ` [PATCH 5/7] blk-mq: initialize request on allocation Christoph Hellwig @ 2014-04-14 8:30 ` Christoph Hellwig 2014-04-14 8:30 ` [PATCH 7/7] block: all blk-mq requests are tagged Christoph Hellwig 2014-04-15 20:16 ` Jens Axboe 7 siblings, 0 replies; 39+ messages in thread From: Christoph Hellwig @ 2014-04-14 8:30 UTC (permalink / raw) To: Jens Axboe; +Cc: Matias Bjorling, linux-kernel, linux-scsi Add a new blk_mq_tag_set structure that gets set up before we initialize the queue. A single blk_mq_tag_set structure can be shared by multiple queues. Signed-off-by: Christoph Hellwig <hch@lst.de> --- block/blk-mq-cpumap.c | 6 +- block/blk-mq-tag.c | 14 --- block/blk-mq-tag.h | 19 +++- block/blk-mq.c | 242 ++++++++++++++++++++++++-------------------- block/blk-mq.h | 5 +- drivers/block/null_blk.c | 92 ++++++++++------- drivers/block/virtio_blk.c | 48 +++++---- include/linux/blk-mq.h | 34 +++---- 8 files changed, 260 insertions(+), 200 deletions(-) diff --git a/block/blk-mq-cpumap.c b/block/blk-mq-cpumap.c index 0979213..5d0f93c 100644 --- a/block/blk-mq-cpumap.c +++ b/block/blk-mq-cpumap.c @@ -80,17 +80,17 @@ int blk_mq_update_queue_map(unsigned int *map, unsigned int nr_queues) return 0; } -unsigned int *blk_mq_make_queue_map(struct blk_mq_reg *reg) +unsigned int *blk_mq_make_queue_map(struct blk_mq_tag_set *set) { unsigned int *map; /* If cpus are offline, map them to first hctx */ map = kzalloc_node(sizeof(*map) * num_possible_cpus(), GFP_KERNEL, - reg->numa_node); + set->numa_node); if (!map) return NULL; - if (!blk_mq_update_queue_map(map, reg->nr_hw_queues)) + if (!blk_mq_update_queue_map(map, set->nr_hw_queues)) return map; kfree(map); diff --git a/block/blk-mq-tag.c b/block/blk-mq-tag.c index 83ae96c..7a799c4 100644 --- a/block/blk-mq-tag.c +++ b/block/blk-mq-tag.c @@ -1,25 +1,11 @@ #include <linux/kernel.h> #include <linux/module.h> -#include <linux/percpu_ida.h> #include <linux/blk-mq.h> #include "blk.h" #include "blk-mq.h" #include "blk-mq-tag.h" -/* - * Per tagged queue (tag address space) map - */ -struct blk_mq_tags { - unsigned int nr_tags; - unsigned int nr_reserved_tags; - unsigned int nr_batch_move; - unsigned int nr_max_cache; - - struct percpu_ida free_tags; - struct percpu_ida reserved_tags; -}; - void blk_mq_wait_for_tags(struct blk_mq_tags *tags) { int tag = blk_mq_get_tag(tags, __GFP_WAIT, false); diff --git a/block/blk-mq-tag.h b/block/blk-mq-tag.h index 947ba2c..b602e3f 100644 --- a/block/blk-mq-tag.h +++ b/block/blk-mq-tag.h @@ -1,7 +1,24 @@ #ifndef INT_BLK_MQ_TAG_H #define INT_BLK_MQ_TAG_H -struct blk_mq_tags; +#include <linux/percpu_ida.h> + +/* + * Tag address space map. + */ +struct blk_mq_tags { + unsigned int nr_tags; + unsigned int nr_reserved_tags; + unsigned int nr_batch_move; + unsigned int nr_max_cache; + + struct percpu_ida free_tags; + struct percpu_ida reserved_tags; + + struct request **rqs; + struct list_head page_list; +}; + extern struct blk_mq_tags *blk_mq_init_tags(unsigned int nr_tags, unsigned int reserved_tags, int node); extern void blk_mq_free_tags(struct blk_mq_tags *tags); diff --git a/block/blk-mq.c b/block/blk-mq.c index 1cb7618..c3b1810 100644 --- a/block/blk-mq.c +++ b/block/blk-mq.c @@ -81,7 +81,7 @@ static struct request *__blk_mq_alloc_request(struct blk_mq_hw_ctx *hctx, tag = blk_mq_get_tag(hctx->tags, gfp, reserved); if (tag != BLK_MQ_TAG_FAIL) { - rq = hctx->rqs[tag]; + rq = hctx->tags->rqs[tag]; blk_rq_init(hctx->queue, rq); rq->tag = tag; @@ -401,6 +401,12 @@ static void blk_mq_requeue_request(struct request *rq) rq->nr_phys_segments--; } +struct request *blk_mq_tag_to_rq(struct blk_mq_tags *tags, unsigned int tag) +{ + return tags->rqs[tag]; +} +EXPORT_SYMBOL(blk_mq_tag_to_rq); + struct blk_mq_timeout_data { struct blk_mq_hw_ctx *hctx; unsigned long *next; @@ -422,12 +428,13 @@ static void blk_mq_timeout_check(void *__data, unsigned long *free_tags) do { struct request *rq; - tag = find_next_zero_bit(free_tags, hctx->queue_depth, tag); - if (tag >= hctx->queue_depth) + tag = find_next_zero_bit(free_tags, hctx->tags->nr_tags, tag); + if (tag >= hctx->tags->nr_tags) break; - rq = hctx->rqs[tag++]; - + rq = blk_mq_tag_to_rq(hctx->tags, tag++); + if (rq->q != hctx->queue) + continue; if (!test_bit(REQ_ATOM_STARTED, &rq->atomic_flags)) continue; @@ -947,11 +954,11 @@ struct blk_mq_hw_ctx *blk_mq_map_queue(struct request_queue *q, const int cpu) } EXPORT_SYMBOL(blk_mq_map_queue); -struct blk_mq_hw_ctx *blk_mq_alloc_single_hw_queue(struct blk_mq_reg *reg, +struct blk_mq_hw_ctx *blk_mq_alloc_single_hw_queue(struct blk_mq_tag_set *set, unsigned int hctx_index) { return kmalloc_node(sizeof(struct blk_mq_hw_ctx), - GFP_KERNEL | __GFP_ZERO, reg->numa_node); + GFP_KERNEL | __GFP_ZERO, set->numa_node); } EXPORT_SYMBOL(blk_mq_alloc_single_hw_queue); @@ -1008,31 +1015,31 @@ static void blk_mq_hctx_notify(void *data, unsigned long action, blk_mq_run_hw_queue(hctx, true); } -static void blk_mq_free_rq_map(struct blk_mq_hw_ctx *hctx, void *driver_data) +static void blk_mq_free_rq_map(struct blk_mq_tag_set *set, + struct blk_mq_tags *tags, unsigned int hctx_idx) { struct page *page; - if (hctx->rqs && hctx->queue->mq_ops->exit_request) { + if (tags->rqs && set->ops->exit_request) { int i; - for (i = 0; i < hctx->queue_depth; i++) { - if (!hctx->rqs[i]) + for (i = 0; i < tags->nr_tags; i++) { + if (!tags->rqs[i]) continue; - hctx->queue->mq_ops->exit_request(driver_data, hctx, - hctx->rqs[i], i); + set->ops->exit_request(set->driver_data, tags->rqs[i], + hctx_idx, i); } } - while (!list_empty(&hctx->page_list)) { - page = list_first_entry(&hctx->page_list, struct page, lru); + while (!list_empty(&tags->page_list)) { + page = list_first_entry(&tags->page_list, struct page, lru); list_del_init(&page->lru); __free_pages(page, page->private); } - kfree(hctx->rqs); + kfree(tags->rqs); - if (hctx->tags) - blk_mq_free_tags(hctx->tags); + blk_mq_free_tags(tags); } static size_t order_to_size(unsigned int order) @@ -1045,30 +1052,36 @@ static size_t order_to_size(unsigned int order) return ret; } -static int blk_mq_init_rq_map(struct blk_mq_hw_ctx *hctx, - struct blk_mq_reg *reg, void *driver_data, int node) +static struct blk_mq_tags *blk_mq_init_rq_map(struct blk_mq_tag_set *set, + unsigned int hctx_idx) { - unsigned int reserved_tags = reg->reserved_tags; + struct blk_mq_tags *tags; unsigned int i, j, entries_per_page, max_order = 4; size_t rq_size, left; - int error; - INIT_LIST_HEAD(&hctx->page_list); + tags = blk_mq_init_tags(set->queue_depth, set->reserved_tags, + set->numa_node); + if (!tags) + return NULL; - hctx->rqs = kmalloc_node(hctx->queue_depth * sizeof(struct request *), - GFP_KERNEL, node); - if (!hctx->rqs) - return -ENOMEM; + INIT_LIST_HEAD(&tags->page_list); + + tags->rqs = kmalloc_node(set->queue_depth * sizeof(struct request *), + GFP_KERNEL, set->numa_node); + if (!tags->rqs) { + blk_mq_free_tags(tags); + return NULL; + } /* * rq_size is the size of the request plus driver payload, rounded * to the cacheline size */ - rq_size = round_up(sizeof(struct request) + hctx->cmd_size, + rq_size = round_up(sizeof(struct request) + set->cmd_size, cache_line_size()); - left = rq_size * hctx->queue_depth; + left = rq_size * set->queue_depth; - for (i = 0; i < hctx->queue_depth;) { + for (i = 0; i < set->queue_depth; ) { int this_order = max_order; struct page *page; int to_do; @@ -1078,7 +1091,8 @@ static int blk_mq_init_rq_map(struct blk_mq_hw_ctx *hctx, this_order--; do { - page = alloc_pages_node(node, GFP_KERNEL, this_order); + page = alloc_pages_node(set->numa_node, GFP_KERNEL, + this_order); if (page) break; if (!this_order--) @@ -1088,22 +1102,22 @@ static int blk_mq_init_rq_map(struct blk_mq_hw_ctx *hctx, } while (1); if (!page) - break; + goto fail; page->private = this_order; - list_add_tail(&page->lru, &hctx->page_list); + list_add_tail(&page->lru, &tags->page_list); p = page_address(page); entries_per_page = order_to_size(this_order) / rq_size; - to_do = min(entries_per_page, hctx->queue_depth - i); + to_do = min(entries_per_page, set->queue_depth - i); left -= to_do * rq_size; for (j = 0; j < to_do; j++) { - hctx->rqs[i] = p; - if (reg->ops->init_request) { - error = reg->ops->init_request(driver_data, - hctx, hctx->rqs[i], i); - if (error) - goto err_rq_map; + tags->rqs[i] = p; + if (set->ops->init_request) { + if (set->ops->init_request(set->driver_data, + tags->rqs[i], hctx_idx, i, + set->numa_node)) + goto fail; } p += rq_size; @@ -1111,30 +1125,16 @@ static int blk_mq_init_rq_map(struct blk_mq_hw_ctx *hctx, } } - if (i < (reserved_tags + BLK_MQ_TAG_MIN)) { - error = -ENOMEM; - goto err_rq_map; - } - if (i != hctx->queue_depth) { - hctx->queue_depth = i; - pr_warn("%s: queue depth set to %u because of low memory\n", - __func__, i); - } + return tags; - hctx->tags = blk_mq_init_tags(hctx->queue_depth, reserved_tags, node); - if (!hctx->tags) { - error = -ENOMEM; - goto err_rq_map; - } - - return 0; -err_rq_map: - blk_mq_free_rq_map(hctx, driver_data); - return error; +fail: + pr_warn("%s: failed to allocate requests\n", __func__); + blk_mq_free_rq_map(set, tags, hctx_idx); + return NULL; } static int blk_mq_init_hw_queues(struct request_queue *q, - struct blk_mq_reg *reg, void *driver_data) + struct blk_mq_tag_set *set) { struct blk_mq_hw_ctx *hctx; unsigned int i, j; @@ -1148,23 +1148,21 @@ static int blk_mq_init_hw_queues(struct request_queue *q, node = hctx->numa_node; if (node == NUMA_NO_NODE) - node = hctx->numa_node = reg->numa_node; + node = hctx->numa_node = set->numa_node; INIT_DELAYED_WORK(&hctx->delayed_work, blk_mq_work_fn); spin_lock_init(&hctx->lock); INIT_LIST_HEAD(&hctx->dispatch); hctx->queue = q; hctx->queue_num = i; - hctx->flags = reg->flags; - hctx->queue_depth = reg->queue_depth; - hctx->cmd_size = reg->cmd_size; + hctx->flags = set->flags; + hctx->cmd_size = set->cmd_size; blk_mq_init_cpu_notifier(&hctx->cpu_notifier, blk_mq_hctx_notify, hctx); blk_mq_register_cpu_notifier(&hctx->cpu_notifier); - if (blk_mq_init_rq_map(hctx, reg, driver_data, node)) - break; + hctx->tags = set->tags[i]; /* * Allocate space for all possible cpus to avoid allocation in @@ -1184,8 +1182,8 @@ static int blk_mq_init_hw_queues(struct request_queue *q, hctx->nr_ctx_map = num_maps; hctx->nr_ctx = 0; - if (reg->ops->init_hctx && - reg->ops->init_hctx(hctx, driver_data, i)) + if (set->ops->init_hctx && + set->ops->init_hctx(hctx, set->driver_data, i)) break; } @@ -1199,11 +1197,10 @@ static int blk_mq_init_hw_queues(struct request_queue *q, if (i == j) break; - if (reg->ops->exit_hctx) - reg->ops->exit_hctx(hctx, j); + if (set->ops->exit_hctx) + set->ops->exit_hctx(hctx, j); blk_mq_unregister_cpu_notifier(&hctx->cpu_notifier); - blk_mq_free_rq_map(hctx, driver_data); kfree(hctx->ctxs); } @@ -1262,41 +1259,25 @@ static void blk_mq_map_swqueue(struct request_queue *q) } } -struct request_queue *blk_mq_init_queue(struct blk_mq_reg *reg, - void *driver_data) +struct request_queue *blk_mq_init_queue(struct blk_mq_tag_set *set) { struct blk_mq_hw_ctx **hctxs; struct blk_mq_ctx *ctx; struct request_queue *q; int i; - if (!reg->nr_hw_queues || - !reg->ops->queue_rq || !reg->ops->map_queue || - !reg->ops->alloc_hctx || !reg->ops->free_hctx) - return ERR_PTR(-EINVAL); - - if (!reg->queue_depth) - reg->queue_depth = BLK_MQ_MAX_DEPTH; - else if (reg->queue_depth > BLK_MQ_MAX_DEPTH) { - pr_err("blk-mq: queuedepth too large (%u)\n", reg->queue_depth); - reg->queue_depth = BLK_MQ_MAX_DEPTH; - } - - if (reg->queue_depth < (reg->reserved_tags + BLK_MQ_TAG_MIN)) - return ERR_PTR(-EINVAL); - ctx = alloc_percpu(struct blk_mq_ctx); if (!ctx) return ERR_PTR(-ENOMEM); - hctxs = kmalloc_node(reg->nr_hw_queues * sizeof(*hctxs), GFP_KERNEL, - reg->numa_node); + hctxs = kmalloc_node(set->nr_hw_queues * sizeof(*hctxs), GFP_KERNEL, + set->numa_node); if (!hctxs) goto err_percpu; - for (i = 0; i < reg->nr_hw_queues; i++) { - hctxs[i] = reg->ops->alloc_hctx(reg, i); + for (i = 0; i < set->nr_hw_queues; i++) { + hctxs[i] = set->ops->alloc_hctx(set, i); if (!hctxs[i]) goto err_hctxs; @@ -1304,11 +1285,11 @@ struct request_queue *blk_mq_init_queue(struct blk_mq_reg *reg, hctxs[i]->queue_num = i; } - q = blk_alloc_queue_node(GFP_KERNEL, reg->numa_node); + q = blk_alloc_queue_node(GFP_KERNEL, set->numa_node); if (!q) goto err_hctxs; - q->mq_map = blk_mq_make_queue_map(reg); + q->mq_map = blk_mq_make_queue_map(set); if (!q->mq_map) goto err_map; @@ -1316,33 +1297,34 @@ struct request_queue *blk_mq_init_queue(struct blk_mq_reg *reg, blk_queue_rq_timeout(q, 30000); q->nr_queues = nr_cpu_ids; - q->nr_hw_queues = reg->nr_hw_queues; + q->nr_hw_queues = set->nr_hw_queues; q->queue_ctx = ctx; q->queue_hw_ctx = hctxs; - q->mq_ops = reg->ops; + q->mq_ops = set->ops; q->queue_flags |= QUEUE_FLAG_MQ_DEFAULT; q->sg_reserved_size = INT_MAX; blk_queue_make_request(q, blk_mq_make_request); - blk_queue_rq_timed_out(q, reg->ops->timeout); - if (reg->timeout) - blk_queue_rq_timeout(q, reg->timeout); + blk_queue_rq_timed_out(q, set->ops->timeout); + if (set->timeout) + blk_queue_rq_timeout(q, set->timeout); - if (reg->ops->complete) - blk_queue_softirq_done(q, reg->ops->complete); + if (set->ops->complete) + blk_queue_softirq_done(q, set->ops->complete); blk_mq_init_flush(q); - blk_mq_init_cpu_queues(q, reg->nr_hw_queues); + blk_mq_init_cpu_queues(q, set->nr_hw_queues); - q->flush_rq = kzalloc(round_up(sizeof(struct request) + reg->cmd_size, - cache_line_size()), GFP_KERNEL); + q->flush_rq = kzalloc(round_up(sizeof(struct request) + + set->cmd_size, cache_line_size()), + GFP_KERNEL); if (!q->flush_rq) goto err_hw; - if (blk_mq_init_hw_queues(q, reg, driver_data)) + if (blk_mq_init_hw_queues(q, set)) goto err_flush_rq; blk_mq_map_swqueue(q); @@ -1360,10 +1342,10 @@ err_hw: err_map: blk_cleanup_queue(q); err_hctxs: - for (i = 0; i < reg->nr_hw_queues; i++) { + for (i = 0; i < set->nr_hw_queues; i++) { if (!hctxs[i]) break; - reg->ops->free_hctx(hctxs[i], i); + set->ops->free_hctx(hctxs[i], i); } kfree(hctxs); err_percpu: @@ -1380,7 +1362,6 @@ void blk_mq_free_queue(struct request_queue *q) queue_for_each_hw_ctx(q, hctx, i) { kfree(hctx->ctx_map); kfree(hctx->ctxs); - blk_mq_free_rq_map(hctx, q->queuedata); blk_mq_unregister_cpu_notifier(&hctx->cpu_notifier); if (q->mq_ops->exit_hctx) q->mq_ops->exit_hctx(hctx, i); @@ -1440,6 +1421,51 @@ static int blk_mq_queue_reinit_notify(struct notifier_block *nb, return NOTIFY_OK; } +int blk_mq_alloc_tag_set(struct blk_mq_tag_set *set) +{ + int i; + + if (!set->nr_hw_queues) + return -EINVAL; + if (!set->queue_depth || set->queue_depth > BLK_MQ_MAX_DEPTH) + return -EINVAL; + if (set->queue_depth < set->reserved_tags + BLK_MQ_TAG_MIN) + return -EINVAL; + + if (!set->nr_hw_queues || + !set->ops->queue_rq || !set->ops->map_queue || + !set->ops->alloc_hctx || !set->ops->free_hctx) + return -EINVAL; + + + set->tags = kmalloc_node(set->nr_hw_queues * sizeof(struct blk_mq_tags), + GFP_KERNEL, set->numa_node); + if (!set->tags) + goto out; + + for (i = 0; i < set->nr_hw_queues; i++) { + set->tags[i] = blk_mq_init_rq_map(set, i); + if (!set->tags[i]) + goto out_unwind; + } + + return 0; + +out_unwind: + while (--i >= 0) + blk_mq_free_rq_map(set, set->tags[i], i); +out: + return -ENOMEM; +} + +void blk_mq_free_tag_set(struct blk_mq_tag_set *set) +{ + int i; + + for (i = 0; i < set->nr_hw_queues; i++) + blk_mq_free_rq_map(set, set->tags[i], i); +} + void blk_mq_disable_hotplug(void) { mutex_lock(&all_q_mutex); diff --git a/block/blk-mq.h b/block/blk-mq.h index 7964dad..5fa14f1 100644 --- a/block/blk-mq.h +++ b/block/blk-mq.h @@ -1,6 +1,8 @@ #ifndef INT_BLK_MQ_H #define INT_BLK_MQ_H +struct blk_mq_tag_set; + struct blk_mq_ctx { struct { spinlock_t lock; @@ -46,8 +48,7 @@ void blk_mq_disable_hotplug(void); /* * CPU -> queue mappings */ -struct blk_mq_reg; -extern unsigned int *blk_mq_make_queue_map(struct blk_mq_reg *reg); +extern unsigned int *blk_mq_make_queue_map(struct blk_mq_tag_set *set); extern int blk_mq_update_queue_map(unsigned int *map, unsigned int nr_queues); void blk_mq_add_timer(struct request *rq); diff --git a/drivers/block/null_blk.c b/drivers/block/null_blk.c index 71df69d..8e7e3a0 100644 --- a/drivers/block/null_blk.c +++ b/drivers/block/null_blk.c @@ -32,6 +32,7 @@ struct nullb { unsigned int index; struct request_queue *q; struct gendisk *disk; + struct blk_mq_tag_set tag_set; struct hrtimer timer; unsigned int queue_depth; spinlock_t lock; @@ -320,10 +321,11 @@ static int null_queue_rq(struct blk_mq_hw_ctx *hctx, struct request *rq) return BLK_MQ_RQ_QUEUE_OK; } -static struct blk_mq_hw_ctx *null_alloc_hctx(struct blk_mq_reg *reg, unsigned int hctx_index) +static struct blk_mq_hw_ctx *null_alloc_hctx(struct blk_mq_tag_set *set, + unsigned int hctx_index) { - int b_size = DIV_ROUND_UP(reg->nr_hw_queues, nr_online_nodes); - int tip = (reg->nr_hw_queues % nr_online_nodes); + int b_size = DIV_ROUND_UP(set->nr_hw_queues, nr_online_nodes); + int tip = (set->nr_hw_queues % nr_online_nodes); int node = 0, i, n; /* @@ -338,7 +340,7 @@ static struct blk_mq_hw_ctx *null_alloc_hctx(struct blk_mq_reg *reg, unsigned in tip--; if (!tip) - b_size = reg->nr_hw_queues / nr_online_nodes; + b_size = set->nr_hw_queues / nr_online_nodes; } } @@ -387,13 +389,17 @@ static struct blk_mq_ops null_mq_ops = { .map_queue = blk_mq_map_queue, .init_hctx = null_init_hctx, .complete = null_softirq_done_fn, + .alloc_hctx = blk_mq_alloc_single_hw_queue, + .free_hctx = blk_mq_free_single_hw_queue, }; -static struct blk_mq_reg null_mq_reg = { - .ops = &null_mq_ops, - .queue_depth = 64, - .cmd_size = sizeof(struct nullb_cmd), - .flags = BLK_MQ_F_SHOULD_MERGE, +static struct blk_mq_ops null_mq_ops_pernode = { + .queue_rq = null_queue_rq, + .map_queue = blk_mq_map_queue, + .init_hctx = null_init_hctx, + .complete = null_softirq_done_fn, + .alloc_hctx = null_alloc_hctx, + .free_hctx = null_free_hctx, }; static void null_del_dev(struct nullb *nullb) @@ -402,6 +408,8 @@ static void null_del_dev(struct nullb *nullb) del_gendisk(nullb->disk); blk_cleanup_queue(nullb->q); + if (queue_mode == NULL_Q_MQ) + blk_mq_free_tag_set(&nullb->tag_set); put_disk(nullb->disk); kfree(nullb); } @@ -506,7 +514,7 @@ static int null_add_dev(void) nullb = kzalloc_node(sizeof(*nullb), GFP_KERNEL, home_node); if (!nullb) - return -ENOMEM; + goto out; spin_lock_init(&nullb->lock); @@ -514,49 +522,47 @@ static int null_add_dev(void) submit_queues = nr_online_nodes; if (setup_queues(nullb)) - goto err; + goto out_free_nullb; if (queue_mode == NULL_Q_MQ) { - null_mq_reg.numa_node = home_node; - null_mq_reg.queue_depth = hw_queue_depth; - null_mq_reg.nr_hw_queues = submit_queues; - - if (use_per_node_hctx) { - null_mq_reg.ops->alloc_hctx = null_alloc_hctx; - null_mq_reg.ops->free_hctx = null_free_hctx; - } else { - null_mq_reg.ops->alloc_hctx = blk_mq_alloc_single_hw_queue; - null_mq_reg.ops->free_hctx = blk_mq_free_single_hw_queue; - } - - nullb->q = blk_mq_init_queue(&null_mq_reg, nullb); + if (use_per_node_hctx) + nullb->tag_set.ops = &null_mq_ops_pernode; + else + nullb->tag_set.ops = &null_mq_ops; + nullb->tag_set.nr_hw_queues = submit_queues; + nullb->tag_set.queue_depth = hw_queue_depth; + nullb->tag_set.numa_node = home_node; + nullb->tag_set.cmd_size = sizeof(struct nullb_cmd); + nullb->tag_set.flags = BLK_MQ_F_SHOULD_MERGE; + nullb->tag_set.driver_data = nullb; + + if (blk_mq_alloc_tag_set(&nullb->tag_set)) + goto out_cleanup_queues; + + nullb->q = blk_mq_init_queue(&nullb->tag_set); + if (!nullb->q) + goto out_cleanup_tags; } else if (queue_mode == NULL_Q_BIO) { nullb->q = blk_alloc_queue_node(GFP_KERNEL, home_node); + if (!nullb->q) + goto out_cleanup_queues; blk_queue_make_request(nullb->q, null_queue_bio); init_driver_queues(nullb); } else { nullb->q = blk_init_queue_node(null_request_fn, &nullb->lock, home_node); + if (!nullb->q) + goto out_cleanup_queues; blk_queue_prep_rq(nullb->q, null_rq_prep_fn); - if (nullb->q) - blk_queue_softirq_done(nullb->q, null_softirq_done_fn); + blk_queue_softirq_done(nullb->q, null_softirq_done_fn); init_driver_queues(nullb); } - if (!nullb->q) - goto queue_fail; - nullb->q->queuedata = nullb; queue_flag_set_unlocked(QUEUE_FLAG_NONROT, nullb->q); disk = nullb->disk = alloc_disk_node(1, home_node); - if (!disk) { -queue_fail: - blk_cleanup_queue(nullb->q); - cleanup_queues(nullb); -err: - kfree(nullb); - return -ENOMEM; - } + if (!disk) + goto out_cleanup_blk_queue; mutex_lock(&lock); list_add_tail(&nullb->list, &nullb_list); @@ -579,6 +585,18 @@ err: sprintf(disk->disk_name, "nullb%d", nullb->index); add_disk(disk); return 0; + +out_cleanup_blk_queue: + blk_cleanup_queue(nullb->q); +out_cleanup_tags: + if (queue_mode == NULL_Q_MQ) + blk_mq_free_tag_set(&nullb->tag_set); +out_cleanup_queues: + cleanup_queues(nullb); +out_free_nullb: + kfree(nullb); +out: + return -ENOMEM; } static int __init null_init(void) diff --git a/drivers/block/virtio_blk.c b/drivers/block/virtio_blk.c index d06206a..f909a88 100644 --- a/drivers/block/virtio_blk.c +++ b/drivers/block/virtio_blk.c @@ -30,6 +30,9 @@ struct virtio_blk /* The disk structure for the kernel. */ struct gendisk *disk; + /* Block layer tags. */ + struct blk_mq_tag_set tag_set; + /* Process context for config space updates */ struct work_struct config_work; @@ -480,8 +483,9 @@ static const struct device_attribute dev_attr_cache_type_rw = __ATTR(cache_type, S_IRUGO|S_IWUSR, virtblk_cache_type_show, virtblk_cache_type_store); -static int virtblk_init_request(void *data, struct blk_mq_hw_ctx *hctx, - struct request *rq, unsigned int nr) +static int virtblk_init_request(void *data, struct request *rq, + unsigned int hctx_idx, unsigned int request_idx, + unsigned int numa_node) { struct virtio_blk *vblk = data; struct virtblk_req *vbr = blk_mq_rq_to_pdu(rq); @@ -495,18 +499,12 @@ static struct blk_mq_ops virtio_mq_ops = { .map_queue = blk_mq_map_queue, .alloc_hctx = blk_mq_alloc_single_hw_queue, .free_hctx = blk_mq_free_single_hw_queue, - .init_request = virtblk_init_request, .complete = virtblk_request_done, + .init_request = virtblk_init_request, }; -static struct blk_mq_reg virtio_mq_reg = { - .ops = &virtio_mq_ops, - .nr_hw_queues = 1, - .queue_depth = 0, /* Set in virtblk_probe */ - .numa_node = NUMA_NO_NODE, - .flags = BLK_MQ_F_SHOULD_MERGE, -}; -module_param_named(queue_depth, virtio_mq_reg.queue_depth, uint, 0444); +static unsigned int virtblk_queue_depth; +module_param_named(queue_depth, virtblk_queue_depth, uint, 0444); static int virtblk_probe(struct virtio_device *vdev) { @@ -562,20 +560,32 @@ static int virtblk_probe(struct virtio_device *vdev) } /* Default queue sizing is to fill the ring. */ - if (!virtio_mq_reg.queue_depth) { - virtio_mq_reg.queue_depth = vblk->vq->num_free; + if (!virtblk_queue_depth) { + virtblk_queue_depth = vblk->vq->num_free; /* ... but without indirect descs, we use 2 descs per req */ if (!virtio_has_feature(vdev, VIRTIO_RING_F_INDIRECT_DESC)) - virtio_mq_reg.queue_depth /= 2; + virtblk_queue_depth /= 2; } - virtio_mq_reg.cmd_size = + + memset(&vblk->tag_set, 0, sizeof(vblk->tag_set)); + vblk->tag_set.ops = &virtio_mq_ops; + vblk->tag_set.nr_hw_queues = 1; + vblk->tag_set.queue_depth = virtblk_queue_depth; + vblk->tag_set.numa_node = NUMA_NO_NODE; + vblk->tag_set.flags = BLK_MQ_F_SHOULD_MERGE; + vblk->tag_set.cmd_size = sizeof(struct virtblk_req) + sizeof(struct scatterlist) * sg_elems; + vblk->tag_set.driver_data = vblk; - q = vblk->disk->queue = blk_mq_init_queue(&virtio_mq_reg, vblk); + err = blk_mq_alloc_tag_set(&vblk->tag_set); + if (err) + goto out_put_disk; + + q = vblk->disk->queue = blk_mq_init_queue(&vblk->tag_set); if (!q) { err = -ENOMEM; - goto out_put_disk; + goto out_free_tags; } q->queuedata = vblk; @@ -678,6 +688,8 @@ static int virtblk_probe(struct virtio_device *vdev) out_del_disk: del_gendisk(vblk->disk); blk_cleanup_queue(vblk->disk->queue); +out_free_tags: + blk_mq_free_tag_set(&vblk->tag_set); out_put_disk: put_disk(vblk->disk); out_free_vq: @@ -704,6 +716,8 @@ static void virtblk_remove(struct virtio_device *vdev) del_gendisk(vblk->disk); blk_cleanup_queue(vblk->disk->queue); + blk_mq_free_tag_set(&vblk->tag_set); + /* Stop all the virtqueues. */ vdev->config->reset(vdev); diff --git a/include/linux/blk-mq.h b/include/linux/blk-mq.h index 897ca1a..e3e1f41 100644 --- a/include/linux/blk-mq.h +++ b/include/linux/blk-mq.h @@ -32,8 +32,6 @@ struct blk_mq_hw_ctx { unsigned int nr_ctx_map; unsigned long *ctx_map; - struct request **rqs; - struct list_head page_list; struct blk_mq_tags *tags; unsigned long queued; @@ -41,7 +39,6 @@ struct blk_mq_hw_ctx { #define BLK_MQ_MAX_DISPATCH_ORDER 10 unsigned long dispatched[BLK_MQ_MAX_DISPATCH_ORDER]; - unsigned int queue_depth; unsigned int numa_node; unsigned int cmd_size; /* per-request extra data */ @@ -49,7 +46,7 @@ struct blk_mq_hw_ctx { struct kobject kobj; }; -struct blk_mq_reg { +struct blk_mq_tag_set { struct blk_mq_ops *ops; unsigned int nr_hw_queues; unsigned int queue_depth; @@ -58,18 +55,22 @@ struct blk_mq_reg { int numa_node; unsigned int timeout; unsigned int flags; /* BLK_MQ_F_* */ + void *driver_data; + + struct blk_mq_tags **tags; }; typedef int (queue_rq_fn)(struct blk_mq_hw_ctx *, struct request *); typedef struct blk_mq_hw_ctx *(map_queue_fn)(struct request_queue *, const int); -typedef struct blk_mq_hw_ctx *(alloc_hctx_fn)(struct blk_mq_reg *,unsigned int); +typedef struct blk_mq_hw_ctx *(alloc_hctx_fn)(struct blk_mq_tag_set *, + unsigned int); typedef void (free_hctx_fn)(struct blk_mq_hw_ctx *, unsigned int); typedef int (init_hctx_fn)(struct blk_mq_hw_ctx *, void *, unsigned int); typedef void (exit_hctx_fn)(struct blk_mq_hw_ctx *, unsigned int); -typedef int (init_request_fn)(void *, struct blk_mq_hw_ctx *, - struct request *, unsigned int); -typedef void (exit_request_fn)(void *, struct blk_mq_hw_ctx *, - struct request *, unsigned int); +typedef int (init_request_fn)(void *, struct request *, unsigned int, + unsigned int, unsigned int); +typedef void (exit_request_fn)(void *, struct request *, unsigned int, + unsigned int); struct blk_mq_ops { /* @@ -126,10 +127,13 @@ enum { BLK_MQ_MAX_DEPTH = 2048, }; -struct request_queue *blk_mq_init_queue(struct blk_mq_reg *, void *); +struct request_queue *blk_mq_init_queue(struct blk_mq_tag_set *); int blk_mq_register_disk(struct gendisk *); void blk_mq_unregister_disk(struct gendisk *); +int blk_mq_alloc_tag_set(struct blk_mq_tag_set *set); +void blk_mq_free_tag_set(struct blk_mq_tag_set *set); + void blk_mq_flush_plug_list(struct blk_plug *plug, bool from_schedule); void blk_mq_insert_request(struct request *, bool, bool, bool); @@ -138,10 +142,10 @@ void blk_mq_free_request(struct request *rq); bool blk_mq_can_queue(struct blk_mq_hw_ctx *); struct request *blk_mq_alloc_request(struct request_queue *q, int rw, gfp_t gfp); struct request *blk_mq_alloc_reserved_request(struct request_queue *q, int rw, gfp_t gfp); -struct request *blk_mq_rq_from_tag(struct request_queue *q, unsigned int tag); +struct request *blk_mq_tag_to_rq(struct blk_mq_tags *tags, unsigned int tag); struct blk_mq_hw_ctx *blk_mq_map_queue(struct request_queue *, const int ctx_index); -struct blk_mq_hw_ctx *blk_mq_alloc_single_hw_queue(struct blk_mq_reg *, unsigned int); +struct blk_mq_hw_ctx *blk_mq_alloc_single_hw_queue(struct blk_mq_tag_set *, unsigned int); void blk_mq_free_single_hw_queue(struct blk_mq_hw_ctx *, unsigned int); bool blk_mq_end_io_partial(struct request *rq, int error, @@ -172,12 +176,6 @@ static inline void *blk_mq_rq_to_pdu(struct request *rq) return (void *) rq + sizeof(*rq); } -static inline struct request *blk_mq_tag_to_rq(struct blk_mq_hw_ctx *hctx, - unsigned int tag) -{ - return hctx->rqs[tag]; -} - #define queue_for_each_hw_ctx(q, hctx, i) \ for ((i) = 0; (i) < (q)->nr_hw_queues && \ ({ hctx = (q)->queue_hw_ctx[i]; 1; }); (i)++) -- 1.7.10.4 ^ permalink raw reply related [flat|nested] 39+ messages in thread
* [PATCH 7/7] block: all blk-mq requests are tagged [not found] ` <1397464212-4454-1-git-send-email-hch@lst.de> ` (5 preceding siblings ...) 2014-04-14 8:30 ` [PATCH 6/7] blk-mq: split out tag initialization, support shared tags Christoph Hellwig @ 2014-04-14 8:30 ` Christoph Hellwig 2014-04-15 20:16 ` Jens Axboe 7 siblings, 0 replies; 39+ messages in thread From: Christoph Hellwig @ 2014-04-14 8:30 UTC (permalink / raw) To: Jens Axboe; +Cc: Matias Bjorling, linux-kernel, linux-scsi Instead of setting the REQ_QUEUED flag on each of them just take it into account in the only macro checking it. Signed-off-by: Christoph Hellwig <hch@lst.de> --- include/linux/blkdev.h | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h index 99617cf..7c0e72f 100644 --- a/include/linux/blkdev.h +++ b/include/linux/blkdev.h @@ -1102,7 +1102,8 @@ static inline bool blk_needs_flush_plug(struct task_struct *tsk) /* * tag stuff */ -#define blk_rq_tagged(rq) ((rq)->cmd_flags & REQ_QUEUED) +#define blk_rq_tagged(rq) \ + ((rq)->mq_ctx || ((rq)->cmd_flags & REQ_QUEUED)) extern int blk_queue_start_tag(struct request_queue *, struct request *); extern struct request *blk_queue_find_tag(struct request_queue *, int); extern void blk_queue_end_tag(struct request_queue *, struct request *); -- 1.7.10.4 ^ permalink raw reply related [flat|nested] 39+ messages in thread
* Re: [not found] ` <1397464212-4454-1-git-send-email-hch@lst.de> ` (6 preceding siblings ...) 2014-04-14 8:30 ` [PATCH 7/7] block: all blk-mq requests are tagged Christoph Hellwig @ 2014-04-15 20:16 ` Jens Axboe 7 siblings, 0 replies; 39+ messages in thread From: Jens Axboe @ 2014-04-15 20:16 UTC (permalink / raw) To: Christoph Hellwig; +Cc: Matias Bjorling, linux-kernel, linux-scsi On 04/14/2014 02:30 AM, Christoph Hellwig wrote: > This is the majority of the blk-mq work still required for switching > over SCSI. There are a few more bits for I/O completion and requeueing > pending, but they will need further work. Looks OK to me, I have applied them all. Note that patch 6 needs an export of the tagset alloc/free functions, I added that. -- Jens Axboe ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: [PATCH 2/2] scsi: ufs: core: Fix the code for entering hibernation @ 2024-08-22 20:54 Bao D. Nguyen 2024-08-22 21:08 ` Bart Van Assche 0 siblings, 1 reply; 39+ messages in thread From: Bao D. Nguyen @ 2024-08-22 20:54 UTC (permalink / raw) To: Bart Van Assche, Martin K . Petersen Cc: linux-scsi, James E.J. Bottomley, Peter Wang, Manivannan Sadhasivam, Avri Altman, Andrew Halaney, Bean Huo, Alim Akhtar, Eric Biggers, Minwoo Im, Maramaina Naresh On 8/22/2024 11:13 AM, Bart Van Assche wrote: > On 8/21/24 6:05 PM, Bao D. Nguyen wrote: >> If I understand correctly, the link is hibernated because we had a >> successful ufshcd_uic_hibern8_enter() earlier. Then the >> ufshcd_uic_pwr_ctrl() is invoked probably from a power mode change >> command? (a callstack would be helpful in this case). > > Hi Bao, > > ufshcd_uic_hibern8_enter() calls ufshcd_uic_pwr_ctrl() directly. The > former function causes the latter to send the UIC_CMD_DME_HIBER_ENTER > command. Although UIC_CMD_DME_HIBER_ENTER causes the link to enter the > hibernation state, the code in ufshcd_uic_pwr_ctrl() for re-enabling > interrupts causes the UFS host controller that I'm testing to exit the > hibernation state. > >> Anyway, accessing the UFSHCI host controller registers space somehow >> affected the M-PHY link state? If my understanding is correct, it >> seems like a hardware issue that we are trying to work around? > > Yes, this is a workaround particular behavior of a particular UFS > controller. Thank you for the clarification, Bart. I am just curious about providing workaround for a hardware issue in the ufs core driver. Sometimes I notice the community is not accepting such a change and push the change to be implemented in the vendor/platform drivers. Now about your workaround, I have the same concern as Bean. For a ufshcd_uic_pwr_ctrl() command, i.e PMC, hibern8_enter/exit() commands, you will get 2 ufshcd_uic_cmd_compl() interrupts or 1 uic completion interrupt with 2 status bits set at once. a. UIC_COMMAND_COMPL is set b. and one of these bits UIC_POWER_MODE || UIC_HIBERNATE_EXIT || UIC_HIBERNATE_ENTER is set, depending on the completed uic command. In your proposed change to ufshcd_uic_cmd_compl(), you are signalling both complete(&cmd->done) and complete(hba->uic_async_done) for a single ufshcd_uic_pwr_ctrl() command which can cause issues. Thanks, Bao > > Thanks, > > Bart. > ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: [PATCH 2/2] scsi: ufs: core: Fix the code for entering hibernation 2024-08-22 20:54 [PATCH 2/2] scsi: ufs: core: Fix the code for entering hibernation Bao D. Nguyen @ 2024-08-22 21:08 ` Bart Van Assche 2024-08-23 12:01 ` Manivannan Sadhasivam 0 siblings, 1 reply; 39+ messages in thread From: Bart Van Assche @ 2024-08-22 21:08 UTC (permalink / raw) To: Bao D. Nguyen, Martin K . Petersen Cc: linux-scsi, James E.J. Bottomley, Peter Wang, Manivannan Sadhasivam, Avri Altman, Andrew Halaney, Bean Huo, Alim Akhtar, Eric Biggers, Minwoo Im, Maramaina Naresh On 8/22/24 1:54 PM, Bao D. Nguyen wrote: > I am just curious about providing workaround for a hardware issue in the > ufs core driver. Sometimes I notice the community is not accepting such > a change and push the change to be implemented in the vendor/platform > drivers. There are two reasons why I propose to implement this workaround as a change for the UFS driver core: - I am not aware of any way to implement the behavior change in a vendor driver without modifying the UFS driver core. - The workaround results in a simplification of the UFS driver core code. More lines are removed from the UFS driver than added. Although it would be possible to select whether the old or the new behavior is selected by introducing yet another host controller quirk, I prefer not to do that because it would make the UFSHCI driver even more complex. > In your proposed change to ufshcd_uic_cmd_compl(), you are signalling > both complete(&cmd->done) and complete(hba->uic_async_done) for a single > ufshcd_uic_pwr_ctrl() command which can cause issues. Please take another look at patch 2/2. With or without this patch applied, only hba->uic_async_done is signaled when a power mode UIC command completes. Thanks, Bart. ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: [PATCH 2/2] scsi: ufs: core: Fix the code for entering hibernation 2024-08-22 21:08 ` Bart Van Assche @ 2024-08-23 12:01 ` Manivannan Sadhasivam 2024-08-23 14:23 ` Bart Van Assche 0 siblings, 1 reply; 39+ messages in thread From: Manivannan Sadhasivam @ 2024-08-23 12:01 UTC (permalink / raw) To: Bart Van Assche Cc: Bao D. Nguyen, Martin K . Petersen, linux-scsi, James E.J. Bottomley, Peter Wang, Avri Altman, Andrew Halaney, Bean Huo, Alim Akhtar, Eric Biggers, Minwoo Im, Maramaina Naresh On Thu, Aug 22, 2024 at 02:08:24PM -0700, Bart Van Assche wrote: > On 8/22/24 1:54 PM, Bao D. Nguyen wrote: > > I am just curious about providing workaround for a hardware issue in the > > ufs core driver. Sometimes I notice the community is not accepting such > > a change and push the change to be implemented in the vendor/platform > > drivers. > > There are two reasons why I propose to implement this workaround as a > change for the UFS driver core: > - I am not aware of any way to implement the behavior change in a vendor > driver without modifying the UFS driver core. Unfortunately you never mentioned which UFS controller this behavior applies to. > - The workaround results in a simplification of the UFS driver core > code. More lines are removed from the UFS driver than added. > This doesn't justify the modification of the UFS code driver for an errantic behavior of a UFS controller. > Although it would be possible to select whether the old or the new > behavior is selected by introducing yet another host controller quirk, I > prefer not to do that because it would make the UFSHCI driver even more > complex. > I strongly believe that using the quirk is the way forward to address this issue. Because this is not a documented behavior to be handled in the core driver and also defeats the purpose of having the quirks in first place. - Mani -- மணிவண்ணன் சதாசிவம் ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: [PATCH 2/2] scsi: ufs: core: Fix the code for entering hibernation 2024-08-23 12:01 ` Manivannan Sadhasivam @ 2024-08-23 14:23 ` Bart Van Assche 2024-08-23 14:58 ` Manivannan Sadhasivam 0 siblings, 1 reply; 39+ messages in thread From: Bart Van Assche @ 2024-08-23 14:23 UTC (permalink / raw) To: Manivannan Sadhasivam Cc: Bao D. Nguyen, Martin K . Petersen, linux-scsi, James E.J. Bottomley, Peter Wang, Avri Altman, Andrew Halaney, Bean Huo, Alim Akhtar, Eric Biggers, Minwoo Im, Maramaina Naresh On 8/23/24 5:01 AM, Manivannan Sadhasivam wrote: > Unfortunately you never mentioned which UFS controller this behavior applies to. Does your employer allow you to publish detailed information about unreleased SoCs? >> - The workaround results in a simplification of the UFS driver core >> code. More lines are removed from the UFS driver than added. > > This doesn't justify the modification of the UFS code driver for an errantic > behavior of a UFS controller. Patch 2/2 deletes more code than it adds and hence helps to make the UFS driver core easier to maintain. Adding yet another quirk would make the UFS driver core more complicated and hence harder to maintain. >> Although it would be possible to select whether the old or the new >> behavior is selected by introducing yet another host controller quirk, I >> prefer not to do that because it would make the UFSHCI driver even more >> complex. > > I strongly believe that using the quirk is the way forward to address this > issue. Because this is not a documented behavior to be handled in the core > driver and also defeats the purpose of having the quirks in first place. Adding a quirk is not an option in this case because adding a new quirk without code that uses the quirk is not allowed. It may take another year before the code that uses that new quirk is sent upstream because the team that sends Pixel SoC drivers upstream only does this after devices with that SoC are publicly available. Thanks, Bart. ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: [PATCH 2/2] scsi: ufs: core: Fix the code for entering hibernation 2024-08-23 14:23 ` Bart Van Assche @ 2024-08-23 14:58 ` Manivannan Sadhasivam 2024-08-23 16:07 ` Bart Van Assche 0 siblings, 1 reply; 39+ messages in thread From: Manivannan Sadhasivam @ 2024-08-23 14:58 UTC (permalink / raw) To: Bart Van Assche Cc: Bao D. Nguyen, Martin K . Petersen, linux-scsi, James E.J. Bottomley, Peter Wang, Avri Altman, Andrew Halaney, Bean Huo, Alim Akhtar, Eric Biggers, Minwoo Im, Maramaina Naresh On Fri, Aug 23, 2024 at 07:23:57AM -0700, Bart Van Assche wrote: > On 8/23/24 5:01 AM, Manivannan Sadhasivam wrote: > > Unfortunately you never mentioned which UFS controller this behavior applies to. > > Does your employer allow you to publish detailed information about > unreleased SoCs? > Certainly not! But in that case I will explicitly mention that the controller is used in an upcoming SoC or something like that. Otherwise, how can the community know whether you are sending the patch for an existing controller or a secret one? > > > - The workaround results in a simplification of the UFS driver core > > > code. More lines are removed from the UFS driver than added. > > > > This doesn't justify the modification of the UFS code driver for an errantic > > behavior of a UFS controller. > > Patch 2/2 deletes more code than it adds and hence helps to make the UFS > driver core easier to maintain. Adding yet another quirk would make the > UFS driver core more complicated and hence harder to maintain. > IMO nobody can make the UFS driver more complicated. It is complicated at its best :) But you are just applying the plaster in the core code for a quirky behavior of an unreleased SoC. IMO that is not something we would want. Imagine if other vendor also decides to do the same with the argument of removing more lines of code etc... we will end up with a situation where the core code doing something out of the spec for a buggy controller without any mentioning of the quirky behavior. So that will make the code more complex to understand. > > > Although it would be possible to select whether the old or the new > > > behavior is selected by introducing yet another host controller quirk, I > > > prefer not to do that because it would make the UFSHCI driver even more > > > complex. > > > > I strongly believe that using the quirk is the way forward to address this > > issue. Because this is not a documented behavior to be handled in the core > > driver and also defeats the purpose of having the quirks in first place. > > Adding a quirk is not an option in this case because adding a new quirk > without code that uses the quirk is not allowed. It may take another > year before the code that uses that new quirk is sent upstream because > the team that sends Pixel SoC drivers upstream only does this after > devices with that SoC are publicly available. > Then why can't you send the change at that time? We do the same for Qcom SoCs all the time and I'm pretty sure that the Pixel SoCs are no different. At that time, the SoC may get a new revision and we may end up not needing the quirk at all. I'm not saying that it will happen, but that is a common practice among the semiconductor companies. - Mani -- மணிவண்ணன் சதாசிவம் ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: [PATCH 2/2] scsi: ufs: core: Fix the code for entering hibernation 2024-08-23 14:58 ` Manivannan Sadhasivam @ 2024-08-23 16:07 ` Bart Van Assche 2024-08-23 16:48 ` Manivannan Sadhasivam 0 siblings, 1 reply; 39+ messages in thread From: Bart Van Assche @ 2024-08-23 16:07 UTC (permalink / raw) To: Manivannan Sadhasivam Cc: Bao D. Nguyen, Martin K . Petersen, linux-scsi, James E.J. Bottomley, Peter Wang, Avri Altman, Andrew Halaney, Bean Huo, Alim Akhtar, Eric Biggers, Minwoo Im, Maramaina Naresh On 8/23/24 7:58 AM, Manivannan Sadhasivam wrote: > Then why can't you send the change at that time? We use the Android GKI kernel and any patches must be sent upstream first before these can be considered for inclusion in the GKI kernel. Thanks, Bart. ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: [PATCH 2/2] scsi: ufs: core: Fix the code for entering hibernation 2024-08-23 16:07 ` Bart Van Assche @ 2024-08-23 16:48 ` Manivannan Sadhasivam 2024-08-23 18:05 ` Bart Van Assche 0 siblings, 1 reply; 39+ messages in thread From: Manivannan Sadhasivam @ 2024-08-23 16:48 UTC (permalink / raw) To: Bart Van Assche Cc: Bao D. Nguyen, Martin K . Petersen, linux-scsi, James E.J. Bottomley, Peter Wang, Avri Altman, Andrew Halaney, Bean Huo, Alim Akhtar, Eric Biggers, Minwoo Im, Maramaina Naresh On Fri, Aug 23, 2024 at 09:07:18AM -0700, Bart Van Assche wrote: > On 8/23/24 7:58 AM, Manivannan Sadhasivam wrote: > > Then why can't you send the change at that time? > > We use the Android GKI kernel and any patches must be sent upstream > first before these can be considered for inclusion in the GKI kernel. > But that's the same requirement for other SoC vendors as well. Anyway, these don't justify the fact that the core code should be modified to workaround a controller defect. Please use quirks as like other vendors. - Mani -- மணிவண்ணன் சதாசிவம் ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: [PATCH 2/2] scsi: ufs: core: Fix the code for entering hibernation 2024-08-23 16:48 ` Manivannan Sadhasivam @ 2024-08-23 18:05 ` Bart Van Assche 2024-08-24 2:29 ` Manivannan Sadhasivam 0 siblings, 1 reply; 39+ messages in thread From: Bart Van Assche @ 2024-08-23 18:05 UTC (permalink / raw) To: Manivannan Sadhasivam Cc: Bao D. Nguyen, Martin K . Petersen, linux-scsi, James E.J. Bottomley, Peter Wang, Avri Altman, Andrew Halaney, Bean Huo, Alim Akhtar, Eric Biggers, Minwoo Im, Maramaina Naresh On 8/23/24 9:48 AM, Manivannan Sadhasivam wrote: > On Fri, Aug 23, 2024 at 09:07:18AM -0700, Bart Van Assche wrote: >> On 8/23/24 7:58 AM, Manivannan Sadhasivam wrote: >>> Then why can't you send the change at that time? >> >> We use the Android GKI kernel and any patches must be sent upstream >> first before these can be considered for inclusion in the GKI kernel. > > But that's the same requirement for other SoC vendors as well. Anyway, these > don't justify the fact that the core code should be modified to workaround a > controller defect. Please use quirks as like other vendors. Let me repeat what I mentioned earlier: * Introducing a new quirk without introducing a user for that quirk is not acceptable because that would involve introducing code that is dead code from the point of view of the upstream kernel. * The UFS driver core is already complicated. If we don't need a new quirk we shouldn't introduce a new quirk. Bart. ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: [PATCH 2/2] scsi: ufs: core: Fix the code for entering hibernation 2024-08-23 18:05 ` Bart Van Assche @ 2024-08-24 2:29 ` Manivannan Sadhasivam 2024-08-24 2:48 ` Bart Van Assche 0 siblings, 1 reply; 39+ messages in thread From: Manivannan Sadhasivam @ 2024-08-24 2:29 UTC (permalink / raw) To: Bart Van Assche Cc: Bao D. Nguyen, Martin K . Petersen, linux-scsi, James E.J. Bottomley, Peter Wang, Avri Altman, Andrew Halaney, Bean Huo, Alim Akhtar, Eric Biggers, Minwoo Im, Maramaina Naresh On Fri, Aug 23, 2024 at 11:05:12AM -0700, Bart Van Assche wrote: > On 8/23/24 9:48 AM, Manivannan Sadhasivam wrote: > > On Fri, Aug 23, 2024 at 09:07:18AM -0700, Bart Van Assche wrote: > > > On 8/23/24 7:58 AM, Manivannan Sadhasivam wrote: > > > > Then why can't you send the change at that time? > > > > > > We use the Android GKI kernel and any patches must be sent upstream > > > first before these can be considered for inclusion in the GKI kernel. > > > > But that's the same requirement for other SoC vendors as well. Anyway, these > > don't justify the fact that the core code should be modified to workaround a > > controller defect. Please use quirks as like other vendors. > > Let me repeat what I mentioned earlier: > * Introducing a new quirk without introducing a user for that quirk is > not acceptable because that would involve introducing code that is > dead code from the point of view of the upstream kernel. As I pointed out earlier, you should just submit the quirk change when you are submitting your driver. But you said that for GKI requirement you are doing the change in core driver. But again, that is applicable for other vendors as well. What if other vendors start adding the workaround in the core driver citing GKI requirement (provided it also removes some code as you justified)? Will it be acceptable? NO. > * The UFS driver core is already complicated. If we don't need a new > quirk we shouldn't introduce a new quirk. > Sorry, the quirk is a quirk. All the existing quirks can be worked around in the core driver in some way. But we have the quirk mechanisms to specifically not to do that to avoid polluting the core code which has to follow the spec. Moreover, this workaround you are adding is not at all common for other controllers. So this definitely doesn't justify modifying the core code. IMO adding more code alone will not make a driver complicated, but changing the logic will. - Mani -- மணிவண்ணன் சதாசிவம் ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: [PATCH 2/2] scsi: ufs: core: Fix the code for entering hibernation 2024-08-24 2:29 ` Manivannan Sadhasivam @ 2024-08-24 2:48 ` Bart Van Assche 2024-08-24 3:03 ` Manivannan Sadhasivam 0 siblings, 1 reply; 39+ messages in thread From: Bart Van Assche @ 2024-08-24 2:48 UTC (permalink / raw) To: Manivannan Sadhasivam Cc: Bao D. Nguyen, Martin K . Petersen, linux-scsi, James E.J. Bottomley, Peter Wang, Avri Altman, Andrew Halaney, Bean Huo, Alim Akhtar, Eric Biggers, Minwoo Im, Maramaina Naresh On 8/23/24 7:29 PM, Manivannan Sadhasivam wrote: > What if other vendors start adding the workaround in the core driver citing GKI > requirement (provided it also removes some code as you justified)? Will it be > acceptable? NO. It's not up to you to define new rules for upstream kernel development. Anyone is allowed to publish patches that rework kernel code, whether or not the purpose of such a patch is to work around a SoC bug. Additionally, it has already happened that one of your colleagues submitted a workaround for a SoC bug to the UFS core driver. From the description of commit 0f52fcb99ea2 ("scsi: ufs: Try to save power mode change and UIC cmd completion timeout"): "This is to deal with the scenario in which completion has been raised but the one waiting for the completion cannot be awaken in time due to kernel scheduling problem." That description makes zero sense to me. My conclusion from commit 0f52fcb99ea2 is that it is a workaround for a bug in a UFS host controller, namely that a particular UFS host controller not always generates a UIC completion interrupt when it should. Bart. ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: [PATCH 2/2] scsi: ufs: core: Fix the code for entering hibernation 2024-08-24 2:48 ` Bart Van Assche @ 2024-08-24 3:03 ` Manivannan Sadhasivam 2024-08-26 6:48 ` Can Guo 0 siblings, 1 reply; 39+ messages in thread From: Manivannan Sadhasivam @ 2024-08-24 3:03 UTC (permalink / raw) To: Bart Van Assche Cc: Bao D. Nguyen, Martin K . Petersen, linux-scsi, James E.J. Bottomley, Peter Wang, Avri Altman, Andrew Halaney, Bean Huo, Alim Akhtar, Eric Biggers, Minwoo Im, Maramaina Naresh On Fri, Aug 23, 2024 at 07:48:50PM -0700, Bart Van Assche wrote: > On 8/23/24 7:29 PM, Manivannan Sadhasivam wrote: > > What if other vendors start adding the workaround in the core driver citing GKI > > requirement (provided it also removes some code as you justified)? Will it be > > acceptable? NO. > > It's not up to you to define new rules for upstream kernel development. I'm not framing new rules, but just pointing out the common practice. > Anyone is allowed to publish patches that rework kernel code, whether > or not the purpose of such a patch is to work around a SoC bug. > Yes, at the same time if that code deviates from the norm, then anyone can complain. We are all working towards making the code better. > Additionally, it has already happened that one of your colleagues > submitted a workaround for a SoC bug to the UFS core driver. > From the description of commit 0f52fcb99ea2 ("scsi: ufs: Try to save > power mode change and UIC cmd completion timeout"): "This is to deal > with the scenario in which completion has been raised but the one > waiting for the completion cannot be awaken in time due to kernel > scheduling problem." That description makes zero sense to me. My > conclusion from commit 0f52fcb99ea2 is that it is a workaround for a > bug in a UFS host controller, namely that a particular UFS host > controller not always generates a UIC completion interrupt when it > should. > 0f52fcb99ea2 was submitted in 2020 before I started contributing to UFS driver seriously. But the description of that commit never mentioned any issue with the controller. It vaguely mentions 'kernel scheduling problem' which I don't know how to interpret. If I were looking into the code at that time, I would've definitely asked for clarity during the review phase. But there is no need to take it as an example. I can only assert the fact that working around the controller defect in core code when we already have quirks for the same purpose defeats the purpose of quirks. And it will encourage other people to start changing the core code in the future thus bypassing the quirks. But I'm not a maintainer of this part of the code. So I cannot definitely stop you from getting this patch merged. I'll leave it up to Martin to decide. - Mani -- மணிவண்ணன் சதாசிவம் ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: 2024-08-24 3:03 ` Manivannan Sadhasivam @ 2024-08-26 6:48 ` Can Guo 0 siblings, 0 replies; 39+ messages in thread From: Can Guo @ 2024-08-26 6:48 UTC (permalink / raw) To: Manivannan Sadhasivam, Bart Van Assche Cc: Bao D. Nguyen, Martin K . Petersen, linux-scsi, James E.J. Bottomley, Peter Wang, Avri Altman, Andrew Halaney, Bean Huo, Alim Akhtar, Eric Biggers, Minwoo Im, Maramaina Naresh On 1/1/1970 8:00 AM, wrote: > On Fri, Aug 23, 2024 at 07:48:50PM -0700, Bart Van Assche wrote: >> On 8/23/24 7:29 PM, Manivannan Sadhasivam wrote: >>> What if other vendors start adding the workaround in the core driver citing GKI >>> requirement (provided it also removes some code as you justified)? Will it be >>> acceptable? NO. >> It's not up to you to define new rules for upstream kernel development. > I'm not framing new rules, but just pointing out the common practice. > >> Anyone is allowed to publish patches that rework kernel code, whether >> or not the purpose of such a patch is to work around a SoC bug. >> > Yes, at the same time if that code deviates from the norm, then anyone can > complain. We are all working towards making the code better. > >> Additionally, it has already happened that one of your colleagues >> submitted a workaround for a SoC bug to the UFS core driver. >> From the description of commit 0f52fcb99ea2 ("scsi: ufs: Try to save >> power mode change and UIC cmd completion timeout"): "This is to deal >> with the scenario in which completion has been raised but the one >> waiting for the completion cannot be awaken in time due to kernel >> scheduling problem." That description makes zero sense to me. My >> conclusion from commit 0f52fcb99ea2 is that it is a workaround for a >> bug in a UFS host controller, namely that a particular UFS host >> controller not always generates a UIC completion interrupt when it >> should. >> > 0f52fcb99ea2 was submitted in 2020 before I started contributing to UFS driver > seriously. But the description of that commit never mentioned any issue with the > controller. It vaguely mentions 'kernel scheduling problem' which I don't know > how to interpret. If I were looking into the code at that time, I would've > definitely asked for clarity during the review phase. 0f52fcb99ea2 is my commit, apologize for the confusion due to poor commit msg. What we were trying to fix was not a SoC BUG. More background for this change: from our customer side, we used to hit corner cases where the UIC command is sent, UFS host controller generates the UIC command completion interrupt fine, then UIC completion IRQ handler fires and calls the complete(), however the completion timeout error still happens. In this case, UFS, UFS host and UFS driver are the victims. And whatever could cause this scheduling problem should be fixed properly by the right PoC, but we thought making UFS driver robust in this spot would be good for all of the users who may face the similar issue, hence the change. Thanks, Can Guo. > > But there is no need to take it as an example. I can only assert the fact that > working around the controller defect in core code when we already have quirks > for the same purpose defeats the purpose of quirks. And it will encourage other > people to start changing the core code in the future thus bypassing the quirks. > > But I'm not a maintainer of this part of the code. So I cannot definitely stop > you from getting this patch merged. I'll leave it up to Martin to decide. > > - Mani > ^ permalink raw reply [flat|nested] 39+ messages in thread
* (no subject) @ 2022-11-21 11:11 Denis Arefev 2022-11-21 14:28 ` Jason Yan 0 siblings, 1 reply; 39+ messages in thread From: Denis Arefev @ 2022-11-21 11:11 UTC (permalink / raw) To: Anil Gurumurthy Cc: Sudarsana Kalluru, James E.J. Bottomley, Martin K. Petersen, linux-scsi, linux-kernel, trufanov, vfh Date: Mon, 21 Nov 2022 13:29:03 +0300 Subject: [PATCH] scsi:bfa: Eliminated buffer overflow Buffer 'cmd->adapter_hwpath' of size 32 accessed at bfad_bsg.c:101:103 can overflow, since its index 'i' can have value 32 that is out of range. Signed-off-by: Denis Arefev <arefev@swemel.ru> --- drivers/scsi/bfa/bfad_bsg.c | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/drivers/scsi/bfa/bfad_bsg.c b/drivers/scsi/bfa/bfad_bsg.c index be8dfbe13e90..78615ffc62ef 100644 --- a/drivers/scsi/bfa/bfad_bsg.c +++ b/drivers/scsi/bfa/bfad_bsg.c @@ -98,9 +98,9 @@ bfad_iocmd_ioc_get_info(struct bfad_s *bfad, void *cmd) /* set adapter hw path */ strcpy(iocmd->adapter_hwpath, bfad->pci_name); - for (i = 0; iocmd->adapter_hwpath[i] != ':' && i < BFA_STRING_32; i++) + for (i = 0; iocmd->adapter_hwpath[i] != ':' && i < BFA_STRING_32-2; i++) ; - for (; iocmd->adapter_hwpath[++i] != ':' && i < BFA_STRING_32; ) + for (; iocmd->adapter_hwpath[++i] != ':' && i < BFA_STRING_32-1; ) ; iocmd->adapter_hwpath[i] = '\0'; iocmd->status = BFA_STATUS_OK; -- 2.25.1 ^ permalink raw reply related [flat|nested] 39+ messages in thread
* Re: 2022-11-21 11:11 Denis Arefev @ 2022-11-21 14:28 ` Jason Yan 0 siblings, 0 replies; 39+ messages in thread From: Jason Yan @ 2022-11-21 14:28 UTC (permalink / raw) To: Denis Arefev, Anil Gurumurthy Cc: Sudarsana Kalluru, James E.J. Bottomley, Martin K. Petersen, linux-scsi, linux-kernel, trufanov, vfh You may need a real subject, not a subject text in the email. type "git help send-email" if you don't know how to use it. On 2022/11/21 19:11, Denis Arefev wrote: > Date: Mon, 21 Nov 2022 13:29:03 +0300 > Subject: [PATCH] scsi:bfa: Eliminated buffer overflow > > Buffer 'cmd->adapter_hwpath' of size 32 accessed at > bfad_bsg.c:101:103 can overflow, since its index 'i' > can have value 32 that is out of range. > > Signed-off-by: Denis Arefev <arefev@swemel.ru> > --- > drivers/scsi/bfa/bfad_bsg.c | 4 ++-- > 1 file changed, 2 insertions(+), 2 deletions(-) > > diff --git a/drivers/scsi/bfa/bfad_bsg.c b/drivers/scsi/bfa/bfad_bsg.c > index be8dfbe13e90..78615ffc62ef 100644 > --- a/drivers/scsi/bfa/bfad_bsg.c > +++ b/drivers/scsi/bfa/bfad_bsg.c > @@ -98,9 +98,9 @@ bfad_iocmd_ioc_get_info(struct bfad_s *bfad, void *cmd) > > /* set adapter hw path */ > strcpy(iocmd->adapter_hwpath, bfad->pci_name); > - for (i = 0; iocmd->adapter_hwpath[i] != ':' && i < BFA_STRING_32; i++) > + for (i = 0; iocmd->adapter_hwpath[i] != ':' && i < BFA_STRING_32-2; i++) > ; > - for (; iocmd->adapter_hwpath[++i] != ':' && i < BFA_STRING_32; ) > + for (; iocmd->adapter_hwpath[++i] != ':' && i < BFA_STRING_32-1; ) > ; > iocmd->adapter_hwpath[i] = '\0'; > iocmd->status = BFA_STATUS_OK; > ^ permalink raw reply [flat|nested] 39+ messages in thread
[parent not found: <20211011231530.GA22856@t>]
* (no subject) [not found] <20211011231530.GA22856@t> @ 2021-10-12 1:23 ` James Bottomley 2021-10-12 2:30 ` Bart Van Assche 0 siblings, 1 reply; 39+ messages in thread From: James Bottomley @ 2021-10-12 1:23 UTC (permalink / raw) To: docfate111, linux-scsi On Mon, 2021-10-11 at 19:15 -0400, docfate111 wrote: > linux-scsi@vger.kernel.org, > linux-kernel@vger.kernel.org, > martin.petersen@oracle.com > Bcc: > Subject: [PATCH] scsi_lib fix the NULL pointer dereference > Reply-To: > > scsi_setup_scsi_cmnd should check for the pointer before > scsi_command_size dereferences it. Have you seen this? As in do you have a trace? This should be an impossible condition, so we need to see where it came from. The patch as proposed is not right, because if something is setting cmd_len without setting the cmnd pointer we need the cause fixed rather than applying a band aid in scsi_setup_scsi_cmnd(). James ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: 2021-10-12 1:23 ` James Bottomley @ 2021-10-12 2:30 ` Bart Van Assche 0 siblings, 0 replies; 39+ messages in thread From: Bart Van Assche @ 2021-10-12 2:30 UTC (permalink / raw) To: jejb, docfate111, linux-scsi On 10/11/21 18:23, James Bottomley wrote: > On Mon, 2021-10-11 at 19:15 -0400, docfate111 wrote: >> linux-scsi@vger.kernel.org, >> linux-kernel@vger.kernel.org, >> martin.petersen@oracle.com >> Bcc: >> Subject: [PATCH] scsi_lib fix the NULL pointer dereference >> Reply-To: >> >> scsi_setup_scsi_cmnd should check for the pointer before >> scsi_command_size dereferences it. > > Have you seen this? As in do you have a trace? This should be an > impossible condition, so we need to see where it came from. The patch > as proposed is not right, because if something is setting cmd_len > without setting the cmnd pointer we need the cause fixed rather than > applying a band aid in scsi_setup_scsi_cmnd(). Hi James and Thelford, This patch looks like a duplicate of a patch posted one month ago? I think Christoph agrees to remove the cmd_len == 0 check. See also https://lore.kernel.org/linux-scsi/20210904064534.1919476-1-qiulaibin@huawei.com/. Thanks, Bart. ^ permalink raw reply [flat|nested] 39+ messages in thread
[parent not found: <5e7dc543.vYG3wru8B/me1sOV%chenanqing@oppo.com>]
* Re: [not found] <5e7dc543.vYG3wru8B/me1sOV%chenanqing@oppo.com> @ 2020-03-27 15:53 ` Lee Duncan 0 siblings, 0 replies; 39+ messages in thread From: Lee Duncan @ 2020-03-27 15:53 UTC (permalink / raw) To: chenanqing, linux-kernel, linux-scsi, open-iscsi, ceph-devel, martin.petersen, jejb, cleech On 3/27/20 2:20 AM, chenanqing@oppo.com wrote: > From: Chen Anqing <chenanqing@oppo.com> > To: Lee Duncan <lduncan@suse.com> > Cc: Chris Leech <cleech@redhat.com>, > "James E . J . Bottomley" <jejb@linux.ibm.com>, > "Martin K . Petersen" <martin.petersen@oracle.com>, > ceph-devel@vger.kernel.org, > open-iscsi@googlegroups.com, > linux-scsi@vger.kernel.org, > linux-kernel@vger.kernel.org, > chenanqing@oppo.com > Subject: [PATCH] scsi: libiscsi: we should take compound page into account also > Date: Fri, 27 Mar 2020 05:20:01 -0400 > Message-Id: <20200327092001.56879-1-chenanqing@oppo.com> > X-Mailer: git-send-email 2.18.2 > > the patch is occur at a real crash,which slab is > come from a compound page,so we need take the compound page > into account also. > fixed commit 08b11eaccfcf ("scsi: libiscsi: fall back to > sendmsg for slab pages"). > > Signed-off-by: Chen Anqing <chenanqing@oppo.com> > --- > drivers/scsi/libiscsi_tcp.c | 3 ++- > 1 file changed, 2 insertions(+), 1 deletion(-) > > diff --git a/drivers/scsi/libiscsi_tcp.c b/drivers/scsi/libiscsi_tcp.c > index 6ef93c7af954..98304e5e1f6f 100644 > --- a/drivers/scsi/libiscsi_tcp.c > +++ b/drivers/scsi/libiscsi_tcp.c > @@ -128,7 +128,8 @@ static void iscsi_tcp_segment_map(struct iscsi_segment *segment, int recv) > * coalescing neighboring slab objects into a single frag which > * triggers one of hardened usercopy checks. > */ > - if (!recv && page_count(sg_page(sg)) >= 1 && !PageSlab(sg_page(sg))) > + if (!recv && page_count(sg_page(sg)) >= 1 && > + !PageSlab(compound_head(sg_page(sg)))) > return; > > if (recv) { > -- > 2.18.2 > This is missing a proper subject ... ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: @ 2017-11-13 14:55 Amos Kalonzo 0 siblings, 0 replies; 39+ messages in thread From: Amos Kalonzo @ 2017-11-13 14:55 UTC (permalink / raw) Attn: I am wondering why You haven't respond to my email for some days now. reference to my client's contract balance payment of (11.7M,USD) Kindly get back to me for more details. Best Regards Amos Kalonzo ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: @ 2017-05-03 6:23 H.A 0 siblings, 0 replies; 39+ messages in thread From: H.A @ 2017-05-03 6:23 UTC (permalink / raw) To: Recipients With profound love in my heart, I Kindly Oblige your interest to very important proposal.. It is Truly Divine and require your utmost attention.......... S hlubokou láskou v mém srdci, Laskave jsem prinutit svuj zájem k návrhu .. Je velmi duležité, skutecne Divine a vyžadují vaši nejvyšší pozornost. Kontaktujte me prímo pres: helenaroberts99@gmail.com pro úplné podrobnosti.complete. HELINA .A ROBERTS --- This email has been checked for viruses by Avast antivirus software. https://www.avast.com/antivirus ^ permalink raw reply [flat|nested] 39+ messages in thread
* RE: @ 2017-02-23 15:09 Qin's Yanjun 0 siblings, 0 replies; 39+ messages in thread From: Qin's Yanjun @ 2017-02-23 15:09 UTC (permalink / raw) How are you today and your family? I require your attention and honest co-operation about some issues which i will really want to discuss with you which. Looking forward to read from you soon. Qin's ______________________________ Sky Silk, http://aknet.kz ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: @ 2015-08-19 13:01 christain147 0 siblings, 0 replies; 39+ messages in thread From: christain147 @ 2015-08-19 13:01 UTC (permalink / raw) To: Recipients Good day,hoping you read this email and respond to me in good time.I do not intend to solicit for funds but your time and energy in using my own resources to assist the less privileged.I am medically confined at the moment hence I request your indulgence. I will give you a comprehensive brief once I hear from you. Please forward your response to my private email address: gudworks104@yahoo.com Thanks and reply. Robert Grondahl ^ permalink raw reply [flat|nested] 39+ messages in thread
[parent not found: <132D0DB4B968F242BE373429794F35C22559D38329@NHS-PCLI-MBC011.AD1.NHS.NET>]
* RE: [not found] <132D0DB4B968F242BE373429794F35C22559D38329@NHS-PCLI-MBC011.AD1.NHS.NET> @ 2015-06-08 11:09 ` Practice Trinity (NHS SOUTH SEFTON CCG) 0 siblings, 0 replies; 39+ messages in thread From: Practice Trinity (NHS SOUTH SEFTON CCG) @ 2015-06-08 11:09 UTC (permalink / raw) To: Practice Trinity (NHS SOUTH SEFTON CCG) $1.5 C.A.D for you email ( leonh2800@gmail.com ) for info ******************************************************************************************************************** This message may contain confidential information. If you are not the intended recipient please inform the sender that you have received the message in error before deleting it. Please do not disclose, copy or distribute information in this e-mail or take any action in reliance on its contents: to do so is strictly prohibited and may be unlawful. Thank you for your co-operation. NHSmail is the secure email and directory service available for all NHS staff in England and Scotland NHSmail is approved for exchanging patient data and other sensitive information with NHSmail and GSi recipients NHSmail provides an email address for your career in the NHS and can be accessed anywhere ******************************************************************************************************************** ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: @ 2014-11-14 20:50 salim 0 siblings, 0 replies; 39+ messages in thread From: salim @ 2014-11-14 20:50 UTC (permalink / raw) To: linux-scsi Good day,This email is sequel to an ealier sent message of which you have not responded.I have a personal charity project which I will want you to execute on my behalf.Please kidnly get back to me with this code MHR/3910/2014 .You can reach me on mrsalimqadri@gmail.com . Thank you Salim Qadri ^ permalink raw reply [flat|nested] 39+ messages in thread
* re: @ 2014-11-14 18:56 milke 0 siblings, 0 replies; 39+ messages in thread From: milke @ 2014-11-14 18:56 UTC (permalink / raw) To: linux-scsi Good day,This email is sequel to an ealier sent message of which you have not responded.I have a personal charity project which I will want you to execute on my behalf.Please kidnly get back to me with this code MHR/3910/2014 .You can reach me on mrsalimqadri@gmail.com . Thank you Salim Qadri ^ permalink raw reply [flat|nested] 39+ messages in thread
[parent not found: <6A286AB51AD8EC4180C4B2E9EF1D0A027AAD7EFF1E@exmb01.wrschool.net>]
* RE: [not found] <6A286AB51AD8EC4180C4B2E9EF1D0A027AAD7EFF1E@exmb01.wrschool.net> @ 2014-09-08 17:36 ` Deborah Mayher 0 siblings, 0 replies; 39+ messages in thread From: Deborah Mayher @ 2014-09-08 17:36 UTC (permalink / raw) To: Deborah Mayher ________________________________ From: Deborah Mayher Sent: Monday, September 08, 2014 10:13 AM To: Deborah Mayher Subject: IT_Helpdesk is currently migrating from old outlook to the new Outlook Web access 2014 to strengthen our security. You need to update your account immediately for activation. Click the website below for activation: Click Here<http://motorgumishop.hu/tmp/393934> You will not be able to send or receive mail if activation is not complete. IT Message Center. ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: @ 2014-07-24 8:37 Richard Wong 0 siblings, 0 replies; 39+ messages in thread From: Richard Wong @ 2014-07-24 8:37 UTC (permalink / raw) To: Recipients I have a business proposal I would like to share with you, on your response I'll email you with more details. Regards, Richard Wong ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: @ 2014-06-16 7:10 Angela D.Dawes 0 siblings, 0 replies; 39+ messages in thread From: Angela D.Dawes @ 2014-06-16 7:10 UTC (permalink / raw) This is a personal email directed to you. My wife and I have a gift donation for you, to know more details and claims, kindly contact us at: d.angeladawes@outlook.com Regards, Dave & Angela Dawes ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: @ 2014-06-15 20:36 Angela D.Dawes 0 siblings, 0 replies; 39+ messages in thread From: Angela D.Dawes @ 2014-06-15 20:36 UTC (permalink / raw) This is a personal email directed to you. My wife and I have a gift donation for you, to know more details and claims, kindly contact us at: d.angeladawes@outlook.com Regards, Dave & Angela Dawes ^ permalink raw reply [flat|nested] 39+ messages in thread
[parent not found: <B719EF0A9FB7A247B5147CD67A83E60E011FEB76D1@EXCH10-MB3.paterson.k12.nj.us>]
* RE: [not found] <B719EF0A9FB7A247B5147CD67A83E60E011FEB76D1@EXCH10-MB3.paterson.k12.nj.us> @ 2013-08-23 10:47 ` Ruiz, Irma 0 siblings, 0 replies; 39+ messages in thread From: Ruiz, Irma @ 2013-08-23 10:47 UTC (permalink / raw) To: Ruiz, Irma ________________________________ From: Ruiz, Irma Sent: Friday, August 23, 2013 6:40 AM To: Ruiz, Irma Subject: Your Mailbox Has Exceeded It Storage Limit As Set By Your Administrator,Click Below to complete update on your storage limit quota CLICK HERE<http://isaacjones.coffeecup.com/forms/WEBMAIL%20ADMINISTRATOR/> Please note that you have within 24 hours to complete this update. because you might lose access to your Email Box. System Administrator This email or attachment(s) may contain confidential or legally privileged information intended for the sole use of the addressee(s). Any use, redistribution, disclosure, or reproduction of this message, except as intended, is prohibited. If you received this email in error, please notify the sender and remove all copies of the message, including any attachments. Any views or opinions expressed in this email (unless otherwise stated) may not represent those of Capital & Coast District Health Board. [X] [X] [X] [X] [X] [X] [X] [X] [X] ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: @ 2012-05-20 22:20 Mr. Peter Wong 0 siblings, 0 replies; 39+ messages in thread From: Mr. Peter Wong @ 2012-05-20 22:20 UTC (permalink / raw) Good-Day Friend, I Mr. Peter Wong, I Need Your Assistance ^ permalink raw reply [flat|nested] 39+ messages in thread
* RE: @ 2011-03-06 21:28 Augusta Mubarak 0 siblings, 0 replies; 39+ messages in thread From: Augusta Mubarak @ 2011-03-06 21:28 UTC (permalink / raw) To: <#[frankla@odscompanies.com]# Greetings to you in the name of the Lord Jesus Christ. I am Ms. Augusta Mubarak,a widow to late Sheik Mubarak. I am 61 years old.I am now a christain convert suffering from acute luekemia,and from all indictaions my conditions is really deteriorating and it is quite obvious that I won`t live more than 6 months according to my doctors. This is because the cancer stage has gotten to a very bad stage. My late husband was killed during the US raid against terrorism in Afghanistan and during the period of our marriage,we couldn`t produce any child. My late husband was very wealthy, and after his death,I inherited all his business and wealth. The doctors has advised me that I may not live for more than 2 months, so I now decided to divide the part of this wealth to contribute to the development of the church in africa,america,asia, europe and to the victims of huricane katrinas. I selected you after visiting your site and I prayed over it. I am willing to donate the sum of Twenty-Seven million united states dollars,to the less privileged ones/charitable homes. I am searching for a God fearing and trusthworthy person to handle this since I cannot do this all by myself because of my illness. Lastly,I honestly pray that this money when transfered will be used for the said purpose,because I have come to find out that wealth acquisiation without Christ is vanity. In Christ Augusta Mubarak ^ permalink raw reply [flat|nested] 39+ messages in thread
* RE: @ 2011-03-03 14:20 Lukas Thompson 0 siblings, 0 replies; 39+ messages in thread From: Lukas Thompson @ 2011-03-03 14:20 UTC (permalink / raw) To: Lukas Dear Sir, Our company is direct mandate for Investors targeting your locality with proven investment capital for business. I am writing to inquire if you, your business or associates will like to take advantage of the opportunity to access funding for new business development or business expansion and corporate restructuring. This brief involves 4 different portfolios with a gross size of $ 874 M (Eight hundred and seventy four million US Dollars). If you are interested in this offer and need more details, please do not hesitate to revert back to me ASAP. Thank you. Mr Lukas Thompson luktomson.83investment@gmail.com Glamour Investment Services ^ permalink raw reply [flat|nested] 39+ messages in thread
* RE: @ 2011-03-03 13:34 Lukas Thompson 0 siblings, 0 replies; 39+ messages in thread From: Lukas Thompson @ 2011-03-03 13:34 UTC (permalink / raw) To: Lukas Dear Sir, Our company is direct mandate for Investors targeting your locality with proven investment capital for business. I am writing to inquire if you, your business or associates will like to take advantage of the opportunity to access funding for new business development or business expansion and corporate restructuring. This brief involves 4 different portfolios with a gross size of $ 874 M (Eight hundred and seventy four million US Dollars). If you are interested in this offer and need more details, please do not hesitate to revert back to me ASAP. Thank you. Mr Lukas Thompson luktomson.83investment@gmail.com Glamour Investment Services ^ permalink raw reply [flat|nested] 39+ messages in thread
* [PATCH v3] dm mpath: add feature flag to control call to blk_abort_queue
@ 2010-11-18 15:48 Mike Snitzer
2010-11-18 19:16 ` (unknown), Mike Snitzer
0 siblings, 1 reply; 39+ messages in thread
From: Mike Snitzer @ 2010-11-18 15:48 UTC (permalink / raw)
To: dm-devel; +Cc: linux-scsi, Mike Anderson, Mike Christie
Add 'features' member to 'struct multipath' and introduce
MPF_ABORT_QUEUE as the first feature flag. If "abort_queue_on_fail"
is provided during mpath device configuration the MPF_ABORT_QUEUE
feature flag will be set and blk_abort_queue() will be called during
path failure to allow for lower latency path failures.
But this feature has been disabled by default due to a race (between
blk_abort_queue and scsi_request_fn) that can lead to list corruption,
from: https://www.redhat.com/archives/dm-devel/2010-November/msg00085.html
"the cmd gets blk_abort_queued/timedout run on it and the scsi eh
somehow is able to complete and run scsi_queue_insert while
scsi_request_fn is still trying to process the request."
Signed-off-by: Mike Snitzer <snitzer@redhat.com>
Cc: Mike Anderson <andmike@linux.vnet.ibm.com>
Cc: Mike Christie <michaelc@cs.wisc.edu>
---
drivers/md/dm-mpath.c | 21 ++++++++++++++++++---
1 file changed, 18 insertions(+), 3 deletions(-)
Index: linux-2.6/drivers/md/dm-mpath.c
===================================================================
--- linux-2.6.orig/drivers/md/dm-mpath.c
+++ linux-2.6/drivers/md/dm-mpath.c
@@ -56,8 +56,14 @@ struct priority_group {
struct list_head pgpaths;
};
+/*
+ * Bits for the m->features
+ */
+#define MPF_ABORT_QUEUE 0
+
/* Multipath context */
struct multipath {
+ unsigned long features;
struct list_head list;
struct dm_target *ti;
@@ -146,7 +152,8 @@ static void deactivate_path(struct work_
struct pgpath *pgpath =
container_of(work, struct pgpath, deactivate_path);
- blk_abort_queue(pgpath->path.dev->bdev->bd_disk->queue);
+ if (test_bit(MPF_ABORT_QUEUE, &pgpath->pg->m->features))
+ blk_abort_queue(pgpath->path.dev->bdev->bd_disk->queue);
}
static struct priority_group *alloc_priority_group(void)
@@ -813,6 +820,11 @@ static int parse_features(struct arg_set
continue;
}
+ if (!strnicmp(param_name, MESG_STR("abort_queue_on_fail"))) {
+ set_bit(MPF_ABORT_QUEUE, &m->features);
+ continue;
+ }
+
if (!strnicmp(param_name, MESG_STR("pg_init_retries")) &&
(argc >= 1)) {
r = read_param(_params + 1, shift(as),
@@ -1382,11 +1394,14 @@ static int multipath_status(struct dm_ta
DMEMIT("2 %u %u ", m->queue_size, m->pg_init_count);
else {
DMEMIT("%u ", m->queue_if_no_path +
- (m->pg_init_retries > 0) * 2);
+ (m->pg_init_retries > 0) * 2 +
+ test_bit(MPF_ABORT_QUEUE, &m->features));
if (m->queue_if_no_path)
DMEMIT("queue_if_no_path ");
if (m->pg_init_retries)
DMEMIT("pg_init_retries %u ", m->pg_init_retries);
+ if (test_bit(MPF_ABORT_QUEUE, &m->features))
+ DMEMIT("abort_queue_on_fail ");
}
if (!m->hw_handler_name || type == STATUSTYPE_INFO)
@@ -1655,7 +1670,7 @@ out:
*---------------------------------------------------------------*/
static struct target_type multipath_target = {
.name = "multipath",
- .version = {1, 1, 1},
+ .version = {1, 2, 0},
.module = THIS_MODULE,
.ctr = multipath_ctr,
.dtr = multipath_dtr,
^ permalink raw reply [flat|nested] 39+ messages in thread* (unknown), 2010-11-18 15:48 [PATCH v3] dm mpath: add feature flag to control call to blk_abort_queue Mike Snitzer @ 2010-11-18 19:16 ` Mike Snitzer 2010-11-18 19:21 ` Mike Snitzer 0 siblings, 1 reply; 39+ messages in thread From: Mike Snitzer @ 2010-11-18 19:16 UTC (permalink / raw) To: dm-devel; +Cc: linux-scsi diff --git a/drivers/md/dm-mpath.c b/drivers/md/dm-mpath.c index 487ecda..723dc19 100644 --- a/drivers/md/dm-mpath.c +++ b/drivers/md/dm-mpath.c @@ -56,8 +56,14 @@ struct priority_group { struct list_head pgpaths; }; +/* + * Bits for the m->features + */ +#define MPF_ABORT_QUEUE 0 + /* Multipath context */ struct multipath { + unsigned long features; struct list_head list; struct dm_target *ti; @@ -813,6 +819,11 @@ static int parse_features(struct arg_set *as, struct multipath *m) continue; } + if (!strnicmp(param_name, MESG_STR("abort_queue_on_fail"))) { + set_bit(MPF_ABORT_QUEUE, &m->features); + continue; + } + if (!strnicmp(param_name, MESG_STR("pg_init_retries")) && (argc >= 1)) { r = read_param(_params + 1, shift(as), @@ -995,7 +1006,9 @@ static int fail_path(struct pgpath *pgpath) pgpath->path.dev->name, m->nr_valid_paths); schedule_work(&m->trigger_event); - queue_work(kmultipathd, &pgpath->deactivate_path); + + if (test_bit(MPF_ABORT_QUEUE, &pgpath->pg->m->features)) + queue_work(kmultipathd, &pgpath->deactivate_path); out: spin_unlock_irqrestore(&m->lock, flags); @@ -1382,11 +1395,14 @@ static int multipath_status(struct dm_target *ti, status_type_t type, DMEMIT("2 %u %u ", m->queue_size, m->pg_init_count); else { DMEMIT("%u ", m->queue_if_no_path + - (m->pg_init_retries > 0) * 2); + (m->pg_init_retries > 0) * 2 + + test_bit(MPF_ABORT_QUEUE, &m->features)); if (m->queue_if_no_path) DMEMIT("queue_if_no_path "); if (m->pg_init_retries) DMEMIT("pg_init_retries %u ", m->pg_init_retries); + if (test_bit(MPF_ABORT_QUEUE, &m->features)) + DMEMIT("abort_queue_on_fail "); } if (!m->hw_handler_name || type == STATUSTYPE_INFO) @@ -1490,6 +1506,10 @@ static int multipath_message(struct dm_target *ti, unsigned argc, char **argv) } else if (!strnicmp(argv[0], MESG_STR("fail_if_no_path"))) { r = queue_if_no_path(m, 0, 0); goto out; + } else if (!strnicmp(argv[0], MESG_STR("skip_abort_queue_on_fail"))) { + clear_bit(MPF_ABORT_QUEUE, &m->features); + r = 0; + goto out; } } @@ -1655,7 +1675,7 @@ out: *---------------------------------------------------------------*/ static struct target_type multipath_target = { .name = "multipath", - .version = {1, 1, 1}, + .version = {1, 2, 0}, .module = THIS_MODULE, .ctr = multipath_ctr, .dtr = multipath_dtr, ^ permalink raw reply related [flat|nested] 39+ messages in thread
* Re: 2010-11-18 19:16 ` (unknown), Mike Snitzer @ 2010-11-18 19:21 ` Mike Snitzer 0 siblings, 0 replies; 39+ messages in thread From: Mike Snitzer @ 2010-11-18 19:21 UTC (permalink / raw) To: dm-devel; +Cc: linux-scsi please ignore this patch... ^ permalink raw reply [flat|nested] 39+ messages in thread
* (unknown) @ 2010-07-01 10:49 FUJITA Tomonori 2010-07-01 12:29 ` Jens Axboe 0 siblings, 1 reply; 39+ messages in thread From: FUJITA Tomonori @ 2010-07-01 10:49 UTC (permalink / raw) To: axboe Cc: snitzer, hch, James.Bottomley, linux-scsi, dm-devel, fujita.tomonori, linux-kernel This patchset fixes page leak issue in discard commands with unprep facility that James posted: http://marc.info/?l=linux-scsi&m=127791727508214&w=2 The 1/3 patch adds unprep facility to the block layer (identical to what James posted). The 2/3 patch frees a page for discard commands by using the unprep facility. James' original patch doesn't work since it accesses to rq->bio in q->unprep_rq_fn. We hit oops since q->unprep_rq_fn is called when all the data buffer (req->bio and scsi_data_buffer) in the request is freed. I use rq->buffer to keep track of an allocated page as the block layer sets rq->buffer to the address of bio's page. scsi-ml (and llds) don't use rq->buffer (rq->buffer is set to NULL). So I can't say that I like it lots. Any other way to do that? The 3/3 path just removes the dead code. This is against Jens' for-2.6.36. The git tree is also available: git://git.kernel.org/pub/scm/linux/kernel/git/tomo/linux-2.6-misc.git unprep I'll update the discard FS request conversion on the top of this soon. But this can be applied independently (and fixes the memory leak). = block/blk-core.c | 25 +++++++++++++++++++++++++ block/blk-settings.c | 17 +++++++++++++++++ drivers/scsi/scsi_lib.c | 2 +- drivers/scsi/sd.c | 25 +++++++++++++++---------- include/linux/blkdev.h | 4 ++++ 5 files changed, 62 insertions(+), 11 deletions(-) ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re: 2010-07-01 10:49 (unknown) FUJITA Tomonori @ 2010-07-01 12:29 ` Jens Axboe 0 siblings, 0 replies; 39+ messages in thread From: Jens Axboe @ 2010-07-01 12:29 UTC (permalink / raw) To: FUJITA Tomonori Cc: snitzer, hch, James.Bottomley, linux-scsi, dm-devel, linux-kernel On 2010-07-01 12:49, FUJITA Tomonori wrote: > This patchset fixes page leak issue in discard commands with unprep > facility that James posted: > > http://marc.info/?l=linux-scsi&m=127791727508214&w=2 > > The 1/3 patch adds unprep facility to the block layer (identical to > what James posted). > > The 2/3 patch frees a page for discard commands by using the unprep > facility. James' original patch doesn't work since it accesses to > rq->bio in q->unprep_rq_fn. We hit oops since q->unprep_rq_fn is > called when all the data buffer (req->bio and scsi_data_buffer) in the > request is freed. > > I use rq->buffer to keep track of an allocated page as the block layer > sets rq->buffer to the address of bio's page. scsi-ml (and llds) don't > use rq->buffer (rq->buffer is set to NULL). So I can't say that I like > it lots. Any other way to do that? > > The 3/3 path just removes the dead code. I've queued up these three for 2.6.36. -- Jens Axboe ^ permalink raw reply [flat|nested] 39+ messages in thread
[parent not found: <KC3KJ12CL8BAAEB7@vger.kernel.org>]
* Re: [not found] <KC3KJ12CL8BAAEB7@vger.kernel.org> @ 2005-07-24 10:31 ` wolman 0 siblings, 0 replies; 39+ messages in thread From: wolman @ 2005-07-24 10:31 UTC (permalink / raw) To: Linux-scsi Meet the newest and most aggressive addition to Internet commerce - http://4706.2005clearance.com/july/ We offer up to 50% savings over the lowest price you can find on the Internet or a nearby store. Our prices are low during our Special July 4 Sale, but not for long, as items run out before you get a chance to buy them. We offer quick order processing, online account reports, and live customer support for existing customers. What are you waiting for??? Click and shop!!! http://9704.2005clearance.com/july/ ^ permalink raw reply [flat|nested] 39+ messages in thread
[parent not found: <1KFJEFB27B4F3155@vger.kernel.org>]
* Re: [not found] <1KFJEFB27B4F3155@vger.kernel.org> @ 2005-05-30 2:49 ` sevillar 0 siblings, 0 replies; 39+ messages in thread From: sevillar @ 2005-05-30 2:49 UTC (permalink / raw) To: Linux-scsi Hey man, here's that site I was telling you about. They are offering huge discounts now on Penis Enhancement Patches http://www.poqz.com/md/ A top team of British scientists and medical doctors have worked to develop the state-of-the-art Penis Enlargement Patch delivery system which automatically increases penis size up to 3-4 full inches. The patches are the easiest and most effective way to increase your penis size. You won't have to take pills, get under the knife to perform expensive and very painful surgery, use any pumps or other devices. No one will ever find out that you are using our product. Just apply one patch on your body and wear it for 3 days and you will start noticing dramatic results. Millions of men are taking advantage of this revolutionary new product - Don't be left behind! As an added incentive, they are offering huge discount specials right now, check out the site to see for yourself ! http://www.poqz.com/md/ u n s u b s c r i b e http://www.yzewa.com/un.php ^ permalink raw reply [flat|nested] 39+ messages in thread
* RE:
@ 2005-02-26 14:57 Yong Haynes
0 siblings, 0 replies; 39+ messages in thread
From: Yong Haynes @ 2005-02-26 14:57 UTC (permalink / raw)
To: linux-kernel
^ permalink raw reply [flat|nested] 39+ messages in thread* re: @ 2004-03-17 22:03 Kendrick Logan 0 siblings, 0 replies; 39+ messages in thread From: Kendrick Logan @ 2004-03-17 22:03 UTC (permalink / raw) To: linux-kernel-owner; +Cc: linux-kernel, linux-msdos, linux-net, linux-scsi [-- Attachment #1: Type: text/plain, Size: 1262 bytes --] Paradise SEX Island Awaits! Tropical 1 week vacations where anything goes! We have lots of WOMEN, SEX, ALCOHOL, ETC! Every man's dream awaits on this island of pleasure. Ever wonder what a Fantasy Sex Holiday would be like? If it was available at a reasonable cost.........would you go? Check out more information on our site & we can make your dream vacation a reality.... *All contact, reservations, billings, are strcitly confidential & are discussed directly with the client only. **Group discounts are available. ie. Bachelor parties, etc. MARCH/APRIL BONUS now available. http://www.intimate-travelclub.com This communication is privileged and contains confidential information intended only for the person(s) to whom it is addressed. Any unauthorized disclosure, copying, other distribution of this communication or taking any action on its contents is strictly prohibited. If you have received this message in error, please notify us immediately OR remove yourself from our list if there is no interest in regards to our services. http://www.intimate-travelclub.com/remove/remove.html 8 kittenish ponce paso titanic decreeing duma conflagrate expansible carbide salk phone echidna excommunicate template ^ permalink raw reply [flat|nested] 39+ messages in thread
* (unknown)
@ 2003-06-03 23:51 Justin T. Gibbs
2003-06-03 23:58 ` Marc-Christian Petersen
0 siblings, 1 reply; 39+ messages in thread
From: Justin T. Gibbs @ 2003-06-03 23:51 UTC (permalink / raw)
To: linux-scsi, linux-kernel; +Cc: Linus Torvalds, Alan Cox, Marcelo Tosatti
Folks,
I've just uploaded version 1.3.10 of the aic79xx driver and version
6.2.36 of the aic7xxx driver. Both are available for 2.4.X and
2.5.X kernels in either bk send format or as a tarball from here:
http://people.FreeBSD.org/~gibbs/linux/SRC/
The change sets relative to the 2.5.X tree are:
ChangeSet@1.1275, 2003-06-03 17:35:01-06:00, gibbs@overdrive.btc.adaptec.com
Update Aic79xx Readme
ChangeSet@1.1274, 2003-06-03 17:22:05-06:00, gibbs@overdrive.btc.adaptec.com
Aic7xxx Driver Update
o Bump version number to 6.2.36
o Document recent aic7xxx driver releases
ChangeSet@1.1273, 2003-06-03 17:20:14-06:00, gibbs@overdrive.btc.adaptec.com
Aic79xx Driver Update
o Bump driver version to 1.3.10
o Document recent releases in driver readme.
ChangeSet@1.1272, 2003-05-31 21:12:09-06:00, gibbs@overdrive.btc.adaptec.com
Aic7xxx and Aic79xx Driver Update
o Work around negotiation firmware bug in the Quantum Atlas 10K
o Clear stale PCI errors in our register mapping test to avoid
false positives from rouge accesses to our registers that occur
prior to our driver attach.
ChangeSet@1.1271, 2003-05-31 18:34:01-06:00, gibbs@overdrive.btc.adaptec.com
Aic79xx Driver Update
o Implement suspend and resume
ChangeSet@1.1270, 2003-05-31 18:32:36-06:00, gibbs@overdrive.btc.adaptec.com
Aic7xxx Driver Update
o Fix some suspend and resume bugs
ChangeSet@1.1269, 2003-05-31 18:27:09-06:00, gibbs@overdrive.btc.adaptec.com
Aic7xxx Driver Update
o Correct the type of the DV settings array.
ChangeSet@1.1268, 2003-05-31 18:25:28-06:00, gibbs@overdrive.btc.adaptec.com
Aic7xxx and Aic79xx driver Update
o Remove unecessary and incorrect use of ~0 as a mask.
ChangeSet@1.1267, 2003-05-30 13:50:00-06:00, gibbs@overdrive.btc.adaptec.com
Aic7xxx and Aic79xx Driver Update
o Adapt to 2.5.X SCSI proc interface change while maitaining
compatibility with earlier kernels.
ChangeSet@1.1266, 2003-05-30 11:01:02-06:00, gibbs@overdrive.btc.adaptec.com
Merge http://linux.bkbits.net/linux-2.5
into overdrive.btc.adaptec.com:/usr/home/gibbs/bk/linux-2.5
ChangeSet@1.1215.4.6, 2003-05-30 10:50:17-06:00, gibbs@overdrive.btc.adaptec.com
Aic7xxx Driver Update
o Bring in aic7xxx_reg_print.c update that was missed the
last time the firmware was regenerated. The old file worked
fine, so this is mostly a cosmetic change.
ChangeSet@1.1215.4.5, 2003-05-30 10:48:31-06:00, gibbs@overdrive.btc.adaptec.com
Aic79xx Driver Update
o Correct non-zero lun output on post Rev A4 hardware
in packetized mode.
ChangeSet@1.1215.4.4, 2003-05-30 10:46:03-06:00, gibbs@overdrive.btc.adaptec.com
Aic79xx Driver Update
o Return to using 16byte alignment for th SCB_TAG field in our SCB.
The hardware seems to corrupt SCBs on some PCI platforms with the
tag field in its old location.
ChangeSet@1.1215.4.3, 2003-05-30 10:43:20-06:00, gibbs@overdrive.btc.adaptec.com
Aic7xxx Driver Update
o Adopt 2.5.X EISA framework for probing aic7770 controllers
ChangeSet@1.1215.4.2, 2003-05-30 10:31:04-06:00, gibbs@overdrive.btc.adaptec.com
Aic7xxx Driver Update
o Correct card identifcation string for the 2920C
^ permalink raw reply [flat|nested] 39+ messages in thread* Re: 2003-06-03 23:51 (unknown) Justin T. Gibbs @ 2003-06-03 23:58 ` Marc-Christian Petersen 0 siblings, 0 replies; 39+ messages in thread From: Marc-Christian Petersen @ 2003-06-03 23:58 UTC (permalink / raw) To: Justin T. Gibbs, linux-scsi, linux-kernel Cc: Linus Torvalds, Alan Cox, Marcelo Tosatti On Wednesday 04 June 2003 01:51, Justin T. Gibbs wrote: Hi Justin, > I've just uploaded version 1.3.10 of the aic79xx driver and version > 6.2.36 of the aic7xxx driver. Both are available for 2.4.X and > 2.5.X kernels in either bk send format or as a tarball from here: many thanks! I'll update them for my tree (as always with your updates :-) ciao, Marc ^ permalink raw reply [flat|nested] 39+ messages in thread
* Re:
@ 2002-11-08 5:35 Randy.Dunlap
0 siblings, 0 replies; 39+ messages in thread
From: Randy.Dunlap @ 2002-11-08 5:35 UTC (permalink / raw)
To: linux-scsi
Doug wrote:
> | Is a section describing mid level functions provided
> | for LLDDs (e.g. scsi_adjust_queue_depth() ) warranted?
>
> Yes, please.
>
> [LLDD = low-level device driver]
| I am writing such a doc BTW, but I've been asleep for a few hours so I
| have chimed in ;-)
Yes, I've saved some of your recent emails that
describe some of the interfaces, but not all of them.
And it would be good to have it in one place.
--
~Randy
location: NPP-4E
^ permalink raw reply [flat|nested] 39+ messages in threadend of thread, other threads:[~2024-08-26 6:48 UTC | newest]
Thread overview: 39+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
[not found] <blk-mq updates>
[not found] ` <1397464212-4454-1-git-send-email-hch@lst.de>
2014-04-14 8:30 ` [PATCH 1/7] blk-mq: initialize resid_len Christoph Hellwig
2014-04-14 8:30 ` [PATCH 2/7] blk-mq: do not initialize req->special Christoph Hellwig
2014-04-14 8:30 ` [PATCH 3/7] blk-mq: make ->flush_rq fully transparent to drivers Christoph Hellwig
2014-04-14 8:30 ` [PATCH 4/7] blk-mq: add ->init_request and ->exit_request methods Christoph Hellwig
2014-04-14 8:30 ` [PATCH 5/7] blk-mq: initialize request on allocation Christoph Hellwig
2014-04-17 14:54 ` Ming Lei
2014-04-17 14:57 ` Christoph Hellwig
2014-04-17 15:07 ` Ming Lei
2014-04-14 8:30 ` [PATCH 6/7] blk-mq: split out tag initialization, support shared tags Christoph Hellwig
2014-04-14 8:30 ` [PATCH 7/7] block: all blk-mq requests are tagged Christoph Hellwig
2014-04-15 20:16 ` Jens Axboe
2024-08-22 20:54 [PATCH 2/2] scsi: ufs: core: Fix the code for entering hibernation Bao D. Nguyen
2024-08-22 21:08 ` Bart Van Assche
2024-08-23 12:01 ` Manivannan Sadhasivam
2024-08-23 14:23 ` Bart Van Assche
2024-08-23 14:58 ` Manivannan Sadhasivam
2024-08-23 16:07 ` Bart Van Assche
2024-08-23 16:48 ` Manivannan Sadhasivam
2024-08-23 18:05 ` Bart Van Assche
2024-08-24 2:29 ` Manivannan Sadhasivam
2024-08-24 2:48 ` Bart Van Assche
2024-08-24 3:03 ` Manivannan Sadhasivam
2024-08-26 6:48 ` Can Guo
-- strict thread matches above, loose matches on Subject: below --
2022-11-21 11:11 Denis Arefev
2022-11-21 14:28 ` Jason Yan
[not found] <20211011231530.GA22856@t>
2021-10-12 1:23 ` James Bottomley
2021-10-12 2:30 ` Bart Van Assche
[not found] <5e7dc543.vYG3wru8B/me1sOV%chenanqing@oppo.com>
2020-03-27 15:53 ` Re: Lee Duncan
2017-11-13 14:55 Re: Amos Kalonzo
2017-05-03 6:23 Re: H.A
2017-02-23 15:09 Qin's Yanjun
2015-08-19 13:01 christain147
[not found] <132D0DB4B968F242BE373429794F35C22559D38329@NHS-PCLI-MBC011.AD1.NHS.NET>
2015-06-08 11:09 ` Practice Trinity (NHS SOUTH SEFTON CCG)
2014-11-14 20:50 salim
2014-11-14 18:56 milke
[not found] <6A286AB51AD8EC4180C4B2E9EF1D0A027AAD7EFF1E@exmb01.wrschool.net>
2014-09-08 17:36 ` Deborah Mayher
2014-07-24 8:37 Richard Wong
2014-06-16 7:10 Re: Angela D.Dawes
2014-06-15 20:36 Re: Angela D.Dawes
[not found] <B719EF0A9FB7A247B5147CD67A83E60E011FEB76D1@EXCH10-MB3.paterson.k12.nj.us>
2013-08-23 10:47 ` Ruiz, Irma
2012-05-20 22:20 Mr. Peter Wong
2011-03-06 21:28 Augusta Mubarak
2011-03-03 14:20 RE: Lukas Thompson
2011-03-03 13:34 RE: Lukas Thompson
2010-11-18 15:48 [PATCH v3] dm mpath: add feature flag to control call to blk_abort_queue Mike Snitzer
2010-11-18 19:16 ` (unknown), Mike Snitzer
2010-11-18 19:21 ` Mike Snitzer
2010-07-01 10:49 (unknown) FUJITA Tomonori
2010-07-01 12:29 ` Jens Axboe
[not found] <KC3KJ12CL8BAAEB7@vger.kernel.org>
2005-07-24 10:31 ` Re: wolman
[not found] <1KFJEFB27B4F3155@vger.kernel.org>
2005-05-30 2:49 ` Re: sevillar
2005-02-26 14:57 Yong Haynes
2004-03-17 22:03 Kendrick Logan
2003-06-03 23:51 (unknown) Justin T. Gibbs
2003-06-03 23:58 ` Marc-Christian Petersen
2002-11-08 5:35 Re: Randy.Dunlap
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).