From: Douglas Gilbert <dgilbert@interlog.com>
To: linux-scsi@vger.kernel.org
Cc: martin.petersen@oracle.com, jejb@linux.vnet.ibm.com, hare@suse.de
Subject: [PATCH v18 80/83] sg: expand bvec usage; re-use bio_s
Date: Tue, 27 Apr 2021 17:57:30 -0400 [thread overview]
Message-ID: <20210427215733.417746-82-dgilbert@interlog.com> (raw)
In-Reply-To: <20210427215733.417746-1-dgilbert@interlog.com>
Rework sg_rq_map_kern() to use the newer bio_add_pc_page() and
blk_rq_append_bio() functions instead of blk_rq_map_kern(). Build
single bio with multiple bvec elements, each holding 1 or more
pages. This requires direct manipulation of this request
object field: nr_phys_segments .
Re-use bio_s. Having built a complex bio why throw it away after
one use if the driver knows (e.g. in mrq svb mode) that it will
be building exactly the same bio again and again? This requires
manipulating bio_get() and bio_put() plus remembering about
5 bio fields that are cleared by bio_reset().
More clearly mark that a request belongs to a larger multiple
requests (mrq) submission with the SG_FRQ_PC_PART_MRQ bit field.
Change the error strategy in sg_svb_mrq_first_come() and
sg_svb_mrq_ordered(): once started they must continue to move
forward, even in the face of errors. If not, then there will be
very hard to detect memory leaks or worse. Read-side requests
especially must not be left stranded.
Signed-off-by: Douglas Gilbert <dgilbert@interlog.com>
---
drivers/scsi/sg.c | 543 ++++++++++++++++++++++++++++------------------
1 file changed, 328 insertions(+), 215 deletions(-)
diff --git a/drivers/scsi/sg.c b/drivers/scsi/sg.c
index 0a0b40a8ab65..26047a8ff1e2 100644
--- a/drivers/scsi/sg.c
+++ b/drivers/scsi/sg.c
@@ -145,6 +145,7 @@ enum sg_shr_var {
#define SG_FRQ_PC_COUNT_ACTIVE 8 /* sfp->submitted + waiting active */
#define SG_FRQ_PC_ISSUED 9 /* blk_execute_rq_nowait() finished */
#define SG_FRQ_POLL_SLEPT 10 /* stop re-entry of hybrid_sleep() */
+#define SG_FRQ_PC_PART_MRQ 11 /* this cmd part of mrq array */
/* Bit positions (flags) for sg_fd::ffd_bm bitmask follow */
#define SG_FFD_FORCE_PACKID 0 /* receive only given pack_id/tag */
@@ -189,7 +190,6 @@ static int sg_allow_dio = SG_ALLOW_DIO_DEF; /* ignored by code */
static int scatter_elem_sz = SG_SCATTER_SZ;
static bool no_attach_msg;
-static atomic_t sg_tmp_count_reused_bios;
#define SG_DEF_SECTOR_SZ 512
@@ -226,10 +226,14 @@ struct sg_slice_hdr3 {
struct sg_slice_hdr4 { /* parts of sg_io_v4 object needed in async usage */
void __user *sbp; /* derived from sg_io_v4::response */
+ bio_end_io_t *bi_end_io;
u64 usr_ptr; /* hold sg_io_v4::usr_ptr as given (u64) */
int out_resid;
u32 wr_offset; /* from v4::spare_in when flagged; in bytes */
u32 wr_len; /* for shared reqs maybe < read-side */
+ unsigned int bi_size; /* reuse_bio: from original bio */
+ unsigned short bi_opf; /* reuse_bio: from original bio */
+ unsigned short bi_vcnt; /* reuse_bio: from original bio */
s16 dir; /* data xfer direction; SG_DXFER_* */
u16 cmd_len; /* truncated of sg_io_v4::request_len */
u16 max_sb_len; /* truncated of sg_io_v4::max_response_len */
@@ -260,6 +264,7 @@ struct sg_request { /* active SCSI command or inactive request */
u32 rq_idx; /* my index within parent's srp_arr */
u32 rq_info; /* info supplied by v3 and v4 interfaces */
u32 rq_result; /* packed scsi request result from LLD */
+ u32 rsv_arr_idx; /* my index in parentfp->rsv_arr */
int in_resid; /* requested-actual byte count on data-in */
int pack_id; /* v3 pack_id or in v4 request_extra field */
int sense_len; /* actual sense buffer length (data-in) */
@@ -304,9 +309,9 @@ struct sg_fd { /* holds the state of a file descriptor */
struct fasync_struct *async_qp; /* used by asynchronous notification */
struct eventfd_ctx *efd_ctxp; /* eventfd context or NULL */
struct xarray srp_arr; /* xarray of sg_request object pointers */
- struct sg_request *rsv_arr[SG_MAX_RSV_REQS];
struct kref f_ref;
struct execute_work ew_fd; /* harvest all fd resources and lists */
+ struct sg_request *rsv_arr[SG_MAX_RSV_REQS];
};
struct sg_device { /* holds the state of each scsi generic device */
@@ -354,6 +359,7 @@ struct sg_mrq_hold { /* for passing context between multiple requests (mrq) func
unsigned ordered_wr:1;
int id_of_mrq;
int s_res; /* secondary error: some-good-then-error; in co.spare_out */
+ int dtd_errs; /* incremented for each driver/transport/device error */
u32 cdb_mxlen; /* cdb length in cdb_ap, actual be may less */
u32 tot_reqs; /* total number of requests and cdb_s */
struct sg_comm_wr_t *cwrp; /* cwrp->h4p is mrq control object */
@@ -397,8 +403,7 @@ static struct sg_request *sg_mk_srp_sgat(struct sg_fd *sfp, bool first, int db_l
static int sg_abort_req(struct sg_fd *sfp, struct sg_request *srp);
static int sg_rq_chg_state(struct sg_request *srp, enum sg_rq_state old_st,
enum sg_rq_state new_st);
-static int sg_finish_rs_rq(struct sg_fd *sfp, struct sg_request *rs_srp,
- bool even_if_in_ws);
+static int sg_finish_rs_rq(struct sg_fd *sfp, struct sg_request *rs_srp, bool even_if_in_ws);
static void sg_rq_chg_state_force(struct sg_request *srp, enum sg_rq_state new_st);
static int sg_sfp_blk_poll(struct sg_fd *sfp, int loop_count);
static int sg_srp_q_blk_poll(struct sg_request *srp, struct request_queue *q,
@@ -1010,6 +1015,14 @@ sg_v4h_partial_zero(struct sg_io_v4 *h4p)
memset((u8 *)h4p + off, 0, SZ_SG_IO_V4 - off);
}
+static inline bool
+sg_v4_cmd_good(struct sg_io_v4 *h4p)
+{
+ return (scsi_status_is_good(h4p->device_status) &&
+ (h4p->driver_status & 0xf) == 0 &&
+ (h4p->transport_status & 0xff) == 0);
+}
+
static void
sg_sgat_zero(struct sg_scatter_hold *sgatp, int off, int nbytes)
{
@@ -1198,33 +1211,28 @@ sg_mrq_get_ready_srp(struct sg_fd *sfp, struct sg_request **srpp)
/* N.B. After this function is completed what srp points to should be considered invalid. */
static int
-sg_mrq_1complet(struct sg_mrq_hold *mhp, struct sg_fd *sfp, struct sg_request *srp)
+sg_mrq_1complet(struct sg_mrq_hold *mhp, struct sg_request *srp)
{
- int s_res, indx;
+ int res, indx;
int tot_reqs = mhp->tot_reqs;
+ struct sg_fd *sfp = srp->parentfp;
struct sg_io_v4 *hp;
struct sg_io_v4 *a_hds = mhp->a_hds;
struct sg_io_v4 *cop = mhp->cwrp->h4p;
- if (unlikely(!srp))
- return -EPROTO;
indx = srp->s_hdr4.mrq_ind;
- if (unlikely(srp->parentfp != sfp)) {
- SG_LOG(1, sfp, "%s: mrq_ind=%d, sfp out-of-sync\n", __func__, indx);
- return -EPROTO;
- }
SG_LOG(3, sfp, "%s: %s, mrq_ind=%d, pack_id=%d\n", __func__, sg_side_str(srp), indx,
srp->pack_id);
if (unlikely(indx < 0 || indx >= tot_reqs))
return -EPROTO;
hp = a_hds + indx;
- s_res = sg_receive_v4(sfp, srp, NULL, hp);
- if (unlikely(!sg_result_is_good(srp->rq_result)))
+ res = sg_receive_v4(sfp, srp, NULL, hp);
+ if (unlikely(res))
+ return res;
+ if (unlikely(!sg_v4_cmd_good(hp)))
SG_LOG(2, sfp, "%s: %s, bad status: drv/tran/scsi=0x%x/0x%x/0x%x\n",
__func__, sg_side_str(srp), hp->driver_status,
hp->transport_status, hp->device_status);
- if (unlikely(s_res == -EFAULT))
- return s_res;
hp->info |= SG_INFO_MRQ_FINI;
++cop->info;
if (cop->din_xfer_len > 0)
@@ -1239,9 +1247,9 @@ sg_mrq_1complet(struct sg_mrq_hold *mhp, struct sg_fd *sfp, struct sg_request *s
pr_info("%s: eventfd_signal problem\n", __func__);
}
} else if (sfp->async_qp && (hp->flags & SGV4_FLAG_SIGNAL)) {
- s_res = sg_mrq_arr_flush(mhp);
- if (unlikely(s_res)) /* can only be -EFAULT */
- return s_res;
+ res = sg_mrq_arr_flush(mhp);
+ if (unlikely(res))
+ return res;
kill_fasync(&sfp->async_qp, SIGPOLL, POLL_IN);
}
return 0;
@@ -1382,8 +1390,6 @@ sg_mrq_poll_either(struct sg_fd *sfp, struct sg_fd *sec_sfp, bool *on_sfp)
if (sfp) {
if (sg_mrq_get_ready_srp(sfp, &srp)) {
__set_current_state(TASK_RUNNING);
- if (!srp)
- return ERR_PTR(-ENODEV);
*on_sfp = true;
return srp;
}
@@ -1391,8 +1397,6 @@ sg_mrq_poll_either(struct sg_fd *sfp, struct sg_fd *sec_sfp, bool *on_sfp)
if (sec_sfp && sfp != sec_sfp) {
if (sg_mrq_get_ready_srp(sec_sfp, &srp)) {
__set_current_state(TASK_RUNNING);
- if (!srp)
- return ERR_PTR(-ENODEV);
*on_sfp = false;
return srp;
}
@@ -1420,14 +1424,18 @@ sg_mrq_complets(struct sg_mrq_hold *mhp, struct sg_fd *sfp, struct sg_fd *sec_sf
SG_LOG(3, sfp, "%s: mreqs=%d, sec_reqs=%d\n", __func__, mreqs, sec_reqs);
while (mreqs + sec_reqs > 0) {
while (mreqs > 0 && sg_mrq_get_ready_srp(sfp, &srp)) {
+ if (IS_ERR(srp))
+ return PTR_ERR(srp);
--mreqs;
- res = sg_mrq_1complet(mhp, sfp, srp);
+ res = sg_mrq_1complet(mhp, srp);
if (unlikely(res))
return res;
}
while (sec_reqs > 0 && sg_mrq_get_ready_srp(sec_sfp, &srp)) {
+ if (IS_ERR(srp))
+ return PTR_ERR(srp);
--sec_reqs;
- res = sg_mrq_1complet(mhp, sec_sfp, srp);
+ res = sg_mrq_1complet(mhp, srp);
if (unlikely(res))
return res;
}
@@ -1443,7 +1451,7 @@ sg_mrq_complets(struct sg_mrq_hold *mhp, struct sg_fd *sfp, struct sg_fd *sec_sf
mreqs = 0;
} else {
--mreqs;
- res = sg_mrq_1complet(mhp, sfp, srp);
+ res = sg_mrq_1complet(mhp, srp);
if (unlikely(res))
return res;
}
@@ -1456,7 +1464,7 @@ sg_mrq_complets(struct sg_mrq_hold *mhp, struct sg_fd *sfp, struct sg_fd *sec_sf
sec_reqs = 0;
} else {
--sec_reqs;
- res = sg_mrq_1complet(mhp, sec_sfp, srp);
+ res = sg_mrq_1complet(mhp, srp);
if (unlikely(res))
return res;
}
@@ -1470,12 +1478,12 @@ sg_mrq_complets(struct sg_mrq_hold *mhp, struct sg_fd *sfp, struct sg_fd *sec_sf
return PTR_ERR(srp);
if (on_sfp) {
--mreqs;
- res = sg_mrq_1complet(mhp, sfp, srp);
+ res = sg_mrq_1complet(mhp, srp);
if (unlikely(res))
return res;
} else {
--sec_reqs;
- res = sg_mrq_1complet(mhp, sec_sfp, srp);
+ res = sg_mrq_1complet(mhp, srp);
if (unlikely(res))
return res;
}
@@ -1649,6 +1657,7 @@ sg_mrq_submit(struct sg_fd *rq_sfp, struct sg_mrq_hold *mhp, int pos_in_rq_arr,
ul_timeout = msecs_to_jiffies(hp->timeout);
__assign_bit(SG_FRQ_PC_SYNC_INVOC, r_cwrp->frq_pc_bm, (int)mhp->from_sg_io);
__set_bit(SG_FRQ_PC_IS_V4I, r_cwrp->frq_pc_bm);
+ __set_bit(SG_FRQ_PC_PART_MRQ, r_cwrp->frq_pc_bm);
r_cwrp->h4p = hp;
r_cwrp->dlen = hp->din_xfer_len ? hp->din_xfer_len : hp->dout_xfer_len;
r_cwrp->timeout = min_t(unsigned long, ul_timeout, INT_MAX);
@@ -1704,15 +1713,18 @@ sg_process_most_mrq(struct sg_fd *fp, struct sg_fd *o_sfp, struct sg_mrq_hold *m
++other_fp_sent;
continue; /* defer completion until all submitted */
}
- mhp->s_res = sg_wait_poll_for_given_srp(rq_sfp, srp, mhp->hipri);
- if (unlikely(mhp->s_res)) {
- if (mhp->s_res == -ERESTARTSYS || mhp->s_res == -ENODEV)
- return mhp->s_res;
+ res = sg_wait_poll_for_given_srp(rq_sfp, srp, mhp->hipri);
+ if (unlikely(res)) {
+ mhp->s_res = res;
+ if (res == -ERESTARTSYS || res == -ENODEV)
+ return res;
break;
}
- res = sg_mrq_1complet(mhp, rq_sfp, srp);
- if (unlikely(res))
+ res = sg_mrq_1complet(mhp, srp);
+ if (unlikely(res)) {
+ mhp->s_res = res;
break;
+ }
++num_cmpl;
} /* end of dispatch request and optionally wait response loop */
cop->dout_resid = mhp->tot_reqs - num_subm;
@@ -1725,28 +1737,54 @@ sg_process_most_mrq(struct sg_fd *fp, struct sg_fd *o_sfp, struct sg_mrq_hold *m
if (mhp->immed)
return res;
if (likely(res == 0 && (this_fp_sent + other_fp_sent) > 0)) {
- mhp->s_res = sg_mrq_complets(mhp, fp, o_sfp, this_fp_sent, other_fp_sent);
- if (unlikely(mhp->s_res == -EFAULT || mhp->s_res == -ERESTARTSYS))
- res = mhp->s_res; /* this may leave orphans */
+ res = sg_mrq_complets(mhp, fp, o_sfp, this_fp_sent, other_fp_sent);
+ if (res)
+ mhp->s_res = res; /* this may leave orphans */
}
if (mhp->id_of_mrq) /* can no longer do a mrq abort */
atomic_set(&fp->mrq_id_abort, 0);
return res;
}
+static bool
+sg_svb_err_process(struct sg_mrq_hold *mhp, int m_ind, struct sg_fd *fp, int res, bool rs)
+{
+ __maybe_unused const char *ss = rs ? "read-side" : "write-side";
+
+ if (res) {
+ if (mhp->s_res == 0)
+ mhp->s_res = res;
+ SG_LOG(1, fp, "%s: %s failure, res=%d\n", __func__, ss, res);
+ } else {
+ struct sg_io_v4 *hp = mhp->a_hds + m_ind;
+
+ ++mhp->dtd_errs;
+ SG_LOG(2, fp, "%s: %s, bad status: drv/trans_host/scsi=0x%x/0x%x/0x%x\n",
+ __func__, ss, hp->driver_status, hp->transport_status, hp->device_status);
+ }
+ return mhp->stop_if;
+}
+
+static inline void
+sg_svb_zero_elem(struct sg_svb_elem *svb_ap, int m)
+{
+ svb_ap[m].rs_srp = NULL;
+ svb_ap[m].prev_ws_srp = NULL;
+}
+
/* For multiple requests (mrq) share variable blocking (svb) with no SGV4_FLAG_ORDERED_WR */
static int
sg_svb_mrq_first_come(struct sg_fd *fp, struct sg_fd *o_sfp, struct sg_mrq_hold *mhp, int ra_ind,
int *num_submp)
{
bool chk_oth_first = false;
+ bool stop_triggered = false;
bool is_first = true;
+ bool rs_fail;
enum sg_rq_state rq_st;
int this_fp_sent = 0;
int other_fp_sent = 0;
- int res = 0;
- int first_err = 0;
- int k, m, idx, ws_pos, num_reads, sent, dir;
+ int k, m, res, idx, ws_pos, num_reads, sent, dir, m_ind;
struct sg_io_v4 *hp = mhp->a_hds + ra_ind;
struct sg_request *srp;
struct sg_request *rs_srp;
@@ -1778,15 +1816,13 @@ sg_svb_mrq_first_come(struct sg_fd *fp, struct sg_fd *o_sfp, struct sg_mrq_hold
}
if (IS_ERR(srp)) {
mhp->s_res = PTR_ERR(srp);
- if (first_err == 0)
- first_err = mhp->s_res;
SG_LOG(1, fp, "%s: sg_mrq_submit() err: %d\n", __func__, mhp->s_res);
break; /* stop doing rs submits */
}
++*num_submp;
+ srp->s_hdr4.mrq_ind = ra_ind;
if (hp->din_xfer_len > 0)
svb_arr[k].rs_srp = srp;
- srp->s_hdr4.mrq_ind = ra_ind;
if (mhp->chk_abort)
atomic_set(&srp->s_hdr4.pack_id_of_mrq, mhp->id_of_mrq);
} /* end of read-side submission, write-side defer loop */
@@ -1794,26 +1830,41 @@ sg_svb_mrq_first_come(struct sg_fd *fp, struct sg_fd *o_sfp, struct sg_mrq_hold
num_reads = k;
sent = this_fp_sent + other_fp_sent;
+ /*
+ * It is important _not_ to break out of this loop as that will lead to hard to detect
+ * memory leaks. We must wait for inflight requests to complete before final cleanup.
+ */
for (k = 0; k < sent; ++k) {
if (other_fp_sent > 0 && sg_mrq_get_ready_srp(o_sfp, &srp)) {
+ if (IS_ERR(srp)) {
+ mhp->s_res = PTR_ERR(srp);
+ continue;
+ }
other_found:
--other_fp_sent;
- res = sg_mrq_1complet(mhp, o_sfp, srp);
- if (unlikely(res))
- break;
+ m_ind = srp->s_hdr4.mrq_ind;
+ res = sg_mrq_1complet(mhp, srp);
+ if (unlikely(res || !sg_v4_cmd_good(mhp->a_hds + m_ind)))
+ stop_triggered = sg_svb_err_process(mhp, m_ind, o_sfp, res, false);
continue; /* do available submits first */
}
if (this_fp_sent > 0 && sg_mrq_get_ready_srp(fp, &srp)) {
+ if (IS_ERR(srp)) {
+ mhp->s_res = PTR_ERR(srp);
+ continue;
+ }
this_found:
--this_fp_sent;
dir = srp->s_hdr4.dir;
- res = sg_mrq_1complet(mhp, fp, srp);
- if (unlikely(res))
- break;
+ rs_fail = false;
+ m_ind = srp->s_hdr4.mrq_ind;
+ res = sg_mrq_1complet(mhp, srp);
+ if (unlikely(res || !sg_v4_cmd_good(mhp->a_hds + m_ind))) {
+ rs_fail = true;
+ stop_triggered = sg_svb_err_process(mhp, m_ind, fp, res, true);
+ }
if (dir != SG_DXFER_FROM_DEV)
continue;
- if (test_bit(SG_FFD_READ_SIDE_ERR, fp->ffd_bm))
- continue;
/* read-side req completed, submit its write-side(s) */
rs_srp = srp;
for (m = 0; m < num_reads; ++m) {
@@ -1825,37 +1876,27 @@ sg_svb_mrq_first_come(struct sg_fd *fp, struct sg_fd *o_sfp, struct sg_mrq_hold
srp->pack_id);
continue;
}
+ if (stop_triggered || rs_fail) {
+ sg_svb_zero_elem(svb_arr, m);
+ continue;
+ }
rq_st = atomic_read(&rs_srp->rq_st);
- if (rq_st == SG_RQ_INACTIVE)
- continue; /* probably an error, bypass paired write-side rq */
- else if (rq_st != SG_RQ_SHR_SWAP) {
+ if (rq_st == SG_RQ_INACTIVE) {
+ sg_svb_zero_elem(svb_arr, m);
+ continue; /* probably an error, bypass paired write-side rq */
+ } else if (rq_st != SG_RQ_SHR_SWAP) {
SG_LOG(1, fp, "%s: expect rs_srp to be in shr_swap\n", __func__);
- res = -EPROTO;
- break;
+ mhp->s_res = -EPROTO;
+ sg_svb_zero_elem(svb_arr, m);
+ continue;
}
ws_pos = svb_arr[m].ws_pos;
- for (idx = 0; idx < SG_MAX_RSV_REQS; ++idx) {
- if (fp->rsv_arr[idx] == srp)
- break;
- }
- if (idx >= SG_MAX_RSV_REQS) {
- SG_LOG(1, fp, "%s: srp not in rsv_arr\n", __func__);
- res = -EPROTO;
- break;
- }
+ idx = srp->rsv_arr_idx;
SG_LOG(6, o_sfp, "%s: ws_pos=%d, rs_idx=%d\n", __func__, ws_pos, idx);
srp = sg_mrq_submit(o_sfp, mhp, ws_pos, idx, svb_arr[m].prev_ws_srp);
if (IS_ERR(srp)) {
+ sg_svb_zero_elem(svb_arr, m);
mhp->s_res = PTR_ERR(srp);
- if (mhp->s_res == -EFBIG) { /* out of reserve slots */
- if (first_err)
- break;
- res = mhp->s_res;
- break;
- }
- if (first_err == 0)
- first_err = mhp->s_res;
- svb_arr[m].prev_ws_srp = NULL;
SG_LOG(1, o_sfp, "%s: sg_mrq_submit(oth)->%d\n", __func__,
mhp->s_res);
continue;
@@ -1876,8 +1917,11 @@ sg_svb_mrq_first_come(struct sg_fd *fp, struct sg_fd *o_sfp, struct sg_mrq_hold
if (this_fp_sent > 0) {
res = sg_wait_any_mrq(fp, &srp);
if (unlikely(res))
- break;
- goto this_found;
+ mhp->s_res = res;
+ else if (IS_ERR(srp))
+ mhp->s_res = PTR_ERR(srp);
+ else
+ goto this_found;
}
if (chk_oth_first)
continue;
@@ -1885,24 +1929,31 @@ sg_svb_mrq_first_come(struct sg_fd *fp, struct sg_fd *o_sfp, struct sg_mrq_hold
if (other_fp_sent > 0) {
res = sg_wait_any_mrq(o_sfp, &srp);
if (unlikely(res))
- break;
- goto other_found;
+ mhp->s_res = res;
+ else if (IS_ERR(srp))
+ mhp->s_res = PTR_ERR(srp);
+ else
+ goto other_found;
}
if (chk_oth_first)
goto this_second;
} /* end of loop for deferred ws submits and rs+ws responses */
- if (res == 0 && first_err)
- res = first_err;
- return res;
+ if (mhp->s_res) {
+ if (mhp->stop_if)
+ stop_triggered = true;
+ }
+ return stop_triggered ? -ECANCELED : 0;
}
static int
sg_svb_mrq_ordered(struct sg_fd *fp, struct sg_fd *o_sfp, struct sg_mrq_hold *mhp, int ra_ind,
int *num_submp)
{
+ bool stop_triggered = false;
+ bool rs_fail;
enum sg_rq_state rq_st;
- int k, m, res, idx, ws_pos, num_reads;
+ int k, m, res, idx, ws_pos, num_reads, m_ind;
int this_fp_sent = 0;
int other_fp_sent = 0;
struct sg_io_v4 *hp = mhp->a_hds + ra_ind;
@@ -1949,45 +2000,46 @@ sg_svb_mrq_ordered(struct sg_fd *fp, struct sg_fd *o_sfp, struct sg_mrq_hold *mh
num_reads = k;
if (this_fp_sent + other_fp_sent <= 0)
- return 0;
+ goto fini;
+
for (m = 0; m < num_reads; ++m) {
rs_srp = svb_arr[m].rs_srp;
if (!rs_srp)
continue;
res = sg_wait_poll_for_given_srp(fp, rs_srp, mhp->hipri);
if (unlikely(res))
- return res;
+ mhp->s_res = res;
--this_fp_sent;
- res = sg_mrq_1complet(mhp, fp, rs_srp);
- if (unlikely(res))
- return res;
- if (test_bit(SG_FFD_READ_SIDE_ERR, fp->ffd_bm))
+ rs_fail = false;
+ m_ind = rs_srp->s_hdr4.mrq_ind;
+ res = sg_mrq_1complet(mhp, rs_srp);
+ if (unlikely(res || !sg_v4_cmd_good(mhp->a_hds + m_ind))) {
+ rs_fail = true;
+ stop_triggered = sg_svb_err_process(mhp, m_ind, fp, res, true);
+ }
+ if (unlikely(stop_triggered || rs_fail)) {
+ sg_svb_zero_elem(svb_arr, m);
continue;
+ }
rq_st = atomic_read(&rs_srp->rq_st);
- if (rq_st == SG_RQ_INACTIVE)
+ if (rq_st == SG_RQ_INACTIVE) {
+ sg_svb_zero_elem(svb_arr, m);
continue; /* probably an error, bypass paired write-side rq */
- else if (rq_st != SG_RQ_SHR_SWAP) {
+ } else if (rq_st != SG_RQ_SHR_SWAP) {
SG_LOG(1, fp, "%s: expect rs_srp to be in shr_swap\n", __func__);
- res = -EPROTO;
- break;
+ mhp->s_res = -EPROTO;
+ sg_svb_zero_elem(svb_arr, m);
+ continue;
}
ws_pos = svb_arr[m].ws_pos;
- for (idx = 0; idx < SG_MAX_RSV_REQS; ++idx) {
- if (fp->rsv_arr[idx] == rs_srp)
- break;
- }
- if (idx >= SG_MAX_RSV_REQS) {
- SG_LOG(1, rs_srp->parentfp, "%s: srp not in rsv_arr\n", __func__);
- res = -EPROTO;
- return res;
- }
+ idx = rs_srp->rsv_arr_idx;
SG_LOG(6, o_sfp, "%s: ws_pos=%d, rs_idx=%d\n", __func__, ws_pos, idx);
srp = sg_mrq_submit(o_sfp, mhp, ws_pos, idx, svb_arr[m].prev_ws_srp);
if (IS_ERR(srp)) {
+ sg_svb_zero_elem(svb_arr, m);
mhp->s_res = PTR_ERR(srp);
- res = mhp->s_res;
- SG_LOG(1, o_sfp, "%s: mrq_submit(oth)->%d\n", __func__, res);
- return res;
+ SG_LOG(1, o_sfp, "%s: mrq_submit(oth)->%d\n", __func__, mhp->s_res);
+ continue;
}
svb_arr[m].prev_ws_srp = srp;
++*num_submp;
@@ -1997,24 +2049,54 @@ sg_svb_mrq_ordered(struct sg_fd *fp, struct sg_fd *o_sfp, struct sg_mrq_hold *mh
atomic_set(&srp->s_hdr4.pack_id_of_mrq, mhp->id_of_mrq);
}
while (this_fp_sent > 0) { /* non-data requests */
- res = sg_wait_any_mrq(fp, &srp);
- if (unlikely(res))
- return res;
--this_fp_sent;
- res = sg_mrq_1complet(mhp, fp, srp);
+ res = sg_wait_any_mrq(fp, &srp);
+ if (unlikely(res)) {
+ mhp->s_res = res;
+ continue;
+ }
+ if (IS_ERR(srp)) {
+ mhp->s_res = PTR_ERR(srp);
+ continue;
+ }
+ m_ind = srp->s_hdr4.mrq_ind;
+ res = sg_mrq_1complet(mhp, srp);
if (unlikely(res))
- return res;
+ mhp->s_res = res;
}
while (other_fp_sent > 0) {
- res = sg_wait_any_mrq(o_sfp, &srp);
- if (unlikely(res))
- return res;
--other_fp_sent;
- res = sg_mrq_1complet(mhp, o_sfp, srp);
+ res = sg_wait_any_mrq(o_sfp, &srp);
+ if (unlikely(res)) {
+ mhp->s_res = res;
+ continue;
+ }
+ if (IS_ERR(srp)) {
+ mhp->s_res = PTR_ERR(srp);
+ continue;
+ }
+ m_ind = srp->s_hdr4.mrq_ind;
+ res = sg_mrq_1complet(mhp, srp);
if (unlikely(res))
- return res;
+ mhp->s_res = res;
+ }
+fini:
+ if (mhp->s_res) {
+ if (mhp->stop_if)
+ stop_triggered = true;
+ }
+ return stop_triggered ? -ECANCELED : 0;
+}
+
+static inline void
+sg_svb_srp_cleanup(struct sg_request *srp)
+{
+ if (test_and_clear_bit(SG_FRQ_LT_REUSE_BIO, srp->frq_lt_bm)) {
+ if (srp->bio) {
+ bio_put(srp->bio); /* _get() near end of sg_start_req() */
+ srp->bio = NULL;
+ }
}
- return 0;
}
static void
@@ -2024,12 +2106,8 @@ sg_svb_cleanup(struct sg_fd *sfp)
struct xarray *xafp = &sfp->srp_arr;
struct sg_request *srp;
- xa_for_each_marked(xafp, idx, srp, SG_XA_RQ_INACTIVE) {
- if (test_and_clear_bit(SG_FRQ_LT_REUSE_BIO, srp->frq_lt_bm)) {
- bio_put(srp->bio); /* _get() near end of sg_start_req() */
- srp->bio = NULL;
- }
- }
+ xa_for_each_marked(xafp, idx, srp, SG_XA_RQ_INACTIVE)
+ sg_svb_srp_cleanup(srp);
}
/*
@@ -2089,7 +2167,7 @@ sg_process_svb_mrq(struct sg_fd *fp, struct sg_fd *o_sfp, struct sg_mrq_hold *mh
atomic_set(&fp->mrq_id_abort, 0);
if (test_and_clear_bit(SG_FFD_CAN_REUSE_BIO, fp->ffd_bm))
sg_svb_cleanup(fp);
- return res;
+ return res == -ECANCELED ? 0 : res;
}
#if IS_ENABLED(SG_LOG_ACTIVE)
@@ -2156,6 +2234,7 @@ sg_do_multi_req(struct sg_comm_wr_t *cwrp, bool from_sg_io)
mhp->id_of_mrq = (int)cop->request_extra;
mhp->tot_reqs = tot_reqs;
mhp->s_res = 0;
+ mhp->dtd_errs = 0;
if (mhp->id_of_mrq) {
existing_id = atomic_cmpxchg(&fp->mrq_id_abort, 0, mhp->id_of_mrq);
if (existing_id && existing_id != mhp->id_of_mrq) {
@@ -2205,11 +2284,15 @@ sg_do_multi_req(struct sg_comm_wr_t *cwrp, bool from_sg_io)
else
return -EPROTO;
}
+ if (din_len > 0) {
+ if (unlikely(din_len > SG_MAX_MULTI_REQ_SZ))
+ return -E2BIG;
+ } else if (dout_len > 0) {
+ if (unlikely(dout_len > SG_MAX_MULTI_REQ_SZ))
+ return -E2BIG;
+ }
if (unlikely(tot_reqs > U16_MAX)) {
return -ERANGE;
- } else if (unlikely(dout_len > SG_MAX_MULTI_REQ_SZ || din_len > SG_MAX_MULTI_REQ_SZ ||
- cdb_alen > SG_MAX_MULTI_REQ_SZ)) {
- return -E2BIG;
} else if (unlikely(mhp->immed && mhp->stop_if)) {
return -ERANGE;
} else if (unlikely(tot_reqs == 0)) {
@@ -2217,6 +2300,8 @@ sg_do_multi_req(struct sg_comm_wr_t *cwrp, bool from_sg_io)
} else if (unlikely(!!cdb_alen != !!cop->request)) {
return -ERANGE; /* both must be zero or both non-zero */
} else if (cdb_alen) {
+ if (unlikely(cdb_alen > SG_MAX_MULTI_REQ_SZ))
+ return -E2BIG;
if (unlikely(cdb_alen % tot_reqs))
return -ERANGE;
cdb_mxlen = cdb_alen / tot_reqs;
@@ -2681,7 +2766,6 @@ sg_get_rsv_str(struct sg_request *srp, const char *leadin, const char *leadout,
{
int k, i_len, o_len, len;
struct sg_fd *sfp;
- struct sg_request **rapp;
if (!b || b_len < 1)
return b;
@@ -2696,13 +2780,9 @@ sg_get_rsv_str(struct sg_request *srp, const char *leadin, const char *leadout,
sfp = srp->parentfp;
if (!sfp)
goto blank;
- rapp = sfp->rsv_arr;
- for (k = 0; k < SG_MAX_RSV_REQS; ++k, ++rapp) {
- if (srp == *rapp)
- break;
- }
- if (k >= SG_MAX_RSV_REQS)
+ if (!test_bit(SG_FRQ_LT_RESERVED, srp->frq_lt_bm))
goto blank;
+ k = srp->rsv_arr_idx;
scnprintf(b, b_len, "%srsv%d%s", leadin, k, leadout);
return b;
blank:
@@ -2803,8 +2883,6 @@ sg_common_write(struct sg_comm_wr_t *cwrp)
rq_flags = hi_p->flags;
pack_id = hi_p->pack_id;
}
- if (unlikely(rq_flags & SGV4_FLAG_MULTIPLE_REQS))
- return ERR_PTR(-ERANGE); /* only control object sets this */
if (sg_fd_is_shared(fp)) {
res = sg_share_chk_flags(fp, rq_flags, dlen, dir, &sh_var);
if (unlikely(res < 0))
@@ -2852,6 +2930,12 @@ sg_common_write(struct sg_comm_wr_t *cwrp)
sg_execute_cmd(fp, srp);
return srp;
err_out:
+ if (srp->sh_var == SG_SHR_WS_RQ) {
+ struct sg_request *rs_srp = srp->sh_srp;
+
+ if (rs_srp)
+ sg_finish_rs_rq(NULL, rs_srp, true);
+ }
sg_deact_request(fp, srp);
return ERR_PTR(res);
}
@@ -3644,70 +3728,81 @@ sg_finish_rs_rq(struct sg_fd *sfp, struct sg_request *rs_srp, bool even_if_in_ws
bool found_one = false;
int res = -EINVAL;
int k;
- enum sg_rq_state sr_st;
+ enum sg_rq_state rq_st;
unsigned long iflags;
struct sg_fd *rs_sfp;
- struct sg_request *rs_rsv_srp;
- struct sg_device *sdp = sfp->parentdp;
-
- rs_sfp = sg_fd_share_ptr(sfp);
- if (unlikely(!rs_sfp))
- goto fini;
- if (xa_get_mark(&sdp->sfp_arr, sfp->idx, SG_XA_FD_RS_SHARE))
- rs_sfp = sfp;
- for (k = 0; k < SG_MAX_RSV_REQS; ++k) {
- res = -EINVAL;
- rs_rsv_srp = rs_sfp->rsv_arr[k];
- if (rs_srp) {
- if (rs_srp != rs_rsv_srp)
+ if (rs_srp) {
+ if (rs_srp->sh_var != SG_SHR_RS_RQ) {
+ res = -EPROTO;
+ goto err;
+ }
+ rs_sfp = rs_srp->parentfp;
+ } else {
+ if (!sfp)
+ goto err;
+ if (xa_get_mark(&sfp->parentdp->sfp_arr, sfp->idx, SG_XA_FD_RS_SHARE)) {
+ rs_sfp = sfp;
+ } else if (sg_fd_is_shared(sfp)) { /* must be write-side */
+ rs_sfp = sg_fd_share_ptr(sfp);
+ } else {
+ pr_warn("%s: non sharing fd given\n", __func__);
+ res = -EINVAL;
+ goto err;
+ }
+ for (k = 0; k < SG_MAX_RSV_REQS; ++k) {
+ rs_srp = rs_sfp->rsv_arr[k];
+ if (IS_ERR_OR_NULL(rs_srp))
continue;
+ if (atomic_read(&rs_srp->rq_st) == SG_RQ_SHR_SWAP)
+ break;
}
- if (IS_ERR_OR_NULL(rs_rsv_srp))
- continue;
- xa_lock_irqsave(&rs_sfp->srp_arr, iflags);
- sr_st = atomic_read(&rs_rsv_srp->rq_st);
- switch (sr_st) {
- case SG_RQ_SHR_SWAP:
- found_one = true;
- break;
- case SG_RQ_SHR_IN_WS:
- if (even_if_in_ws)
- found_one = true;
- else
- res = -EBUSY;
- break;
- case SG_RQ_BUSY:
- res = -EBUSY;
- break;
- default:
+ if (k >= SG_MAX_RSV_REQS) {
res = -EINVAL;
- break;
+ goto fini;
}
- if (found_one)
- goto found;
- xa_unlock_irqrestore(&rs_sfp->srp_arr, iflags);
- if (rs_srp)
- return res; /* found rs_srp but was in wrong state */
}
-fini:
+ xa_lock_irqsave(&rs_sfp->srp_arr, iflags);
+ rq_st = atomic_read(&rs_srp->rq_st);
+ switch (rq_st) {
+ case SG_RQ_SHR_SWAP:
+ found_one = true;
+ break;
+ case SG_RQ_SHR_IN_WS:
+ case SG_RQ_BUSY:
+ if (even_if_in_ws)
+ found_one = true;
+ else
+ res = -EBUSY;
+ break;
+ default:
+ res = -EINVAL;
+ break;
+ }
+ if (found_one)
+ goto found;
+ xa_unlock_irqrestore(&rs_sfp->srp_arr, iflags);
+err:
if (unlikely(res))
- SG_LOG(1, sfp, "%s: err=%d\n", __func__, -res);
+ SG_LOG(1, rs_sfp, "%s: err=%d\n", __func__, -res);
+ if (rs_srp)
+ goto fini;
return res;
found:
- res = sg_rq_chg_state_ulck(rs_rsv_srp, sr_st, SG_RQ_BUSY);
+ res = sg_rq_chg_state_ulck(rs_srp, rq_st, SG_RQ_BUSY);
if (!res)
atomic_inc(&rs_sfp->inactives);
- rs_rsv_srp->tag = SG_TAG_WILDCARD;
- rs_rsv_srp->sh_var = SG_SHR_NONE;
- rs_rsv_srp->in_resid = 0;
- rs_rsv_srp->rq_info = 0;
- rs_rsv_srp->sense_len = 0;
- rs_rsv_srp->sh_srp = NULL;
+ rs_srp->tag = SG_TAG_WILDCARD;
+ rs_srp->sh_var = SG_SHR_NONE;
+ rs_srp->in_resid = 0;
+ rs_srp->rq_info = 0;
+ rs_srp->sense_len = 0;
+ rs_srp->sh_srp = NULL;
xa_unlock_irqrestore(&rs_sfp->srp_arr, iflags);
- sg_finish_scsi_blk_rq(rs_rsv_srp);
- sg_deact_request(rs_rsv_srp->parentfp, rs_rsv_srp);
- return 0;
+fini:
+ sg_finish_scsi_blk_rq(rs_srp);
+ sg_deact_request(rs_sfp, rs_srp);
+ return res;
}
static void
@@ -4126,6 +4221,8 @@ sg_match_first_mrq_after(struct sg_fd *sfp, int pack_id, struct sg_request *afte
look_for_after = false;
continue;
}
+ if (!test_bit(SG_FRQ_PC_PART_MRQ, srp->frq_pc_bm))
+ continue;
id = atomic_read(&srp->s_hdr4.pack_id_of_mrq);
if (id == 0) /* mrq_pack_ids cannot be zero */
continue;
@@ -4584,6 +4681,7 @@ sg_set_reserved_sz(struct sg_fd *sfp, int want_rsv_sz)
__clear_bit(SG_FRQ_LT_RESERVED, o_srp->frq_lt_bm);
__set_bit(SG_FRQ_LT_RESERVED, t_srp->frq_lt_bm);
__assign_bit(SG_FRQ_LT_REUSE_BIO, t_srp->frq_lt_bm, is_reuse_bio);
+ o_srp->rsv_arr_idx = 0;
*rapp = t_srp;
xa_unlock_irqrestore(xafp, iflags);
sg_remove_srp(n_srp);
@@ -5871,7 +5969,8 @@ sg_rq_end_io(struct request *rqq, blk_status_t status)
u64 n = eventfd_signal(sfp->efd_ctxp, 1);
if (n != 1)
- pr_info("%s: srp=%pK eventfd_signal problem\n", __func__, srp);
+ pr_info("%s: srp->pack_id=%d eventfd_signal problem\n", __func__,
+ srp->pack_id);
}
kref_put(&sfp->f_ref, sg_remove_sfp); /* get in: sg_execute_cmd() */
}
@@ -6259,7 +6358,7 @@ sg_rq_map_kern(struct sg_request *srp, struct request_queue *q, struct request *
static int
sg_start_req(struct sg_request *srp, struct sg_comm_wr_t *cwrp, int dxfer_dir)
{
- bool bump_bio_get = false;
+ bool first_reuse_bio = false;
bool no_dxfer, us_xfer;
int res = 0;
int dlen = cwrp->dlen;
@@ -6271,7 +6370,7 @@ sg_start_req(struct sg_request *srp, struct sg_comm_wr_t *cwrp, int dxfer_dir)
struct scsi_request *scsi_rp;
struct sg_fd *sfp = cwrp->sfp;
struct sg_device *sdp;
- struct sg_scatter_hold *req_schp;
+ struct sg_scatter_hold *schp;
struct request_queue *q;
struct rq_map_data *md = (void *)srp; /* want any non-NULL value */
u8 *long_cmdp = NULL;
@@ -6349,7 +6448,7 @@ sg_start_req(struct sg_request *srp, struct sg_comm_wr_t *cwrp, int dxfer_dir)
__assign_bit(SG_FRQ_PC_US_XFER, srp->frq_pc_bm, !no_dxfer && us_xfer);
rqq->end_io_data = srp;
scsi_rp->retries = SG_DEFAULT_RETRIES;
- req_schp = srp->sgatp;
+ schp = srp->sgatp;
if (no_dxfer) {
SG_LOG(4, sfp, "%s: no data xfer [0x%pK]\n", __func__, srp);
@@ -6361,22 +6460,33 @@ sg_start_req(struct sg_request *srp, struct sg_comm_wr_t *cwrp, int dxfer_dir)
md = NULL;
if (IS_ENABLED(CONFIG_SCSI_PROC_FS))
cp = "direct_io, ";
- } else if (test_bit(SG_FFD_CAN_REUSE_BIO, sfp->ffd_bm)) {
+ } else if (test_bit(SG_FFD_CAN_REUSE_BIO, sfp->ffd_bm) &&
+ test_bit(SG_FRQ_PC_PART_MRQ, srp->frq_pc_bm)) {
if (test_bit(SG_FRQ_LT_REUSE_BIO, srp->frq_lt_bm)) {
- if (srp->bio) {
+ if (srp->bio) { /* 2,3,4 ... reuse bio handling */
+ bio_reset(srp->bio);
+ srp->bio->bi_iter.bi_size = srp->s_hdr4.bi_size;
+ srp->bio->bi_opf = srp->s_hdr4.bi_opf;
+ srp->bio->bi_vcnt = srp->s_hdr4.bi_vcnt;
+ srp->bio->bi_end_io = srp->s_hdr4.bi_end_io;
res = blk_rq_append_bio(rqq, &srp->bio);
+ rqq->nr_phys_segments = (1 << schp->page_order) * schp->num_sgat;
+ bio_get(rqq->bio);
+ /*
+ * balancing bio_put() is either in:
+ * - normal case: in sg_mk_kern_bio(), or
+ * - error case: in sg_common_write() after err_out label
+ */
if (res)
SG_LOG(1, sfp, "%s: blk_rq_append_bio err=%d\n", __func__,
res);
- md = NULL;
- atomic_inc(&sg_tmp_count_reused_bios);
} else {
res = -EPROTO;
}
goto fini;
} else { /* first use of bio, almost normal setup */
md = &map_data;
- bump_bio_get = true;
+ first_reuse_bio = true;
}
} else { /* normal indirect IO */
md = &map_data;
@@ -6389,16 +6499,16 @@ sg_start_req(struct sg_request *srp, struct sg_comm_wr_t *cwrp, int dxfer_dir)
struct sg_request *r_srp = sfp->rsv_arr[0];
reserve0 = (r_srp == srp);
- if (unlikely(!reserve0 || dlen > req_schp->buflen))
+ if (unlikely(!reserve0 || dlen > schp->buflen))
res = reserve0 ? -ENOMEM : -EBUSY;
- } else if (req_schp->buflen == 0) {
+ } else if (schp->buflen == 0) {
res = sg_mk_sgat(srp, sfp, max_t(int, dlen, sfp->sgat_elem_sz));
}
if (unlikely(res))
goto fini;
- md->pages = req_schp->pages;
- md->page_order = req_schp->page_order;
- md->nr_entries = req_schp->num_sgat;
+ md->pages = schp->pages;
+ md->page_order = schp->page_order;
+ md->nr_entries = schp->num_sgat;
md->offset = 0;
md->null_mapped = !up;
md->from_user = (dxfer_dir == SG_DXFER_TO_FROM_DEV);
@@ -6434,9 +6544,13 @@ sg_start_req(struct sg_request *srp, struct sg_comm_wr_t *cwrp, int dxfer_dir)
res = sg_rq_map_kern(srp, q, rqq, r0w);
if (res)
goto fini;
- if (bump_bio_get) { /* keep bio alive to re-use next time */
+ if (first_reuse_bio) { /* keep bio alive to re-use, hold some bio fields */
+ srp->s_hdr4.bi_size = rqq->bio->bi_iter.bi_size;
+ srp->s_hdr4.bi_opf = rqq->bio->bi_opf;
+ srp->s_hdr4.bi_vcnt = rqq->bio->bi_vcnt;
+ srp->s_hdr4.bi_end_io = rqq->bio->bi_end_io;
set_bit(SG_FRQ_LT_REUSE_BIO, srp->frq_lt_bm);
- bio_get(rqq->bio); /* _put() in sg_svb_cleanup() */
+ bio_get(rqq->bio); /* _put() in sg_svb_srp_cleanup() */
}
}
fini:
@@ -6478,9 +6592,8 @@ sg_finish_scsi_blk_rq(struct sg_request *srp)
clear_bit(SG_FFD_HIPRI_SEEN, sfp->ffd_bm);
atomic_dec_return_release(&sfp->waiting);
}
-
/* Expect blk_put_request(rqq) already called in sg_rq_end_io() */
- if (rqq) { /* blk_get_request() may have failed */
+ if (unlikely(rqq)) {
WRITE_ONCE(srp->rqq, NULL);
if (scsi_req(rqq))
scsi_req_free_cmd(scsi_req(rqq));
@@ -6494,10 +6607,10 @@ sg_finish_scsi_blk_rq(struct sg_request *srp)
ret = blk_rq_unmap_user(bio);
if (unlikely(ret)) /* -EINTR (-4) can be ignored */
SG_LOG(6, sfp, "%s: blk_rq_unmap_user() --> %d\n", __func__, ret);
+ }
+ if (!test_bit(SG_FRQ_LT_REUSE_BIO, srp->frq_lt_bm))
srp->bio = NULL;
- } else if (!test_bit(SG_FRQ_LT_REUSE_BIO, srp->frq_lt_bm)) {
- srp->bio = NULL;
- } /* else may be able to re-use this bio [mrq, uniform svb] */
+ /* else may be able to re-use this bio [mrq, uniform svb] */
}
/* In worst case, READ data returned to user space by this point */
}
@@ -6936,6 +7049,7 @@ sg_setup_req_new_srp(struct sg_comm_wr_t *cwrp, bool new_rsv_srp, bool no_reqs,
r_srp);
if (new_rsv_srp) {
fp->rsv_arr[ra_idx] = r_srp;
+ r_srp->rsv_arr_idx = ra_idx;
set_bit(SG_FRQ_LT_RESERVED, r_srp->frq_lt_bm);
r_srp->sh_srp = NULL;
}
@@ -6970,7 +7084,6 @@ sg_setup_req(struct sg_comm_wr_t *cwrp, enum sg_shr_var sh_var)
bool no_reqs = false;
bool ws_rq = false;
bool try_harder = false;
- bool keep_frq_pc_bm = false;
bool second = false;
int res, ra_idx, l_used_idx;
int dlen = cwrp->dlen;
@@ -6993,7 +7106,6 @@ sg_setup_req(struct sg_comm_wr_t *cwrp, enum sg_shr_var sh_var)
r_srp = NULL;
} else {
atomic_dec(&fp->inactives);
- keep_frq_pc_bm = true;
r_srp->sh_srp = NULL;
goto final_setup;
}
@@ -7048,7 +7160,6 @@ sg_setup_req(struct sg_comm_wr_t *cwrp, enum sg_shr_var sh_var)
if (likely(res == 0)) {
/* possible_srp bypasses loop to find candidate */
mk_new_srp = false;
- keep_frq_pc_bm = true;
goto final_setup;
}
}
@@ -7178,11 +7289,10 @@ sg_setup_req(struct sg_comm_wr_t *cwrp, enum sg_shr_var sh_var)
goto start_again;
}
final_setup:
- if (!keep_frq_pc_bm)
- r_srp->frq_pc_bm[0] = cwrp->frq_pc_bm[0];
+ r_srp->frq_pc_bm[0] = cwrp->frq_pc_bm[0];
r_srp->sgatp->dlen = dlen; /* must be <= r_srp->sgat_h.buflen */
r_srp->sh_var = sh_var;
- r_srp->cmd_opcode = 0xff; /* set invalid opcode (VS), 0x0 is TUR */
+ r_srp->cmd_opcode = cwrp->cmdp ? cwrp->cmdp[0] : 0xff; /* get later if in user space */
/* If setup stalls (e.g. blk_get_request()) debug shows 'elap=1 ns' */
if (test_bit(SG_FFD_TIME_IN_NS, fp->ffd_bm))
r_srp->start_ns = S64_MAX;
@@ -7754,6 +7864,8 @@ sg_proc_debug_sreq(struct sg_request *srp, int to, bool t_in_ns, bool inactive,
sg_shr_str(srp->sh_var, false));
if (srp->sgatp->num_sgat > 1)
n += scnprintf(obp + n, len - n, " sgat=%d", srp->sgatp->num_sgat);
+ if (test_bit(SG_FRQ_LT_REUSE_BIO, srp->frq_lt_bm))
+ n += scnprintf(obp + n, len - n, " re-use_bio");
cp = (srp->rq_flags & SGV4_FLAG_HIPRI) ? "hipri " : "";
n += scnprintf(obp + n, len - n, " %sop=0x%02x\n", cp, srp->cmd_opcode);
if (inactive && rq_st != SG_RQ_INACTIVE)
@@ -7800,6 +7912,8 @@ sg_proc_debug_fd(struct sg_fd *fp, char *obp, int len, unsigned long idx, bool r
n += scnprintf(obp + n, len - n, " excl_waitq");
if (test_bit(SG_FFD_SVB_ACTIVE, fp->ffd_bm))
n += scnprintf(obp + n, len - n, " svb");
+ if (test_bit(SG_FFD_CAN_REUSE_BIO, fp->ffd_bm))
+ n += scnprintf(obp + n, len - n, " can_reuse_bio");
n += scnprintf(obp + n, len - n, " fd_bm=0x%lx\n", fp->ffd_bm[0]);
n += scnprintf(obp + n, len - n,
" mmap_sz=%d low_used_idx=%d low_await_idx=%d sum_fd_dlens=%u\n",
@@ -7899,9 +8013,8 @@ sg_proc_seq_show_debug(struct seq_file *s, void *v, bool reduced)
b1[0] = '\0';
if (it && it->index == 0)
- seq_printf(s, "max_active_device=%d def_reserved_size=%d num_reused_bios=%d\n",
- (int)it->max, def_reserved_size,
- atomic_read(&sg_tmp_count_reused_bios));
+ seq_printf(s, "max_active_device=%d def_reserved_size=%d\n",
+ (int)it->max, def_reserved_size);
fdi_p = it ? &it->fd_index : &k;
bp = kzalloc(bp_len, __GFP_NOWARN | GFP_KERNEL);
if (unlikely(!bp)) {
--
2.25.1
next prev parent reply other threads:[~2021-04-27 22:00 UTC|newest]
Thread overview: 86+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-04-27 21:56 [PATCH v18 00/83] sg: add v4 interface, request sharing Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 00/45] sg: add v4 interface Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 01/83] sg: move functions around Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 02/83] sg: remove typedefs, type+formatting cleanup Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 03/83] sg: sg_log and is_enabled Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 04/83] sg: rework sg_poll(), minor changes Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 05/83] sg: bitops in sg_device Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 06/83] sg: make open count an atomic Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 07/83] sg: move header to uapi section Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 08/83] sg: speed sg_poll and sg_get_num_waiting Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 09/83] sg: sg_allow_if_err_recovery and renames Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 10/83] sg: improve naming Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 11/83] sg: change rwlock to spinlock Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 12/83] sg: ioctl handling Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 13/83] sg: split sg_read Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 14/83] sg: sg_common_write add structure for arguments Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 15/83] sg: rework sg_vma_fault Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 16/83] sg: rework sg_mmap Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 17/83] sg: replace sg_allow_access Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 18/83] sg: rework scatter gather handling Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 19/83] sg: introduce request state machine Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 20/83] sg: sg_find_srp_by_id Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 21/83] sg: sg_fill_request_element Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 22/83] sg: printk change %p to %pK Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 23/83] sg: xarray for fds in device Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 24/83] sg: xarray for reqs in fd Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 25/83] sg: replace rq array with xarray Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 26/83] sg: sense buffer rework Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 27/83] sg: add sg v4 interface support Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 28/83] sg: rework debug info Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 29/83] sg: add 8 byte SCSI LUN to sg_scsi_id Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 30/83] sg: expand sg_comm_wr_t Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 31/83] sg: add sg_iosubmit_v3 and sg_ioreceive_v3 ioctls Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 32/83] sg: add some __must_hold macros Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 33/83] sg: move procfs objects to avoid forward decls Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 34/83] sg: protect multiple receivers Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 35/83] sg: first debugfs support Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 36/83] sg: rework mmap support Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 37/83] sg: defang allow_dio Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 38/83] sg: warn v3 write system call users Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 39/83] sg: add mmap_sz tracking Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 40/83] sg: remove rcv_done request state Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 41/83] sg: track lowest inactive and await indexes Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 42/83] sg: remove unit attention check for device changed Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 43/83] sg: no_dxfer: move to/from kernel buffers Douglas Gilbert
2021-04-28 7:07 ` Hannes Reinecke
2021-04-27 21:56 ` [PATCH v18 44/83] sg: add blk_poll support Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 45/83] sg: bump version to 4.0.12 Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 46/83] sg: add sg_ioabort ioctl Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 47/83] sg: add sg_set_get_extended ioctl Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 48/83] sg: sgat_elem_sz and sum_fd_dlens Douglas Gilbert
2021-04-27 21:56 ` [PATCH v18 49/83] sg: tag and more_async Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 50/83] sg: add fd sharing , change, unshare Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 51/83] sg: add shared requests Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 52/83] sg: add multiple request support Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 53/83] sg: rename some mrq variables Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 54/83] sg: unlikely likely Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 55/83] sg: mrq abort Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 56/83] sg: reduce atomic operations Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 57/83] sg: add excl_wait flag Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 58/83] sg: tweak sg_find_sfp_by_fd() Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 59/83] sg: add snap_dev flag and snapped in debugfs Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 60/83] sg: compress usercontext to uc Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 61/83] sg: optionally output sg_request.frq_bm flags Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 62/83] sg: work on sg_mrq_sanity() Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 63/83] sg: shared variable blocking Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 64/83] sg: device timestamp Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 65/83] sg: condition met is not an error Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 66/83] sg: split sg_setup_req Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 67/83] sg: finish after read-side request Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 68/83] sg: keep share and dout offset flags Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 69/83] sg: add dlen to sg_comm_wr_t Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 70/83] sg: make use of struct sg_mrq_hold Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 71/83] sg: add mmap IO option for mrq metadata Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 72/83] sg: add eventfd support Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 73/83] sg: table of error number explanations Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 74/83] sg: add ordered write flag Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 75/83] sg: expand source line length to 100 characters Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 76/83] sg: add no_attach_msg parameter Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 77/83] sg: add SGV4_FLAG_REC_ORDER Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 78/83] sg: max to read for mrq sg_ioreceive Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 79/83] sg: mrq: if uniform svb then re-use bio_s Douglas Gilbert
2021-04-27 21:57 ` Douglas Gilbert [this message]
2021-04-27 21:57 ` [PATCH v18 81/83] sg: blk_poll/hipri work for mrq Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 82/83] sg: pollable and non-pollable requests Douglas Gilbert
2021-04-27 21:57 ` [PATCH v18 83/83] sg: bump version to 4.0.47 Douglas Gilbert
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20210427215733.417746-82-dgilbert@interlog.com \
--to=dgilbert@interlog.com \
--cc=hare@suse.de \
--cc=jejb@linux.vnet.ibm.com \
--cc=linux-scsi@vger.kernel.org \
--cc=martin.petersen@oracle.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox