From: Pierre Ossman <drzeus-list@drzeus.cx>
To: linux-fsdevel@vger.kernel.org
Cc: David Woodhouse <dwmw2@infradead.org>, Jens Axboe <axboe@kernel.dk>
Subject: [PATCH 1/2] mmc_block: factor out the mmc request handling
Date: Sat, 16 Aug 2008 19:11:27 +0200 [thread overview]
Message-ID: <20080816191127.2c8f9829@mjolnir.drzeus.cx> (raw)
In-Reply-To: <20080816190858.4d150ea1@mjolnir.drzeus.cx>
From: Pierre Ossman <drzeus@drzeus.cx>
Move the handling of the actual MMC request to its own function so
that the main request handler can be extended to handle other types
of requests than simple reads and writes.
Signed-off-by: Pierre Ossman <drzeus@drzeus.cx>
diff --git a/drivers/mmc/card/block.c b/drivers/mmc/card/block.c
index 86dbb36..3140e92 100644
--- a/drivers/mmc/card/block.c
+++ b/drivers/mmc/card/block.c
@@ -208,195 +208,215 @@ static u32 mmc_sd_num_wr_blocks(struct mmc_card *card)
return blocks;
}
-static int mmc_blk_issue_rq(struct mmc_queue *mq, struct request *req)
+static int mmc_blk_xfer_rq(struct mmc_blk_data *md,
+ struct request *req, unsigned int *bytes_xfered)
{
- struct mmc_blk_data *md = mq->data;
- struct mmc_card *card = md->queue.card;
+ struct mmc_card *card;
+
struct mmc_blk_request brq;
- int ret = 1, data_size, i;
+ int ret, data_size, i;
struct scatterlist *sg;
+ u32 readcmd, writecmd;
- mmc_claim_host(card->host);
+ BUG_ON(!bytes_xfered);
- do {
- struct mmc_command cmd;
- u32 readcmd, writecmd;
-
- memset(&brq, 0, sizeof(struct mmc_blk_request));
- brq.mrq.cmd = &brq.cmd;
- brq.mrq.data = &brq.data;
-
- brq.cmd.arg = req->sector;
- if (!mmc_card_blockaddr(card))
- brq.cmd.arg <<= 9;
- brq.cmd.flags = MMC_RSP_SPI_R1 | MMC_RSP_R1 | MMC_CMD_ADTC;
- brq.data.blksz = 1 << md->block_bits;
- brq.stop.opcode = MMC_STOP_TRANSMISSION;
- brq.stop.arg = 0;
- brq.stop.flags = MMC_RSP_SPI_R1B | MMC_RSP_R1B | MMC_CMD_AC;
- brq.data.blocks = req->nr_sectors >> (md->block_bits - 9);
- if (brq.data.blocks > card->host->max_blk_count)
- brq.data.blocks = card->host->max_blk_count;
-
- if (brq.data.blocks > 1) {
- /* SPI multiblock writes terminate using a special
- * token, not a STOP_TRANSMISSION request.
- */
- if (!mmc_host_is_spi(card->host)
- || rq_data_dir(req) == READ)
- brq.mrq.stop = &brq.stop;
- readcmd = MMC_READ_MULTIPLE_BLOCK;
- writecmd = MMC_WRITE_MULTIPLE_BLOCK;
- } else {
- brq.mrq.stop = NULL;
- readcmd = MMC_READ_SINGLE_BLOCK;
- writecmd = MMC_WRITE_BLOCK;
- }
-
- if (rq_data_dir(req) == READ) {
- brq.cmd.opcode = readcmd;
- brq.data.flags |= MMC_DATA_READ;
- } else {
- brq.cmd.opcode = writecmd;
- brq.data.flags |= MMC_DATA_WRITE;
- }
+ card = md->queue.card;
- mmc_set_data_timeout(&brq.data, card);
+ memset(&brq, 0, sizeof(struct mmc_blk_request));
+ brq.mrq.cmd = &brq.cmd;
+ brq.mrq.data = &brq.data;
- brq.data.sg = mq->sg;
- brq.data.sg_len = mmc_queue_map_sg(mq);
-
- mmc_queue_bounce_pre(mq);
+ brq.cmd.arg = req->sector;
+ if (!mmc_card_blockaddr(card))
+ brq.cmd.arg <<= 9;
+ brq.cmd.flags = MMC_RSP_SPI_R1 | MMC_RSP_R1 | MMC_CMD_ADTC;
+ brq.data.blksz = 1 << md->block_bits;
+ brq.stop.opcode = MMC_STOP_TRANSMISSION;
+ brq.stop.arg = 0;
+ brq.stop.flags = MMC_RSP_SPI_R1B | MMC_RSP_R1B | MMC_CMD_AC;
+ brq.data.blocks = req->nr_sectors >> (md->block_bits - 9);
+ if (brq.data.blocks > card->host->max_blk_count)
+ brq.data.blocks = card->host->max_blk_count;
+ if (brq.data.blocks > 1) {
/*
- * Adjust the sg list so it is the same size as the
- * request.
+ * SPI multiblock writes terminate using a special
+ * token, not a STOP_TRANSMISSION request.
*/
- if (brq.data.blocks !=
- (req->nr_sectors >> (md->block_bits - 9))) {
- data_size = brq.data.blocks * brq.data.blksz;
- for_each_sg(brq.data.sg, sg, brq.data.sg_len, i) {
- data_size -= sg->length;
- if (data_size <= 0) {
- sg->length += data_size;
- i++;
- break;
- }
- }
- brq.data.sg_len = i;
- }
+ if (!mmc_host_is_spi(card->host)
+ || rq_data_dir(req) == READ)
+ brq.mrq.stop = &brq.stop;
+ readcmd = MMC_READ_MULTIPLE_BLOCK;
+ writecmd = MMC_WRITE_MULTIPLE_BLOCK;
+ } else {
+ brq.mrq.stop = NULL;
+ readcmd = MMC_READ_SINGLE_BLOCK;
+ writecmd = MMC_WRITE_BLOCK;
+ }
- mmc_wait_for_req(card->host, &brq.mrq);
+ if (rq_data_dir(req) == READ) {
+ brq.cmd.opcode = readcmd;
+ brq.data.flags |= MMC_DATA_READ;
+ } else {
+ brq.cmd.opcode = writecmd;
+ brq.data.flags |= MMC_DATA_WRITE;
+ }
- mmc_queue_bounce_post(mq);
+ mmc_set_data_timeout(&brq.data, card);
- /*
- * Check for errors here, but don't jump to cmd_err
- * until later as we need to wait for the card to leave
- * programming mode even when things go wrong.
- */
- if (brq.cmd.error) {
- printk(KERN_ERR "%s: error %d sending read/write command\n",
- req->rq_disk->disk_name, brq.cmd.error);
- }
+ brq.data.sg = md->queue.sg;
+ brq.data.sg_len = mmc_queue_map_sg(&md->queue);
- if (brq.data.error) {
- printk(KERN_ERR "%s: error %d transferring data\n",
- req->rq_disk->disk_name, brq.data.error);
- }
+ mmc_queue_bounce_pre(&md->queue);
- if (brq.stop.error) {
- printk(KERN_ERR "%s: error %d sending stop command\n",
- req->rq_disk->disk_name, brq.stop.error);
+ /*
+ * Adjust the sg list so it is the same size as the
+ * request.
+ */
+ if (brq.data.blocks !=
+ (req->nr_sectors >> (md->block_bits - 9))) {
+ data_size = brq.data.blocks * brq.data.blksz;
+ for_each_sg(brq.data.sg, sg, brq.data.sg_len, i) {
+ data_size -= sg->length;
+ if (data_size <= 0) {
+ sg->length += data_size;
+ i++;
+ break;
+ }
}
+ brq.data.sg_len = i;
+ }
- if (!mmc_host_is_spi(card->host) && rq_data_dir(req) != READ) {
- do {
- int err;
-
- cmd.opcode = MMC_SEND_STATUS;
- cmd.arg = card->rca << 16;
- cmd.flags = MMC_RSP_R1 | MMC_CMD_AC;
- err = mmc_wait_for_cmd(card->host, &cmd, 5);
- if (err) {
- printk(KERN_ERR "%s: error %d requesting status\n",
- req->rq_disk->disk_name, err);
- goto cmd_err;
- }
- /*
- * Some cards mishandle the status bits,
- * so make sure to check both the busy
- * indication and the card state.
- */
- } while (!(cmd.resp[0] & R1_READY_FOR_DATA) ||
- (R1_CURRENT_STATE(cmd.resp[0]) == 7));
-
-#if 0
- if (cmd.resp[0] & ~0x00000900)
- printk(KERN_ERR "%s: status = %08x\n",
- req->rq_disk->disk_name, cmd.resp[0]);
- if (mmc_decode_status(cmd.resp))
- goto cmd_err;
-#endif
- }
+ mmc_wait_for_req(card->host, &brq.mrq);
- if (brq.cmd.error || brq.data.error || brq.stop.error)
- goto cmd_err;
+ mmc_queue_bounce_post(&md->queue);
- /*
- * A block was successfully transferred.
- */
- spin_lock_irq(&md->lock);
- ret = __blk_end_request(req, 0, brq.data.bytes_xfered);
- spin_unlock_irq(&md->lock);
- } while (ret);
+ ret = 0;
+ *bytes_xfered = brq.data.bytes_xfered;
- mmc_release_host(card->host);
+ if (brq.cmd.error) {
+ ret = brq.cmd.error;
+ printk(KERN_ERR "%s: error %d sending read/write command\n",
+ req->rq_disk->disk_name, brq.cmd.error);
+ }
- return 1;
+ if (brq.data.error) {
+ ret = brq.cmd.error;
+ printk(KERN_ERR "%s: error %d transferring data\n",
+ req->rq_disk->disk_name, brq.data.error);
+ }
- cmd_err:
- /*
- * If this is an SD card and we're writing, we can first
- * mark the known good sectors as ok.
- *
- * If the card is not SD, we can still ok written sectors
- * as reported by the controller (which might be less than
- * the real number of written sectors, but never more).
- *
- * For reads we just fail the entire chunk as that should
- * be safe in all cases.
+ if (brq.stop.error) {
+ ret = brq.cmd.error;
+ printk(KERN_ERR "%s: error %d sending stop command\n",
+ req->rq_disk->disk_name, brq.stop.error);
+ }
+
+ /*
+ * We need to wait for the card to leave programming mode
+ * even when things go wrong.
*/
- if (rq_data_dir(req) != READ) {
- if (mmc_card_sd(card)) {
+ if (!mmc_host_is_spi(card->host) && rq_data_dir(req) != READ) {
+ int cmd_ret;
+ struct mmc_command cmd;
+
+ do {
+ cmd.opcode = MMC_SEND_STATUS;
+ cmd.arg = card->rca << 16;
+ cmd.flags = MMC_RSP_R1 | MMC_CMD_AC;
+ cmd_ret = mmc_wait_for_cmd(card->host, &cmd, 5);
+ if (cmd_ret) {
+ printk(KERN_ERR "%s: error %d requesting status\n",
+ req->rq_disk->disk_name, cmd_ret);
+ ret = cmd_ret;
+ break;
+ }
+ /*
+ * Some cards mishandle the status bits,
+ * so make sure to check both the busy
+ * indication and the card state.
+ */
+ } while (!(cmd.resp[0] & R1_READY_FOR_DATA) ||
+ (R1_CURRENT_STATE(cmd.resp[0]) == 7));
+ }
+
+ /*
+ * Adjust the number of bytes transferred if there has been
+ * an error...
+ */
+ if (ret) {
+ /*
+ * For reads we just fail the entire chunk as that should
+ * be safe in all cases.
+ *
+ * If this is an SD card and we're writing, we can ask the
+ * card for known good sectors.
+ *
+ * If the card is not SD, we can still ok written sectors
+ * as reported by the controller (which might be less than
+ * the real number of written sectors, but never more).
+ */
+ if (rq_data_dir(req) == READ)
+ *bytes_xfered = 0;
+ else if (mmc_card_sd(card)) {
u32 blocks;
- unsigned int bytes;
blocks = mmc_sd_num_wr_blocks(card);
- if (blocks != (u32)-1) {
+ if (blocks == (u32)-1)
+ *bytes_xfered = 0;
+ else {
if (card->csd.write_partial)
- bytes = blocks << md->block_bits;
+ *bytes_xfered = blocks << md->block_bits;
else
- bytes = blocks << 9;
- spin_lock_irq(&md->lock);
- ret = __blk_end_request(req, 0, bytes);
- spin_unlock_irq(&md->lock);
+ *bytes_xfered = blocks << 9;
}
- } else {
+ }
+ }
+
+ return ret;
+}
+
+static int mmc_blk_issue_rq(struct mmc_queue *mq, struct request *req)
+{
+ struct mmc_blk_data *md = mq->data;
+ struct mmc_card *card = md->queue.card;
+ int ret, err, bytes_xfered;
+
+ mmc_claim_host(card->host);
+
+ do {
+ err = mmc_blk_xfer_rq(md, req, &bytes_xfered);
+
+ /*
+ * First handle the sectors that got transferred
+ * successfully...
+ */
+ spin_lock_irq(&md->lock);
+ ret = __blk_end_request(req, 0, bytes_xfered);
+ spin_unlock_irq(&md->lock);
+
+ /*
+ * ...then check if things went south.
+ */
+ if (err) {
+ mmc_release_host(card->host);
+
+ /*
+ * Kill of the rest of the request...
+ */
spin_lock_irq(&md->lock);
- ret = __blk_end_request(req, 0, brq.data.bytes_xfered);
+ while (ret)
+ ret = __blk_end_request(req, -EIO,
+ blk_rq_cur_bytes(req));
spin_unlock_irq(&md->lock);
+
+ return 0;
}
- }
+ } while (ret);
mmc_release_host(card->host);
- spin_lock_irq(&md->lock);
- while (ret)
- ret = __blk_end_request(req, -EIO, blk_rq_cur_bytes(req));
- spin_unlock_irq(&md->lock);
-
- return 0;
+ return 1;
}
--
-- Pierre Ossman
Linux kernel, MMC maintainer http://www.kernel.org
rdesktop, core developer http://www.rdesktop.org
WARNING: This correspondence is being monitored by the
Swedish government. Make sure your server uses encryption
for SMTP traffic and consider using PGP for end-to-end
encryption.
next prev parent reply other threads:[~2008-08-16 17:11 UTC|newest]
Thread overview: 88+ messages / expand[flat|nested] mbox.gz Atom feed top
2008-08-09 16:26 [PATCH 0/7] Discard requests, v2 David Woodhouse
2008-08-09 16:29 ` [PATCH 1/7] [BLOCK] Fix typo causing compile error in blk_queue_bounce() David Woodhouse
2008-08-09 16:30 ` [PATCH 2/7] [BLOCK] Fix up comments about matching flags between bio and rq David Woodhouse
2008-08-09 16:30 ` [PATCH 3/7] [BLOCK] Add 'discard' request handling David Woodhouse
2008-08-09 20:39 ` OGAWA Hirofumi
2008-08-09 21:37 ` David Woodhouse
2008-08-10 6:32 ` David Woodhouse
2008-08-09 16:31 ` [PATCH 4/7] [FAT] Let the block device know when sectors can be discarded David Woodhouse
2008-08-09 16:32 ` [PATCH 5/7] [MTD] Support 'discard sectors' operation in translation layer support core David Woodhouse
2008-08-09 16:33 ` [PATCH 6/7] [MTD] [FTL] Support 'discard sectors' operation David Woodhouse
2008-08-09 16:33 ` [PATCH 7/7] [BLOCK] Allow elevators to sort/merge discard requests David Woodhouse
2008-10-03 20:29 ` Andrew Morton
2008-10-07 12:07 ` Jens Axboe
2008-08-09 22:48 ` [PATCH 0/7] Discard requests, v2 OGAWA Hirofumi
2008-08-10 10:25 ` David Woodhouse
2008-08-10 16:37 ` Jamie Lokier
2008-08-10 17:55 ` OGAWA Hirofumi
2008-08-10 20:07 ` David Woodhouse
2008-08-10 21:40 ` OGAWA Hirofumi
2008-08-11 9:40 ` David Woodhouse
2008-08-11 10:25 ` OGAWA Hirofumi
2008-08-11 13:17 ` David Woodhouse
2008-08-11 14:21 ` OGAWA Hirofumi
2008-08-10 10:29 ` [PATCH 8/7] blktrace: support discard requests David Woodhouse
2008-08-10 10:35 ` [USERSPACE PATCH] " David Woodhouse
2008-08-15 8:43 ` Jens Axboe
2008-08-15 9:01 ` David Woodhouse
2008-08-15 9:08 ` Jens Axboe
2008-08-10 10:41 ` [PATCH 8/7] " David Woodhouse
2008-08-13 11:17 ` Jens Axboe
2008-08-10 11:48 ` [PATCH 9/7] blktrace: simplify flags handling in __blk_add_trace David Woodhouse
2008-08-10 11:50 ` David Woodhouse
2008-08-11 15:11 ` [PATCH 10/7] [BLOCK] Add BLKDISCARD ioctl to allow userspace to discard sectors David Woodhouse
2008-08-11 18:27 ` Matthew Wilcox
2008-08-11 20:52 ` David Woodhouse
2008-08-12 9:14 ` [PATCH 0/7] Discard requests, v2 Jens Axboe
2008-08-12 10:00 ` David Woodhouse
2008-08-12 10:54 ` Jens Axboe
2008-08-12 11:16 ` David Woodhouse
2008-08-12 12:19 ` David Woodhouse
2008-08-12 12:53 ` Jens Axboe
2008-08-12 13:04 ` David Woodhouse
2008-08-12 15:47 ` David Woodhouse
2008-08-12 18:04 ` Jamie Lokier
2008-08-13 10:22 ` David Woodhouse
2008-08-13 12:19 ` Jamie Lokier
2008-08-13 12:26 ` David Woodhouse
2008-08-13 11:15 ` Jens Axboe
2008-08-13 11:23 ` David Woodhouse
2008-08-13 11:32 ` Jens Axboe
2008-08-13 11:34 ` David Woodhouse
2008-08-13 12:07 ` David Woodhouse
2008-08-14 7:49 ` Jens Axboe
2008-08-14 7:52 ` David Woodhouse
2008-08-14 7:25 ` David Woodhouse
2008-08-14 7:33 ` Stephen Rothwell
2008-08-14 7:37 ` David Woodhouse
2008-08-14 7:42 ` Jens Axboe
2008-08-14 7:46 ` Stephen Rothwell
2008-08-12 18:10 ` Jamie Lokier
2008-08-13 10:20 ` David Woodhouse
2008-08-12 11:42 ` Matthew Wilcox
2008-08-12 11:46 ` David Woodhouse
2008-08-12 19:53 ` OGAWA Hirofumi
2008-08-12 20:11 ` OGAWA Hirofumi
2008-08-13 11:39 ` [PATCH 11/7] Kill REQ_TYPE_FLUSH David Woodhouse
2008-08-13 11:58 ` Geert Uytterhoeven
2008-08-13 12:43 ` David Woodhouse
2008-08-13 15:40 ` Jens Axboe
2008-08-13 15:46 ` David Woodhouse
2008-08-16 17:08 ` [PATCH 0/2] MMC discard support (was [PATCH 0/7] Discard requests, v2) Pierre Ossman
2008-08-16 17:11 ` Pierre Ossman [this message]
2008-08-16 17:12 ` [PATCH 2/2] mmc_block: erase discarded blocks Pierre Ossman
2008-08-16 17:38 ` [PATCH 0/2] MMC discard support (was [PATCH 0/7] Discard requests, v2) David Woodhouse
2008-08-16 17:51 ` Pierre Ossman
2008-08-22 9:24 ` Jens Axboe
2008-08-22 9:45 ` David Woodhouse
2008-08-22 10:50 ` Jens Axboe
2008-08-22 10:58 ` David Woodhouse
2008-08-22 11:11 ` Pierre Ossman
2008-08-22 11:19 ` Jens Axboe
2008-08-22 11:13 ` Pierre Ossman
2008-08-22 11:20 ` Jens Axboe
2008-08-22 14:49 ` OGAWA Hirofumi
2008-08-22 23:02 ` Pierre Ossman
2008-08-22 23:59 ` OGAWA Hirofumi
2008-08-24 11:23 ` Pierre Ossman
2008-08-24 13:39 ` OGAWA Hirofumi
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20080816191127.2c8f9829@mjolnir.drzeus.cx \
--to=drzeus-list@drzeus.cx \
--cc=axboe@kernel.dk \
--cc=dwmw2@infradead.org \
--cc=linux-fsdevel@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).