Commit ee8a43a5 authored by Per Forlin's avatar Per Forlin Committed by Chris Ball

mmc: block: add handling for two parallel block requests in issue_rw_rq

Change mmc_blk_issue_rw_rq() to become asynchronous.
The execution flow looks like this:

* The mmc-queue calls issue_rw_rq(), which sends the request
  to the host and returns back to the mmc-queue.
* The mmc-queue calls issue_rw_rq() again with a new request.
* This new request is prepared in issue_rw_rq(), then it waits for
  the active request to complete before pushing it to the host.
* When the mmc-queue is empty it will call issue_rw_rq() with a NULL
  req to finish off the active request without starting a new request.
Signed-off-by: default avatarPer Forlin <per.forlin@linaro.org>
Acked-by: default avatarKyungmin Park <kyungmin.park@samsung.com>
Acked-by: default avatarArnd Bergmann <arnd@arndb.de>
Reviewed-by: default avatarVenkatraman S <svenkatr@ti.com>
Tested-by: default avatarSourav Poddar <sourav.poddar@ti.com>
Tested-by: default avatarLinus Walleij <linus.walleij@linaro.org>
Signed-off-by: default avatarChris Ball <cjb@laptop.org>
parent 04296b7b
...@@ -822,12 +822,14 @@ static inline void mmc_apply_rel_rw(struct mmc_blk_request *brq, ...@@ -822,12 +822,14 @@ static inline void mmc_apply_rel_rw(struct mmc_blk_request *brq,
R1_CC_ERROR | /* Card controller error */ \ R1_CC_ERROR | /* Card controller error */ \
R1_ERROR) /* General/unknown error */ R1_ERROR) /* General/unknown error */
int mmc_blk_err_check(struct mmc_blk_request *brq, static int mmc_blk_err_check(struct mmc_card *card,
struct request *req, struct mmc_async_req *areq)
struct mmc_card *card,
struct mmc_blk_data *md)
{ {
int ret = MMC_BLK_SUCCESS; enum mmc_blk_status ret = MMC_BLK_SUCCESS;
struct mmc_queue_req *mq_mrq = container_of(areq, struct mmc_queue_req,
mmc_active);
struct mmc_blk_request *brq = &mq_mrq->brq;
struct request *req = mq_mrq->req;
/* /*
* sbc.error indicates a problem with the set block count * sbc.error indicates a problem with the set block count
...@@ -1038,24 +1040,41 @@ static void mmc_blk_rw_rq_prep(struct mmc_queue_req *mqrq, ...@@ -1038,24 +1040,41 @@ static void mmc_blk_rw_rq_prep(struct mmc_queue_req *mqrq,
brq->data.sg_len = i; brq->data.sg_len = i;
} }
mqrq->mmc_active.mrq = &brq->mrq;
mqrq->mmc_active.err_check = mmc_blk_err_check;
mmc_queue_bounce_pre(mqrq); mmc_queue_bounce_pre(mqrq);
} }
static int mmc_blk_issue_rw_rq(struct mmc_queue *mq, struct request *req) static int mmc_blk_issue_rw_rq(struct mmc_queue *mq, struct request *rqc)
{ {
struct mmc_blk_data *md = mq->data; struct mmc_blk_data *md = mq->data;
struct mmc_card *card = md->queue.card; struct mmc_card *card = md->queue.card;
struct mmc_blk_request *brq = &mq->mqrq_cur->brq; struct mmc_blk_request *brq = &mq->mqrq_cur->brq;
int ret = 1, disable_multi = 0, retry = 0; int ret = 1, disable_multi = 0, retry = 0;
enum mmc_blk_status status; enum mmc_blk_status status;
struct mmc_queue_req *mq_rq;
struct request *req;
struct mmc_async_req *areq;
if (!rqc && !mq->mqrq_prev->req)
return 0;
do { do {
mmc_blk_rw_rq_prep(mq->mqrq_cur, card, disable_multi, mq); if (rqc) {
mmc_wait_for_req(card->host, &brq->mrq); mmc_blk_rw_rq_prep(mq->mqrq_cur, card, 0, mq);
areq = &mq->mqrq_cur->mmc_active;
} else
areq = NULL;
areq = mmc_start_req(card->host, areq, (int *) &status);
if (!areq)
return 0;
mmc_queue_bounce_post(mq->mqrq_cur); mq_rq = container_of(areq, struct mmc_queue_req, mmc_active);
brq = &mq_rq->brq;
req = mq_rq->req;
mmc_queue_bounce_post(mq_rq);
status = mmc_blk_err_check(brq, req, card, md);
switch (status) { switch (status) {
case MMC_BLK_SUCCESS: case MMC_BLK_SUCCESS:
case MMC_BLK_PARTIAL: case MMC_BLK_PARTIAL:
...@@ -1066,6 +1085,19 @@ static int mmc_blk_issue_rw_rq(struct mmc_queue *mq, struct request *req) ...@@ -1066,6 +1085,19 @@ static int mmc_blk_issue_rw_rq(struct mmc_queue *mq, struct request *req)
ret = __blk_end_request(req, 0, ret = __blk_end_request(req, 0,
brq->data.bytes_xfered); brq->data.bytes_xfered);
spin_unlock_irq(&md->lock); spin_unlock_irq(&md->lock);
if (status == MMC_BLK_SUCCESS && ret) {
/*
* The blk_end_request has returned non zero
* even though all data is transfered and no
* erros returned by host.
* If this happen it's a bug.
*/
printk(KERN_ERR "%s BUG rq_tot %d d_xfer %d\n",
__func__, blk_rq_bytes(req),
brq->data.bytes_xfered);
rqc = NULL;
goto cmd_abort;
}
break; break;
case MMC_BLK_CMD_ERR: case MMC_BLK_CMD_ERR:
goto cmd_err; goto cmd_err;
...@@ -1087,9 +1119,19 @@ static int mmc_blk_issue_rw_rq(struct mmc_queue *mq, struct request *req) ...@@ -1087,9 +1119,19 @@ static int mmc_blk_issue_rw_rq(struct mmc_queue *mq, struct request *req)
ret = __blk_end_request(req, -EIO, ret = __blk_end_request(req, -EIO,
brq->data.blksz); brq->data.blksz);
spin_unlock_irq(&md->lock); spin_unlock_irq(&md->lock);
if (!ret)
goto start_new_req;
break; break;
} }
if (ret) {
/*
* In case of a none complete request
* prepare it again and resend.
*/
mmc_blk_rw_rq_prep(mq_rq, card, disable_multi, mq);
mmc_start_req(card->host, &mq_rq->mmc_active, NULL);
}
} while (ret); } while (ret);
return 1; return 1;
...@@ -1124,6 +1166,12 @@ static int mmc_blk_issue_rw_rq(struct mmc_queue *mq, struct request *req) ...@@ -1124,6 +1166,12 @@ static int mmc_blk_issue_rw_rq(struct mmc_queue *mq, struct request *req)
ret = __blk_end_request(req, -EIO, blk_rq_cur_bytes(req)); ret = __blk_end_request(req, -EIO, blk_rq_cur_bytes(req));
spin_unlock_irq(&md->lock); spin_unlock_irq(&md->lock);
start_new_req:
if (rqc) {
mmc_blk_rw_rq_prep(mq->mqrq_cur, card, 0, mq);
mmc_start_req(card->host, &mq->mqrq_cur->mmc_active, NULL);
}
return 0; return 0;
} }
...@@ -1133,25 +1181,33 @@ static int mmc_blk_issue_rq(struct mmc_queue *mq, struct request *req) ...@@ -1133,25 +1181,33 @@ static int mmc_blk_issue_rq(struct mmc_queue *mq, struct request *req)
struct mmc_blk_data *md = mq->data; struct mmc_blk_data *md = mq->data;
struct mmc_card *card = md->queue.card; struct mmc_card *card = md->queue.card;
if (req && !mq->mqrq_prev->req)
/* claim host only for the first request */
mmc_claim_host(card->host); mmc_claim_host(card->host);
ret = mmc_blk_part_switch(card, md); ret = mmc_blk_part_switch(card, md);
if (ret) { if (ret) {
ret = 0; ret = 0;
goto out; goto out;
} }
if (req->cmd_flags & REQ_DISCARD) { if (req && req->cmd_flags & REQ_DISCARD) {
/* complete ongoing async transfer before issuing discard */
if (card->host->areq)
mmc_blk_issue_rw_rq(mq, NULL);
if (req->cmd_flags & REQ_SECURE) if (req->cmd_flags & REQ_SECURE)
ret = mmc_blk_issue_secdiscard_rq(mq, req); ret = mmc_blk_issue_secdiscard_rq(mq, req);
else else
ret = mmc_blk_issue_discard_rq(mq, req); ret = mmc_blk_issue_discard_rq(mq, req);
} else if (req->cmd_flags & REQ_FLUSH) { } else if (req && req->cmd_flags & REQ_FLUSH) {
ret = mmc_blk_issue_flush(mq, req); ret = mmc_blk_issue_flush(mq, req);
} else { } else {
ret = mmc_blk_issue_rw_rq(mq, req); ret = mmc_blk_issue_rw_rq(mq, req);
} }
out: out:
if (!req)
/* release host only when there are no more requests */
mmc_release_host(card->host); mmc_release_host(card->host);
return ret; return ret;
} }
......
...@@ -52,6 +52,7 @@ static int mmc_queue_thread(void *d) ...@@ -52,6 +52,7 @@ static int mmc_queue_thread(void *d)
down(&mq->thread_sem); down(&mq->thread_sem);
do { do {
struct request *req = NULL; struct request *req = NULL;
struct mmc_queue_req *tmp;
spin_lock_irq(q->queue_lock); spin_lock_irq(q->queue_lock);
set_current_state(TASK_INTERRUPTIBLE); set_current_state(TASK_INTERRUPTIBLE);
...@@ -59,7 +60,10 @@ static int mmc_queue_thread(void *d) ...@@ -59,7 +60,10 @@ static int mmc_queue_thread(void *d)
mq->mqrq_cur->req = req; mq->mqrq_cur->req = req;
spin_unlock_irq(q->queue_lock); spin_unlock_irq(q->queue_lock);
if (!req) { if (req || mq->mqrq_prev->req) {
set_current_state(TASK_RUNNING);
mq->issue_fn(mq, req);
} else {
if (kthread_should_stop()) { if (kthread_should_stop()) {
set_current_state(TASK_RUNNING); set_current_state(TASK_RUNNING);
break; break;
...@@ -67,11 +71,14 @@ static int mmc_queue_thread(void *d) ...@@ -67,11 +71,14 @@ static int mmc_queue_thread(void *d)
up(&mq->thread_sem); up(&mq->thread_sem);
schedule(); schedule();
down(&mq->thread_sem); down(&mq->thread_sem);
continue;
} }
set_current_state(TASK_RUNNING);
mq->issue_fn(mq, req); /* Current request becomes previous request and vice versa. */
mq->mqrq_prev->brq.mrq.data = NULL;
mq->mqrq_prev->req = NULL;
tmp = mq->mqrq_prev;
mq->mqrq_prev = mq->mqrq_cur;
mq->mqrq_cur = tmp;
} while (1); } while (1);
up(&mq->thread_sem); up(&mq->thread_sem);
...@@ -97,7 +104,7 @@ static void mmc_request(struct request_queue *q) ...@@ -97,7 +104,7 @@ static void mmc_request(struct request_queue *q)
return; return;
} }
if (!mq->mqrq_cur->req) if (!mq->mqrq_cur->req && !mq->mqrq_prev->req)
wake_up_process(mq->thread); wake_up_process(mq->thread);
} }
......
...@@ -19,6 +19,7 @@ struct mmc_queue_req { ...@@ -19,6 +19,7 @@ struct mmc_queue_req {
char *bounce_buf; char *bounce_buf;
struct scatterlist *bounce_sg; struct scatterlist *bounce_sg;
unsigned int bounce_sg_len; unsigned int bounce_sg_len;
struct mmc_async_req mmc_active;
}; };
struct mmc_queue { struct mmc_queue {
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment