This commit was manufactured by cvs2svn to create branch 'vserver'.
[linux-2.6.git] / drivers / mmc / mmc_queue.c
1 /*
2  *  linux/drivers/mmc/mmc_queue.c
3  *
4  *  Copyright (C) 2003 Russell King, All Rights Reserved.
5  *
6  * This program is free software; you can redistribute it and/or modify
7  * it under the terms of the GNU General Public License version 2 as
8  * published by the Free Software Foundation.
9  *
10  */
11 #include <linux/module.h>
12 #include <linux/blkdev.h>
13
14 #include <linux/mmc/card.h>
15 #include <linux/mmc/host.h>
16 #include "mmc_queue.h"
17
18 #define MMC_QUEUE_EXIT          (1 << 0)
19 #define MMC_QUEUE_SUSPENDED     (1 << 1)
20
21 /*
22  * Prepare a MMC request.  Essentially, this means passing the
23  * preparation off to the media driver.  The media driver will
24  * create a mmc_io_request in req->special.
25  */
26 static int mmc_prep_request(struct request_queue *q, struct request *req)
27 {
28         struct mmc_queue *mq = q->queuedata;
29         int ret = BLKPREP_KILL;
30
31         if (req->flags & REQ_SPECIAL) {
32                 /*
33                  * Special commands already have the command
34                  * blocks already setup in req->special.
35                  */
36                 BUG_ON(!req->special);
37
38                 ret = BLKPREP_OK;
39         } else if (req->flags & (REQ_CMD | REQ_BLOCK_PC)) {
40                 /*
41                  * Block I/O requests need translating according
42                  * to the protocol.
43                  */
44                 ret = mq->prep_fn(mq, req);
45         } else {
46                 /*
47                  * Everything else is invalid.
48                  */
49                 blk_dump_rq_flags(req, "MMC bad request");
50         }
51
52         if (ret == BLKPREP_OK)
53                 req->flags |= REQ_DONTPREP;
54
55         return ret;
56 }
57
58 static int mmc_queue_thread(void *d)
59 {
60         struct mmc_queue *mq = d;
61         struct request_queue *q = mq->queue;
62         DECLARE_WAITQUEUE(wait, current);
63
64         /*
65          * Set iothread to ensure that we aren't put to sleep by
66          * the process freezing.  We handle suspension ourselves.
67          */
68         current->flags |= PF_MEMALLOC|PF_NOFREEZE;
69
70         daemonize("mmcqd");
71
72         complete(&mq->thread_complete);
73
74         down(&mq->thread_sem);
75         add_wait_queue(&mq->thread_wq, &wait);
76         do {
77                 struct request *req = NULL;
78
79                 spin_lock_irq(q->queue_lock);
80                 set_current_state(TASK_INTERRUPTIBLE);
81                 if (!blk_queue_plugged(q))
82                         mq->req = req = elv_next_request(q);
83                 spin_unlock(q->queue_lock);
84
85                 if (!req) {
86                         if (mq->flags & MMC_QUEUE_EXIT)
87                                 break;
88                         up(&mq->thread_sem);
89                         schedule();
90                         down(&mq->thread_sem);
91                         continue;
92                 }
93                 set_current_state(TASK_RUNNING);
94
95                 mq->issue_fn(mq, req);
96         } while (1);
97         remove_wait_queue(&mq->thread_wq, &wait);
98         up(&mq->thread_sem);
99
100         complete_and_exit(&mq->thread_complete, 0);
101         return 0;
102 }
103
104 /*
105  * Generic MMC request handler.  This is called for any queue on a
106  * particular host.  When the host is not busy, we look for a request
107  * on any queue on this host, and attempt to issue it.  This may
108  * not be the queue we were asked to process.
109  */
110 static void mmc_request(request_queue_t *q)
111 {
112         struct mmc_queue *mq = q->queuedata;
113
114         if (!mq->req)
115                 wake_up(&mq->thread_wq);
116 }
117
118 /**
119  * mmc_init_queue - initialise a queue structure.
120  * @mq: mmc queue
121  * @card: mmc card to attach this queue
122  * @lock: queue lock
123  *
124  * Initialise a MMC card request queue.
125  */
126 int mmc_init_queue(struct mmc_queue *mq, struct mmc_card *card, spinlock_t *lock)
127 {
128         struct mmc_host *host = card->host;
129         u64 limit = BLK_BOUNCE_HIGH;
130         int ret;
131
132         if (host->dev->dma_mask && *host->dev->dma_mask)
133                 limit = *host->dev->dma_mask;
134
135         mq->card = card;
136         mq->queue = blk_init_queue(mmc_request, lock);
137         if (!mq->queue)
138                 return -ENOMEM;
139
140         blk_queue_prep_rq(mq->queue, mmc_prep_request);
141         blk_queue_bounce_limit(mq->queue, limit);
142         blk_queue_max_sectors(mq->queue, host->max_sectors);
143         blk_queue_max_phys_segments(mq->queue, host->max_phys_segs);
144         blk_queue_max_hw_segments(mq->queue, host->max_hw_segs);
145         blk_queue_max_segment_size(mq->queue, host->max_seg_size);
146
147         mq->queue->queuedata = mq;
148         mq->req = NULL;
149
150         init_completion(&mq->thread_complete);
151         init_waitqueue_head(&mq->thread_wq);
152         init_MUTEX(&mq->thread_sem);
153
154         ret = kernel_thread(mmc_queue_thread, mq, CLONE_KERNEL);
155         if (ret < 0) {
156                 blk_cleanup_queue(mq->queue);
157         } else {
158                 wait_for_completion(&mq->thread_complete);
159                 init_completion(&mq->thread_complete);
160                 ret = 0;
161         }
162
163         return ret;
164 }
165 EXPORT_SYMBOL(mmc_init_queue);
166
167 void mmc_cleanup_queue(struct mmc_queue *mq)
168 {
169         mq->flags |= MMC_QUEUE_EXIT;
170         wake_up(&mq->thread_wq);
171         wait_for_completion(&mq->thread_complete);
172         blk_cleanup_queue(mq->queue);
173
174         mq->card = NULL;
175 }
176 EXPORT_SYMBOL(mmc_cleanup_queue);
177
178 /**
179  * mmc_queue_suspend - suspend a MMC request queue
180  * @mq: MMC queue to suspend
181  *
182  * Stop the block request queue, and wait for our thread to
183  * complete any outstanding requests.  This ensures that we
184  * won't suspend while a request is being processed.
185  */
186 void mmc_queue_suspend(struct mmc_queue *mq)
187 {
188         request_queue_t *q = mq->queue;
189         unsigned long flags;
190
191         if (!(mq->flags & MMC_QUEUE_SUSPENDED)) {
192                 mq->flags |= MMC_QUEUE_SUSPENDED;
193
194                 spin_lock_irqsave(q->queue_lock, flags);
195                 blk_stop_queue(q);
196                 spin_unlock_irqrestore(q->queue_lock, flags);
197
198                 down(&mq->thread_sem);
199         }
200 }
201 EXPORT_SYMBOL(mmc_queue_suspend);
202
203 /**
204  * mmc_queue_resume - resume a previously suspended MMC request queue
205  * @mq: MMC queue to resume
206  */
207 void mmc_queue_resume(struct mmc_queue *mq)
208 {
209         request_queue_t *q = mq->queue;
210         unsigned long flags;
211
212         if (mq->flags & MMC_QUEUE_SUSPENDED) {
213                 mq->flags &= ~MMC_QUEUE_SUSPENDED;
214
215                 up(&mq->thread_sem);
216
217                 spin_lock_irqsave(q->queue_lock, flags);
218                 blk_start_queue(q);
219                 spin_unlock_irqrestore(q->queue_lock, flags);
220         }
221 }
222 EXPORT_SYMBOL(mmc_queue_resume);