blk: centralize non-request unplug handling.
Both md and umem has similar code for getting notified on an blk_finish_plug event. Centralize this code in block/ and allow each driver to provide its distinctive difference. Signed-off-by: NeilBrown <neilb@suse.de> Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
@ -2927,6 +2927,31 @@ static void flush_plug_callbacks(struct blk_plug *plug)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
struct blk_plug_cb *blk_check_plugged(blk_plug_cb_fn unplug, void *data,
|
||||||
|
int size)
|
||||||
|
{
|
||||||
|
struct blk_plug *plug = current->plug;
|
||||||
|
struct blk_plug_cb *cb;
|
||||||
|
|
||||||
|
if (!plug)
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
list_for_each_entry(cb, &plug->cb_list, list)
|
||||||
|
if (cb->callback == unplug && cb->data == data)
|
||||||
|
return cb;
|
||||||
|
|
||||||
|
/* Not currently on the callback list */
|
||||||
|
BUG_ON(size < sizeof(*cb));
|
||||||
|
cb = kzalloc(size, GFP_ATOMIC);
|
||||||
|
if (cb) {
|
||||||
|
cb->data = data;
|
||||||
|
cb->callback = unplug;
|
||||||
|
list_add(&cb->list, &plug->cb_list);
|
||||||
|
}
|
||||||
|
return cb;
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(blk_check_plugged);
|
||||||
|
|
||||||
void blk_flush_plug_list(struct blk_plug *plug, bool from_schedule)
|
void blk_flush_plug_list(struct blk_plug *plug, bool from_schedule)
|
||||||
{
|
{
|
||||||
struct request_queue *q;
|
struct request_queue *q;
|
||||||
|
@ -513,42 +513,19 @@ static void process_page(unsigned long data)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
struct mm_plug_cb {
|
|
||||||
struct blk_plug_cb cb;
|
|
||||||
struct cardinfo *card;
|
|
||||||
};
|
|
||||||
|
|
||||||
static void mm_unplug(struct blk_plug_cb *cb)
|
static void mm_unplug(struct blk_plug_cb *cb)
|
||||||
{
|
{
|
||||||
struct mm_plug_cb *mmcb = container_of(cb, struct mm_plug_cb, cb);
|
struct cardinfo *card = cb->data;
|
||||||
|
|
||||||
spin_lock_irq(&mmcb->card->lock);
|
spin_lock_irq(&card->lock);
|
||||||
activate(mmcb->card);
|
activate(card);
|
||||||
spin_unlock_irq(&mmcb->card->lock);
|
spin_unlock_irq(&card->lock);
|
||||||
kfree(mmcb);
|
kfree(cb);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int mm_check_plugged(struct cardinfo *card)
|
static int mm_check_plugged(struct cardinfo *card)
|
||||||
{
|
{
|
||||||
struct blk_plug *plug = current->plug;
|
return !!blk_check_plugged(mm_unplug, card, sizeof(struct blk_plug_cb));
|
||||||
struct mm_plug_cb *mmcb;
|
|
||||||
|
|
||||||
if (!plug)
|
|
||||||
return 0;
|
|
||||||
|
|
||||||
list_for_each_entry(mmcb, &plug->cb_list, cb.list) {
|
|
||||||
if (mmcb->cb.callback == mm_unplug && mmcb->card == card)
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
/* Not currently on the callback list */
|
|
||||||
mmcb = kmalloc(sizeof(*mmcb), GFP_ATOMIC);
|
|
||||||
if (!mmcb)
|
|
||||||
return 0;
|
|
||||||
|
|
||||||
mmcb->card = card;
|
|
||||||
mmcb->cb.callback = mm_unplug;
|
|
||||||
list_add(&mmcb->cb.list, &plug->cb_list);
|
|
||||||
return 1;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void mm_make_request(struct request_queue *q, struct bio *bio)
|
static void mm_make_request(struct request_queue *q, struct bio *bio)
|
||||||
|
@ -498,59 +498,13 @@ void md_flush_request(struct mddev *mddev, struct bio *bio)
|
|||||||
}
|
}
|
||||||
EXPORT_SYMBOL(md_flush_request);
|
EXPORT_SYMBOL(md_flush_request);
|
||||||
|
|
||||||
/* Support for plugging.
|
void md_unplug(struct blk_plug_cb *cb)
|
||||||
* This mirrors the plugging support in request_queue, but does not
|
|
||||||
* require having a whole queue or request structures.
|
|
||||||
* We allocate an md_plug_cb for each md device and each thread it gets
|
|
||||||
* plugged on. This links tot the private plug_handle structure in the
|
|
||||||
* personality data where we keep a count of the number of outstanding
|
|
||||||
* plugs so other code can see if a plug is active.
|
|
||||||
*/
|
|
||||||
struct md_plug_cb {
|
|
||||||
struct blk_plug_cb cb;
|
|
||||||
struct mddev *mddev;
|
|
||||||
};
|
|
||||||
|
|
||||||
static void plugger_unplug(struct blk_plug_cb *cb)
|
|
||||||
{
|
{
|
||||||
struct md_plug_cb *mdcb = container_of(cb, struct md_plug_cb, cb);
|
struct mddev *mddev = cb->data;
|
||||||
md_wakeup_thread(mdcb->mddev->thread);
|
md_wakeup_thread(mddev->thread);
|
||||||
kfree(mdcb);
|
kfree(cb);
|
||||||
}
|
}
|
||||||
|
EXPORT_SYMBOL(md_unplug);
|
||||||
/* Check that an unplug wakeup will come shortly.
|
|
||||||
* If not, wakeup the md thread immediately
|
|
||||||
*/
|
|
||||||
int mddev_check_plugged(struct mddev *mddev)
|
|
||||||
{
|
|
||||||
struct blk_plug *plug = current->plug;
|
|
||||||
struct md_plug_cb *mdcb;
|
|
||||||
|
|
||||||
if (!plug)
|
|
||||||
return 0;
|
|
||||||
|
|
||||||
list_for_each_entry(mdcb, &plug->cb_list, cb.list) {
|
|
||||||
if (mdcb->cb.callback == plugger_unplug &&
|
|
||||||
mdcb->mddev == mddev) {
|
|
||||||
/* Already on the list, move to top */
|
|
||||||
if (mdcb != list_first_entry(&plug->cb_list,
|
|
||||||
struct md_plug_cb,
|
|
||||||
cb.list))
|
|
||||||
list_move(&mdcb->cb.list, &plug->cb_list);
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
/* Not currently on the callback list */
|
|
||||||
mdcb = kmalloc(sizeof(*mdcb), GFP_ATOMIC);
|
|
||||||
if (!mdcb)
|
|
||||||
return 0;
|
|
||||||
|
|
||||||
mdcb->mddev = mddev;
|
|
||||||
mdcb->cb.callback = plugger_unplug;
|
|
||||||
list_add(&mdcb->cb.list, &plug->cb_list);
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL_GPL(mddev_check_plugged);
|
|
||||||
|
|
||||||
static inline struct mddev *mddev_get(struct mddev *mddev)
|
static inline struct mddev *mddev_get(struct mddev *mddev)
|
||||||
{
|
{
|
||||||
|
@ -627,6 +627,12 @@ extern struct bio *bio_clone_mddev(struct bio *bio, gfp_t gfp_mask,
|
|||||||
struct mddev *mddev);
|
struct mddev *mddev);
|
||||||
extern struct bio *bio_alloc_mddev(gfp_t gfp_mask, int nr_iovecs,
|
extern struct bio *bio_alloc_mddev(gfp_t gfp_mask, int nr_iovecs,
|
||||||
struct mddev *mddev);
|
struct mddev *mddev);
|
||||||
extern int mddev_check_plugged(struct mddev *mddev);
|
|
||||||
extern void md_trim_bio(struct bio *bio, int offset, int size);
|
extern void md_trim_bio(struct bio *bio, int offset, int size);
|
||||||
|
|
||||||
|
extern void md_unplug(struct blk_plug_cb *cb);
|
||||||
|
static inline int mddev_check_plugged(struct mddev *mddev)
|
||||||
|
{
|
||||||
|
return !!blk_check_plugged(md_unplug, mddev,
|
||||||
|
sizeof(struct blk_plug_cb));
|
||||||
|
}
|
||||||
#endif /* _MD_MD_H */
|
#endif /* _MD_MD_H */
|
||||||
|
@ -922,11 +922,15 @@ struct blk_plug {
|
|||||||
};
|
};
|
||||||
#define BLK_MAX_REQUEST_COUNT 16
|
#define BLK_MAX_REQUEST_COUNT 16
|
||||||
|
|
||||||
|
struct blk_plug_cb;
|
||||||
|
typedef void (*blk_plug_cb_fn)(struct blk_plug_cb *);
|
||||||
struct blk_plug_cb {
|
struct blk_plug_cb {
|
||||||
struct list_head list;
|
struct list_head list;
|
||||||
void (*callback)(struct blk_plug_cb *);
|
blk_plug_cb_fn callback;
|
||||||
|
void *data;
|
||||||
};
|
};
|
||||||
|
extern struct blk_plug_cb *blk_check_plugged(blk_plug_cb_fn unplug,
|
||||||
|
void *data, int size);
|
||||||
extern void blk_start_plug(struct blk_plug *);
|
extern void blk_start_plug(struct blk_plug *);
|
||||||
extern void blk_finish_plug(struct blk_plug *);
|
extern void blk_finish_plug(struct blk_plug *);
|
||||||
extern void blk_flush_plug_list(struct blk_plug *, bool);
|
extern void blk_flush_plug_list(struct blk_plug *, bool);
|
||||||
|
Reference in New Issue
Block a user