Lines Matching +full:background +full:- +full:layer
1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * Interface to Linux block layer for MTD 'translation layers'.
5 * Copyright © 2003-2010 David Woodhouse <dwmw2@infradead.org>
16 #include <linux/blk-mq.h>
32 put_disk(dev->disk); in blktrans_dev_release()
33 blk_mq_free_tag_set(dev->tag_set); in blktrans_dev_release()
34 kfree(dev->tag_set); in blktrans_dev_release()
35 list_del(&dev->list); in blktrans_dev_release()
41 kref_put(&dev->ref, blktrans_dev_release); in blktrans_dev_put()
54 block = blk_rq_pos(req) << 9 >> tr->blkshift; in do_blktrans_request()
55 nsect = blk_rq_cur_bytes(req) >> tr->blkshift; in do_blktrans_request()
59 if (tr->flush(dev)) in do_blktrans_request()
63 if (tr->discard(dev, block, nsect)) in do_blktrans_request()
67 buf = kmap(bio_page(req->bio)) + bio_offset(req->bio); in do_blktrans_request()
68 for (; nsect > 0; nsect--, block++, buf += tr->blksize) { in do_blktrans_request()
69 if (tr->readsect(dev, block, buf)) { in do_blktrans_request()
70 kunmap(bio_page(req->bio)); in do_blktrans_request()
74 kunmap(bio_page(req->bio)); in do_blktrans_request()
80 if (!tr->writesect) in do_blktrans_request()
86 buf = kmap(bio_page(req->bio)) + bio_offset(req->bio); in do_blktrans_request()
87 for (; nsect > 0; nsect--, block++, buf += tr->blksize) { in do_blktrans_request()
88 if (tr->writesect(dev, block, buf)) { in do_blktrans_request()
89 kunmap(bio_page(req->bio)); in do_blktrans_request()
93 kunmap(bio_page(req->bio)); in do_blktrans_request()
102 return dev->bg_stop; in mtd_blktrans_cease_background()
110 rq = list_first_entry_or_null(&dev->rq_list, struct request, queuelist); in mtd_next_request()
112 list_del_init(&rq->queuelist); in mtd_next_request()
121 __releases(&dev->queue_lock) in mtd_blktrans_work()
122 __acquires(&dev->queue_lock) in mtd_blktrans_work()
124 struct mtd_blktrans_ops *tr = dev->tr; in mtd_blktrans_work()
131 dev->bg_stop = false; in mtd_blktrans_work()
133 if (tr->background && !background_done) { in mtd_blktrans_work()
134 spin_unlock_irq(&dev->queue_lock); in mtd_blktrans_work()
135 mutex_lock(&dev->lock); in mtd_blktrans_work()
136 tr->background(dev); in mtd_blktrans_work()
137 mutex_unlock(&dev->lock); in mtd_blktrans_work()
138 spin_lock_irq(&dev->queue_lock); in mtd_blktrans_work()
140 * Do background processing just once per idle in mtd_blktrans_work()
143 background_done = !dev->bg_stop; in mtd_blktrans_work()
149 spin_unlock_irq(&dev->queue_lock); in mtd_blktrans_work()
151 mutex_lock(&dev->lock); in mtd_blktrans_work()
152 res = do_blktrans_request(dev->tr, dev, req); in mtd_blktrans_work()
153 mutex_unlock(&dev->lock); in mtd_blktrans_work()
162 spin_lock_irq(&dev->queue_lock); in mtd_blktrans_work()
171 dev = hctx->queue->queuedata; in mtd_queue_rq()
173 blk_mq_start_request(bd->rq); in mtd_queue_rq()
177 spin_lock_irq(&dev->queue_lock); in mtd_queue_rq()
178 list_add_tail(&bd->rq->queuelist, &dev->rq_list); in mtd_queue_rq()
180 spin_unlock_irq(&dev->queue_lock); in mtd_queue_rq()
187 struct mtd_blktrans_dev *dev = disk->private_data; in blktrans_open()
190 kref_get(&dev->ref); in blktrans_open()
192 mutex_lock(&dev->lock); in blktrans_open()
194 if (dev->open) in blktrans_open()
197 __module_get(dev->tr->owner); in blktrans_open()
199 if (!dev->mtd) in blktrans_open()
202 if (dev->tr->open) { in blktrans_open()
203 ret = dev->tr->open(dev); in blktrans_open()
208 ret = __get_mtd_device(dev->mtd); in blktrans_open()
211 dev->writable = mode & BLK_OPEN_WRITE; in blktrans_open()
214 dev->open++; in blktrans_open()
215 mutex_unlock(&dev->lock); in blktrans_open()
219 if (dev->tr->release) in blktrans_open()
220 dev->tr->release(dev); in blktrans_open()
222 module_put(dev->tr->owner); in blktrans_open()
223 mutex_unlock(&dev->lock); in blktrans_open()
230 struct mtd_blktrans_dev *dev = disk->private_data; in blktrans_release()
232 mutex_lock(&dev->lock); in blktrans_release()
234 if (--dev->open) in blktrans_release()
237 module_put(dev->tr->owner); in blktrans_release()
239 if (dev->mtd) { in blktrans_release()
240 if (dev->tr->release) in blktrans_release()
241 dev->tr->release(dev); in blktrans_release()
242 __put_mtd_device(dev->mtd); in blktrans_release()
245 mutex_unlock(&dev->lock); in blktrans_release()
251 struct mtd_blktrans_dev *dev = bdev->bd_disk->private_data; in blktrans_getgeo()
252 int ret = -ENXIO; in blktrans_getgeo()
254 mutex_lock(&dev->lock); in blktrans_getgeo()
256 if (!dev->mtd) in blktrans_getgeo()
259 ret = dev->tr->getgeo ? dev->tr->getgeo(dev, geo) : -ENOTTY; in blktrans_getgeo()
261 mutex_unlock(&dev->lock); in blktrans_getgeo()
278 struct mtd_blktrans_ops *tr = new->tr; in add_mtd_blktrans_dev()
280 int last_devnum = -1; in add_mtd_blktrans_dev()
286 list_for_each_entry(d, &tr->devs, list) { in add_mtd_blktrans_dev()
287 if (new->devnum == -1) { in add_mtd_blktrans_dev()
289 if (d->devnum != last_devnum+1) { in add_mtd_blktrans_dev()
291 new->devnum = last_devnum+1; in add_mtd_blktrans_dev()
292 list_add_tail(&new->list, &d->list); in add_mtd_blktrans_dev()
295 } else if (d->devnum == new->devnum) { in add_mtd_blktrans_dev()
297 return -EBUSY; in add_mtd_blktrans_dev()
298 } else if (d->devnum > new->devnum) { in add_mtd_blktrans_dev()
300 list_add_tail(&new->list, &d->list); in add_mtd_blktrans_dev()
303 last_devnum = d->devnum; in add_mtd_blktrans_dev()
306 ret = -EBUSY; in add_mtd_blktrans_dev()
307 if (new->devnum == -1) in add_mtd_blktrans_dev()
308 new->devnum = last_devnum+1; in add_mtd_blktrans_dev()
313 if (new->devnum > (MINORMASK >> tr->part_bits) || in add_mtd_blktrans_dev()
314 (tr->part_bits && new->devnum >= 27 * 26)) in add_mtd_blktrans_dev()
317 list_add_tail(&new->list, &tr->devs); in add_mtd_blktrans_dev()
320 mutex_init(&new->lock); in add_mtd_blktrans_dev()
321 kref_init(&new->ref); in add_mtd_blktrans_dev()
322 if (!tr->writesect) in add_mtd_blktrans_dev()
323 new->readonly = 1; in add_mtd_blktrans_dev()
325 ret = -ENOMEM; in add_mtd_blktrans_dev()
326 new->tag_set = kzalloc(sizeof(*new->tag_set), GFP_KERNEL); in add_mtd_blktrans_dev()
327 if (!new->tag_set) in add_mtd_blktrans_dev()
330 ret = blk_mq_alloc_sq_tag_set(new->tag_set, &mtd_mq_ops, 2, in add_mtd_blktrans_dev()
336 gd = blk_mq_alloc_disk(new->tag_set, new); in add_mtd_blktrans_dev()
342 new->disk = gd; in add_mtd_blktrans_dev()
343 new->rq = new->disk->queue; in add_mtd_blktrans_dev()
344 gd->private_data = new; in add_mtd_blktrans_dev()
345 gd->major = tr->major; in add_mtd_blktrans_dev()
346 gd->first_minor = (new->devnum) << tr->part_bits; in add_mtd_blktrans_dev()
347 gd->minors = 1 << tr->part_bits; in add_mtd_blktrans_dev()
348 gd->fops = &mtd_block_ops; in add_mtd_blktrans_dev()
350 if (tr->part_bits) { in add_mtd_blktrans_dev()
351 if (new->devnum < 26) in add_mtd_blktrans_dev()
352 snprintf(gd->disk_name, sizeof(gd->disk_name), in add_mtd_blktrans_dev()
353 "%s%c", tr->name, 'a' + new->devnum); in add_mtd_blktrans_dev()
355 snprintf(gd->disk_name, sizeof(gd->disk_name), in add_mtd_blktrans_dev()
356 "%s%c%c", tr->name, in add_mtd_blktrans_dev()
357 'a' - 1 + new->devnum / 26, in add_mtd_blktrans_dev()
358 'a' + new->devnum % 26); in add_mtd_blktrans_dev()
360 snprintf(gd->disk_name, sizeof(gd->disk_name), in add_mtd_blktrans_dev()
361 "%s%d", tr->name, new->devnum); in add_mtd_blktrans_dev()
362 gd->flags |= GENHD_FL_NO_PART; in add_mtd_blktrans_dev()
365 set_capacity(gd, ((u64)new->size * tr->blksize) >> 9); in add_mtd_blktrans_dev()
368 spin_lock_init(&new->queue_lock); in add_mtd_blktrans_dev()
369 INIT_LIST_HEAD(&new->rq_list); in add_mtd_blktrans_dev()
371 if (tr->flush) in add_mtd_blktrans_dev()
372 blk_queue_write_cache(new->rq, true, false); in add_mtd_blktrans_dev()
374 blk_queue_logical_block_size(new->rq, tr->blksize); in add_mtd_blktrans_dev()
376 blk_queue_flag_set(QUEUE_FLAG_NONROT, new->rq); in add_mtd_blktrans_dev()
377 blk_queue_flag_clear(QUEUE_FLAG_ADD_RANDOM, new->rq); in add_mtd_blktrans_dev()
379 if (tr->discard) in add_mtd_blktrans_dev()
380 blk_queue_max_discard_sectors(new->rq, UINT_MAX); in add_mtd_blktrans_dev()
382 gd->queue = new->rq; in add_mtd_blktrans_dev()
384 if (new->readonly) in add_mtd_blktrans_dev()
387 ret = device_add_disk(&new->mtd->dev, gd, NULL); in add_mtd_blktrans_dev()
391 if (new->disk_attributes) { in add_mtd_blktrans_dev()
392 ret = sysfs_create_group(&disk_to_dev(gd)->kobj, in add_mtd_blktrans_dev()
393 new->disk_attributes); in add_mtd_blktrans_dev()
399 put_disk(new->disk); in add_mtd_blktrans_dev()
401 blk_mq_free_tag_set(new->tag_set); in add_mtd_blktrans_dev()
403 kfree(new->tag_set); in add_mtd_blktrans_dev()
405 list_del(&new->list); in add_mtd_blktrans_dev()
415 if (old->disk_attributes) in del_mtd_blktrans_dev()
416 sysfs_remove_group(&disk_to_dev(old->disk)->kobj, in del_mtd_blktrans_dev()
417 old->disk_attributes); in del_mtd_blktrans_dev()
420 del_gendisk(old->disk); in del_mtd_blktrans_dev()
423 spin_lock_irqsave(&old->queue_lock, flags); in del_mtd_blktrans_dev()
424 old->rq->queuedata = NULL; in del_mtd_blktrans_dev()
425 spin_unlock_irqrestore(&old->queue_lock, flags); in del_mtd_blktrans_dev()
428 blk_mq_freeze_queue(old->rq); in del_mtd_blktrans_dev()
429 blk_mq_quiesce_queue(old->rq); in del_mtd_blktrans_dev()
430 blk_mq_unquiesce_queue(old->rq); in del_mtd_blktrans_dev()
431 blk_mq_unfreeze_queue(old->rq); in del_mtd_blktrans_dev()
435 mutex_lock(&old->lock); in del_mtd_blktrans_dev()
436 if (old->open) { in del_mtd_blktrans_dev()
437 if (old->tr->release) in del_mtd_blktrans_dev()
438 old->tr->release(old); in del_mtd_blktrans_dev()
439 __put_mtd_device(old->mtd); in del_mtd_blktrans_dev()
442 old->mtd = NULL; in del_mtd_blktrans_dev()
444 mutex_unlock(&old->lock); in del_mtd_blktrans_dev()
455 list_for_each_entry_safe(dev, next, &tr->devs, list) in blktrans_notify_remove()
456 if (dev->mtd == mtd) in blktrans_notify_remove()
457 tr->remove_dev(dev); in blktrans_notify_remove()
464 if (mtd->type == MTD_ABSENT || mtd->type == MTD_UBIVOLUME) in blktrans_notify_add()
468 tr->add_mtd(tr, mtd); in blktrans_notify_add()
487 ret = register_blkdev(tr->major, tr->name); in register_mtd_blktrans()
490 tr->name, tr->major, ret); in register_mtd_blktrans()
495 tr->major = ret; in register_mtd_blktrans()
497 tr->blkshift = ffs(tr->blksize) - 1; in register_mtd_blktrans()
499 INIT_LIST_HEAD(&tr->devs); in register_mtd_blktrans()
502 list_add(&tr->list, &blktrans_majors); in register_mtd_blktrans()
504 if (mtd->type != MTD_ABSENT && mtd->type != MTD_UBIVOLUME) in register_mtd_blktrans()
505 tr->add_mtd(tr, mtd); in register_mtd_blktrans()
517 list_del(&tr->list); in deregister_mtd_blktrans()
519 list_for_each_entry_safe(dev, next, &tr->devs, list) in deregister_mtd_blktrans()
520 tr->remove_dev(dev); in deregister_mtd_blktrans()
523 unregister_blkdev(tr->major, tr->name); in deregister_mtd_blktrans()
525 BUG_ON(!list_empty(&tr->devs)); in deregister_mtd_blktrans()
531 /* No race here -- if someone's currently in register_mtd_blktrans in mtd_blktrans_exit()
546 MODULE_DESCRIPTION("Common interface to block layer for MTD 'translation layers'");