Lines Matching +full:data +full:- +full:mirror

1 // SPDX-License-Identifier: GPL-2.0
12 #include "async-thread.h"
13 #include "dev-replace.h"
14 #include "rcu-string.h"
16 #include "file-item.h"
17 #include "raid-stripe-tree.h"
30 /* Is this a data path I/O that needs storage layer checksum and repair? */
33 return bbio->inode && is_data_inode(&bbio->inode->vfs_inode); in is_data_bbio()
38 return is_data_bbio(bbio) && btrfs_op(&bbio->bio) == BTRFS_MAP_WRITE; in bbio_has_ordered_extent()
49 bbio->fs_info = fs_info; in btrfs_bio_init()
50 bbio->end_io = end_io; in btrfs_bio_init()
51 bbio->private = private; in btrfs_bio_init()
52 atomic_set(&bbio->pending_ios, 1); in btrfs_bio_init()
85 bio = bio_split_rw(&orig_bbio->bio, &fs_info->limits, &nr_segs, in btrfs_split_bio()
88 bio = bio_split(&orig_bbio->bio, map_length >> SECTOR_SHIFT, in btrfs_split_bio()
93 bbio->inode = orig_bbio->inode; in btrfs_split_bio()
94 bbio->file_offset = orig_bbio->file_offset; in btrfs_split_bio()
95 orig_bbio->file_offset += map_length; in btrfs_split_bio()
97 refcount_inc(&orig_bbio->ordered->refs); in btrfs_split_bio()
98 bbio->ordered = orig_bbio->ordered; in btrfs_split_bio()
100 atomic_inc(&orig_bbio->pending_ios); in btrfs_split_bio()
108 btrfs_put_ordered_extent(bbio->ordered); in btrfs_cleanup_bio()
109 bio_put(&bbio->bio); in btrfs_cleanup_bio()
115 struct btrfs_ordered_extent *ordered = bbio->ordered; in __btrfs_bio_end_io()
117 bbio->end_io(bbio); in __btrfs_bio_end_io()
120 bbio->end_io(bbio); in __btrfs_bio_end_io()
126 bbio->bio.bi_status = status; in btrfs_bio_end_io()
136 * For writes we tolerate nr_mirrors - 1 write failures, so we can't in btrfs_bbio_propagate_error()
141 if (bbio->bio.bi_end_io == &btrfs_orig_write_end_io) { in btrfs_bbio_propagate_error()
142 struct btrfs_io_stripe *orig_stripe = orig_bbio->bio.bi_private; in btrfs_bbio_propagate_error()
143 struct btrfs_io_context *orig_bioc = orig_stripe->bioc; in btrfs_bbio_propagate_error()
145 atomic_add(orig_bioc->max_errors, &orig_bioc->error); in btrfs_bbio_propagate_error()
147 orig_bbio->bio.bi_status = bbio->bio.bi_status; in btrfs_bbio_propagate_error()
153 if (bbio->bio.bi_pool == &btrfs_clone_bioset) { in btrfs_orig_bbio_end_io()
154 struct btrfs_bio *orig_bbio = bbio->private; in btrfs_orig_bbio_end_io()
156 if (bbio->bio.bi_status) in btrfs_orig_bbio_end_io()
162 if (atomic_dec_and_test(&bbio->pending_ios)) in btrfs_orig_bbio_end_io()
168 if (cur_mirror == fbio->num_copies) in next_repair_mirror()
169 return cur_mirror + 1 - fbio->num_copies; in next_repair_mirror()
176 return fbio->num_copies; in prev_repair_mirror()
177 return cur_mirror - 1; in prev_repair_mirror()
182 if (atomic_dec_and_test(&fbio->repair_count)) { in btrfs_repair_done()
183 btrfs_orig_bbio_end_io(fbio->bbio); in btrfs_repair_done()
191 struct btrfs_failed_bio *fbio = repair_bbio->private; in btrfs_end_repair_bio()
192 struct btrfs_inode *inode = repair_bbio->inode; in btrfs_end_repair_bio()
193 struct btrfs_fs_info *fs_info = inode->root->fs_info; in btrfs_end_repair_bio()
194 struct bio_vec *bv = bio_first_bvec_all(&repair_bbio->bio); in btrfs_end_repair_bio()
195 int mirror = repair_bbio->mirror_num; in btrfs_end_repair_bio() local
198 * We can only trigger this for data bio, which doesn't support larger in btrfs_end_repair_bio()
201 ASSERT(folio_order(page_folio(bv->bv_page)) == 0); in btrfs_end_repair_bio()
203 if (repair_bbio->bio.bi_status || in btrfs_end_repair_bio()
205 bio_reset(&repair_bbio->bio, NULL, REQ_OP_READ); in btrfs_end_repair_bio()
206 repair_bbio->bio.bi_iter = repair_bbio->saved_iter; in btrfs_end_repair_bio()
208 mirror = next_repair_mirror(fbio, mirror); in btrfs_end_repair_bio()
209 if (mirror == fbio->bbio->mirror_num) { in btrfs_end_repair_bio()
210 btrfs_debug(fs_info, "no mirror left"); in btrfs_end_repair_bio()
211 fbio->bbio->bio.bi_status = BLK_STS_IOERR; in btrfs_end_repair_bio()
215 btrfs_submit_bio(repair_bbio, mirror); in btrfs_end_repair_bio()
220 mirror = prev_repair_mirror(fbio, mirror); in btrfs_end_repair_bio()
222 repair_bbio->file_offset, fs_info->sectorsize, in btrfs_end_repair_bio()
223 repair_bbio->saved_iter.bi_sector << SECTOR_SHIFT, in btrfs_end_repair_bio()
224 page_folio(bv->bv_page), bv->bv_offset, mirror); in btrfs_end_repair_bio()
225 } while (mirror != fbio->bbio->mirror_num); in btrfs_end_repair_bio()
229 bio_put(&repair_bbio->bio); in btrfs_end_repair_bio()
233 * Try to kick off a repair read to the next available mirror for a bad sector.
235 * This primarily tries to recover good data to serve the actual read request,
236 * but also tries to write the good data back to the bad mirror(s) when a
244 struct btrfs_inode *inode = failed_bbio->inode; in repair_one_sector()
245 struct btrfs_fs_info *fs_info = inode->root->fs_info; in repair_one_sector()
246 const u32 sectorsize = fs_info->sectorsize; in repair_one_sector()
247 const u64 logical = (failed_bbio->saved_iter.bi_sector << SECTOR_SHIFT); in repair_one_sector()
251 int mirror; in repair_one_sector() local
254 failed_bbio->file_offset + bio_offset); in repair_one_sector()
259 failed_bbio->bio.bi_status = BLK_STS_IOERR; in repair_one_sector()
265 fbio->bbio = failed_bbio; in repair_one_sector()
266 fbio->num_copies = num_copies; in repair_one_sector()
267 atomic_set(&fbio->repair_count, 1); in repair_one_sector()
270 atomic_inc(&fbio->repair_count); in repair_one_sector()
274 repair_bio->bi_iter.bi_sector = failed_bbio->saved_iter.bi_sector; in repair_one_sector()
275 __bio_add_page(repair_bio, bv->bv_page, bv->bv_len, bv->bv_offset); in repair_one_sector()
279 repair_bbio->inode = failed_bbio->inode; in repair_one_sector()
280 repair_bbio->file_offset = failed_bbio->file_offset + bio_offset; in repair_one_sector()
282 mirror = next_repair_mirror(fbio, failed_bbio->mirror_num); in repair_one_sector()
283 btrfs_debug(fs_info, "submitting repair read to mirror %d", mirror); in repair_one_sector()
284 btrfs_submit_bio(repair_bbio, mirror); in repair_one_sector()
290 struct btrfs_inode *inode = bbio->inode; in btrfs_check_read_bio()
291 struct btrfs_fs_info *fs_info = inode->root->fs_info; in btrfs_check_read_bio()
292 u32 sectorsize = fs_info->sectorsize; in btrfs_check_read_bio()
293 struct bvec_iter *iter = &bbio->saved_iter; in btrfs_check_read_bio()
294 blk_status_t status = bbio->bio.bi_status; in btrfs_check_read_bio()
298 /* Read-repair requires the inode field to be set by the submitter. */ in btrfs_check_read_bio()
305 if (bbio->bio.bi_pool == &btrfs_repair_bioset) { in btrfs_check_read_bio()
311 bbio->bio.bi_status = BLK_STS_OK; in btrfs_check_read_bio()
313 while (iter->bi_size) { in btrfs_check_read_bio()
314 struct bio_vec bv = bio_iter_iovec(&bbio->bio, *iter); in btrfs_check_read_bio()
320 bio_advance_iter_single(&bbio->bio, iter, sectorsize); in btrfs_check_read_bio()
324 if (bbio->csum != bbio->csum_inline) in btrfs_check_read_bio()
325 kfree(bbio->csum); in btrfs_check_read_bio()
335 if (!dev || !dev->bdev) in btrfs_log_dev_io_error()
337 if (bio->bi_status != BLK_STS_IOERR && bio->bi_status != BLK_STS_TARGET) in btrfs_log_dev_io_error()
342 else if (!(bio->bi_opf & REQ_RAHEAD)) in btrfs_log_dev_io_error()
344 if (bio->bi_opf & REQ_PREFLUSH) in btrfs_log_dev_io_error()
351 if (bio->bi_opf & REQ_META) in btrfs_end_io_wq()
352 return fs_info->endio_meta_workers; in btrfs_end_io_wq()
353 return fs_info->endio_workers; in btrfs_end_io_wq()
362 btrfs_check_read_bio(bbio, bbio->bio.bi_private); in btrfs_end_bio_work()
370 struct btrfs_device *dev = bio->bi_private; in btrfs_simple_end_io()
371 struct btrfs_fs_info *fs_info = bbio->fs_info; in btrfs_simple_end_io()
375 if (bio->bi_status) in btrfs_simple_end_io()
379 INIT_WORK(&bbio->end_io_work, btrfs_end_bio_work); in btrfs_simple_end_io()
380 queue_work(btrfs_end_io_wq(fs_info, bio), &bbio->end_io_work); in btrfs_simple_end_io()
382 if (bio_op(bio) == REQ_OP_ZONE_APPEND && !bio->bi_status) in btrfs_simple_end_io()
390 struct btrfs_io_context *bioc = bio->bi_private; in btrfs_raid56_end_io()
393 btrfs_bio_counter_dec(bioc->fs_info); in btrfs_raid56_end_io()
394 bbio->mirror_num = bioc->mirror_num; in btrfs_raid56_end_io()
405 struct btrfs_io_stripe *stripe = bio->bi_private; in btrfs_orig_write_end_io()
406 struct btrfs_io_context *bioc = stripe->bioc; in btrfs_orig_write_end_io()
409 btrfs_bio_counter_dec(bioc->fs_info); in btrfs_orig_write_end_io()
411 if (bio->bi_status) { in btrfs_orig_write_end_io()
412 atomic_inc(&bioc->error); in btrfs_orig_write_end_io()
413 btrfs_log_dev_io_error(bio, stripe->dev); in btrfs_orig_write_end_io()
420 if (atomic_read(&bioc->error) > bioc->max_errors) in btrfs_orig_write_end_io()
421 bio->bi_status = BLK_STS_IOERR; in btrfs_orig_write_end_io()
423 bio->bi_status = BLK_STS_OK; in btrfs_orig_write_end_io()
425 if (bio_op(bio) == REQ_OP_ZONE_APPEND && !bio->bi_status) in btrfs_orig_write_end_io()
426 stripe->physical = bio->bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_orig_write_end_io()
434 struct btrfs_io_stripe *stripe = bio->bi_private; in btrfs_clone_write_end_io()
436 if (bio->bi_status) { in btrfs_clone_write_end_io()
437 atomic_inc(&stripe->bioc->error); in btrfs_clone_write_end_io()
438 btrfs_log_dev_io_error(bio, stripe->dev); in btrfs_clone_write_end_io()
440 stripe->physical = bio->bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_clone_write_end_io()
444 bio_endio(stripe->bioc->orig_bio); in btrfs_clone_write_end_io()
450 if (!dev || !dev->bdev || in btrfs_submit_dev_bio()
451 test_bit(BTRFS_DEV_STATE_MISSING, &dev->dev_state) || in btrfs_submit_dev_bio()
453 !test_bit(BTRFS_DEV_STATE_WRITEABLE, &dev->dev_state))) { in btrfs_submit_dev_bio()
458 bio_set_dev(bio, dev->bdev); in btrfs_submit_dev_bio()
465 u64 physical = bio->bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_submit_dev_bio()
466 u64 zone_start = round_down(physical, dev->fs_info->zone_size); in btrfs_submit_dev_bio()
469 bio->bi_iter.bi_sector = zone_start >> SECTOR_SHIFT; in btrfs_submit_dev_bio()
471 btrfs_debug_in_rcu(dev->fs_info, in btrfs_submit_dev_bio()
473 __func__, bio_op(bio), bio->bi_opf, bio->bi_iter.bi_sector, in btrfs_submit_dev_bio()
474 (unsigned long)dev->bdev->bd_dev, btrfs_dev_name(dev), in btrfs_submit_dev_bio()
475 dev->devid, bio->bi_iter.bi_size); in btrfs_submit_dev_bio()
477 if (bio->bi_opf & REQ_BTRFS_CGROUP_PUNT) in btrfs_submit_dev_bio()
485 struct bio *orig_bio = bioc->orig_bio, *bio; in btrfs_submit_mirrored_bio()
489 /* Reuse the bio embedded into the btrfs_bio for the last mirror */ in btrfs_submit_mirrored_bio()
490 if (dev_nr == bioc->num_stripes - 1) { in btrfs_submit_mirrored_bio()
492 bio->bi_end_io = btrfs_orig_write_end_io; in btrfs_submit_mirrored_bio()
496 bio->bi_end_io = btrfs_clone_write_end_io; in btrfs_submit_mirrored_bio()
499 bio->bi_private = &bioc->stripes[dev_nr]; in btrfs_submit_mirrored_bio()
500 bio->bi_iter.bi_sector = bioc->stripes[dev_nr].physical >> SECTOR_SHIFT; in btrfs_submit_mirrored_bio()
501 bioc->stripes[dev_nr].bioc = bioc; in btrfs_submit_mirrored_bio()
502 bioc->size = bio->bi_iter.bi_size; in btrfs_submit_mirrored_bio()
503 btrfs_submit_dev_bio(bioc->stripes[dev_nr].dev, bio); in btrfs_submit_mirrored_bio()
510 /* Single mirror read/write fast path. */ in __btrfs_submit_bio()
511 btrfs_bio(bio)->mirror_num = mirror_num; in __btrfs_submit_bio()
513 btrfs_bio(bio)->orig_physical = smap->physical; in __btrfs_submit_bio()
514 bio->bi_iter.bi_sector = smap->physical >> SECTOR_SHIFT; in __btrfs_submit_bio()
516 btrfs_bio(bio)->orig_physical = smap->physical; in __btrfs_submit_bio()
517 bio->bi_private = smap->dev; in __btrfs_submit_bio()
518 bio->bi_end_io = btrfs_simple_end_io; in __btrfs_submit_bio()
519 btrfs_submit_dev_bio(smap->dev, bio); in __btrfs_submit_bio()
520 } else if (bioc->map_type & BTRFS_BLOCK_GROUP_RAID56_MASK) { in __btrfs_submit_bio()
522 bio->bi_private = bioc; in __btrfs_submit_bio()
523 bio->bi_end_io = btrfs_raid56_end_io; in __btrfs_submit_bio()
530 int total_devs = bioc->num_stripes; in __btrfs_submit_bio()
532 bioc->orig_bio = bio; in __btrfs_submit_bio()
540 if (bbio->bio.bi_opf & REQ_META) in btrfs_bio_csum()
571 ret = btrfs_bio_csum(async->bbio); in run_one_async_start()
573 async->bbio->bio.bi_status = ret; in run_one_async_start()
590 struct bio *bio = &async->bbio->bio; in run_one_async_done()
598 if (bio->bi_status) { in run_one_async_done()
599 btrfs_orig_bbio_end_io(async->bbio); in run_one_async_done()
608 bio->bi_opf |= REQ_BTRFS_CGROUP_PUNT; in run_one_async_done()
609 __btrfs_submit_bio(bio, async->bioc, &async->smap, async->mirror_num); in run_one_async_done()
615 if (test_bit(BTRFS_FS_CSUM_IMPL_FAST, &bbio->fs_info->flags)) in should_async_write()
622 if (op_is_sync(bbio->bio.bi_opf)) in should_async_write()
626 if ((bbio->bio.bi_opf & REQ_META) && btrfs_is_zoned(bbio->fs_info)) in should_async_write()
641 struct btrfs_fs_info *fs_info = bbio->fs_info; in btrfs_wq_submit_bio()
648 async->bbio = bbio; in btrfs_wq_submit_bio()
649 async->bioc = bioc; in btrfs_wq_submit_bio()
650 async->smap = *smap; in btrfs_wq_submit_bio()
651 async->mirror_num = mirror_num; in btrfs_wq_submit_bio()
653 btrfs_init_work(&async->work, run_one_async_start, run_one_async_done); in btrfs_wq_submit_bio()
654 btrfs_queue_work(fs_info->workers, &async->work); in btrfs_wq_submit_bio()
660 struct btrfs_inode *inode = bbio->inode; in btrfs_submit_chunk()
661 struct btrfs_fs_info *fs_info = bbio->fs_info; in btrfs_submit_chunk()
663 struct bio *bio = &bbio->bio; in btrfs_submit_chunk()
664 u64 logical = bio->bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_submit_chunk()
665 u64 length = bio->bi_iter.bi_size; in btrfs_submit_chunk()
673 smap.is_scrub = !bbio->inode; in btrfs_submit_chunk()
685 map_length = min(map_length, fs_info->max_zone_append_size); in btrfs_submit_chunk()
689 bio = &bbio->bio; in btrfs_submit_chunk()
694 * data reads. in btrfs_submit_chunk()
697 bbio->saved_iter = bio->bi_iter; in btrfs_submit_chunk()
705 bio->bi_opf &= ~REQ_OP_WRITE; in btrfs_submit_chunk()
706 bio->bi_opf |= REQ_OP_ZONE_APPEND; in btrfs_submit_chunk()
710 btrfs_need_stripe_tree_update(bioc->fs_info, bioc->map_type)) { in btrfs_submit_chunk()
718 list_add_tail(&bioc->rst_ordered_entry, &bbio->ordered->bioc_list); in btrfs_submit_chunk()
723 * point, so they are handled as part of the no-checksum case. in btrfs_submit_chunk()
725 if (inode && !(inode->flags & BTRFS_INODE_NODATASUM) && in btrfs_submit_chunk()
726 !test_bit(BTRFS_FS_STATE_NO_CSUMS, &fs_info->fs_state) && in btrfs_submit_chunk()
727 !btrfs_is_data_reloc_root(inode->root)) { in btrfs_submit_chunk()
758 /* If bbio->inode is not populated, its file_offset must be 0. */ in btrfs_submit_bio()
759 ASSERT(bbio->inode || bbio->file_offset == 0); in btrfs_submit_bio()
784 ASSERT(!(fs_info->sb->s_flags & SB_RDONLY)); in btrfs_repair_io_failure()
800 if (!smap.dev->bdev || in btrfs_repair_io_failure()
801 !test_bit(BTRFS_DEV_STATE_WRITEABLE, &smap.dev->dev_state)) { in btrfs_repair_io_failure()
802 ret = -EIO; in btrfs_repair_io_failure()
806 bio_init(&bio, smap.dev->bdev, &bvec, 1, REQ_OP_WRITE | REQ_SYNC); in btrfs_repair_io_failure()
833 * If @dev_replace is true, the write would be submitted to dev-replace target.
837 struct btrfs_fs_info *fs_info = bbio->fs_info; in btrfs_submit_repair_write()
838 u64 logical = bbio->bio.bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_submit_repair_write()
839 u64 length = bbio->bio.bi_iter.bi_size; in btrfs_submit_repair_write()
845 ASSERT(btrfs_op(&bbio->bio) == BTRFS_MAP_WRITE); in btrfs_submit_repair_write()
846 ASSERT(!bbio->inode); in btrfs_submit_repair_write()
854 ASSERT(smap.dev == fs_info->dev_replace.srcdev); in btrfs_submit_repair_write()
855 smap.dev = fs_info->dev_replace.tgtdev; in btrfs_submit_repair_write()
857 __btrfs_submit_bio(&bbio->bio, NULL, &smap, mirror_num); in btrfs_submit_repair_write()
870 return -ENOMEM; in btrfs_bioset_init()
889 return -ENOMEM; in btrfs_bioset_init()