| /linux/drivers/md/ |
| H A D | md-bitmap.h | 111 sector_t *blocks, bool degraded); 150 sector_t *blocks, bool degraded) in md_bitmap_start_sync() argument 158 return mddev->bitmap_ops->start_sync(mddev, offset, blocks, degraded); in md_bitmap_start_sync()
|
| H A D | raid0.c | 648 if (mddev->degraded != 1) { in raid0_takeover_raid45() 651 mddev->degraded); in raid0_takeover_raid45() 701 if (mddev->degraded != (mddev->raid_disks>>1)) { in raid0_takeover_raid10() 713 mddev->degraded = 0; in raid0_takeover_raid10() 730 if ((mddev->raid_disks - 1) != mddev->degraded) { in raid0_takeover_raid1()
|
| H A D | raid1.c | 413 if (r1_bio->mddev->degraded == conf->raid_disks || in raid1_end_read_request() 414 (r1_bio->mddev->degraded == conf->raid_disks-1 && in raid1_end_read_request() 1680 conf->raid_disks - mddev->degraded > 1) in raid1_write_request() 1747 conf->raid_disks - mddev->degraded); in raid1_status() 1781 (conf->raid_disks - mddev->degraded) == 1) { in raid1_error() 1791 mddev->degraded++; in raid1_error() 1803 mdname(mddev), conf->raid_disks - mddev->degraded); in raid1_error() 1815 pr_debug(" --- wd:%d rd:%d\n", conf->raid_disks - conf->mddev->degraded, in print_conf() 1886 mddev->degraded -= count; in raid1_spare_active() 1928 rdev->mddev->degraded < conf->raid_disks) in raid1_remove_conf() [all …]
|
| H A D | md-llbitmap.c | 468 if (llbitmap->mddev->degraded || !infect) { in llbitmap_set_page_dirty() 760 need_resync = !mddev->degraded; in llbitmap_state_machine() 764 if (unlikely(mddev->degraded)) { in llbitmap_state_machine() 790 need_resync = !mddev->degraded; in llbitmap_state_machine() 1060 if (llbitmap->mddev->degraded) in md_llbitmap_daemon_fn() 1378 if (mddev->degraded) in llbitmap_skip_sync_blocks() 1391 sector_t *blocks, bool degraded) in llbitmap_start_sync() argument
|
| H A D | raid10.c | 1921 conf->geo.raid_disks - mddev->degraded); in raid10_status() 2011 mddev->degraded++; in raid10_error() 2022 mdname(mddev), conf->geo.raid_disks - mddev->degraded); in raid10_error() 2035 pr_debug(" --- wd:%d rd:%d\n", conf->geo.raid_disks - conf->mddev->degraded, in print_conf() 2098 mddev->degraded -= count; in raid10_spare_active() 3522 mddev->degraded) && in raid10_sync_request() 4042 mddev->degraded = 0; in raid10_run() 4060 mddev->degraded++; in raid10_run() 4077 mdname(mddev), conf->geo.raid_disks - mddev->degraded, in raid10_run() 4190 if (mddev->degraded > 0) { in raid10_takeover_raid0() [all …]
|
| H A D | raid5.c | 697 int degraded, degraded2; in raid5_calc_degraded() local 700 degraded = 0; in raid5_calc_degraded() 707 degraded++; in raid5_calc_degraded() 721 degraded++; in raid5_calc_degraded() 724 return degraded; in raid5_calc_degraded() 744 if (degraded2 > degraded) in raid5_calc_degraded() 746 return degraded; in raid5_calc_degraded() 751 int degraded = conf->mddev->degraded; in has_failed() local 757 degraded = raid5_calc_degraded(conf); in has_failed() 759 return degraded > conf->max_degraded; in has_failed() [all …]
|
| H A D | md-bitmap.c | 1751 if (!bitmap->mddev->degraded) { in bitmap_end_write() 1780 sector_t *blocks, bool degraded) in __bitmap_start_sync() argument 1793 if (!degraded) { /* don't set/clear bits if degraded */ in __bitmap_start_sync() 1805 sector_t *blocks, bool degraded) in bitmap_start_sync() argument 1820 &blocks1, degraded); in bitmap_start_sync() 2233 if (mddev->degraded == 0 in bitmap_load() 2915 if (mddev->degraded) in can_clear_store()
|
| H A D | md.h | 529 int degraded; /* whether md should consider member
|
| H A D | dm-raid.c | 1748 if (rs->md.degraded) { in rs_check_takeover() 1818 mddev->degraded = 1; in rs_check_takeover() 2047 else if (mddev->degraded) in rs_check_reshape() 2820 mddev->degraded = rs->delta_disks < 0 ? -rs->delta_disks : rs->delta_disks; in rs_prepare_reshape()
|
| H A D | md.c | 2876 if (mddev->degraded) in md_update_sb() 3001 if (mddev->degraded) in add_bound_rdev() 4156 mddev->degraded = 0; in level_store() 5461 return sprintf(page, "%d\n", mddev->degraded); in degraded_show() 5463 static struct md_sysfs_entry md_degraded = __ATTR_RO(degraded); 6795 if (mddev->degraded && md_is_rdwr(mddev)) in md_run() 6963 mddev->degraded = 0; in md_clean() 8721 if (mddev->degraded && !test_bit(MD_BROKEN, &mddev->flags)) in md_error() 10450 mddev->degraded != mddev->raid_disks) { in md_reap_sync_thread() 10462 if (!mddev->degraded) in md_reap_sync_thread()
|
| /linux/fs/btrfs/tests/ |
| H A D | zoned-tests.c | 26 bool degraded; member 84 if (test->degraded) in test_load_zone_info() 244 .degraded = true, 347 .degraded = true, 539 .degraded = true,
|
| /linux/Documentation/driver-api/md/ |
| H A D | raid5-ppl.rst | 8 in degraded state, there is no way to recalculate parity, because one of the 10 array or using it is as degraded - data calculated from parity for array blocks 13 this, md by default does not allow starting a dirty degraded array.
|
| H A D | raid5-cache.rst | 29 unclean shutdown. We call an array degraded if it has inconsistent data. MD 37 shutdown even the array is degraded. Thus the cache can close the 'write hole'.
|
| /linux/Documentation/admin-guide/ |
| H A D | md.rst | 89 Boot time assembly of degraded/dirty arrays 92 If a raid5 or raid6 array is both dirty and degraded, it could have 95 is degraded means that some datablocks are missing and cannot reliably 107 when set to 1, bypassed the checks and will allows dirty degraded 110 So, to boot with a root filesystem of a dirty degraded raid 5 or 6, use:: 457 degraded array, or if the array becomes degraded during a write. 459 once the array becomes non-degraded, and this fact has been 709 degraded 711 arrays is degraded. So an optimal array will show ``0``. A
|
| /linux/arch/arm/boot/dts/nxp/imx/ |
| H A D | imx7d-flex-concentrator.dts | 55 label = "hmi-b:red:heartbeat-degraded";
|
| /linux/Documentation/networking/ |
| H A D | ipvlan.rst | 131 puts the NIC in promiscuous mode and degraded performance is a concern.
|
| H A D | nexthop-group-resilient.rst | 44 degraded application performance.
|
| /linux/Documentation/scsi/ |
| H A D | ChangeLog.megaraid_sas | 105 1. Fix reglockFlags for degraded raid5/6 for MR 9360/9380. 130 1. Fix FastPath I/O to work with degraded RAID 1.
|
| /linux/Documentation/PCI/ |
| H A D | pcieaer-howto.rst | 129 of those errors, which may result in degraded performance or function
|
| /linux/Documentation/networking/device_drivers/ethernet/chelsio/ |
| H A D | cxgb.rst | 341 is handling 10Gb networking can switch between CPUs causing degraded
|
| /linux/Documentation/admin-guide/device-mapper/ |
| H A D | cache.rst | 212 passthrough a degraded mode useful for various cache coherency
|
| H A D | vdo-design.rst | 260 available, the target can resume normal operation in a degraded mode. Read 261 and write requests can be serviced, perhaps with degraded performance,
|
| /linux/Documentation/driver-api/ |
| H A D | device_link.rst | 139 in a degraded mode (reduced feature set or performance) when those resources
|
| /linux/arch/x86/math-emu/ |
| H A D | README | 289 give much degraded precision. For example, the integer number
|
| /linux/Documentation/arch/arm64/ |
| H A D | acpi_object_usage.rst | 728 system can continue correct operation, even if possibly degraded. The NMI is
|