Lines Matching refs:ubi
68 unsigned long long ubi_next_sqnum(struct ubi_device *ubi) in ubi_next_sqnum() argument
72 spin_lock(&ubi->ltree_lock); in ubi_next_sqnum()
73 sqnum = ubi->global_sqnum++; in ubi_next_sqnum()
74 spin_unlock(&ubi->ltree_lock); in ubi_next_sqnum()
87 static int ubi_get_compat(const struct ubi_device *ubi, int vol_id) in ubi_get_compat() argument
206 static struct ubi_ltree_entry *ltree_lookup(struct ubi_device *ubi, int vol_id, in ltree_lookup() argument
211 p = ubi->ltree.rb_node; in ltree_lookup()
245 static struct ubi_ltree_entry *ltree_add_entry(struct ubi_device *ubi, in ltree_add_entry() argument
259 spin_lock(&ubi->ltree_lock); in ltree_add_entry()
260 le1 = ltree_lookup(ubi, vol_id, lnum); in ltree_add_entry()
278 p = &ubi->ltree.rb_node; in ltree_add_entry()
297 rb_insert_color(&le->rb, &ubi->ltree); in ltree_add_entry()
300 spin_unlock(&ubi->ltree_lock); in ltree_add_entry()
315 static int leb_read_lock(struct ubi_device *ubi, int vol_id, int lnum) in leb_read_lock() argument
319 le = ltree_add_entry(ubi, vol_id, lnum); in leb_read_lock()
332 static void leb_read_unlock(struct ubi_device *ubi, int vol_id, int lnum) in leb_read_unlock() argument
336 spin_lock(&ubi->ltree_lock); in leb_read_unlock()
337 le = ltree_lookup(ubi, vol_id, lnum); in leb_read_unlock()
342 rb_erase(&le->rb, &ubi->ltree); in leb_read_unlock()
345 spin_unlock(&ubi->ltree_lock); in leb_read_unlock()
357 static int leb_write_lock(struct ubi_device *ubi, int vol_id, int lnum) in leb_write_lock() argument
361 le = ltree_add_entry(ubi, vol_id, lnum); in leb_write_lock()
379 static int leb_write_trylock(struct ubi_device *ubi, int vol_id, int lnum) in leb_write_trylock() argument
383 le = ltree_add_entry(ubi, vol_id, lnum); in leb_write_trylock()
390 spin_lock(&ubi->ltree_lock); in leb_write_trylock()
394 rb_erase(&le->rb, &ubi->ltree); in leb_write_trylock()
397 spin_unlock(&ubi->ltree_lock); in leb_write_trylock()
408 static void leb_write_unlock(struct ubi_device *ubi, int vol_id, int lnum) in leb_write_unlock() argument
412 spin_lock(&ubi->ltree_lock); in leb_write_unlock()
413 le = ltree_lookup(ubi, vol_id, lnum); in leb_write_unlock()
418 rb_erase(&le->rb, &ubi->ltree); in leb_write_unlock()
421 spin_unlock(&ubi->ltree_lock); in leb_write_unlock()
446 int ubi_eba_unmap_leb(struct ubi_device *ubi, struct ubi_volume *vol, in ubi_eba_unmap_leb() argument
451 if (ubi->ro_mode) in ubi_eba_unmap_leb()
454 err = leb_write_lock(ubi, vol_id, lnum); in ubi_eba_unmap_leb()
465 down_read(&ubi->fm_eba_sem); in ubi_eba_unmap_leb()
467 up_read(&ubi->fm_eba_sem); in ubi_eba_unmap_leb()
468 err = ubi_wl_put_peb(ubi, vol_id, lnum, pnum, 0); in ubi_eba_unmap_leb()
471 leb_write_unlock(ubi, vol_id, lnum); in ubi_eba_unmap_leb()
493 static int check_mapping(struct ubi_device *ubi, struct ubi_volume *vol, int lnum, in check_mapping() argument
500 if (!ubi->fast_attach) in check_mapping()
506 vidb = ubi_alloc_vid_buf(ubi, GFP_NOFS); in check_mapping()
510 err = ubi_io_read_vid_hdr(ubi, *pnum, vidb, 0); in check_mapping()
527 down_read(&ubi->fm_eba_sem); in check_mapping()
529 up_read(&ubi->fm_eba_sem); in check_mapping()
530 ubi_wl_put_peb(ubi, vol->vol_id, lnum, *pnum, torture); in check_mapping()
534 ubi_err(ubi, "unable to read VID header back from PEB %i: %i", in check_mapping()
548 ubi_err(ubi, "EBA mismatch! PEB %i is LEB %i:%i instead of LEB %i:%i", in check_mapping()
550 ubi_ro_mode(ubi); in check_mapping()
565 static int check_mapping(struct ubi_device *ubi, struct ubi_volume *vol, int lnum, in check_mapping() argument
591 int ubi_eba_read_leb(struct ubi_device *ubi, struct ubi_volume *vol, int lnum, in ubi_eba_read_leb() argument
599 err = leb_read_lock(ubi, vol_id, lnum); in ubi_eba_read_leb()
605 err = check_mapping(ubi, vol, lnum, &pnum); in ubi_eba_read_leb()
618 leb_read_unlock(ubi, vol_id, lnum); in ubi_eba_read_leb()
632 vidb = ubi_alloc_vid_buf(ubi, GFP_NOFS); in ubi_eba_read_leb()
640 err = ubi_io_read_vid_hdr(ubi, pnum, vidb, 1); in ubi_eba_read_leb()
653 ubi_warn(ubi, "corrupted VID header at PEB %d, LEB %d:%d", in ubi_eba_read_leb()
670 if (ubi->fast_attach) { in ubi_eba_read_leb()
674 ubi_ro_mode(ubi); in ubi_eba_read_leb()
689 err = ubi_io_read_data(ubi, buf, pnum, offset, len); in ubi_eba_read_leb()
698 ubi_msg(ubi, "force data checking"); in ubi_eba_read_leb()
709 ubi_warn(ubi, "CRC error: calculated %#08x, must be %#08x", in ubi_eba_read_leb()
717 err = ubi_wl_scrub_peb(ubi, pnum); in ubi_eba_read_leb()
719 leb_read_unlock(ubi, vol_id, lnum); in ubi_eba_read_leb()
725 leb_read_unlock(ubi, vol_id, lnum); in ubi_eba_read_leb()
743 int ubi_eba_read_leb_sg(struct ubi_device *ubi, struct ubi_volume *vol, in ubi_eba_read_leb_sg() argument
759 ret = ubi_eba_read_leb(ubi, vol, lnum, in ubi_eba_read_leb_sg()
806 struct ubi_device *ubi = vol->ubi; in try_recover_peb() local
813 new_pnum = ubi_wl_get_peb(ubi); in try_recover_peb()
819 ubi_msg(ubi, "recover PEB %d, move data to PEB %d", in try_recover_peb()
822 err = ubi_io_read_vid_hdr(ubi, pnum, vidb, 1); in try_recover_peb()
832 mutex_lock(&ubi->buf_mutex); in try_recover_peb()
833 memset(ubi->peb_buf + offset, 0xFF, len); in try_recover_peb()
837 err = ubi_io_read_data(ubi, ubi->peb_buf, pnum, 0, offset); in try_recover_peb()
844 memcpy(ubi->peb_buf + offset, buf, len); in try_recover_peb()
847 crc = crc32(UBI_CRC32_INIT, ubi->peb_buf, data_size); in try_recover_peb()
848 vid_hdr->sqnum = cpu_to_be64(ubi_next_sqnum(ubi)); in try_recover_peb()
852 err = ubi_io_write_vid_hdr(ubi, new_pnum, vidb); in try_recover_peb()
856 err = ubi_io_write_data(ubi, ubi->peb_buf, new_pnum, 0, data_size); in try_recover_peb()
859 mutex_unlock(&ubi->buf_mutex); in try_recover_peb()
865 up_read(&ubi->fm_eba_sem); in try_recover_peb()
868 ubi_wl_put_peb(ubi, vol_id, lnum, pnum, 1); in try_recover_peb()
869 ubi_msg(ubi, "data was successfully recovered"); in try_recover_peb()
875 ubi_wl_put_peb(ubi, vol_id, lnum, new_pnum, 1); in try_recover_peb()
876 ubi_warn(ubi, "failed to write to PEB %d", new_pnum); in try_recover_peb()
898 static int recover_peb(struct ubi_device *ubi, int pnum, int vol_id, int lnum, in recover_peb() argument
901 int err, idx = vol_id2idx(ubi, vol_id), tries; in recover_peb()
902 struct ubi_volume *vol = ubi->volumes[idx]; in recover_peb()
905 vidb = ubi_alloc_vid_buf(ubi, GFP_NOFS); in recover_peb()
917 ubi_msg(ubi, "try again"); in recover_peb()
944 struct ubi_device *ubi = vol->ubi; in try_write_vid_and_data() local
947 pnum = ubi_wl_get_peb(ubi); in try_write_vid_and_data()
958 err = ubi_io_write_vid_hdr(ubi, pnum, vidb); in try_write_vid_and_data()
960 ubi_warn(ubi, "failed to write VID header to LEB %d:%d, PEB %d", in try_write_vid_and_data()
966 err = ubi_io_write_data(ubi, buf, pnum, offset, len); in try_write_vid_and_data()
968 ubi_warn(ubi, in try_write_vid_and_data()
978 up_read(&ubi->fm_eba_sem); in try_write_vid_and_data()
981 err2 = ubi_wl_put_peb(ubi, vol_id, lnum, pnum, 1); in try_write_vid_and_data()
983 ubi_warn(ubi, "failed to return physical eraseblock %d, error %d", in try_write_vid_and_data()
987 err2 = ubi_wl_put_peb(ubi, vol_id, lnum, opnum, 0); in try_write_vid_and_data()
989 ubi_warn(ubi, "failed to return physical eraseblock %d, error %d", in try_write_vid_and_data()
1012 int ubi_eba_write_leb(struct ubi_device *ubi, struct ubi_volume *vol, int lnum, in ubi_eba_write_leb() argument
1019 if (ubi->ro_mode) in ubi_eba_write_leb()
1022 err = leb_write_lock(ubi, vol_id, lnum); in ubi_eba_write_leb()
1028 err = check_mapping(ubi, vol, lnum, &pnum); in ubi_eba_write_leb()
1037 err = ubi_io_write_data(ubi, buf, pnum, offset, len); in ubi_eba_write_leb()
1039 ubi_warn(ubi, "failed to write data to PEB %d", pnum); in ubi_eba_write_leb()
1040 if (err == -EIO && ubi->bad_allowed) in ubi_eba_write_leb()
1041 err = recover_peb(ubi, pnum, vol_id, lnum, buf, in ubi_eba_write_leb()
1052 vidb = ubi_alloc_vid_buf(ubi, GFP_NOFS); in ubi_eba_write_leb()
1054 leb_write_unlock(ubi, vol_id, lnum); in ubi_eba_write_leb()
1061 vid_hdr->sqnum = cpu_to_be64(ubi_next_sqnum(ubi)); in ubi_eba_write_leb()
1064 vid_hdr->compat = ubi_get_compat(ubi, vol_id); in ubi_eba_write_leb()
1069 if (err != -EIO || !ubi->bad_allowed) in ubi_eba_write_leb()
1078 vid_hdr->sqnum = cpu_to_be64(ubi_next_sqnum(ubi)); in ubi_eba_write_leb()
1079 ubi_msg(ubi, "try another PEB"); in ubi_eba_write_leb()
1086 ubi_ro_mode(ubi); in ubi_eba_write_leb()
1088 leb_write_unlock(ubi, vol_id, lnum); in ubi_eba_write_leb()
1115 int ubi_eba_write_leb_st(struct ubi_device *ubi, struct ubi_volume *vol, in ubi_eba_write_leb_st() argument
1123 if (ubi->ro_mode) in ubi_eba_write_leb_st()
1128 len = ALIGN(data_size, ubi->min_io_size); in ubi_eba_write_leb_st()
1130 ubi_assert(!(len & (ubi->min_io_size - 1))); in ubi_eba_write_leb_st()
1132 vidb = ubi_alloc_vid_buf(ubi, GFP_NOFS); in ubi_eba_write_leb_st()
1138 err = leb_write_lock(ubi, vol_id, lnum); in ubi_eba_write_leb_st()
1142 vid_hdr->sqnum = cpu_to_be64(ubi_next_sqnum(ubi)); in ubi_eba_write_leb_st()
1145 vid_hdr->compat = ubi_get_compat(ubi, vol_id); in ubi_eba_write_leb_st()
1158 if (err != -EIO || !ubi->bad_allowed) in ubi_eba_write_leb_st()
1161 vid_hdr->sqnum = cpu_to_be64(ubi_next_sqnum(ubi)); in ubi_eba_write_leb_st()
1162 ubi_msg(ubi, "try another PEB"); in ubi_eba_write_leb_st()
1166 ubi_ro_mode(ubi); in ubi_eba_write_leb_st()
1168 leb_write_unlock(ubi, vol_id, lnum); in ubi_eba_write_leb_st()
1193 int ubi_eba_atomic_leb_change(struct ubi_device *ubi, struct ubi_volume *vol, in ubi_eba_atomic_leb_change() argument
1201 if (ubi->ro_mode) in ubi_eba_atomic_leb_change()
1209 err = ubi_eba_unmap_leb(ubi, vol, lnum); in ubi_eba_atomic_leb_change()
1212 return ubi_eba_write_leb(ubi, vol, lnum, NULL, 0, 0); in ubi_eba_atomic_leb_change()
1215 vidb = ubi_alloc_vid_buf(ubi, GFP_NOFS); in ubi_eba_atomic_leb_change()
1221 mutex_lock(&ubi->alc_mutex); in ubi_eba_atomic_leb_change()
1222 err = leb_write_lock(ubi, vol_id, lnum); in ubi_eba_atomic_leb_change()
1226 vid_hdr->sqnum = cpu_to_be64(ubi_next_sqnum(ubi)); in ubi_eba_atomic_leb_change()
1229 vid_hdr->compat = ubi_get_compat(ubi, vol_id); in ubi_eba_atomic_leb_change()
1242 if (err != -EIO || !ubi->bad_allowed) in ubi_eba_atomic_leb_change()
1245 vid_hdr->sqnum = cpu_to_be64(ubi_next_sqnum(ubi)); in ubi_eba_atomic_leb_change()
1246 ubi_msg(ubi, "try another PEB"); in ubi_eba_atomic_leb_change()
1255 ubi_ro_mode(ubi); in ubi_eba_atomic_leb_change()
1257 leb_write_unlock(ubi, vol_id, lnum); in ubi_eba_atomic_leb_change()
1260 mutex_unlock(&ubi->alc_mutex); in ubi_eba_atomic_leb_change()
1306 int ubi_eba_copy_leb(struct ubi_device *ubi, int from, int to, in ubi_eba_copy_leb() argument
1314 ubi_assert(rwsem_is_locked(&ubi->fm_eba_sem)); in ubi_eba_copy_leb()
1323 aldata_size = ALIGN(data_size, ubi->min_io_size); in ubi_eba_copy_leb()
1326 ubi->leb_size - be32_to_cpu(vid_hdr->data_pad); in ubi_eba_copy_leb()
1328 idx = vol_id2idx(ubi, vol_id); in ubi_eba_copy_leb()
1329 spin_lock(&ubi->volumes_lock); in ubi_eba_copy_leb()
1336 vol = ubi->volumes[idx]; in ubi_eba_copy_leb()
1337 spin_unlock(&ubi->volumes_lock); in ubi_eba_copy_leb()
1359 err = leb_write_trylock(ubi, vol_id, lnum); in ubi_eba_copy_leb()
1383 mutex_lock(&ubi->buf_mutex); in ubi_eba_copy_leb()
1385 err = ubi_io_read_data(ubi, ubi->peb_buf, from, 0, aldata_size); in ubi_eba_copy_leb()
1387 ubi_warn(ubi, "error %d while reading data from PEB %d", in ubi_eba_copy_leb()
1405 ubi_calc_data_len(ubi, ubi->peb_buf, data_size); in ubi_eba_copy_leb()
1408 crc = crc32(UBI_CRC32_INIT, ubi->peb_buf, data_size); in ubi_eba_copy_leb()
1422 vid_hdr->sqnum = cpu_to_be64(ubi_next_sqnum(ubi)); in ubi_eba_copy_leb()
1424 err = ubi_io_write_vid_hdr(ubi, to, vidb); in ubi_eba_copy_leb()
1434 err = ubi_io_read_vid_hdr(ubi, to, vidb, 1); in ubi_eba_copy_leb()
1437 ubi_warn(ubi, "error %d while reading VID header back from PEB %d", in ubi_eba_copy_leb()
1447 err = ubi_io_write_data(ubi, ubi->peb_buf, to, 0, aldata_size); in ubi_eba_copy_leb()
1463 spin_lock(&ubi->volumes_lock); in ubi_eba_copy_leb()
1465 spin_unlock(&ubi->volumes_lock); in ubi_eba_copy_leb()
1468 mutex_unlock(&ubi->buf_mutex); in ubi_eba_copy_leb()
1470 leb_write_unlock(ubi, vol_id, lnum); in ubi_eba_copy_leb()
1493 static void print_rsvd_warning(struct ubi_device *ubi, in print_rsvd_warning() argument
1501 int min = ubi->beb_rsvd_level / 10; in print_rsvd_warning()
1505 if (ubi->beb_rsvd_pebs > min) in print_rsvd_warning()
1509 ubi_warn(ubi, "cannot reserve enough PEBs for bad PEB handling, reserved %d, need %d", in print_rsvd_warning()
1510 ubi->beb_rsvd_pebs, ubi->beb_rsvd_level); in print_rsvd_warning()
1511 if (ubi->corr_peb_count) in print_rsvd_warning()
1512 ubi_warn(ubi, "%d PEBs are corrupted and not used", in print_rsvd_warning()
1513 ubi->corr_peb_count); in print_rsvd_warning()
1526 int self_check_eba(struct ubi_device *ubi, struct ubi_attach_info *ai_fastmap, in self_check_eba() argument
1536 num_volumes = ubi->vtbl_slots + UBI_INT_VOL_COUNT; in self_check_eba()
1549 vol = ubi->volumes[i]; in self_check_eba()
1569 av = ubi_find_av(ai_scan, idx2vol_id(ubi, i)); in self_check_eba()
1576 av = ubi_find_av(ai_fastmap, idx2vol_id(ubi, i)); in self_check_eba()
1589 ubi_err(ubi, "LEB:%i:%i is PEB:%i instead of %i!", in self_check_eba()
1599 if (!ubi->volumes[i]) in self_check_eba()
1619 int ubi_eba_init(struct ubi_device *ubi, struct ubi_attach_info *ai) in ubi_eba_init() argument
1629 spin_lock_init(&ubi->ltree_lock); in ubi_eba_init()
1630 mutex_init(&ubi->alc_mutex); in ubi_eba_init()
1631 ubi->ltree = RB_ROOT; in ubi_eba_init()
1633 ubi->global_sqnum = ai->max_sqnum + 1; in ubi_eba_init()
1634 num_volumes = ubi->vtbl_slots + UBI_INT_VOL_COUNT; in ubi_eba_init()
1639 vol = ubi->volumes[i]; in ubi_eba_init()
1653 av = ubi_find_av(ai, idx2vol_id(ubi, i)); in ubi_eba_init()
1673 if (ubi->avail_pebs < EBA_RESERVED_PEBS) { in ubi_eba_init()
1674 ubi_err(ubi, "no enough physical eraseblocks (%d, need %d)", in ubi_eba_init()
1675 ubi->avail_pebs, EBA_RESERVED_PEBS); in ubi_eba_init()
1676 if (ubi->corr_peb_count) in ubi_eba_init()
1677 ubi_err(ubi, "%d PEBs are corrupted and not used", in ubi_eba_init()
1678 ubi->corr_peb_count); in ubi_eba_init()
1682 ubi->avail_pebs -= EBA_RESERVED_PEBS; in ubi_eba_init()
1683 ubi->rsvd_pebs += EBA_RESERVED_PEBS; in ubi_eba_init()
1685 if (ubi->bad_allowed) { in ubi_eba_init()
1686 ubi_calculate_reserved(ubi); in ubi_eba_init()
1688 if (ubi->avail_pebs < ubi->beb_rsvd_level) { in ubi_eba_init()
1690 ubi->beb_rsvd_pebs = ubi->avail_pebs; in ubi_eba_init()
1691 print_rsvd_warning(ubi, ai); in ubi_eba_init()
1693 ubi->beb_rsvd_pebs = ubi->beb_rsvd_level; in ubi_eba_init()
1695 ubi->avail_pebs -= ubi->beb_rsvd_pebs; in ubi_eba_init()
1696 ubi->rsvd_pebs += ubi->beb_rsvd_pebs; in ubi_eba_init()
1704 if (!ubi->volumes[i]) in ubi_eba_init()
1706 ubi_eba_replace_table(ubi->volumes[i], NULL); in ubi_eba_init()