Lines Matching full:if
42 if (bioset_init(&f2fs_bioset, F2FS_BIO_POOL_SIZE, in f2fs_init_bioset()
61 if (noio) { in f2fs_bio_alloc()
66 if (time_to_inject(sbi, FAULT_ALLOC_BIO)) { in f2fs_bio_alloc()
80 if (!mapping) in __is_cp_guaranteed()
83 if (f2fs_is_compressed_page(page)) in __is_cp_guaranteed()
89 if (inode->i_ino == F2FS_META_INO(sbi) || in __is_cp_guaranteed()
103 if (mapping) { in __read_io_type()
107 if (inode->i_ino == F2FS_META_INO(sbi)) in __read_io_type()
110 if (inode->i_ino == F2FS_NODE_INO(sbi)) in __read_io_type()
141 if (compr && f2fs_is_compressed_page(page)) { in __read_end_io()
145 if (verity) in __read_end_io()
149 /* PG_error was set if any post_read step failed */ in __read_end_io()
150 if (bio->bi_status || PageError(page)) { in __read_end_io()
165 if (!compr) in __f2fs_read_end_io()
204 if (dic) { in f2fs_verify_bio()
205 if (atomic_dec_return(&dic->pending_pages)) in f2fs_verify_bio()
213 if (bio->bi_status || PageError(page)) in f2fs_verify_bio()
216 if (fsverity_verify_page(page)) { in f2fs_verify_bio()
251 if (enabled_steps & (1 << STEP_DECOMPRESS)) { in f2fs_verity_work()
267 if (ctx->enabled_steps & (1 << STEP_DECRYPT)) in f2fs_post_read_work()
270 if (ctx->enabled_steps & (1 << STEP_DECOMPRESS)) in f2fs_post_read_work()
273 if (ctx->enabled_steps & (1 << STEP_VERITY)) { in f2fs_post_read_work()
297 if (ctx->enabled_steps & (1 << STEP_DECRYPT) || in bio_post_read_processing()
304 if (ctx->enabled_steps & (1 << STEP_VERITY)) { in bio_post_read_processing()
322 if (time_to_inject(sbi, FAULT_READ_IO)) { in f2fs_read_end_io()
327 if (f2fs_bio_post_read_required(bio)) { in f2fs_read_end_io()
343 if (time_to_inject(sbi, FAULT_WRITE_IO)) { in f2fs_write_end_io()
352 if (IS_DUMMY_WRITTEN_PAGE(page)) { in f2fs_write_end_io()
358 if (unlikely(bio->bi_status)) in f2fs_write_end_io()
366 if (f2fs_is_compressed_page(page)) { in f2fs_write_end_io()
372 if (unlikely(bio->bi_status)) { in f2fs_write_end_io()
374 if (type == F2FS_WB_CP_DATA) in f2fs_write_end_io()
382 if (f2fs_in_warm_node_list(sbi, page)) in f2fs_write_end_io()
387 if (!get_pages(sbi, F2FS_WB_CP_DATA) && in f2fs_write_end_io()
400 if (f2fs_is_multi_device(sbi)) { in f2fs_target_device()
402 if (FDEV(i).start_blk <= blk_addr && in f2fs_target_device()
410 if (bio) { in f2fs_target_device()
421 if (!f2fs_is_multi_device(sbi)) in f2fs_target_device_index()
425 if (FDEV(i).start_blk <= blkaddr && FDEV(i).end_blk >= blkaddr) in f2fs_target_device_index()
431 * Return true, if pre_bio's bdev is same as its target device.
448 if (is_read_io(fio->op)) { in __bio_alloc()
457 if (fio->io_wbc) in __bio_alloc()
472 if (!fio || !fio->encrypted_page) in f2fs_set_bio_crypt_ctx()
484 if (fio && fio->encrypted_page) in f2fs_crypt_mergeable_bio()
493 if (!is_read_io(bio_op(bio))) { in __submit_bio()
496 if (type != DATA && type != NODE) in __submit_bio()
499 if (f2fs_lfs_mode(sbi) && current->plug) in __submit_bio()
502 if (F2FS_IO_ALIGNED(sbi)) in __submit_bio()
508 if (start == 0) in __submit_bio()
522 if (bio_add_page(bio, page, PAGE_SIZE, 0) < PAGE_SIZE) in __submit_bio()
529 if (type == NODE) in __submit_bio()
533 if (is_read_io(bio_op(bio))) in __submit_bio()
552 if (fio->type == DATA) in __attach_io_flag()
554 else if (fio->type == NODE) in __attach_io_flag()
568 if ((1 << fio->temp) & meta_flag) in __attach_io_flag()
570 if ((1 << fio->temp) & fua_flag) in __attach_io_flag()
578 if (!io->bio) in __submit_merged_bio()
584 if (is_read_io(fio->op)) in __submit_merged_bio()
599 if (!bio) in __has_merged_page()
602 if (!inode && !page && !ino) in __has_merged_page()
608 if (fscrypt_is_bounce_page(target)) { in __has_merged_page()
610 if (IS_ERR(target)) in __has_merged_page()
613 if (f2fs_is_compressed_page(target)) { in __has_merged_page()
615 if (IS_ERR(target)) in __has_merged_page()
619 if (inode && inode == target->mapping->host) in __has_merged_page()
621 if (page && page == target) in __has_merged_page()
623 if (ino && ino == ino_of_node(target)) in __has_merged_page()
639 if (type >= META_FLUSH) { in __f2fs_submit_merged_write()
643 if (!test_opt(sbi, NOBARRIER)) in __f2fs_submit_merged_write()
658 if (!force) { in __submit_merged_write_cond()
666 if (ret) in __submit_merged_write_cond()
670 if (type >= META) in __submit_merged_write_cond()
704 if (!f2fs_is_valid_blkaddr(fio->sbi, fio->new_blkaddr, in f2fs_submit_page_bio()
718 if (bio_add_page(bio, page, PAGE_SIZE, 0) < PAGE_SIZE) { in f2fs_submit_page_bio()
723 if (fio->io_wbc && !is_read_io(fio->op)) in f2fs_submit_page_bio()
739 if (last_blkaddr + 1 != cur_blkaddr) in page_is_mergeable()
747 if (io->fio.op != fio->op) in io_type_is_mergeable()
758 if (F2FS_IO_ALIGNED(sbi) && (fio->type == DATA || fio->type == NODE)) { in io_is_mergeable()
765 if (!(filled_blocks % io_size) && left_vecs < io_size) in io_is_mergeable()
768 if (!page_is_mergeable(sbi, bio, last_blkaddr, cur_blkaddr)) in io_is_mergeable()
783 if (bio_add_page(bio, page, PAGE_SIZE, 0) != PAGE_SIZE) in add_bio_entry()
812 if (be->bio != *bio) in add_ipu_page()
820 if (f2fs_crypt_mergeable_bio(*bio, in add_ipu_page()
837 if (ret) { in add_ipu_page()
857 if (list_empty(head)) in f2fs_submit_merged_ipu_write()
862 if (target) in f2fs_submit_merged_ipu_write()
867 if (found) in f2fs_submit_merged_ipu_write()
872 if (!found) in f2fs_submit_merged_ipu_write()
879 if (target) in f2fs_submit_merged_ipu_write()
884 if (found) { in f2fs_submit_merged_ipu_write()
893 if (found) in f2fs_submit_merged_ipu_write()
895 if (bio && *bio) { in f2fs_submit_merged_ipu_write()
907 if (!f2fs_is_valid_blkaddr(fio->sbi, fio->new_blkaddr, in f2fs_merge_page_bio()
914 if (bio && !page_is_mergeable(fio->sbi, bio, *fio->last_block, in f2fs_merge_page_bio()
918 if (!bio) { in f2fs_merge_page_bio()
927 if (add_ipu_page(fio, &bio, page)) in f2fs_merge_page_bio()
931 if (fio->io_wbc) in f2fs_merge_page_bio()
953 if (fio->in_list) { in f2fs_submit_page_write()
955 if (list_empty(&io->io_list)) { in f2fs_submit_page_write()
967 if (fio->encrypted_page) in f2fs_submit_page_write()
969 else if (fio->compressed_page) in f2fs_submit_page_write()
979 if (io->bio && in f2fs_submit_page_write()
986 if (io->bio == NULL) { in f2fs_submit_page_write()
987 if (F2FS_IO_ALIGNED(sbi) && in f2fs_submit_page_write()
1000 if (bio_add_page(io->bio, bio_page, PAGE_SIZE, 0) < PAGE_SIZE) { in f2fs_submit_page_write()
1005 if (fio->io_wbc) in f2fs_submit_page_write()
1013 if (fio->in_list) in f2fs_submit_page_write()
1016 if (is_sbi_flag_set(sbi, SBI_IS_SHUTDOWN) || in f2fs_submit_page_write()
1039 if (!bio) in f2fs_grab_read_bio()
1048 if (fscrypt_inode_uses_fs_layer_crypto(inode)) in f2fs_grab_read_bio()
1050 if (f2fs_compressed_file(inode)) in f2fs_grab_read_bio()
1052 if (f2fs_need_verity(inode, first_idx)) in f2fs_grab_read_bio()
1055 if (post_read_steps) { in f2fs_grab_read_bio()
1069 if (bio->bi_private) in f2fs_release_read_bio()
1083 if (IS_ERR(bio)) in f2fs_submit_page_read()
1089 if (bio_add_page(bio, page, PAGE_SIZE, 0) < PAGE_SIZE) { in f2fs_submit_page_read()
1106 if (IS_INODE(dn->node_page) && f2fs_has_extra_attr(dn->inode)) in __set_data_blkaddr()
1124 if (set_page_dirty(dn->node_page)) in f2fs_set_data_blkaddr()
1141 if (!count) in f2fs_reserve_new_blocks()
1144 if (unlikely(is_inode_flag_set(dn->inode, FI_NO_ALLOC))) in f2fs_reserve_new_blocks()
1146 if (unlikely((err = inc_valid_block_count(sbi, dn->inode, &count)))) in f2fs_reserve_new_blocks()
1156 if (blkaddr == NULL_ADDR) { in f2fs_reserve_new_blocks()
1163 if (set_page_dirty(dn->node_page)) in f2fs_reserve_new_blocks()
1185 if (err) in f2fs_reserve_block()
1188 if (dn->data_blkaddr == NULL_ADDR) in f2fs_reserve_block()
1190 if (err || need_put) in f2fs_reserve_block()
1200 if (f2fs_lookup_extent_cache(inode, index, &ei)) { in f2fs_get_block()
1218 if (!page) in f2fs_get_read_data_page()
1221 if (f2fs_lookup_extent_cache(inode, index, &ei)) { in f2fs_get_read_data_page()
1223 if (!f2fs_is_valid_blkaddr(F2FS_I_SB(inode), dn.data_blkaddr, in f2fs_get_read_data_page()
1233 if (err) in f2fs_get_read_data_page()
1237 if (unlikely(dn.data_blkaddr == NULL_ADDR)) { in f2fs_get_read_data_page()
1241 if (dn.data_blkaddr != NEW_ADDR && in f2fs_get_read_data_page()
1249 if (PageUptodate(page)) { in f2fs_get_read_data_page()
1261 if (dn.data_blkaddr == NEW_ADDR) { in f2fs_get_read_data_page()
1263 if (!PageUptodate(page)) in f2fs_get_read_data_page()
1271 if (err) in f2fs_get_read_data_page()
1286 if (page && PageUptodate(page)) in f2fs_find_data_page()
1291 if (IS_ERR(page)) in f2fs_find_data_page()
1294 if (PageUptodate(page)) in f2fs_find_data_page()
1298 if (unlikely(!PageUptodate(page))) { in f2fs_find_data_page()
1306 * If it tries to access a hole, return an error.
1317 if (IS_ERR(page)) in f2fs_get_lock_data_page()
1322 if (unlikely(page->mapping != mapping)) { in f2fs_get_lock_data_page()
1326 if (unlikely(!PageUptodate(page))) { in f2fs_get_lock_data_page()
1339 * Note that, ipage is set only by make_empty_dir, and if any error occur,
1351 if (!page) { in f2fs_get_new_data_page()
1354 * if any error occur. in f2fs_get_new_data_page()
1362 if (err) { in f2fs_get_new_data_page()
1366 if (!ipage) in f2fs_get_new_data_page()
1369 if (PageUptodate(page)) in f2fs_get_new_data_page()
1372 if (dn.data_blkaddr == NEW_ADDR) { in f2fs_get_new_data_page()
1374 if (!PageUptodate(page)) in f2fs_get_new_data_page()
1379 /* if ipage exists, blkaddr should be NEW_ADDR */ in f2fs_get_new_data_page()
1382 if (IS_ERR(page)) in f2fs_get_new_data_page()
1386 if (new_i_size && i_size_read(inode) < in f2fs_get_new_data_page()
1401 if (unlikely(is_inode_flag_set(dn->inode, FI_NO_ALLOC))) in __allocate_data_block()
1405 if (err) in __allocate_data_block()
1409 if (dn->data_blkaddr != NULL_ADDR) in __allocate_data_block()
1412 if (unlikely((err = inc_valid_block_count(sbi, dn->inode, &count)))) in __allocate_data_block()
1420 if (GET_SEGNO(sbi, old_blkaddr) != NULL_SEGNO) in __allocate_data_block()
1442 if (map.m_len > map.m_lblk) in f2fs_preallocate_blocks()
1452 if (direct_io) { in f2fs_preallocate_blocks()
1459 if (iocb->ki_pos + iov_iter_count(from) > MAX_INLINE_DATA(inode)) { in f2fs_preallocate_blocks()
1461 if (err) in f2fs_preallocate_blocks()
1464 if (f2fs_has_inline_data(inode)) in f2fs_preallocate_blocks()
1471 if (map.m_len > 0 && err == -ENOSPC) { in f2fs_preallocate_blocks()
1472 if (!direct_io) in f2fs_preallocate_blocks()
1481 if (flag == F2FS_GET_BLOCK_PRE_AIO) { in f2fs_do_map_lock()
1482 if (lock) in f2fs_do_map_lock()
1487 if (lock) in f2fs_do_map_lock()
1514 if (!maxblocks) in f2fs_map_blocks()
1524 if (!create && f2fs_lookup_extent_cache(inode, pgofs, &ei)) { in f2fs_map_blocks()
1525 if (f2fs_lfs_mode(sbi) && flag == F2FS_GET_BLOCK_DIO && in f2fs_map_blocks()
1532 if (map->m_next_extent) in f2fs_map_blocks()
1536 if (flag == F2FS_GET_BLOCK_DIO) in f2fs_map_blocks()
1543 if (map->m_may_create) in f2fs_map_blocks()
1549 if (err) { in f2fs_map_blocks()
1550 if (flag == F2FS_GET_BLOCK_BMAP) in f2fs_map_blocks()
1552 if (err == -ENOENT) { in f2fs_map_blocks()
1554 if (map->m_next_pgofs) in f2fs_map_blocks()
1557 if (map->m_next_extent) in f2fs_map_blocks()
1572 if (__is_valid_data_blkaddr(blkaddr) && in f2fs_map_blocks()
1578 if (__is_valid_data_blkaddr(blkaddr)) { in f2fs_map_blocks()
1580 if (f2fs_lfs_mode(sbi) && flag == F2FS_GET_BLOCK_DIO && in f2fs_map_blocks()
1583 if (err) in f2fs_map_blocks()
1589 if (create) { in f2fs_map_blocks()
1590 if (unlikely(f2fs_cp_error(sbi))) { in f2fs_map_blocks()
1594 if (flag == F2FS_GET_BLOCK_PRE_AIO) { in f2fs_map_blocks()
1595 if (blkaddr == NULL_ADDR) { in f2fs_map_blocks()
1604 if (!err) in f2fs_map_blocks()
1607 if (err) in f2fs_map_blocks()
1612 if (flag == F2FS_GET_BLOCK_BMAP) { in f2fs_map_blocks()
1616 if (flag == F2FS_GET_BLOCK_PRECACHE) in f2fs_map_blocks()
1618 if (flag == F2FS_GET_BLOCK_FIEMAP && in f2fs_map_blocks()
1620 if (map->m_next_pgofs) in f2fs_map_blocks()
1624 if (flag != F2FS_GET_BLOCK_FIEMAP) { in f2fs_map_blocks()
1626 if (map->m_next_pgofs) in f2fs_map_blocks()
1633 if (flag == F2FS_GET_BLOCK_PRE_AIO) in f2fs_map_blocks()
1636 if (map->m_len == 0) { in f2fs_map_blocks()
1638 if (blkaddr == NEW_ADDR) in f2fs_map_blocks()
1644 } else if ((map->m_pblk != NEW_ADDR && in f2fs_map_blocks()
1659 if (flag == F2FS_GET_BLOCK_PRE_AIO && in f2fs_map_blocks()
1664 if (err) in f2fs_map_blocks()
1668 if (prealloc && dn.ofs_in_node != last_ofs_in_node + 1) { in f2fs_map_blocks()
1675 if (pgofs >= end) in f2fs_map_blocks()
1677 else if (dn.ofs_in_node < end_offset) in f2fs_map_blocks()
1680 if (flag == F2FS_GET_BLOCK_PRECACHE) { in f2fs_map_blocks()
1681 if (map->m_flags & F2FS_MAP_MAPPED) { in f2fs_map_blocks()
1692 if (map->m_may_create) { in f2fs_map_blocks()
1701 if (flag == F2FS_GET_BLOCK_DIO && map->m_flags & F2FS_MAP_MAPPED) in f2fs_map_blocks()
1705 if (flag == F2FS_GET_BLOCK_PRECACHE) { in f2fs_map_blocks()
1706 if (map->m_flags & F2FS_MAP_MAPPED) { in f2fs_map_blocks()
1713 if (map->m_next_extent) in f2fs_map_blocks()
1718 if (map->m_may_create) { in f2fs_map_blocks()
1733 if (pos + len > i_size_read(inode)) in f2fs_overwrite_io()
1746 if (err || map.m_len == 0) in f2fs_overwrite_io()
1768 if (!err) { in __get_data_block()
1832 if (f2fs_has_inline_xattr(inode)) { in f2fs_xattr_fiemap()
1837 if (!page) in f2fs_xattr_fiemap()
1841 if (err) { in f2fs_xattr_fiemap()
1858 if (!xnid) in f2fs_xattr_fiemap()
1863 if (err || err == 1) in f2fs_xattr_fiemap()
1867 if (xnid) { in f2fs_xattr_fiemap()
1869 if (!page) in f2fs_xattr_fiemap()
1873 if (err) { in f2fs_xattr_fiemap()
1886 if (phys) { in f2fs_xattr_fiemap()
1925 if (fieinfo->fi_flags & FIEMAP_FLAG_CACHE) { in f2fs_fiemap()
1927 if (ret) in f2fs_fiemap()
1932 if (ret) in f2fs_fiemap()
1937 if (fieinfo->fi_flags & FIEMAP_FLAG_XATTR) { in f2fs_fiemap()
1942 if (f2fs_has_inline_data(inode) || f2fs_has_inline_dentry(inode)) { in f2fs_fiemap()
1944 if (ret != -EAGAIN) in f2fs_fiemap()
1948 if (logical_to_blk(inode, len) == 0) in f2fs_fiemap()
1958 if (compr_cluster) in f2fs_fiemap()
1963 if (ret) in f2fs_fiemap()
1967 if (!buffer_mapped(&map_bh)) { in f2fs_fiemap()
1970 if (blk_to_logical(inode, start_blk) < blk_to_logical(inode, in f2fs_fiemap()
1977 if (size) { in f2fs_fiemap()
1978 if (IS_ENCRYPTED(inode)) in f2fs_fiemap()
1984 if (ret) in f2fs_fiemap()
1989 if (start_blk > last_blk) in f2fs_fiemap()
1992 if (compr_cluster) { in f2fs_fiemap()
2004 if (start_blk > last_blk) in f2fs_fiemap()
2010 if (map_bh.b_blocknr == COMPRESS_ADDR) { in f2fs_fiemap()
2020 if (buffer_unwritten(&map_bh)) in f2fs_fiemap()
2027 if (fatal_signal_pending(current)) in f2fs_fiemap()
2032 if (ret == 1) in f2fs_fiemap()
2041 if (IS_ENABLED(CONFIG_FS_VERITY) && in f2fs_readpage_limit()
2068 if (last_block > last_block_in_file) in f2fs_read_single_page()
2072 if (block_in_file >= last_block) in f2fs_read_single_page()
2077 if ((map->m_flags & F2FS_MAP_MAPPED) && in f2fs_read_single_page()
2090 if (ret) in f2fs_read_single_page()
2093 if ((map->m_flags & F2FS_MAP_MAPPED)) { in f2fs_read_single_page()
2097 if (!PageUptodate(page) && (!PageSwapCache(page) && in f2fs_read_single_page()
2103 if (!f2fs_is_valid_blkaddr(F2FS_I_SB(inode), block_nr, in f2fs_read_single_page()
2111 if (f2fs_need_verity(inode, page->index) && in f2fs_read_single_page()
2116 if (!PageUptodate(page)) in f2fs_read_single_page()
2126 if (bio && (!page_is_mergeable(F2FS_I_SB(inode), bio, in f2fs_read_single_page()
2133 if (bio == NULL) { in f2fs_read_single_page()
2137 if (IS_ERR(bio)) { in f2fs_read_single_page()
2145 * If the page is under writeback, we need to wait for in f2fs_read_single_page()
2150 if (bio_add_page(bio, page, blocksize, 0) < blocksize) in f2fs_read_single_page()
2159 if (bio) { in f2fs_read_single_page()
2195 if (!page) in f2fs_read_multi_pages()
2197 if ((sector_t)page->index >= last_block_in_file) { in f2fs_read_multi_pages()
2199 if (!PageUptodate(page)) in f2fs_read_multi_pages()
2201 } else if (!PageUptodate(page)) { in f2fs_read_multi_pages()
2210 if (f2fs_cluster_is_empty(cc)) in f2fs_read_multi_pages()
2215 if (ret) in f2fs_read_multi_pages()
2226 if (!__is_valid_data_blkaddr(blkaddr)) in f2fs_read_multi_pages()
2229 if (!f2fs_is_valid_blkaddr(sbi, blkaddr, DATA_GENERIC)) { in f2fs_read_multi_pages()
2237 if (cc->nr_cpages == 0) { in f2fs_read_multi_pages()
2243 if (IS_ERR(dic)) { in f2fs_read_multi_pages()
2256 if (bio && (!page_is_mergeable(sbi, bio, in f2fs_read_multi_pages()
2264 if (!bio) { in f2fs_read_multi_pages()
2268 if (IS_ERR(bio)) { in f2fs_read_multi_pages()
2271 if (!atomic_sub_return(dic->nr_cpages - i, in f2fs_read_multi_pages()
2286 if (bio_add_page(bio, page, blocksize, 0) < blocksize) in f2fs_read_multi_pages()
2291 if (!(ctx->enabled_steps & (1 << STEP_DECOMPRESS))) in f2fs_read_multi_pages()
2319 * Note that the aops->readpages() function is ONLY used for read-ahead. If
2360 if (rac && readahead_count(rac)) { in f2fs_mpage_readpages()
2361 if (READ_ONCE(F2FS_I(inode)->ra_offset) == readahead_index(rac)) in f2fs_mpage_readpages()
2369 if (rac) { in f2fs_mpage_readpages()
2372 if (drop_ra) { in f2fs_mpage_readpages()
2379 if (f2fs_compressed_file(inode)) { in f2fs_mpage_readpages()
2381 if (!f2fs_cluster_can_merge_page(&cc, page->index)) { in f2fs_mpage_readpages()
2387 if (ret) in f2fs_mpage_readpages()
2391 if (ret < 0) in f2fs_mpage_readpages()
2393 else if (!ret) in f2fs_mpage_readpages()
2397 if (ret) in f2fs_mpage_readpages()
2409 if (ret) { in f2fs_mpage_readpages()
2420 if (rac) in f2fs_mpage_readpages()
2424 if (f2fs_compressed_file(inode)) { in f2fs_mpage_readpages()
2426 if (nr_pages == 1 && !f2fs_cluster_is_empty(&cc)) { in f2fs_mpage_readpages()
2436 if (bio) in f2fs_mpage_readpages()
2439 if (rac && readahead_count(rac) && !drop_ra) in f2fs_mpage_readpages()
2451 if (!f2fs_is_compress_backend_ready(inode)) { in f2fs_read_data_page()
2456 /* If the file has inline data, try to read it directly */ in f2fs_read_data_page()
2457 if (f2fs_has_inline_data(inode)) in f2fs_read_data_page()
2459 if (ret == -EAGAIN) in f2fs_read_data_page()
2470 if (!f2fs_is_compress_backend_ready(inode)) in f2fs_readahead()
2473 /* If the file has inline data, skip readpages */ in f2fs_readahead()
2474 if (f2fs_has_inline_data(inode)) in f2fs_readahead()
2486 if (!f2fs_encrypted_file(inode)) in f2fs_encrypt_one_page()
2494 if (fscrypt_inode_uses_inline_crypto(inode)) in f2fs_encrypt_one_page()
2500 if (IS_ERR(fio->encrypted_page)) { in f2fs_encrypt_one_page()
2502 if (PTR_ERR(fio->encrypted_page) == -ENOMEM) { in f2fs_encrypt_one_page()
2512 if (mpage) { in f2fs_encrypt_one_page()
2513 if (PageUptodate(mpage)) in f2fs_encrypt_one_page()
2527 if (policy & (0x1 << F2FS_IPU_FORCE)) in check_inplace_update_policy()
2529 if (policy & (0x1 << F2FS_IPU_SSR) && f2fs_need_SSR(sbi)) in check_inplace_update_policy()
2531 if (policy & (0x1 << F2FS_IPU_UTIL) && in check_inplace_update_policy()
2534 if (policy & (0x1 << F2FS_IPU_SSR_UTIL) && f2fs_need_SSR(sbi) && in check_inplace_update_policy()
2541 if (policy & (0x1 << F2FS_IPU_ASYNC) && in check_inplace_update_policy()
2548 if (policy & (0x1 << F2FS_IPU_FSYNC) && in check_inplace_update_policy()
2552 if (unlikely(fio && is_sbi_flag_set(sbi, SBI_CP_DISABLED) && in check_inplace_update_policy()
2561 if (f2fs_is_pinned_file(inode)) in f2fs_should_update_inplace()
2564 /* if this is cold file, we should overwrite to avoid fragmentation */ in f2fs_should_update_inplace()
2565 if (file_is_cold(inode)) in f2fs_should_update_inplace()
2575 if (f2fs_lfs_mode(sbi)) in f2fs_should_update_outplace()
2577 if (S_ISDIR(inode->i_mode)) in f2fs_should_update_outplace()
2579 if (IS_NOQUOTA(inode)) in f2fs_should_update_outplace()
2581 if (f2fs_is_atomic_file(inode)) in f2fs_should_update_outplace()
2583 if (fio) { in f2fs_should_update_outplace()
2584 if (is_cold_data(fio->page)) in f2fs_should_update_outplace()
2586 if (IS_ATOMIC_WRITTEN_PAGE(fio->page)) in f2fs_should_update_outplace()
2588 if (unlikely(is_sbi_flag_set(sbi, SBI_CP_DISABLED) && in f2fs_should_update_outplace()
2599 if (f2fs_should_update_outplace(inode, fio)) in need_inplace_update()
2616 if (need_inplace_update(fio) && in f2fs_do_write_data_page()
2620 if (!f2fs_is_valid_blkaddr(fio->sbi, fio->old_blkaddr, in f2fs_do_write_data_page()
2630 if (fio->need_lock == LOCK_REQ && !f2fs_trylock_op(fio->sbi)) in f2fs_do_write_data_page()
2634 if (err) in f2fs_do_write_data_page()
2640 if (fio->old_blkaddr == NULL_ADDR) { in f2fs_do_write_data_page()
2646 if (__is_valid_data_blkaddr(fio->old_blkaddr) && in f2fs_do_write_data_page()
2653 * If current allocation needs SSR, in f2fs_do_write_data_page()
2656 if (ipu_force || in f2fs_do_write_data_page()
2660 if (err) in f2fs_do_write_data_page()
2666 if (fio->need_lock == LOCK_REQ) in f2fs_do_write_data_page()
2669 if (err) { in f2fs_do_write_data_page()
2670 if (fscrypt_inode_uses_fs_layer_crypto(inode)) in f2fs_do_write_data_page()
2672 if (PageWriteback(page)) in f2fs_do_write_data_page()
2681 if (fio->need_lock == LOCK_RETRY) { in f2fs_do_write_data_page()
2682 if (!f2fs_trylock_op(fio->sbi)) { in f2fs_do_write_data_page()
2690 if (err) in f2fs_do_write_data_page()
2696 if (err) in f2fs_do_write_data_page()
2702 if (fio->compr_blocks && fio->old_blkaddr == COMPRESS_ADDR) in f2fs_do_write_data_page()
2709 if (page->index == 0) in f2fs_do_write_data_page()
2714 if (fio->need_lock == LOCK_REQ) in f2fs_do_write_data_page()
2756 if (unlikely(f2fs_cp_error(sbi))) { in f2fs_write_single_data_page()
2762 if (S_ISDIR(inode->i_mode)) in f2fs_write_single_data_page()
2767 if (unlikely(is_sbi_flag_set(sbi, SBI_POR_DOING))) in f2fs_write_single_data_page()
2770 if (page->index < end_index || in f2fs_write_single_data_page()
2776 * If the offset is out-of-range of file size, in f2fs_write_single_data_page()
2780 if ((page->index >= end_index + 1) || !offset) in f2fs_write_single_data_page()
2785 if (f2fs_is_drop_cache(inode)) in f2fs_write_single_data_page()
2788 if (f2fs_is_volatile_file(inode) && (!page->index || in f2fs_write_single_data_page()
2794 if (S_ISDIR(inode->i_mode) || IS_NOQUOTA(inode)) { in f2fs_write_single_data_page()
2800 if (IS_NOQUOTA(inode)) in f2fs_write_single_data_page()
2806 if (IS_NOQUOTA(inode)) in f2fs_write_single_data_page()
2812 if (!wbc->for_reclaim) in f2fs_write_single_data_page()
2814 else if (has_not_enough_free_secs(sbi, 0, 0)) in f2fs_write_single_data_page()
2820 if (f2fs_has_inline_data(inode)) { in f2fs_write_single_data_page()
2822 if (!err) in f2fs_write_single_data_page()
2826 if (err == -EAGAIN) { in f2fs_write_single_data_page()
2828 if (err == -EAGAIN) { in f2fs_write_single_data_page()
2834 if (err) { in f2fs_write_single_data_page()
2838 if (F2FS_I(inode)->last_disk_size < psize) in f2fs_write_single_data_page()
2844 if (err && err != -ENOENT) in f2fs_write_single_data_page()
2849 if (err) { in f2fs_write_single_data_page()
2854 if (wbc->for_reclaim) { in f2fs_write_single_data_page()
2861 if (!S_ISDIR(inode->i_mode) && !IS_NOQUOTA(inode) && in f2fs_write_single_data_page()
2865 if (unlikely(f2fs_cp_error(sbi))) { in f2fs_write_single_data_page()
2871 if (submitted) in f2fs_write_single_data_page()
2884 if (!err || wbc->for_reclaim) in f2fs_write_single_data_page()
2896 if (unlikely(f2fs_cp_error(F2FS_I_SB(inode)))) in f2fs_write_data_page()
2899 if (f2fs_compressed_file(inode)) { in f2fs_write_data_page()
2900 if (f2fs_is_compressed_cluster(inode, page->index)) { in f2fs_write_data_page()
2955 if (get_dirty_pages(mapping->host) <= in f2fs_write_cache_pages()
2961 if (wbc->range_cyclic) { in f2fs_write_cache_pages()
2967 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) in f2fs_write_cache_pages()
2970 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in f2fs_write_cache_pages()
2976 if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) in f2fs_write_cache_pages()
2982 if (nr_pages == 0) in f2fs_write_cache_pages()
2991 if (f2fs_compressed_file(inode)) { in f2fs_write_cache_pages()
2993 if (ret) { in f2fs_write_cache_pages()
2998 if (!f2fs_cluster_can_merge_page(&cc, in f2fs_write_cache_pages()
3002 if (!ret) in f2fs_write_cache_pages()
3007 if (unlikely(f2fs_cp_error(sbi))) in f2fs_write_cache_pages()
3010 if (f2fs_cluster_is_empty(&cc)) { in f2fs_write_cache_pages()
3018 if (ret2 < 0) { in f2fs_write_cache_pages()
3022 } else if (ret2 && in f2fs_write_cache_pages()
3035 if (atomic_read(&sbi->wb_sync_req[DATA]) && in f2fs_write_cache_pages()
3047 if (unlikely(page->mapping != mapping)) { in f2fs_write_cache_pages()
3053 if (!PageDirty(page)) { in f2fs_write_cache_pages()
3058 if (PageWriteback(page)) { in f2fs_write_cache_pages()
3059 if (wbc->sync_mode != WB_SYNC_NONE) in f2fs_write_cache_pages()
3066 if (!clear_page_dirty_for_io(page)) in f2fs_write_cache_pages()
3070 if (f2fs_compressed_file(inode)) { in f2fs_write_cache_pages()
3078 if (ret == AOP_WRITEPAGE_ACTIVATE) in f2fs_write_cache_pages()
3086 if (unlikely(ret)) { in f2fs_write_cache_pages()
3091 if (ret == AOP_WRITEPAGE_ACTIVATE) { in f2fs_write_cache_pages()
3094 } else if (ret == -EAGAIN) { in f2fs_write_cache_pages()
3096 if (wbc->sync_mode == WB_SYNC_ALL) { in f2fs_write_cache_pages()
3109 if (wbc->nr_to_write <= 0 && in f2fs_write_cache_pages()
3115 if (need_readd) in f2fs_write_cache_pages()
3123 if (f2fs_compressed_file(inode) && !f2fs_cluster_is_empty(&cc)) { in f2fs_write_cache_pages()
3127 if (ret) { in f2fs_write_cache_pages()
3132 if (f2fs_compressed_file(inode)) in f2fs_write_cache_pages()
3135 if (retry) { in f2fs_write_cache_pages()
3140 if (wbc->range_cyclic && !done) in f2fs_write_cache_pages()
3142 if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0)) in f2fs_write_cache_pages()
3145 if (nwritten) in f2fs_write_cache_pages()
3149 if (bio) in f2fs_write_cache_pages()
3159 if (F2FS_I(inode)->cp_task) in __should_serialize_io()
3162 if (!S_ISREG(inode->i_mode)) in __should_serialize_io()
3164 if (IS_NOQUOTA(inode)) in __should_serialize_io()
3167 if (f2fs_compressed_file(inode)) in __should_serialize_io()
3169 if (wbc->sync_mode != WB_SYNC_ALL) in __should_serialize_io()
3171 if (get_dirty_pages(inode) >= SM_I(F2FS_I_SB(inode))->min_seq_blocks) in __should_serialize_io()
3187 if (!mapping->a_ops->writepage) in __f2fs_write_data_pages()
3190 /* skip writing if there is no dirty page in this inode */ in __f2fs_write_data_pages()
3191 if (!get_dirty_pages(inode) && wbc->sync_mode == WB_SYNC_NONE) in __f2fs_write_data_pages()
3195 if (unlikely(is_sbi_flag_set(sbi, SBI_POR_DOING))) in __f2fs_write_data_pages()
3198 if ((S_ISDIR(inode->i_mode) || IS_NOQUOTA(inode)) && in __f2fs_write_data_pages()
3205 if (is_inode_flag_set(inode, FI_DO_DEFRAG)) in __f2fs_write_data_pages()
3211 if (wbc->sync_mode == WB_SYNC_ALL) in __f2fs_write_data_pages()
3213 else if (atomic_read(&sbi->wb_sync_req[DATA])) in __f2fs_write_data_pages()
3216 if (__should_serialize_io(inode, wbc)) { in __f2fs_write_data_pages()
3225 if (locked) in __f2fs_write_data_pages()
3228 if (wbc->sync_mode == WB_SYNC_ALL) in __f2fs_write_data_pages()
3231 * if some pages were truncated, we cannot guarantee its mapping->host in __f2fs_write_data_pages()
3259 if (IS_NOQUOTA(inode)) in f2fs_write_failed()
3263 if (to > i_size && !f2fs_verity_in_progress(inode)) { in f2fs_write_failed()
3292 if (!f2fs_has_inline_data(inode) && len == PAGE_SIZE && in prepare_write_begin()
3298 if (f2fs_has_inline_data(inode) && pos + len > MAX_INLINE_DATA(inode)) in prepare_write_begin()
3303 if (f2fs_has_inline_data(inode) || in prepare_write_begin()
3312 if (IS_ERR(ipage)) { in prepare_write_begin()
3319 if (f2fs_has_inline_data(inode)) { in prepare_write_begin()
3320 if (pos + len <= MAX_INLINE_DATA(inode)) { in prepare_write_begin()
3323 if (inode->i_nlink) in prepare_write_begin()
3327 if (err) in prepare_write_begin()
3329 if (dn.data_blkaddr == NULL_ADDR) in prepare_write_begin()
3332 } else if (locked) { in prepare_write_begin()
3335 if (f2fs_lookup_extent_cache(inode, index, &ei)) { in prepare_write_begin()
3340 if (err || dn.data_blkaddr == NULL_ADDR) { in prepare_write_begin()
3357 if (locked) in prepare_write_begin()
3376 if (!f2fs_is_checkpoint_ready(sbi)) { in f2fs_write_begin()
3381 if ((f2fs_is_atomic_file(inode) && in f2fs_write_begin()
3394 if (index != 0) { in f2fs_write_begin()
3396 if (err) in f2fs_write_begin()
3401 if (f2fs_compressed_file(inode)) { in f2fs_write_begin()
3408 if (ret < 0) { in f2fs_write_begin()
3411 } else if (ret) { in f2fs_write_begin()
3424 if (!page) { in f2fs_write_begin()
3435 if (err) in f2fs_write_begin()
3438 if (need_balance && !IS_NOQUOTA(inode) && in f2fs_write_begin()
3443 if (page->mapping != mapping) { in f2fs_write_begin()
3452 if (len == PAGE_SIZE || PageUptodate(page)) in f2fs_write_begin()
3455 if (!(pos & (PAGE_SIZE - 1)) && (pos + len) >= i_size_read(inode) && in f2fs_write_begin()
3461 if (blkaddr == NEW_ADDR) { in f2fs_write_begin()
3465 if (!f2fs_is_valid_blkaddr(sbi, blkaddr, in f2fs_write_begin()
3471 if (err) in f2fs_write_begin()
3475 if (unlikely(page->mapping != mapping)) { in f2fs_write_begin()
3479 if (unlikely(!PageUptodate(page))) { in f2fs_write_begin()
3489 if (drop_atomic) in f2fs_write_begin()
3508 if (!PageUptodate(page)) { in f2fs_write_end()
3509 if (unlikely(copied != len)) in f2fs_write_end()
3517 if (f2fs_compressed_file(inode) && fsdata) { in f2fs_write_end()
3521 if (pos + copied > i_size_read(inode) && in f2fs_write_end()
3528 if (!copied) in f2fs_write_end()
3533 if (pos + copied > i_size_read(inode) && in f2fs_write_end()
3551 if (iov_iter_rw(iter) == READ && offset >= i_size_read(inode)) in check_direct_IO()
3554 if (align & blocksize_mask) { in check_direct_IO()
3555 if (bdev) in check_direct_IO()
3558 if (align & blocksize_mask) in check_direct_IO()
3588 if (!dio) in f2fs_dio_submit_bio()
3624 if (err) in f2fs_direct_IO()
3627 if (f2fs_force_buffered_io(inode, iocb, iter)) in f2fs_direct_IO()
3634 if (rw == WRITE && whint_mode == WHINT_MODE_OFF) in f2fs_direct_IO()
3637 if (iocb->ki_flags & IOCB_NOWAIT) { in f2fs_direct_IO()
3638 if (!down_read_trylock(&fi->i_gc_rwsem[rw])) { in f2fs_direct_IO()
3643 if (do_opu && !down_read_trylock(&fi->i_gc_rwsem[READ])) { in f2fs_direct_IO()
3651 if (do_opu) in f2fs_direct_IO()
3661 if (do_opu) in f2fs_direct_IO()
3666 if (rw == WRITE) { in f2fs_direct_IO()
3667 if (whint_mode == WHINT_MODE_OFF) in f2fs_direct_IO()
3669 if (err > 0) { in f2fs_direct_IO()
3672 if (!do_opu) in f2fs_direct_IO()
3674 } else if (err == -EIOCBQUEUED) { in f2fs_direct_IO()
3677 } else if (err < 0) { in f2fs_direct_IO()
3681 if (err > 0) in f2fs_direct_IO()
3683 else if (err == -EIOCBQUEUED) in f2fs_direct_IO()
3700 if (inode->i_ino >= F2FS_ROOT_INO(sbi) && in f2fs_invalidate_page()
3704 if (PageDirty(page)) { in f2fs_invalidate_page()
3705 if (inode->i_ino == F2FS_META_INO(sbi)) { in f2fs_invalidate_page()
3707 } else if (inode->i_ino == F2FS_NODE_INO(sbi)) { in f2fs_invalidate_page()
3717 if (IS_ATOMIC_WRITTEN_PAGE(page)) in f2fs_invalidate_page()
3725 /* If this is dirty page, keep PagePrivate */ in f2fs_release_page()
3726 if (PageDirty(page)) in f2fs_release_page()
3730 if (IS_ATOMIC_WRITTEN_PAGE(page)) in f2fs_release_page()
3744 if (!PageUptodate(page)) in f2fs_set_data_page_dirty()
3746 if (PageSwapCache(page)) in f2fs_set_data_page_dirty()
3749 if (f2fs_is_atomic_file(inode) && !f2fs_is_commit_atomic_write(inode)) { in f2fs_set_data_page_dirty()
3750 if (!IS_ATOMIC_WRITTEN_PAGE(page)) { in f2fs_set_data_page_dirty()
3761 if (!PageDirty(page)) { in f2fs_set_data_page_dirty()
3781 if (ret) in f2fs_bmap_compress()
3784 if (dn.data_blkaddr != COMPRESS_ADDR) { in f2fs_bmap_compress()
3787 if (!__is_valid_data_blkaddr(blknr)) in f2fs_bmap_compress()
3807 if (f2fs_has_inline_data(inode)) in f2fs_bmap()
3811 if (mapping_tagged(mapping, PAGECACHE_TAG_DIRTY)) in f2fs_bmap()
3815 if (unlikely(block >= F2FS_I_SB(inode)->max_file_blocks)) in f2fs_bmap()
3818 if (f2fs_compressed_file(inode)) { in f2fs_bmap()
3821 if (!get_data_block_bmap(inode, block, &tmp, 0)) in f2fs_bmap()
3842 if (atomic_written) { in f2fs_migrate_page()
3843 if (mode != MIGRATE_SYNC) in f2fs_migrate_page()
3845 if (!mutex_trylock(&fi->inmem_lock)) in f2fs_migrate_page()
3853 if (rc != MIGRATEPAGE_SUCCESS) { in f2fs_migrate_page()
3854 if (atomic_written) in f2fs_migrate_page()
3859 if (atomic_written) { in f2fs_migrate_page()
3862 if (cur->page == page) { in f2fs_migrate_page()
3871 if (PagePrivate(page)) { in f2fs_migrate_page()
3876 if (mode != MIGRATE_SYNC_NO_COPY) in f2fs_migrate_page()
3920 if (ret) in check_swap_activate_fast()
3924 if (!buffer_mapped(&map_bh)) in check_swap_activate_fast()
3930 if (cur_lblock + nr_pblocks >= sis->max) in check_swap_activate_fast()
3933 if (cur_lblock) { /* exclude the header page */ in check_swap_activate_fast()
3934 if (pblock < lowest_pblock) in check_swap_activate_fast()
3936 if (pblock + nr_pblocks - 1 > highest_pblock) in check_swap_activate_fast()
3944 if (ret < 0) in check_swap_activate_fast()
3951 if (cur_lblock == 0) in check_swap_activate_fast()
3979 if (PAGE_SIZE == F2FS_BLKSIZE) in check_swap_activate()
4003 if (err || !block) in check_swap_activate()
4010 if (first_block & (blocks_per_page - 1)) { in check_swap_activate()
4021 if (err || !block) in check_swap_activate()
4024 if (block != first_block + block_in_page) { in check_swap_activate()
4032 if (page_no) { /* exclude the header page */ in check_swap_activate()
4033 if (first_block < lowest_block) in check_swap_activate()
4035 if (first_block > highest_block) in check_swap_activate()
4043 if (ret < 0) in check_swap_activate()
4053 if (page_no == 0) in check_swap_activate()
4071 if (!S_ISREG(inode->i_mode)) in f2fs_swap_activate()
4074 if (f2fs_readonly(F2FS_I_SB(inode)->sb)) in f2fs_swap_activate()
4078 if (ret) in f2fs_swap_activate()
4081 if (!f2fs_disable_compressed_file(inode)) in f2fs_swap_activate()
4085 if (ret < 0) in f2fs_swap_activate()
4147 if (!bio_post_read_ctx_cache) in f2fs_init_post_read_processing()
4152 if (!bio_post_read_ctx_pool) in f2fs_init_post_read_processing()
4170 if (!f2fs_sb_has_encrypt(sbi) && in f2fs_init_post_read_wq()
4178 if (!sbi->post_read_wq) in f2fs_init_post_read_wq()
4185 if (sbi->post_read_wq) in f2fs_destroy_post_read_wq()
4193 if (!bio_entry_slab) in f2fs_init_bio_entry_cache()