| /linux/fs/f2fs/ |
| H A D | inode.c | 71 static void __get_inode_rdev(struct inode *inode, struct folio *node_folio) in __get_inode_rdev() argument 73 __le32 *addr = get_dnode_addr(inode, node_folio); in __get_inode_rdev() 84 static void __set_inode_rdev(struct inode *inode, struct folio *node_folio) in __set_inode_rdev() argument 86 __le32 *addr = get_dnode_addr(inode, node_folio); in __set_inode_rdev() 270 static bool sanity_check_inode(struct inode *inode, struct folio *node_folio) in sanity_check_inode() argument 274 struct f2fs_inode *ri = F2FS_INODE(node_folio); in sanity_check_inode() 277 iblocks = le64_to_cpu(F2FS_INODE(node_folio)->i_blocks); in sanity_check_inode() 284 if (ino_of_node(node_folio) != nid_of_node(node_folio)) { in sanity_check_inode() 287 ino_of_node(node_folio), nid_of_node(node_folio)); in sanity_check_inode() 291 if (ino_of_node(node_folio) == fi->i_xattr_nid) { in sanity_check_inode() [all …]
|
| H A D | node.h | 239 static inline nid_t ino_of_node(const struct folio *node_folio) in ino_of_node() argument 241 struct f2fs_node *rn = F2FS_NODE(node_folio); in ino_of_node() 245 static inline nid_t nid_of_node(const struct folio *node_folio) in nid_of_node() argument 247 struct f2fs_node *rn = F2FS_NODE(node_folio); in nid_of_node() 251 static inline unsigned int ofs_of_node(const struct folio *node_folio) in ofs_of_node() argument 253 struct f2fs_node *rn = F2FS_NODE(node_folio); in ofs_of_node() 258 static inline __u64 cpver_of_node(const struct folio *node_folio) in cpver_of_node() argument 260 struct f2fs_node *rn = F2FS_NODE(node_folio); in cpver_of_node() 264 static inline block_t next_blkaddr_of_node(const struct folio *node_folio) in next_blkaddr_of_node() argument 266 struct f2fs_node *rn = F2FS_NODE(node_folio); in next_blkaddr_of_node() [all …]
|
| H A D | recovery.c | 500 struct folio *sum_folio, *node_folio; in check_index_in_prev_nodes() local 533 max_addrs = ADDRS_PER_PAGE(dn->node_folio, dn->inode); in check_index_in_prev_nodes() 545 tdn.node_folio = dn->inode_folio; in check_index_in_prev_nodes() 554 node_folio = f2fs_get_node_folio(sbi, nid, NODE_TYPE_REGULAR); in check_index_in_prev_nodes() 555 if (IS_ERR(node_folio)) in check_index_in_prev_nodes() 556 return PTR_ERR(node_folio); in check_index_in_prev_nodes() 558 offset = ofs_of_node(node_folio); in check_index_in_prev_nodes() 559 ino = ino_of_node(node_folio); in check_index_in_prev_nodes() 560 f2fs_folio_put(node_folio, true); in check_index_in_prev_nodes() 668 f2fs_folio_wait_writeback(dn.node_folio, NODE, true, true); in do_recover_data() [all …]
|
| H A D | gc.c | 1059 struct folio *node_folio; in gc_node_segment() local 1082 node_folio = f2fs_get_node_folio(sbi, nid, NODE_TYPE_REGULAR); in gc_node_segment() 1083 if (IS_ERR(node_folio)) in gc_node_segment() 1088 f2fs_folio_put(node_folio, true); in gc_node_segment() 1093 f2fs_folio_put(node_folio, true); in gc_node_segment() 1098 f2fs_folio_put(node_folio, true); in gc_node_segment() 1102 err = f2fs_move_node_folio(node_folio, gc_type); in gc_node_segment() 1151 struct folio *node_folio; in is_alive() local 1159 node_folio = f2fs_get_node_folio(sbi, nid, NODE_TYPE_REGULAR); in is_alive() 1160 if (IS_ERR(node_folio)) in is_alive() [all …]
|
| H A D | node.c | 905 dn->node_folio = nfolio[level]; in f2fs_get_dnode_of_data() 926 blkaddr = data_blkaddr(dn->inode, dn->node_folio, ofs_in_node); in f2fs_get_dnode_of_data() 928 blkaddr = data_blkaddr(dn->inode, dn->node_folio, in f2fs_get_dnode_of_data() 943 dn->node_folio = NULL; in f2fs_get_dnode_of_data() 984 clear_node_folio_dirty(dn->node_folio); in truncate_node() 987 index = dn->node_folio->index; in truncate_node() 988 f2fs_folio_put(dn->node_folio, true); in truncate_node() 993 dn->node_folio = NULL; in truncate_node() 1025 dn->node_folio = folio; in truncate_dnode() 1098 dn->node_folio = folio; in truncate_nodes() [all …]
|
| H A D | f2fs.h | 1113 struct folio *node_folio; /* cached direct node folio */ member 1129 dn->node_folio = nfolio; in set_new_dnode() 3089 if (dn->node_folio) in f2fs_put_dnode() 3090 f2fs_folio_put(dn->node_folio, true); in f2fs_put_dnode() 3091 if (dn->inode_folio && dn->node_folio != dn->inode_folio) in f2fs_put_dnode() 3093 dn->node_folio = NULL; in f2fs_put_dnode() 3205 struct folio *node_folio) in get_dnode_base() argument 3207 if (!IS_INODE(node_folio)) in get_dnode_base() 3211 offset_in_addr(&F2FS_NODE(node_folio)->i); in get_dnode_base() 3215 struct folio *node_folio) in get_dnode_addr() argument [all …]
|
| H A D | compress.c | 910 block_t blkaddr = data_blkaddr(dn->inode, dn->node_folio, in f2fs_sanity_check_cluster() 951 block_t blkaddr = data_blkaddr(dn->inode, dn->node_folio, in __f2fs_get_cluster_blocks() 1325 if (data_blkaddr(dn.inode, dn.node_folio, in f2fs_write_compressed_pages() 1357 fio.old_blkaddr = data_blkaddr(dn.inode, dn.node_folio, in f2fs_write_compressed_pages() 1890 bool compressed = data_blkaddr(dn->inode, dn->node_folio, in f2fs_cluster_blocks_are_contiguous() 1893 block_t first_blkaddr = data_blkaddr(dn->inode, dn->node_folio, in f2fs_cluster_blocks_are_contiguous() 1897 block_t blkaddr = data_blkaddr(dn->inode, dn->node_folio, in f2fs_cluster_blocks_are_contiguous()
|
| H A D | file.c | 430 block_t first_blkaddr = data_blkaddr(dn->inode, dn->node_folio, in __found_offset() 500 end_offset = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_seek_block() 662 addr = get_dnode_addr(dn->inode, dn->node_folio) + ofs; in f2fs_truncate_data_blocks_range() 726 fofs = f2fs_start_bidx_of_node(ofs_of_node(dn->node_folio), in f2fs_truncate_data_blocks_range() 834 count = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_do_truncate_blocks() 839 if (dn.ofs_in_node || IS_INODE(dn.node_folio)) { in f2fs_do_truncate_blocks() 1260 end_offset = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_truncate_hole() 1358 done = min((pgoff_t)ADDRS_PER_PAGE(dn.node_folio, inode) - in __read_out_blkaddrs() 1447 ADDRS_PER_PAGE(dn.node_folio, dst_inode) - in __clone_blkaddrs() 1746 end_offset = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_zero_range() [all …]
|
| H A D | data.c | 1171 __le32 *addr = get_dnode_addr(dn->inode, dn->node_folio); in __set_data_blkaddr() 1185 f2fs_folio_wait_writeback(dn->node_folio, NODE, true, true); in f2fs_set_data_blkaddr() 1187 if (folio_mark_dirty(dn->node_folio)) in f2fs_set_data_blkaddr() 1215 f2fs_folio_wait_writeback(dn->node_folio, NODE, true, true); in f2fs_reserve_new_blocks() 1226 if (folio_mark_dirty(dn->node_folio)) in f2fs_reserve_new_blocks() 1676 end_offset = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_map_blocks() 2334 blkaddr = from_dnode ? data_blkaddr(dn.inode, dn.node_folio, in f2fs_read_multi_pages() 2368 blkaddr = from_dnode ? data_blkaddr(dn.inode, dn.node_folio, in f2fs_read_multi_pages()
|
| H A D | extent_cache.c | 952 ei.fofs = f2fs_start_bidx_of_node(ofs_of_node(dn->node_folio), dn->inode) + in __update_extent_cache()
|
| H A D | segment.c | 337 blen = min((pgoff_t)ADDRS_PER_PAGE(dn.node_folio, cow_inode), in __f2fs_commit_atomic_write()
|