Home
last modified time | relevance | path

Searched refs:node_folio (Results 1 – 11 of 11) sorted by relevance

/linux/fs/f2fs/
H A Dinode.c71 static void __get_inode_rdev(struct inode *inode, struct folio *node_folio) in __get_inode_rdev() argument
73 __le32 *addr = get_dnode_addr(inode, node_folio); in __get_inode_rdev()
84 static void __set_inode_rdev(struct inode *inode, struct folio *node_folio) in __set_inode_rdev() argument
86 __le32 *addr = get_dnode_addr(inode, node_folio); in __set_inode_rdev()
270 static bool sanity_check_inode(struct inode *inode, struct folio *node_folio) in sanity_check_inode() argument
274 struct f2fs_inode *ri = F2FS_INODE(node_folio); in sanity_check_inode()
277 iblocks = le64_to_cpu(F2FS_INODE(node_folio)->i_blocks); in sanity_check_inode()
284 if (ino_of_node(node_folio) != nid_of_node(node_folio)) { in sanity_check_inode()
287 ino_of_node(node_folio), nid_of_node(node_folio)); in sanity_check_inode()
291 if (ino_of_node(node_folio) == fi->i_xattr_nid) { in sanity_check_inode()
[all …]
H A Dnode.h239 static inline nid_t ino_of_node(const struct folio *node_folio) in ino_of_node() argument
241 struct f2fs_node *rn = F2FS_NODE(node_folio); in ino_of_node()
245 static inline nid_t nid_of_node(const struct folio *node_folio) in nid_of_node() argument
247 struct f2fs_node *rn = F2FS_NODE(node_folio); in nid_of_node()
251 static inline unsigned int ofs_of_node(const struct folio *node_folio) in ofs_of_node() argument
253 struct f2fs_node *rn = F2FS_NODE(node_folio); in ofs_of_node()
258 static inline __u64 cpver_of_node(const struct folio *node_folio) in cpver_of_node() argument
260 struct f2fs_node *rn = F2FS_NODE(node_folio); in cpver_of_node()
264 static inline block_t next_blkaddr_of_node(const struct folio *node_folio) in next_blkaddr_of_node() argument
266 struct f2fs_node *rn = F2FS_NODE(node_folio); in next_blkaddr_of_node()
[all …]
H A Drecovery.c500 struct folio *sum_folio, *node_folio; in check_index_in_prev_nodes() local
533 max_addrs = ADDRS_PER_PAGE(dn->node_folio, dn->inode); in check_index_in_prev_nodes()
545 tdn.node_folio = dn->inode_folio; in check_index_in_prev_nodes()
554 node_folio = f2fs_get_node_folio(sbi, nid, NODE_TYPE_REGULAR); in check_index_in_prev_nodes()
555 if (IS_ERR(node_folio)) in check_index_in_prev_nodes()
556 return PTR_ERR(node_folio); in check_index_in_prev_nodes()
558 offset = ofs_of_node(node_folio); in check_index_in_prev_nodes()
559 ino = ino_of_node(node_folio); in check_index_in_prev_nodes()
560 f2fs_folio_put(node_folio, true); in check_index_in_prev_nodes()
668 f2fs_folio_wait_writeback(dn.node_folio, NODE, true, true); in do_recover_data()
[all …]
H A Dgc.c1059 struct folio *node_folio; in gc_node_segment() local
1082 node_folio = f2fs_get_node_folio(sbi, nid, NODE_TYPE_REGULAR); in gc_node_segment()
1083 if (IS_ERR(node_folio)) in gc_node_segment()
1088 f2fs_folio_put(node_folio, true); in gc_node_segment()
1093 f2fs_folio_put(node_folio, true); in gc_node_segment()
1098 f2fs_folio_put(node_folio, true); in gc_node_segment()
1102 err = f2fs_move_node_folio(node_folio, gc_type); in gc_node_segment()
1151 struct folio *node_folio; in is_alive() local
1159 node_folio = f2fs_get_node_folio(sbi, nid, NODE_TYPE_REGULAR); in is_alive()
1160 if (IS_ERR(node_folio)) in is_alive()
[all …]
H A Dnode.c905 dn->node_folio = nfolio[level]; in f2fs_get_dnode_of_data()
926 blkaddr = data_blkaddr(dn->inode, dn->node_folio, ofs_in_node); in f2fs_get_dnode_of_data()
928 blkaddr = data_blkaddr(dn->inode, dn->node_folio, in f2fs_get_dnode_of_data()
943 dn->node_folio = NULL; in f2fs_get_dnode_of_data()
984 clear_node_folio_dirty(dn->node_folio); in truncate_node()
987 index = dn->node_folio->index; in truncate_node()
988 f2fs_folio_put(dn->node_folio, true); in truncate_node()
993 dn->node_folio = NULL; in truncate_node()
1025 dn->node_folio = folio; in truncate_dnode()
1098 dn->node_folio = folio; in truncate_nodes()
[all …]
H A Df2fs.h1113 struct folio *node_folio; /* cached direct node folio */ member
1129 dn->node_folio = nfolio; in set_new_dnode()
3089 if (dn->node_folio) in f2fs_put_dnode()
3090 f2fs_folio_put(dn->node_folio, true); in f2fs_put_dnode()
3091 if (dn->inode_folio && dn->node_folio != dn->inode_folio) in f2fs_put_dnode()
3093 dn->node_folio = NULL; in f2fs_put_dnode()
3205 struct folio *node_folio) in get_dnode_base() argument
3207 if (!IS_INODE(node_folio)) in get_dnode_base()
3211 offset_in_addr(&F2FS_NODE(node_folio)->i); in get_dnode_base()
3215 struct folio *node_folio) in get_dnode_addr() argument
[all …]
H A Dcompress.c910 block_t blkaddr = data_blkaddr(dn->inode, dn->node_folio, in f2fs_sanity_check_cluster()
951 block_t blkaddr = data_blkaddr(dn->inode, dn->node_folio, in __f2fs_get_cluster_blocks()
1325 if (data_blkaddr(dn.inode, dn.node_folio, in f2fs_write_compressed_pages()
1357 fio.old_blkaddr = data_blkaddr(dn.inode, dn.node_folio, in f2fs_write_compressed_pages()
1890 bool compressed = data_blkaddr(dn->inode, dn->node_folio, in f2fs_cluster_blocks_are_contiguous()
1893 block_t first_blkaddr = data_blkaddr(dn->inode, dn->node_folio, in f2fs_cluster_blocks_are_contiguous()
1897 block_t blkaddr = data_blkaddr(dn->inode, dn->node_folio, in f2fs_cluster_blocks_are_contiguous()
H A Dfile.c430 block_t first_blkaddr = data_blkaddr(dn->inode, dn->node_folio, in __found_offset()
500 end_offset = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_seek_block()
662 addr = get_dnode_addr(dn->inode, dn->node_folio) + ofs; in f2fs_truncate_data_blocks_range()
726 fofs = f2fs_start_bidx_of_node(ofs_of_node(dn->node_folio), in f2fs_truncate_data_blocks_range()
834 count = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_do_truncate_blocks()
839 if (dn.ofs_in_node || IS_INODE(dn.node_folio)) { in f2fs_do_truncate_blocks()
1260 end_offset = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_truncate_hole()
1358 done = min((pgoff_t)ADDRS_PER_PAGE(dn.node_folio, inode) - in __read_out_blkaddrs()
1447 ADDRS_PER_PAGE(dn.node_folio, dst_inode) - in __clone_blkaddrs()
1746 end_offset = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_zero_range()
[all …]
H A Ddata.c1171 __le32 *addr = get_dnode_addr(dn->inode, dn->node_folio); in __set_data_blkaddr()
1185 f2fs_folio_wait_writeback(dn->node_folio, NODE, true, true); in f2fs_set_data_blkaddr()
1187 if (folio_mark_dirty(dn->node_folio)) in f2fs_set_data_blkaddr()
1215 f2fs_folio_wait_writeback(dn->node_folio, NODE, true, true); in f2fs_reserve_new_blocks()
1226 if (folio_mark_dirty(dn->node_folio)) in f2fs_reserve_new_blocks()
1676 end_offset = ADDRS_PER_PAGE(dn.node_folio, inode); in f2fs_map_blocks()
2334 blkaddr = from_dnode ? data_blkaddr(dn.inode, dn.node_folio, in f2fs_read_multi_pages()
2368 blkaddr = from_dnode ? data_blkaddr(dn.inode, dn.node_folio, in f2fs_read_multi_pages()
H A Dextent_cache.c952 ei.fofs = f2fs_start_bidx_of_node(ofs_of_node(dn->node_folio), dn->inode) + in __update_extent_cache()
H A Dsegment.c337 blen = min((pgoff_t)ADDRS_PER_PAGE(dn.node_folio, cow_inode), in __f2fs_commit_atomic_write()