Lines Matching +full:9 +full:v

147 	bch2_fs_inconsistent_on((s64) inode->v.i_blocks + sectors < 0, c,  in __bch2_i_sectors_acct()
149 inode->v.i_ino, (u64) inode->v.i_blocks, sectors, in __bch2_i_sectors_acct()
151 inode->v.i_blocks += sectors; in __bch2_i_sectors_acct()
194 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bch2_fsync()
200 ret = sync_inode_metadata(&inode->v, 1); in bch2_fsync()
243 struct bch_fs *c = inode->v.i_sb->s_fs_info; in __bch2_truncate_folio()
244 struct address_space *mapping = inode->v.i_mapping; in __bch2_truncate_folio()
261 POS(inode->v.i_ino, (index << PAGE_SECTORS_SHIFT)), in __bch2_truncate_folio()
262 POS(inode->v.i_ino, (index << PAGE_SECTORS_SHIFT) + PAGE_SECTORS)); in __bch2_truncate_folio()
303 for (i = round_up(start_offset, block_bytes(c)) >> 9; in __bch2_truncate_folio()
304 i < round_down(end_offset, block_bytes(c)) >> 9; in __bch2_truncate_folio()
324 WARN_ON_ONCE(folio_pos(folio) >= inode->v.i_size); in __bch2_truncate_folio()
326 if (inode->v.i_size > folio_pos(folio)) in __bch2_truncate_folio()
327 end_pos = min_t(u64, inode->v.i_size, end_pos); in __bch2_truncate_folio()
379 struct address_space *mapping = inode->v.i_mapping; in bch2_extend()
391 truncate_setsize(&inode->v, iattr->ia_size); in bch2_extend()
399 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bchfs_truncate()
400 struct address_space *mapping = inode->v.i_mapping; in bchfs_truncate()
410 if (iattr->ia_size != inode->v.i_size) { in bchfs_truncate()
418 inode_dio_wait(&inode->v); in bchfs_truncate()
435 inode->v.i_size < inode_u.bi_size, in bchfs_truncate()
437 (u64) inode->v.i_size, inode_u.bi_size); in bchfs_truncate()
439 if (iattr->ia_size > inode->v.i_size) { in bchfs_truncate()
450 truncate_setsize(&inode->v, iattr->ia_size); in bchfs_truncate()
484 bch2_fs_inconsistent_on(!inode->v.i_size && inode->v.i_blocks && in bchfs_truncate()
487 inode->v.i_ino, (u64) inode->v.i_blocks, in bchfs_truncate()
502 struct bch_fs *c = inode->v.i_sb->s_fs_info; in inode_update_times_fn()
510 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bchfs_fpunch()
523 truncate_pagecache_range(&inode->v, offset, end - 1); in bchfs_fpunch()
529 block_start >> 9, block_end >> 9, in bchfs_fpunch()
535 if (end >= inode->v.i_size && !truncated_last_page) { in bchfs_fpunch()
536 ret = bch2_write_inode_size(c, inode, inode->v.i_size, in bchfs_fpunch()
551 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bchfs_fcollapse_finsert()
552 struct address_space *mapping = inode->v.i_mapping; in bchfs_fcollapse_finsert()
560 if (offset >= inode->v.i_size) in bchfs_fcollapse_finsert()
563 if (offset + len >= inode->v.i_size) in bchfs_fcollapse_finsert()
572 i_size_write(&inode->v, inode->v.i_size + len); in bchfs_fcollapse_finsert()
574 ret = bch2_fcollapse_finsert(c, inode_inum(inode), offset >> 9, len >> 9, in bchfs_fcollapse_finsert()
577 i_size_write(&inode->v, inode->v.i_size - len); in bchfs_fcollapse_finsert()
586 struct bch_fs *c = inode->v.i_sb->s_fs_info; in __bchfs_fallocate()
589 struct bpos end_pos = POS(inode->v.i_ino, end_sector); in __bchfs_fallocate()
596 POS(inode->v.i_ino, start_sector), in __bchfs_fallocate()
643 if (bch2_clamp_data_hole(&inode->v, in __bchfs_fallocate()
648 (bch2_clamp_data_hole(&inode->v, in __bchfs_fallocate()
707 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bchfs_fallocate()
714 if (!(mode & FALLOC_FL_KEEP_SIZE) && end > inode->v.i_size) { in bchfs_fallocate()
715 ret = inode_newsize_ok(&inode->v, end); in bchfs_fallocate()
727 truncate_pagecache_range(&inode->v, offset, end - 1); in bchfs_fallocate()
733 ret = __bchfs_fallocate(inode, mode, block_start >> 9, block_end >> 9); in bchfs_fallocate()
743 if (mode & FALLOC_FL_KEEP_SIZE && end > inode->v.i_size) in bchfs_fallocate()
744 end = inode->v.i_size; in bchfs_fallocate()
746 if (end >= inode->v.i_size && in bchfs_fallocate()
749 spin_lock(&inode->v.i_lock); in bchfs_fallocate()
750 i_size_write(&inode->v, end); in bchfs_fallocate()
751 spin_unlock(&inode->v.i_lock); in bchfs_fallocate()
765 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bch2_fallocate_dispatch()
771 inode_lock(&inode->v); in bch2_fallocate_dispatch()
772 inode_dio_wait(&inode->v); in bch2_fallocate_dispatch()
791 inode_unlock(&inode->v); in bch2_fallocate_dispatch()
805 struct bch_fs *c = inode->v.i_sb->s_fs_info; in quota_reserve_range()
821 SPOS(inode->v.i_ino, pos, snapshot), 0); in quota_reserve_range()
824 (k = bch2_btree_iter_peek_upto(&iter, POS(inode->v.i_ino, end - 1))).k && in quota_reserve_range()
851 struct bch_fs *c = src->v.i_sb->s_fs_info; in bch2_remap_file_range()
871 lock_two_nondirectories(&src->v, &dst->v); in bch2_remap_file_range()
874 inode_dio_wait(&src->v); in bch2_remap_file_range()
875 inode_dio_wait(&dst->v); in bch2_remap_file_range()
885 ret = bch2_write_invalidate_inode_pages_range(dst->v.i_mapping, in bch2_remap_file_range()
890 ret = quota_reserve_range(dst, &quota_res, pos_dst >> 9, in bch2_remap_file_range()
891 (pos_dst + aligned_len) >> 9); in bch2_remap_file_range()
897 bch2_mark_pagecache_unallocated(src, pos_src >> 9, in bch2_remap_file_range()
898 (pos_src + aligned_len) >> 9); in bch2_remap_file_range()
901 inode_inum(dst), pos_dst >> 9, in bch2_remap_file_range()
902 inode_inum(src), pos_src >> 9, in bch2_remap_file_range()
903 aligned_len >> 9, in bch2_remap_file_range()
911 ret = min((u64) ret << 9, (u64) len); in bch2_remap_file_range()
915 spin_lock(&dst->v.i_lock); in bch2_remap_file_range()
916 if (pos_dst + ret > dst->v.i_size) in bch2_remap_file_range()
917 i_size_write(&dst->v, pos_dst + ret); in bch2_remap_file_range()
918 spin_unlock(&dst->v.i_lock); in bch2_remap_file_range()
926 unlock_two_nondirectories(&src->v, &dst->v); in bch2_remap_file_range()
936 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bch2_seek_data()
945 isize = i_size_read(&inode->v); in bch2_seek_data()
958 SPOS(inode->v.i_ino, offset >> 9, snapshot), in bch2_seek_data()
959 POS(inode->v.i_ino, U64_MAX), in bch2_seek_data()
962 next_data = max(offset, bkey_start_offset(k.k) << 9); in bch2_seek_data()
964 } else if (k.k->p.offset >> 9 > isize) in bch2_seek_data()
977 next_data = bch2_seek_pagecache_data(&inode->v, in bch2_seek_data()
989 struct bch_fs *c = inode->v.i_sb->s_fs_info; in bch2_seek_hole()
998 isize = i_size_read(&inode->v); in bch2_seek_hole()
1011 SPOS(inode->v.i_ino, offset >> 9, snapshot), in bch2_seek_hole()
1013 if (k.k->p.inode != inode->v.i_ino) { in bch2_seek_hole()
1014 next_hole = bch2_seek_pagecache_hole(&inode->v, in bch2_seek_hole()
1018 next_hole = bch2_seek_pagecache_hole(&inode->v, in bch2_seek_hole()
1019 max(offset, bkey_start_offset(k.k) << 9), in bch2_seek_hole()
1020 k.k->p.offset << 9, 0, false); in bch2_seek_hole()
1022 if (next_hole < k.k->p.offset << 9) in bch2_seek_hole()
1025 offset = max(offset, bkey_start_offset(k.k) << 9); in bch2_seek_hole()