Lines Matching refs:async_chunk

762 struct async_chunk {  struct
776 struct async_chunk chunks[]; argument
779 static int add_async_extent(struct async_chunk *cow, u64 start, u64 ram_size, in add_async_extent()
934 struct async_chunk *async_chunk = in compress_file_range() local
935 container_of(work, struct async_chunk, work); in compress_file_range()
936 struct btrfs_inode *inode = async_chunk->inode; in compress_file_range()
942 u64 start = async_chunk->start; in compress_file_range()
943 u64 end = async_chunk->end; in compress_file_range()
1024 compress_level, async_chunk->write_flags); in compress_file_range()
1085 ret = add_async_extent(async_chunk, start, total_in, cb); in compress_file_range()
1098 ret = add_async_extent(async_chunk, start, end - start + 1, NULL); in compress_file_range()
1133 static void submit_one_async_extent(struct async_chunk *async_chunk, in submit_one_async_extent() argument
1137 struct btrfs_inode *inode = async_chunk->inode; in submit_one_async_extent()
1152 if (async_chunk->blkcg_css) in submit_one_async_extent()
1153 kthread_associate_blkcg(async_chunk->blkcg_css); in submit_one_async_extent()
1159 if (async_chunk->locked_folio) { in submit_one_async_extent()
1160 u64 locked_folio_start = folio_pos(async_chunk->locked_folio); in submit_one_async_extent()
1162 folio_size(async_chunk->locked_folio) - 1; in submit_one_async_extent()
1165 locked_folio = async_chunk->locked_folio; in submit_one_async_extent()
1228 if (async_chunk->blkcg_css) in submit_one_async_extent()
1246 if (async_chunk->blkcg_css) in submit_one_async_extent()
1659 struct async_chunk *async_chunk = container_of(work, struct async_chunk, in submit_compressed_extents() local
1669 btrfs_add_delayed_iput(async_chunk->inode); in submit_compressed_extents()
1670 if (async_chunk->blkcg_css) in submit_compressed_extents()
1671 css_put(async_chunk->blkcg_css); in submit_compressed_extents()
1673 async_cow = async_chunk->async_cow; in submit_compressed_extents()
1679 nr_pages = (async_chunk->end - async_chunk->start + PAGE_SIZE) >> in submit_compressed_extents()
1682 while (!list_empty(&async_chunk->extents)) { in submit_compressed_extents()
1683 async_extent = list_first_entry(&async_chunk->extents, in submit_compressed_extents()
1686 submit_one_async_extent(async_chunk, async_extent, &alloc_hint); in submit_compressed_extents()
1702 struct async_chunk *async_chunk; in run_delalloc_compressed() local
1717 async_chunk = ctx->chunks; in run_delalloc_compressed()
1728 async_chunk[i].async_cow = ctx; in run_delalloc_compressed()
1729 async_chunk[i].inode = inode; in run_delalloc_compressed()
1730 async_chunk[i].start = start; in run_delalloc_compressed()
1731 async_chunk[i].end = cur_end; in run_delalloc_compressed()
1732 async_chunk[i].write_flags = write_flags; in run_delalloc_compressed()
1733 INIT_LIST_HEAD(&async_chunk[i].extents); in run_delalloc_compressed()
1757 async_chunk[i].locked_folio = locked_folio; in run_delalloc_compressed()
1760 async_chunk[i].locked_folio = NULL; in run_delalloc_compressed()
1765 async_chunk[i].blkcg_css = blkcg_css; in run_delalloc_compressed()
1766 async_chunk[i].write_flags |= REQ_BTRFS_CGROUP_PUNT; in run_delalloc_compressed()
1768 async_chunk[i].blkcg_css = NULL; in run_delalloc_compressed()
1771 btrfs_init_work(&async_chunk[i].work, compress_file_range, in run_delalloc_compressed()
1777 btrfs_queue_work(fs_info->delalloc_workers, &async_chunk[i].work); in run_delalloc_compressed()