Lines Matching full:mapping
142 file_ra_state_init(struct file_ra_state *ra, struct address_space *mapping) in file_ra_state_init() argument
144 ra->ra_pages = inode_to_bdi(mapping->host)->ra_pages; in file_ra_state_init()
151 const struct address_space_operations *aops = rac->mapping->a_ops; in read_pages()
207 * Context: File is referenced by caller, and ractl->mapping->invalidate_lock
214 struct address_space *mapping = ractl->mapping; in page_cache_ra_unbounded() local
216 gfp_t gfp_mask = readahead_gfp_mask(mapping); in page_cache_ra_unbounded()
218 unsigned int min_nrpages = mapping_min_folio_nrpages(mapping); in page_cache_ra_unbounded()
227 * filesystems already specify __GFP_NOFS in their mapping's in page_cache_ra_unbounded()
232 lockdep_assert_held(&mapping->invalidate_lock); in page_cache_ra_unbounded()
234 trace_page_cache_ra_unbounded(mapping->host, index, nr_to_read, in page_cache_ra_unbounded()
236 index = mapping_align_index(mapping, index); in page_cache_ra_unbounded()
259 struct folio *folio = xa_load(&mapping->i_pages, index + i); in page_cache_ra_unbounded()
278 mapping_min_folio_order(mapping)); in page_cache_ra_unbounded()
282 ret = filemap_add_folio(mapping, folio, index + i, gfp_mask); in page_cache_ra_unbounded()
318 struct address_space *mapping = ractl->mapping; in do_page_cache_ra() local
320 loff_t isize = i_size_read(mapping->host); in do_page_cache_ra()
333 filemap_invalidate_lock_shared(mapping); in do_page_cache_ra()
335 filemap_invalidate_unlock_shared(mapping); in do_page_cache_ra()
345 struct address_space *mapping = ractl->mapping; in force_page_cache_ra() local
347 struct backing_dev_info *bdi = inode_to_bdi(mapping->host); in force_page_cache_ra()
350 if (unlikely(!mapping->a_ops->read_folio && !mapping->a_ops->readahead)) in force_page_cache_ra()
456 err = filemap_add_folio(ractl->mapping, folio, index, gfp); in ra_alloc_folio()
470 struct address_space *mapping = ractl->mapping; in page_cache_ra_order() local
473 unsigned int min_order = mapping_min_folio_order(mapping); in page_cache_ra_order()
474 pgoff_t limit = (i_size_read(mapping->host) - 1) >> PAGE_SHIFT; in page_cache_ra_order()
478 gfp_t gfp = readahead_gfp_mask(mapping); in page_cache_ra_order()
481 trace_page_cache_ra_order(mapping->host, start, ra); in page_cache_ra_order()
482 if (!mapping_large_folio_support(mapping)) { in page_cache_ra_order()
489 new_order = min(mapping_max_folio_order(mapping), new_order); in page_cache_ra_order()
497 filemap_invalidate_lock_shared(mapping); in page_cache_ra_order()
503 ractl->_index = mapping_align_index(mapping, index); in page_cache_ra_order()
522 filemap_invalidate_unlock_shared(mapping); in page_cache_ra_order()
545 struct backing_dev_info *bdi = inode_to_bdi(ractl->mapping->host); in ractl_max_pages()
566 trace_page_cache_sync_ra(ractl->mapping->host, index, ra, req_count); in page_cache_sync_ra()
606 miss = page_cache_prev_miss(ractl->mapping, index - 1, max_pages); in page_cache_sync_ra()
651 trace_page_cache_async_ra(ractl->mapping->host, index, ra, req_count); in page_cache_async_ra()
681 start = page_cache_next_miss(ractl->mapping, index + 1, max_pages); in page_cache_async_ra()
769 struct address_space *mapping = ractl->mapping; in readahead_expand() local
772 gfp_t gfp_mask = readahead_gfp_mask(mapping); in readahead_expand()
773 unsigned long min_nrpages = mapping_min_folio_nrpages(mapping); in readahead_expand()
774 unsigned int min_order = mapping_min_folio_order(mapping); in readahead_expand()
786 struct folio *folio = xa_load(&mapping->i_pages, index); in readahead_expand()
795 index = mapping_align_index(mapping, index); in readahead_expand()
796 if (filemap_add_folio(mapping, folio, index, gfp_mask) < 0) { in readahead_expand()
815 struct folio *folio = xa_load(&mapping->i_pages, index); in readahead_expand()
824 index = mapping_align_index(mapping, index); in readahead_expand()
825 if (filemap_add_folio(mapping, folio, index, gfp_mask) < 0) { in readahead_expand()