/linux/include/linux/ |
H A D | pagemap.h | 21 unsigned long invalidate_mapping_pages(struct address_space *mapping, 30 int invalidate_inode_pages2(struct address_space *mapping); 31 int invalidate_inode_pages2_range(struct address_space *mapping, 35 int filemap_invalidate_pages(struct address_space *mapping, 41 int filemap_fdatawait_keep_errors(struct address_space *mapping); 43 int filemap_fdatawait_range_keep_errors(struct address_space *mapping, 48 static inline int filemap_fdatawait(struct address_space *mapping) in filemap_fdatawait() argument 50 return filemap_fdatawait_range(mapping, 0, LLONG_MAX); in filemap_fdatawait() 54 int filemap_write_and_wait_range(struct address_space *mapping, 56 int __filemap_fdatawrite_range(struct address_space *mapping, 66 filemap_write_and_wait(struct address_space * mapping) filemap_write_and_wait() argument 85 filemap_set_wb_err(struct address_space * mapping,int err) filemap_set_wb_err() argument 102 filemap_check_wb_err(struct address_space * mapping,errseq_t since) filemap_check_wb_err() argument 115 filemap_sample_wb_err(struct address_space * mapping) filemap_sample_wb_err() argument 143 mapping_empty(struct address_space * mapping) mapping_empty() argument 169 mapping_shrinkable(struct address_space * mapping) mapping_shrinkable() argument 239 mapping_set_error(struct address_space * mapping,int error) mapping_set_error() argument 258 mapping_set_unevictable(struct address_space * mapping) mapping_set_unevictable() argument 263 mapping_clear_unevictable(struct address_space * mapping) mapping_clear_unevictable() argument 268 mapping_unevictable(struct address_space * mapping) mapping_unevictable() argument 273 mapping_set_exiting(struct address_space * mapping) mapping_set_exiting() argument 278 mapping_exiting(struct address_space * mapping) mapping_exiting() argument 283 mapping_set_no_writeback_tags(struct address_space * mapping) mapping_set_no_writeback_tags() argument 288 mapping_use_writeback_tags(struct address_space * mapping) mapping_use_writeback_tags() argument 293 mapping_release_always(const struct address_space * mapping) mapping_release_always() argument 298 mapping_set_release_always(struct address_space * mapping) mapping_set_release_always() argument 303 mapping_clear_release_always(struct address_space * mapping) mapping_clear_release_always() argument 308 mapping_stable_writes(const struct address_space * mapping) mapping_stable_writes() argument 313 mapping_set_stable_writes(struct address_space * mapping) mapping_set_stable_writes() argument 318 mapping_clear_stable_writes(struct address_space * mapping) mapping_clear_stable_writes() argument 323 mapping_set_inaccessible(struct address_space * mapping) mapping_set_inaccessible() argument 334 mapping_inaccessible(struct address_space * mapping) mapping_inaccessible() argument 339 mapping_set_writeback_may_deadlock_on_reclaim(struct address_space * mapping) mapping_set_writeback_may_deadlock_on_reclaim() argument 344 mapping_writeback_may_deadlock_on_reclaim(struct address_space * mapping) mapping_writeback_may_deadlock_on_reclaim() argument 349 mapping_gfp_mask(struct address_space * mapping) mapping_gfp_mask() argument 355 mapping_gfp_constraint(struct address_space * mapping,gfp_t gfp_mask) mapping_gfp_constraint() argument 419 mapping_set_folio_order_range(struct address_space * mapping,unsigned int min,unsigned int max) mapping_set_folio_order_range() argument 439 mapping_set_folio_min_order(struct address_space * mapping,unsigned int min) mapping_set_folio_min_order() argument 456 mapping_set_large_folios(struct address_space * mapping) mapping_set_large_folios() argument 462 mapping_max_folio_order(const struct address_space * mapping) mapping_max_folio_order() argument 470 mapping_min_folio_order(const struct address_space * mapping) mapping_min_folio_order() argument 478 mapping_min_folio_nrpages(struct address_space * mapping) mapping_min_folio_nrpages() argument 492 mapping_align_index(struct address_space * mapping,pgoff_t index) mapping_align_index() argument 502 mapping_large_folio_support(struct address_space * mapping) mapping_large_folio_support() argument 512 mapping_max_folio_size(const struct address_space * mapping) mapping_max_folio_size() argument 517 filemap_nr_thps(struct address_space * mapping) filemap_nr_thps() argument 526 filemap_nr_thps_inc(struct address_space * mapping) filemap_nr_thps_inc() argument 536 filemap_nr_thps_dec(struct address_space * mapping) filemap_nr_thps_dec() argument 767 write_begin_get_folio(const struct kiocb * iocb,struct address_space * mapping,pgoff_t index,size_t len) write_begin_get_folio() argument 791 filemap_get_folio(struct address_space * mapping,pgoff_t index) filemap_get_folio() argument 809 filemap_lock_folio(struct address_space * mapping,pgoff_t index) filemap_lock_folio() argument 827 filemap_grab_folio(struct address_space * mapping,pgoff_t index) filemap_grab_folio() argument 845 find_get_page(struct address_space * mapping,pgoff_t offset) find_get_page() argument 851 find_get_page_flags(struct address_space * mapping,pgoff_t offset,fgf_t fgp_flags) find_get_page_flags() argument 870 find_lock_page(struct address_space * mapping,pgoff_t index) find_lock_page() argument 895 find_or_create_page(struct address_space * mapping,pgoff_t index,gfp_t gfp_mask) find_or_create_page() argument 917 grab_cache_page_nowait(struct address_space * mapping,pgoff_t index) grab_cache_page_nowait() argument 982 read_mapping_page(struct address_space * mapping,pgoff_t index,struct file * file) read_mapping_page() argument 988 read_mapping_folio(struct address_space * mapping,pgoff_t index,struct file * file) read_mapping_folio() argument 1291 filemap_range_needs_writeback(struct address_space * mapping,loff_t start_byte,loff_t end_byte) filemap_range_needs_writeback() argument 1321 struct address_space *mapping; global() member 1364 page_cache_sync_readahead(struct address_space * mapping,struct file_ra_state * ra,struct file * file,pgoff_t index,unsigned long req_count) page_cache_sync_readahead() argument 1386 page_cache_async_readahead(struct address_space * mapping,struct file_ra_state * ra,struct file * file,struct folio * folio,unsigned long req_count) page_cache_async_readahead() argument [all...] |
H A D | io-mapping.h | 17 * The io_mapping mechanism provides an abstraction for mapping 20 * See Documentation/driver-api/io-mapping.rst 35 * For small address space machines, mapping large objects 58 io_mapping_fini(struct io_mapping *mapping) in io_mapping_fini() argument 60 iomap_free(mapping->base, mapping->size); in io_mapping_fini() 65 io_mapping_map_atomic_wc(struct io_mapping *mapping, in io_mapping_map_atomic_wc() argument 70 BUG_ON(offset >= mapping->size); in io_mapping_map_atomic_wc() 71 phys_addr = mapping->base + offset; in io_mapping_map_atomic_wc() 77 return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping in io_mapping_map_atomic_wc() 92 io_mapping_map_local_wc(struct io_mapping * mapping,unsigned long offset) io_mapping_map_local_wc() argument 107 io_mapping_map_wc(struct io_mapping * mapping,unsigned long offset,unsigned long size) io_mapping_map_wc() argument 147 io_mapping_fini(struct io_mapping * mapping) io_mapping_fini() argument 154 io_mapping_map_wc(struct io_mapping * mapping,unsigned long offset,unsigned long size) io_mapping_map_wc() argument 168 io_mapping_map_atomic_wc(struct io_mapping * mapping,unsigned long offset) io_mapping_map_atomic_wc() argument 191 io_mapping_map_local_wc(struct io_mapping * mapping,unsigned long offset) io_mapping_map_local_wc() argument [all...] |
H A D | tpm_eventlog.h | 166 void *mapping = NULL; in __calc_tpm2_event_size() local 186 mapping = TPM_MEMREMAP((unsigned long)marker_start, in __calc_tpm2_event_size() 188 if (!mapping) { in __calc_tpm2_event_size() 193 mapping = marker_start; in __calc_tpm2_event_size() 196 event = (struct tcg_pcr_event2_head *)mapping; in __calc_tpm2_event_size() 233 TPM_MEMUNMAP(mapping, mapping_size); in __calc_tpm2_event_size() 235 mapping = TPM_MEMREMAP((unsigned long)marker, in __calc_tpm2_event_size() 237 if (!mapping) { in __calc_tpm2_event_size() 242 mapping = marker; in __calc_tpm2_event_size() 245 memcpy(&halg, mapping, halg_siz in __calc_tpm2_event_size() [all...] |
/linux/mm/ |
H A D | truncate.c | 26 static void clear_shadow_entries(struct address_space *mapping, in clear_shadow_entries() argument 29 XA_STATE(xas, &mapping->i_pages, start); in clear_shadow_entries() 33 if (shmem_mapping(mapping) || dax_mapping(mapping)) in clear_shadow_entries() 38 spin_lock(&mapping->host->i_lock); in clear_shadow_entries() 48 if (mapping_shrinkable(mapping)) in clear_shadow_entries() 49 inode_add_lru(mapping->host); in clear_shadow_entries() 50 spin_unlock(&mapping->host->i_lock); in clear_shadow_entries() 60 static void truncate_folio_batch_exceptionals(struct address_space *mapping, in truncate_folio_batch_exceptionals() argument 63 XA_STATE(xas, &mapping in truncate_folio_batch_exceptionals() 170 truncate_inode_folio(struct address_space * mapping,struct folio * folio) truncate_inode_folio() argument 274 generic_error_remove_folio(struct address_space * mapping,struct folio * folio) generic_error_remove_folio() argument 300 mapping_evict_folio(struct address_space * mapping,struct folio * folio) mapping_evict_folio() argument 341 truncate_inode_pages_range(struct address_space * mapping,loff_t lstart,loff_t lend) truncate_inode_pages_range() argument 458 truncate_inode_pages(struct address_space * mapping,loff_t lstart) truncate_inode_pages() argument 473 truncate_inode_pages_final(struct address_space * mapping) truncate_inode_pages_final() argument 509 mapping_try_invalidate(struct address_space * mapping,pgoff_t start,pgoff_t end,unsigned long * nr_failed) mapping_try_invalidate() argument 574 invalidate_mapping_pages(struct address_space * mapping,pgoff_t start,pgoff_t end) invalidate_mapping_pages() argument 581 folio_launder(struct address_space * mapping,struct folio * folio) folio_launder() argument 597 folio_unmap_invalidate(struct address_space * mapping,struct folio * folio,gfp_t gfp) folio_unmap_invalidate() argument 647 invalidate_inode_pages2_range(struct address_space * mapping,pgoff_t start,pgoff_t end) invalidate_inode_pages2_range() argument 733 invalidate_inode_pages2(struct address_space * mapping) invalidate_inode_pages2() argument 756 struct address_space *mapping = inode->i_mapping; truncate_pagecache() local 877 struct address_space *mapping = inode->i_mapping; truncate_pagecache_range() local [all...] |
H A D | filemap.c | 128 static void page_cache_delete(struct address_space *mapping, in page_cache_delete() argument 131 XA_STATE(xas, &mapping->i_pages, folio->index); in page_cache_delete() 134 mapping_set_update(&xas, mapping); in page_cache_delete() 144 folio->mapping = NULL; in page_cache_delete() 146 mapping->nrpages -= nr; in page_cache_delete() 149 static void filemap_unaccount_folio(struct address_space *mapping, in filemap_unaccount_folio() argument 162 if (mapping_exiting(mapping) && !folio_test_large(folio)) { in filemap_unaccount_folio() 191 filemap_nr_thps_dec(mapping); in filemap_unaccount_folio() 209 mapping_can_writeback(mapping))) in filemap_unaccount_folio() 210 folio_account_cleaned(folio, inode_to_wb(mapping in filemap_unaccount_folio() 220 struct address_space *mapping = folio->mapping; __filemap_remove_folio() local 227 filemap_free_folio(struct address_space * mapping,struct folio * folio) filemap_free_folio() argument 248 struct address_space *mapping = folio->mapping; filemap_remove_folio() local 275 page_cache_delete_batch(struct address_space * mapping,struct folio_batch * fbatch) page_cache_delete_batch() argument 316 delete_from_page_cache_batch(struct address_space * mapping,struct folio_batch * fbatch) delete_from_page_cache_batch() argument 342 filemap_check_errors(struct address_space * mapping) filemap_check_errors() argument 356 filemap_check_and_keep_errors(struct address_space * mapping) filemap_check_and_keep_errors() argument 376 filemap_fdatawrite_wbc(struct address_space * mapping,struct writeback_control * wbc) filemap_fdatawrite_wbc() argument 409 __filemap_fdatawrite_range(struct address_space * mapping,loff_t start,loff_t end,int sync_mode) __filemap_fdatawrite_range() argument 422 __filemap_fdatawrite(struct address_space * mapping,int sync_mode) __filemap_fdatawrite() argument 428 filemap_fdatawrite(struct address_space * mapping) filemap_fdatawrite() argument 434 filemap_fdatawrite_range(struct address_space * mapping,loff_t start,loff_t end) filemap_fdatawrite_range() argument 452 filemap_fdatawrite_range_kick(struct address_space * mapping,loff_t start,loff_t end) filemap_fdatawrite_range_kick() argument 468 filemap_flush(struct address_space * mapping) filemap_flush() argument 486 filemap_range_has_page(struct address_space * mapping,loff_t start_byte,loff_t end_byte) filemap_range_has_page() argument 517 __filemap_fdatawait_range(struct address_space * mapping,loff_t start_byte,loff_t end_byte) __filemap_fdatawait_range() argument 562 filemap_fdatawait_range(struct address_space * mapping,loff_t start_byte,loff_t end_byte) filemap_fdatawait_range() argument 584 filemap_fdatawait_range_keep_errors(struct address_space * mapping,loff_t start_byte,loff_t end_byte) filemap_fdatawait_range_keep_errors() argument 610 struct address_space *mapping = file->f_mapping; file_fdatawait_range() local 631 filemap_fdatawait_keep_errors(struct address_space * mapping) filemap_fdatawait_keep_errors() argument 639 mapping_needs_writeback(struct address_space * mapping) mapping_needs_writeback() argument 644 filemap_range_has_writeback(struct address_space * mapping,loff_t start_byte,loff_t end_byte) filemap_range_has_writeback() argument 682 filemap_write_and_wait_range(struct address_space * mapping,loff_t lstart,loff_t lend) filemap_write_and_wait_range() argument 709 __filemap_set_wb_err(struct address_space * mapping,int err) __filemap_set_wb_err() argument 745 struct address_space *mapping = file->f_mapping; file_check_and_advance_wb_err() local 788 struct address_space *mapping = file->f_mapping; file_write_and_wait_range() local 822 struct address_space *mapping = old->mapping; replace_page_cache_folio() local 857 __filemap_add_folio(struct address_space * mapping,struct folio * folio,pgoff_t index,gfp_t gfp,void ** shadowp) __filemap_add_folio() argument 958 filemap_add_folio(struct address_space * mapping,struct folio * folio,pgoff_t index,gfp_t gfp) filemap_add_folio() argument 1594 struct address_space *mapping = folio->mapping; filemap_end_dropbehind() local 1777 page_cache_next_miss(struct address_space * mapping,pgoff_t index,unsigned long max_scan) page_cache_next_miss() argument 1814 page_cache_prev_miss(struct address_space * mapping,pgoff_t index,unsigned long max_scan) page_cache_prev_miss() argument 1863 filemap_get_entry(struct address_space * mapping,pgoff_t index) filemap_get_entry() argument 1910 __filemap_get_folio(struct address_space * mapping,pgoff_t index,fgf_t fgp_flags,gfp_t gfp) __filemap_get_folio() argument 2085 find_get_entries(struct address_space * mapping,pgoff_t * start,pgoff_t end,struct folio_batch * fbatch,pgoff_t * indices) find_get_entries() argument 2134 find_lock_entries(struct address_space * mapping,pgoff_t * start,pgoff_t end,struct folio_batch * fbatch,pgoff_t * indices) find_lock_entries() argument 2202 filemap_get_folios(struct address_space * mapping,pgoff_t * start,pgoff_t end,struct folio_batch * fbatch) filemap_get_folios() argument 2224 filemap_get_folios_contig(struct address_space * mapping,pgoff_t * start,pgoff_t end,struct folio_batch * fbatch) filemap_get_folios_contig() argument 2300 filemap_get_folios_tag(struct address_space * mapping,pgoff_t * start,pgoff_t end,xa_mark_t tag,struct folio_batch * fbatch) filemap_get_folios_tag() argument 2367 filemap_get_read_batch(struct address_space * mapping,pgoff_t index,pgoff_t max,struct folio_batch * fbatch) filemap_get_read_batch() argument 2429 filemap_range_uptodate(struct address_space * mapping,loff_t pos,size_t count,struct folio * folio,bool need_uptodate) filemap_range_uptodate() argument 2454 filemap_update_page(struct kiocb * iocb,struct address_space * mapping,size_t count,struct folio * folio,bool need_uptodate) filemap_update_page() argument 2511 struct address_space *mapping = iocb->ki_filp->f_mapping; filemap_create_folio() local 2563 filemap_readahead(struct kiocb * iocb,struct file * file,struct address_space * mapping,struct folio * folio,pgoff_t last_index) filemap_readahead() argument 2580 struct address_space *mapping = filp->f_mapping; filemap_get_pages() local 2680 struct address_space *mapping = filp->f_mapping; filemap_read() local 2792 struct address_space *mapping = iocb->ki_filp->f_mapping; kiocb_write_and_wait() local 2806 filemap_invalidate_pages(struct address_space * mapping,loff_t pos,loff_t end,bool nowait) filemap_invalidate_pages() argument 2833 struct address_space *mapping = iocb->ki_filp->f_mapping; kiocb_invalidate_pages() local 2873 struct address_space *mapping = file->f_mapping; generic_file_read_iter() local 3054 folio_seek_hole_data(struct xa_state * xas,struct address_space * mapping,struct folio * folio,loff_t start,loff_t end,bool seek_data) folio_seek_hole_data() argument 3111 mapping_seek_hole_data(struct address_space * mapping,loff_t start,loff_t end,int whence) mapping_seek_hole_data() argument 3216 struct address_space *mapping = file->f_mapping; do_sync_mmap_readahead() local 3408 struct address_space *mapping = file->f_mapping; filemap_fault() local 3593 next_uptodate_folio(struct xa_state * xas,struct address_space * mapping,pgoff_t end_pgoff) next_uptodate_folio() argument 3737 struct address_space *mapping = file->f_mapping; filemap_map_pages() local 3808 struct address_space *mapping = vmf->vma->vm_file->f_mapping; filemap_page_mkwrite() local 3842 struct address_space *mapping = file->f_mapping; generic_file_mmap() local 3854 struct address_space *mapping = file->f_mapping; generic_file_mmap_prepare() local 3908 do_read_cache_folio(struct address_space * mapping,pgoff_t index,filler_t filler,struct file * file,gfp_t gfp) do_read_cache_folio() argument 3986 read_cache_folio(struct address_space * mapping,pgoff_t index,filler_t filler,struct file * file) read_cache_folio() argument 4011 mapping_read_folio_gfp(struct address_space * mapping,pgoff_t index,gfp_t gfp) mapping_read_folio_gfp() argument 4018 do_read_cache_page(struct address_space * mapping,pgoff_t index,filler_t * filler,struct file * file,gfp_t gfp) do_read_cache_page() argument 4029 read_cache_page(struct address_space * mapping,pgoff_t index,filler_t * filler,struct file * file) read_cache_page() argument 4052 read_cache_page_gfp(struct address_space * mapping,pgoff_t index,gfp_t gfp) read_cache_page_gfp() argument 4082 struct address_space *mapping = iocb->ki_filp->f_mapping; kiocb_invalidate_post_direct_write() local 4094 struct address_space *mapping = iocb->ki_filp->f_mapping; generic_file_direct_write() local 4151 struct address_space *mapping = file->f_mapping; generic_perform_write() local 4265 struct address_space *mapping = file->f_mapping; __generic_file_write_iter() local 4346 struct address_space * const mapping = folio->mapping; filemap_release_folio() local 4376 struct address_space *mapping = inode->i_mapping; filemap_invalidate_inode() local 4427 filemap_cachestat(struct address_space * mapping,pgoff_t first_index,pgoff_t last_index,struct cachestat * cs) filemap_cachestat() argument 4576 struct address_space *mapping; SYSCALL_DEFINE4() local [all...] |
H A D | readahead.c | 139 file_ra_state_init(struct file_ra_state *ra, struct address_space *mapping) in file_ra_state_init() argument 141 ra->ra_pages = inode_to_bdi(mapping->host)->ra_pages; in file_ra_state_init() 148 const struct address_space_operations *aops = rac->mapping->a_ops; in read_pages() 210 struct address_space *mapping = ractl->mapping; in page_cache_ra_unbounded() local 212 gfp_t gfp_mask = readahead_gfp_mask(mapping); in page_cache_ra_unbounded() 214 unsigned int min_nrpages = mapping_min_folio_nrpages(mapping); in page_cache_ra_unbounded() 223 * filesystems already specify __GFP_NOFS in their mapping's in page_cache_ra_unbounded() 228 filemap_invalidate_lock_shared(mapping); in page_cache_ra_unbounded() 229 index = mapping_align_index(mapping, inde in page_cache_ra_unbounded() 337 struct address_space *mapping = ractl->mapping; force_page_cache_ra() local 462 struct address_space *mapping = ractl->mapping; page_cache_ra_order() local 758 struct address_space *mapping = ractl->mapping; readahead_expand() local [all...] |
/linux/drivers/gpu/drm/tegra/ |
H A D | uapi.c | 17 struct tegra_drm_mapping *mapping = in tegra_drm_mapping_release() local 20 host1x_bo_unpin(mapping->map); in tegra_drm_mapping_release() 21 host1x_bo_put(mapping->bo); in tegra_drm_mapping_release() 23 kfree(mapping); in tegra_drm_mapping_release() 26 void tegra_drm_mapping_put(struct tegra_drm_mapping *mapping) in tegra_drm_mapping_put() argument 28 kref_put(&mapping->ref, tegra_drm_mapping_release); in tegra_drm_mapping_put() 33 struct tegra_drm_mapping *mapping; in tegra_drm_channel_context_close() local 39 xa_for_each(&context->mappings, id, mapping) in tegra_drm_channel_context_close() 40 tegra_drm_mapping_put(mapping); in tegra_drm_channel_context_close() 189 struct tegra_drm_mapping *mapping; in tegra_drm_ioctl_channel_map() local 276 struct tegra_drm_mapping *mapping; tegra_drm_ioctl_channel_unmap() local [all...] |
/linux/drivers/media/usb/uvc/ |
H A D | uvc_ctrl.c | 388 static bool uvc_ctrl_mapping_is_compound(struct uvc_control_mapping *mapping) in uvc_ctrl_mapping_is_compound() argument 390 return mapping->v4l2_type >= V4L2_CTRL_COMPOUND_TYPES; in uvc_ctrl_mapping_is_compound() 393 static s32 uvc_mapping_get_s32(struct uvc_control_mapping *mapping, in uvc_mapping_get_s32() argument 398 mapping->get(mapping, query, data_in, sizeof(data_out), &data_out); in uvc_mapping_get_s32() 403 static void uvc_mapping_set_s32(struct uvc_control_mapping *mapping, in uvc_mapping_set_s32() argument 406 mapping->set(mapping, sizeof(data_in), &data_in, data_out); in uvc_mapping_set_s32() 412 * device. The custom menu_mapping in the control @mapping is used when 423 static int uvc_mapping_get_menu_value(const struct uvc_control_mapping *mapping, in uvc_mapping_get_menu_value() argument 436 uvc_mapping_get_menu_name(const struct uvc_control_mapping * mapping,u32 idx) uvc_mapping_get_menu_name() argument 447 uvc_ctrl_get_zoom(struct uvc_control_mapping * mapping,u8 query,const void * uvc_in,size_t v4l2_size,void * v4l2_out) uvc_ctrl_get_zoom() argument 473 uvc_ctrl_set_zoom(struct uvc_control_mapping * mapping,size_t v4l2_size,const void * v4l2_in,void * uvc_out) uvc_ctrl_set_zoom() argument 490 uvc_ctrl_get_rel_speed(struct uvc_control_mapping * mapping,u8 query,const void * uvc_in,size_t v4l2_size,void * v4l2_out) uvc_ctrl_get_rel_speed() argument 518 uvc_ctrl_set_rel_speed(struct uvc_control_mapping * mapping,size_t v4l2_size,const void * v4l2_in,void * uvc_out) uvc_ctrl_set_rel_speed() argument 624 uvc_get_rect(struct uvc_control_mapping * mapping,u8 query,const void * uvc_in,size_t v4l2_size,void * v4l2_out) uvc_get_rect() argument 645 uvc_set_rect(struct uvc_control_mapping * mapping,size_t v4l2_size,const void * v4l2_in,void * uvc_out) uvc_set_rect() argument 999 uvc_menu_to_v4l2_menu(struct uvc_control_mapping * mapping,s32 val) uvc_menu_to_v4l2_menu() argument 1024 uvc_get_le_value(struct uvc_control_mapping * mapping,u8 query,const void * uvc_in,size_t v4l2_size,void * v4l2_out) uvc_get_le_value() argument 1079 uvc_set_le_value(struct uvc_control_mapping * mapping,size_t v4l2_size,const void * v4l2_in,void * uvc_out) uvc_set_le_value() argument 1140 __uvc_find_control(struct uvc_entity * entity,u32 v4l2_id,struct uvc_control_mapping ** mapping,struct uvc_control ** control,int next,int next_compound) __uvc_find_control() argument 1174 uvc_find_control(struct uvc_video_chain * chain,u32 v4l2_id,struct uvc_control_mapping ** mapping) uvc_find_control() argument 1296 __uvc_ctrl_get(struct uvc_video_chain * chain,struct uvc_control * ctrl,struct uvc_control_mapping * mapping,s32 * value) __uvc_ctrl_get() argument 1360 uvc_ctrl_is_readable(u32 which,struct uvc_control * ctrl,struct uvc_control_mapping * mapping) uvc_ctrl_is_readable() argument 1406 struct uvc_control_mapping *mapping; uvc_ctrl_is_accessible() local 1468 uvc_get_ctrl_bitmap(struct uvc_control * ctrl,struct uvc_control_mapping * mapping) uvc_get_ctrl_bitmap() argument 1494 __uvc_queryctrl_boundaries(struct uvc_video_chain * chain,struct uvc_control * ctrl,struct uvc_control_mapping * mapping,struct v4l2_query_ext_ctrl * v4l2_ctrl) __uvc_queryctrl_boundaries() argument 1566 uvc_mapping_v4l2_size(struct uvc_control_mapping * mapping) uvc_mapping_v4l2_size() argument 1579 __uvc_query_v4l2_ctrl(struct uvc_video_chain * chain,struct uvc_control * ctrl,struct uvc_control_mapping * mapping,struct v4l2_query_ext_ctrl * v4l2_ctrl) __uvc_query_v4l2_ctrl() argument 1662 struct uvc_control_mapping *mapping; uvc_query_v4l2_ctrl() local 1712 struct uvc_control_mapping *mapping; uvc_query_v4l2_menu() local 1782 uvc_ctrl_fill_event(struct uvc_video_chain * chain,struct v4l2_event * ev,struct uvc_control * ctrl,struct uvc_control_mapping * mapping,s32 value,u32 changes) uvc_ctrl_fill_event() argument 1811 uvc_ctrl_send_event(struct uvc_video_chain * chain,struct uvc_fh * handle,struct uvc_control * ctrl,struct uvc_control_mapping * mapping,s32 value,u32 changes) uvc_ctrl_send_event() argument 1838 struct uvc_control_mapping *mapping = NULL; uvc_ctrl_send_slave_event() local 1902 struct uvc_control_mapping *mapping; uvc_ctrl_status_event() local 2000 struct uvc_control_mapping *mapping; uvc_ctrl_send_events() local 2056 struct uvc_control_mapping *mapping; uvc_ctrl_add_event() local 2234 struct uvc_control_mapping *mapping = NULL; uvc_ctrl_find_ctrl_idx() local 2290 uvc_mapping_get_xctrl_compound(struct uvc_video_chain * chain,struct uvc_control * ctrl,struct uvc_control_mapping * mapping,u32 which,struct v4l2_ext_control * xctrl) uvc_mapping_get_xctrl_compound() argument 2353 uvc_mapping_get_xctrl_std(struct uvc_video_chain * chain,struct uvc_control * ctrl,struct uvc_control_mapping * mapping,u32 which,struct v4l2_ext_control * xctrl) uvc_mapping_get_xctrl_std() argument 2391 uvc_mapping_get_xctrl(struct uvc_video_chain * chain,struct uvc_control * ctrl,struct uvc_control_mapping * mapping,u32 which,struct v4l2_ext_control * xctrl) uvc_mapping_get_xctrl() argument 2404 struct uvc_control_mapping *mapping; uvc_ctrl_get() local 2418 uvc_ctrl_clamp(struct uvc_video_chain * chain,struct uvc_control * ctrl,struct uvc_control_mapping * mapping,s32 * value_in_out) uvc_ctrl_clamp() argument 2500 uvc_mapping_set_xctrl_compound(struct uvc_control * ctrl,struct uvc_control_mapping * mapping,struct v4l2_ext_control * xctrl) uvc_mapping_set_xctrl_compound() argument 2523 uvc_mapping_set_xctrl(struct uvc_control * ctrl,struct uvc_control_mapping * mapping,struct v4l2_ext_control * xctrl) uvc_mapping_set_xctrl() argument 2537 struct uvc_control_mapping *mapping; uvc_ctrl_set() local 2945 __uvc_ctrl_add_mapping(struct uvc_video_chain * chain,struct uvc_control * ctrl,const struct uvc_control_mapping * mapping) __uvc_ctrl_add_mapping() argument 3028 uvc_ctrl_add_mapping(struct uvc_video_chain * chain,const struct uvc_control_mapping * mapping) uvc_ctrl_add_mapping() argument 3215 const struct uvc_control_mapping *mapping = &uvc_ctrl_mappings[i]; uvc_ctrl_init_ctrl() local 3338 struct uvc_control_mapping *mapping, *nm; uvc_ctrl_cleanup_mappings() local [all...] |
/linux/include/trace/events/ |
H A D | filemap.h | 32 __entry->i_ino = folio->mapping->host->i_ino; 34 if (folio->mapping->host->i_sb) 35 __entry->s_dev = folio->mapping->host->i_sb->s_dev; 37 __entry->s_dev = folio->mapping->host->i_rdev; 62 struct address_space *mapping, 67 TP_ARGS(mapping, index, last_index), 77 __entry->i_ino = mapping->host->i_ino; 78 if (mapping->host->i_sb) 80 mapping->host->i_sb->s_dev; 82 __entry->s_dev = mapping [all...] |
/linux/tools/testing/selftests/arm64/mte/ |
H A D | check_mmap_options.c | 47 int mapping; member 113 static int check_anonymous_memory_mapping(int mem_type, int mode, int mapping, in check_anonymous_memory_mapping() argument 126 map_ptr = (char *)mte_allocate_memory(map_size, mem_type, mapping, false); in check_anonymous_memory_mapping() 148 static int check_file_memory_mapping(int mem_type, int mode, int mapping, in check_file_memory_mapping() argument 166 map_ptr = (char *)mte_allocate_file_memory(map_size, mem_type, mapping, false, fd); in check_file_memory_mapping() 191 static int check_clear_prot_mte_flag(int mem_type, int mode, int mapping, int atag_check) in check_clear_prot_mte_flag() argument 201 ptr = (char *)mte_allocate_memory_tag_range(sizes[run], mem_type, mapping, in check_clear_prot_mte_flag() 222 ptr = (char *)mte_allocate_file_memory_tag_range(sizes[run], mem_type, mapping, in check_clear_prot_mte_flag() 300 switch (tc->mapping) { in format_test_name() 337 "Check %s with %s mapping, in format_test_name() [all...] |
/linux/drivers/gpu/drm/panfrost/ |
H A D | panfrost_gem.c | 8 #include <linux/dma-mapping.h> 94 struct panfrost_gem_mapping *iter, *mapping = NULL; in panfrost_gem_mapping_get() local 100 mapping = iter; in panfrost_gem_mapping_get() 106 return mapping; in panfrost_gem_mapping_get() 110 panfrost_gem_teardown_mapping(struct panfrost_gem_mapping *mapping) in panfrost_gem_teardown_mapping() argument 112 if (mapping->active) in panfrost_gem_teardown_mapping() 113 panfrost_mmu_unmap(mapping); in panfrost_gem_teardown_mapping() 115 spin_lock(&mapping->mmu->mm_lock); in panfrost_gem_teardown_mapping() 116 if (drm_mm_node_allocated(&mapping->mmnode)) in panfrost_gem_teardown_mapping() 117 drm_mm_remove_node(&mapping in panfrost_gem_teardown_mapping() 123 struct panfrost_gem_mapping *mapping; panfrost_gem_mapping_release() local 133 panfrost_gem_mapping_put(struct panfrost_gem_mapping * mapping) panfrost_gem_mapping_put() argument 143 struct panfrost_gem_mapping *mapping; panfrost_gem_teardown_mappings_locked() local 157 struct panfrost_gem_mapping *mapping; panfrost_gem_open() local 208 struct panfrost_gem_mapping *mapping = NULL, *iter; panfrost_gem_close() local [all...] |
/linux/arch/arm/mm/ |
H A D | dma-mapping.c | 3 * linux/arch/arm/mm/dma-mapping.c 7 * DMA uncached mapping support. 290 * Clear previous low-memory mapping to ensure that the in dma_contiguous_remap() 603 * Free a buffer as defined by the above mapping. 677 * Use the driver DMA support - see dma-mapping.h (dma_sync_*) 754 static int extend_iommu_mapping(struct dma_iommu_mapping *mapping); 756 static inline dma_addr_t __alloc_iova(struct dma_iommu_mapping *mapping, in __alloc_iova() argument 762 size_t mapping_size = mapping->bits << PAGE_SHIFT; in __alloc_iova() 773 spin_lock_irqsave(&mapping->lock, flags); in __alloc_iova() 774 for (i = 0; i < mapping in __alloc_iova() 814 __free_iova(struct dma_iommu_mapping * mapping,dma_addr_t addr,size_t size) __free_iova() argument 965 struct dma_iommu_mapping *mapping = to_dma_iommu_mapping(dev); __iommu_create_mapping() local 1004 struct dma_iommu_mapping *mapping = to_dma_iommu_mapping(dev); __iommu_remove_mapping() local 1189 struct dma_iommu_mapping *mapping = to_dma_iommu_mapping(dev); __map_sg_chunk() local 1372 struct dma_iommu_mapping *mapping = to_dma_iommu_mapping(dev); arm_iommu_map_page() local 1408 struct dma_iommu_mapping *mapping = to_dma_iommu_mapping(dev); arm_iommu_unmap_page() local 1437 struct dma_iommu_mapping *mapping = to_dma_iommu_mapping(dev); arm_iommu_map_resource() local 1471 struct dma_iommu_mapping *mapping = to_dma_iommu_mapping(dev); arm_iommu_unmap_resource() local 1486 struct dma_iommu_mapping *mapping = to_dma_iommu_mapping(dev); arm_iommu_sync_single_for_cpu() local 1501 struct dma_iommu_mapping *mapping = to_dma_iommu_mapping(dev); arm_iommu_sync_single_for_device() local 1551 struct dma_iommu_mapping *mapping; arm_iommu_create_mapping() local 1610 struct dma_iommu_mapping *mapping = release_iommu_mapping() local 1620 extend_iommu_mapping(struct dma_iommu_mapping * mapping) extend_iommu_mapping() argument 1638 arm_iommu_release_mapping(struct dma_iommu_mapping * mapping) arm_iommu_release_mapping() argument 1646 __arm_iommu_attach_device(struct device * dev,struct dma_iommu_mapping * mapping) __arm_iommu_attach_device() argument 1675 arm_iommu_attach_device(struct device * dev,struct dma_iommu_mapping * mapping) arm_iommu_attach_device() argument 1697 struct dma_iommu_mapping *mapping; arm_iommu_detach_device() local 1716 struct dma_iommu_mapping *mapping; arm_setup_iommu_dma_ops() local 1742 struct dma_iommu_mapping *mapping = to_dma_iommu_mapping(dev); arm_teardown_iommu_dma_ops() local [all...] |
H A D | flush.c | 199 void __flush_dcache_folio(struct address_space *mapping, struct folio *folio) in __flush_dcache_folio() argument 202 * Writeback any data associated with the kernel mapping of this in __flush_dcache_folio() 204 * coherent with the kernels mapping. in __flush_dcache_folio() 234 if (mapping && cache_is_vipt_aliasing()) in __flush_dcache_folio() 238 static void __flush_dcache_aliases(struct address_space *mapping, struct folio *folio) in __flush_dcache_aliases() argument 248 * - aliasing VIPT: we only need to find one mapping of this page. in __flush_dcache_aliases() 253 flush_dcache_mmap_lock(mapping); in __flush_dcache_aliases() 254 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff_end) { in __flush_dcache_aliases() 281 flush_dcache_mmap_unlock(mapping); in __flush_dcache_aliases() 289 struct address_space *mapping; in __sync_icache_dcache() local 336 struct address_space *mapping; flush_dcache_folio() local [all...] |
/linux/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_mmu.c | 6 #include <linux/dma-mapping.h> 53 /* unroll mapping in case something went wrong */ in etnaviv_context_map() 113 struct etnaviv_vram_mapping *mapping) in etnaviv_iommu_remove_mapping() argument 115 struct etnaviv_gem_object *etnaviv_obj = mapping->object; in etnaviv_iommu_remove_mapping() 119 etnaviv_iommu_unmap(context, mapping->vram_node.start, in etnaviv_iommu_remove_mapping() 121 drm_mm_remove_node(&mapping->vram_node); in etnaviv_iommu_remove_mapping() 124 void etnaviv_iommu_reap_mapping(struct etnaviv_vram_mapping *mapping) in etnaviv_iommu_reap_mapping() argument 126 struct etnaviv_iommu_context *context = mapping->context; in etnaviv_iommu_reap_mapping() 129 WARN_ON(mapping->use); in etnaviv_iommu_reap_mapping() 131 etnaviv_iommu_remove_mapping(context, mapping); in etnaviv_iommu_reap_mapping() 266 etnaviv_iommu_map_gem(struct etnaviv_iommu_context * context,struct etnaviv_gem_object * etnaviv_obj,u32 memory_base,struct etnaviv_vram_mapping * mapping,u64 va) etnaviv_iommu_map_gem() argument 318 etnaviv_iommu_unmap_gem(struct etnaviv_iommu_context * context,struct etnaviv_vram_mapping * mapping) etnaviv_iommu_unmap_gem() argument 396 etnaviv_iommu_get_suballoc_va(struct etnaviv_iommu_context * context,struct etnaviv_vram_mapping * mapping,u32 memory_base,dma_addr_t paddr,size_t size) etnaviv_iommu_get_suballoc_va() argument 447 etnaviv_iommu_put_suballoc_va(struct etnaviv_iommu_context * context,struct etnaviv_vram_mapping * mapping) etnaviv_iommu_put_suballoc_va() argument [all...] |
H A D | etnaviv_gem.c | 7 #include <linux/dma-mapping.h> 218 struct etnaviv_vram_mapping *mapping; in etnaviv_gem_get_vram_mapping() local 220 list_for_each_entry(mapping, &obj->vram_list, obj_node) { in etnaviv_gem_get_vram_mapping() 221 if (mapping->context == context) in etnaviv_gem_get_vram_mapping() 222 return mapping; in etnaviv_gem_get_vram_mapping() 228 void etnaviv_gem_mapping_unreference(struct etnaviv_vram_mapping *mapping) in etnaviv_gem_mapping_unreference() argument 230 struct etnaviv_gem_object *etnaviv_obj = mapping->object; in etnaviv_gem_mapping_unreference() 233 WARN_ON(mapping->use == 0); in etnaviv_gem_mapping_unreference() 234 mapping->use -= 1; in etnaviv_gem_mapping_unreference() 245 struct etnaviv_vram_mapping *mapping; in etnaviv_gem_mapping_get() local 504 struct etnaviv_vram_mapping *mapping, *tmp; etnaviv_gem_free_object() local [all...] |
/linux/arch/arm64/kvm/ |
H A D | pkvm.c | 301 struct pkvm_mapping *mapping; in __pkvm_pgtable_stage2_unmap() local 307 for_each_mapping_in_range_safe(pgt, start, end, mapping) { in __pkvm_pgtable_stage2_unmap() 308 ret = kvm_call_hyp_nvhe(__pkvm_host_unshare_guest, handle, mapping->gfn, in __pkvm_pgtable_stage2_unmap() 309 mapping->nr_pages); in __pkvm_pgtable_stage2_unmap() 312 pkvm_mapping_remove(mapping, &pgt->pkvm_mappings); in __pkvm_pgtable_stage2_unmap() 313 kfree(mapping); in __pkvm_pgtable_stage2_unmap() 329 struct pkvm_mapping *mapping = NULL; in pkvm_pgtable_stage2_map() local 345 mapping = pkvm_mapping_iter_first(&pgt->pkvm_mappings, addr, addr + size - 1); in pkvm_pgtable_stage2_map() 346 if (mapping) { in pkvm_pgtable_stage2_map() 347 if (size == (mapping in pkvm_pgtable_stage2_map() 381 struct pkvm_mapping *mapping; pkvm_pgtable_stage2_wrprotect() local 398 struct pkvm_mapping *mapping; pkvm_pgtable_stage2_flush() local 412 struct pkvm_mapping *mapping; pkvm_pgtable_stage2_test_clear_young() local [all...] |
/linux/fs/gfs2/ |
H A D | aops.c | 74 struct inode * const inode = folio->mapping->host; in gfs2_write_jdata_folio() 105 struct inode *inode = folio->mapping->host; in __gfs2_jdata_write_folio() 122 * @mapping: The mapping to write 127 int gfs2_jdata_writeback(struct address_space *mapping, struct writeback_control *wbc) in gfs2_jdata_writeback() argument 129 struct inode *inode = mapping->host; in gfs2_jdata_writeback() 131 struct gfs2_sbd *sdp = GFS2_SB(mapping->host); in gfs2_jdata_writeback() 139 while ((folio = writeback_iter(mapping, wbc, folio, &error))) { in gfs2_jdata_writeback() 153 * @mapping: The mapping t 158 gfs2_writepages(struct address_space * mapping,struct writeback_control * wbc) gfs2_writepages() argument 191 gfs2_write_jdata_batch(struct address_space * mapping,struct writeback_control * wbc,struct folio_batch * fbatch,pgoff_t * done_index) gfs2_write_jdata_batch() argument 283 gfs2_write_cache_jdata(struct address_space * mapping,struct writeback_control * wbc) gfs2_write_cache_jdata() argument 365 gfs2_jdata_writepages(struct address_space * mapping,struct writeback_control * wbc) gfs2_jdata_writepages() argument 456 struct address_space *mapping = ip->i_inode.i_mapping; gfs2_internal_read() local 549 gfs2_jdata_dirty_folio(struct address_space * mapping,struct folio * folio) gfs2_jdata_dirty_folio() argument 565 gfs2_bmap(struct address_space * mapping,sector_t lblock) gfs2_bmap() argument 653 struct address_space *mapping = folio->mapping; gfs2_release_folio() local [all...] |
/linux/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_dma.c | 34 * drm_iommu_attach_device- attach device to iommu mapping 40 * mapping. 57 * Keep the original DMA mapping of the sub-device and in drm_iommu_attach_device() 66 ret = arm_iommu_attach_device(subdrv_dev, priv->mapping); in drm_iommu_attach_device() 68 ret = iommu_attach_device(priv->mapping, subdrv_dev); in drm_iommu_attach_device() 75 * drm_iommu_detach_device -detach device address space mapping from device 81 * mapping 92 iommu_detach_device(priv->mapping, subdrv_dev); in drm_iommu_detach_device() 102 DRM_INFO("Exynos DRM: using %s device for DMA mapping operations\n", in exynos_drm_register_dma() 109 if (!priv->mapping) { in exynos_drm_register_dma() 110 void *mapping = NULL; exynos_drm_register_dma() local [all...] |
/linux/Documentation/arch/powerpc/ |
H A D | vmemmap_dedup.rst | 14 With 2M PMD level mapping, we require 32 struct pages and a single 64K vmemmap 18 With 1G PUD level mapping, we require 16384 struct pages and a single 64K 20 require 16 64K pages in vmemmap to map the struct page for 1G PUD level mapping. 23 +-----------+ ---virt_to_page---> +-----------+ mapping to +-----------+ 35 | mapping | +-----------+ | | 46 With 4K page size, 2M PMD level mapping requires 512 struct pages and a single 48 require 8 4K pages in vmemmap to map the struct page for 2M pmd level mapping. 52 +-----------+ ---virt_to_page---> +-----------+ mapping to +-----------+ 64 | mapping | +-----------+ | | 74 With 1G PUD level mapping, w [all...] |
/linux/Documentation/translations/zh_CN/mm/ |
H A D | page_migration.rst | 143 2. ``int (*migratepage) (struct address_space *mapping,`` 168 void __SetPageMovable(struct page *page, struct address_space *mapping) 175 page->mapping = page->mapping | PAGE_MAPPING_MOVABLE; 177 所以驱动不应该直接访问page->mapping。相反,驱动应该使用page_mapping(),它可 190 mapping->a_ops->isolate_page。lock_page()可以防止突然破坏page->mapping。
|
/linux/drivers/net/wireless/marvell/mwifiex/ |
H A D | util.h | 57 struct mwifiex_dma_mapping *mapping) in mwifiex_store_mapping() argument 61 memcpy(&cb->dma_mapping, mapping, sizeof(*mapping)); in mwifiex_store_mapping() 65 struct mwifiex_dma_mapping *mapping) in mwifiex_get_mapping() argument 69 memcpy(mapping, &cb->dma_mapping, sizeof(*mapping)); in mwifiex_get_mapping() 74 struct mwifiex_dma_mapping mapping; in MWIFIEX_SKB_DMA_ADDR() local 76 mwifiex_get_mapping(skb, &mapping); in MWIFIEX_SKB_DMA_ADDR() 78 return mapping.addr; in MWIFIEX_SKB_DMA_ADDR()
|
/linux/drivers/sh/clk/ |
H A D | core.c | 340 struct clk_mapping *mapping = clk->mapping; in clk_establish_mapping() local 345 if (!mapping) { in clk_establish_mapping() 349 * dummy mapping for root clocks with no specified ranges in clk_establish_mapping() 352 clk->mapping = &dummy_mapping; in clk_establish_mapping() 357 * If we're on a child clock and it provides no mapping of its in clk_establish_mapping() 358 * own, inherit the mapping from its root clock. in clk_establish_mapping() 361 mapping = clkp->mapping; in clk_establish_mapping() 362 BUG_ON(!mapping); in clk_establish_mapping() 390 struct clk_mapping *mapping; clk_destroy_mapping() local 399 struct clk_mapping *mapping = clk->mapping; clk_teardown_mapping() local [all...] |
/linux/Documentation/driver-api/ |
H A D | io-mapping.rst | 8 The io_mapping functions in linux/io-mapping.h provide an abstraction for 9 efficiently mapping small regions of an I/O device to the CPU. The initial 14 A mapping object is created during driver initialization using:: 20 mappable, while 'size' indicates how large a mapping region to 23 This _wc variant provides a mapping which may only be used with 27 With this mapping object, individual pages can be mapped either temporarily 31 void *io_mapping_map_local_wc(struct io_mapping *mapping, 34 void *io_mapping_map_atomic_wc(struct io_mapping *mapping, 37 'offset' is the offset within the defined mapping region. Accessing 46 Temporary mappings are only valid in the context of the caller. The mapping [all...] |
/linux/arch/nios2/include/asm/ |
H A D | cacheflush.h | 54 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) argument 55 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages) argument 56 #define flush_dcache_mmap_lock_irqsave(mapping, flags) \ argument 57 xa_lock_irqsave(&mapping->i_pages, flags) 58 #define flush_dcache_mmap_unlock_irqrestore(mapping, flags) \ argument 59 xa_unlock_irqrestore(&mapping->i_pages, flags)
|
/linux/fs/ |
H A D | dax.c | 181 * @entry may no longer be the entry at the index in the mapping. 346 * A DAX folio is considered shared if it has no mapping set and ->share (which 353 return !folio->mapping && folio->share; in dax_folio_is_shared() 359 * previously been associated with any mappings the ->mapping and ->index 360 * fields will be set. If it has already been associated with a mapping 361 * the mapping will be cleared and the share count set. It's then up to 363 * recover ->mapping and ->index information. For example by implementing 370 * folio->mapping. in dax_folio_make_shared() 372 folio->mapping = NULL; in dax_folio_make_shared() 394 folio->mapping in dax_folio_put() 441 dax_associate_entry(void * entry,struct address_space * mapping,struct vm_area_struct * vma,unsigned long address,bool shared) dax_associate_entry() argument 468 dax_disassociate_entry(void * entry,struct address_space * mapping,bool trunc) dax_disassociate_entry() argument 508 struct address_space *mapping = READ_ONCE(folio->mapping); dax_lock_folio() local 549 struct address_space *mapping = folio->mapping; dax_unlock_folio() local 567 dax_lock_mapping_entry(struct address_space * mapping,pgoff_t index,struct page ** page) dax_lock_mapping_entry() argument 610 dax_unlock_mapping_entry(struct address_space * mapping,pgoff_t index,dax_entry_t cookie) dax_unlock_mapping_entry() argument 651 grab_mapping_entry(struct xa_state * xas,struct address_space * mapping,unsigned int order) grab_mapping_entry() argument 754 dax_layout_busy_page_range(struct address_space * mapping,loff_t start,loff_t end) dax_layout_busy_page_range() argument 809 dax_layout_busy_page(struct address_space * mapping) dax_layout_busy_page() argument 815 __dax_invalidate_entry(struct address_space * mapping,pgoff_t index,bool trunc) __dax_invalidate_entry() argument 840 __dax_clear_dirty_range(struct address_space * mapping,pgoff_t start,pgoff_t end) __dax_clear_dirty_range() argument 873 dax_delete_mapping_entry(struct address_space * mapping,pgoff_t index) dax_delete_mapping_entry() argument 888 dax_delete_mapping_range(struct address_space * mapping,loff_t start,loff_t end) dax_delete_mapping_range() argument 992 dax_invalidate_mapping_entry_sync(struct address_space * mapping,pgoff_t index) dax_invalidate_mapping_entry_sync() argument 1046 struct address_space *mapping = vmf->vma->vm_file->f_mapping; dax_insert_entry() local 1101 dax_writeback_one(struct xa_state * xas,struct dax_device * dax_dev,struct address_space * mapping,void * entry) dax_writeback_one() argument 1199 dax_writeback_mapping_range(struct address_space * mapping,struct dax_device * dax_dev,struct writeback_control * wbc) dax_writeback_mapping_range() argument 1377 struct address_space *mapping = vmf->vma->vm_file->f_mapping; dax_pmd_load_hole() local 1889 struct address_space *mapping = vmf->vma->vm_file->f_mapping; dax_iomap_pte_fault() local 2001 struct address_space *mapping = vmf->vma->vm_file->f_mapping; dax_iomap_pmd_fault() local 2124 struct address_space *mapping = vmf->vma->vm_file->f_mapping; dax_insert_pfn_mkwrite() local [all...] |