| /linux/drivers/android/ |
| H A D | binder_trace.h | 319 TP_PROTO(const struct binder_alloc *alloc, size_t page_index), 320 TP_ARGS(alloc, page_index), 323 __field(size_t, page_index) 327 __entry->page_index = page_index; 330 __entry->proc, __entry->page_index) 334 TP_PROTO(const struct binder_alloc *alloc, size_t page_index), 335 TP_ARGS(alloc, page_index)); 338 TP_PROTO(const struct binder_alloc *alloc, size_t page_index), 339 TP_ARGS(alloc, page_index)); 342 TP_PROTO(const struct binder_alloc *alloc, size_t page_index), [all …]
|
| H A D | binder_alloc.h | 69 unsigned long page_index; member
|
| /linux/arch/x86/kernel/cpu/sgx/ |
| H A D | encl.c | 15 static int sgx_encl_lookup_backing(struct sgx_encl *encl, unsigned long page_index, 111 unsigned long page_index) in sgx_encl_get_backing_page_pcmd_offset() argument 115 return epc_end_off + page_index * sizeof(struct sgx_pcmd); in sgx_encl_get_backing_page_pcmd_offset() 121 static inline void sgx_encl_truncate_backing_page(struct sgx_encl *encl, unsigned long page_index) in sgx_encl_truncate_backing_page() argument 125 shmem_truncate_range(inode, PFN_PHYS(page_index), PFN_PHYS(page_index) + PAGE_SIZE - 1); in sgx_encl_truncate_backing_page() 138 pgoff_t page_index, page_pcmd_off; in __sgx_encl_eldu() local 147 page_index = PFN_DOWN(encl_page->desc - encl_page->encl->base); in __sgx_encl_eldu() 149 page_index = PFN_DOWN(encl->size); in __sgx_encl_eldu() 154 pcmd_first_page = PFN_PHYS(page_index & ~PCMD_FIRST_MASK) + encl->base; in __sgx_encl_eldu() 156 page_pcmd_off = sgx_encl_get_backing_page_pcmd_offset(encl, page_index); in __sgx_encl_eldu() [all …]
|
| H A D | encl.h | 110 int sgx_encl_alloc_backing(struct sgx_encl *encl, unsigned long page_index,
|
| H A D | main.c | 306 pgoff_t page_index; in sgx_reclaim_pages() local 338 page_index = PFN_DOWN(encl_page->desc - encl_page->encl->base); in sgx_reclaim_pages() 341 ret = sgx_encl_alloc_backing(encl_page->encl, page_index, &backing[i]); in sgx_reclaim_pages()
|
| /linux/sound/firewire/ |
| H A D | packets-buffer.c | 26 unsigned int i, page_index, offset_in_page; in iso_packets_buffer_init() local 50 page_index = i / packets_per_page; in iso_packets_buffer_init() 51 p = page_address(b->iso_buffer.pages[page_index]); in iso_packets_buffer_init() 54 b->packets[i].offset = page_index * PAGE_SIZE + offset_in_page; in iso_packets_buffer_init()
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | pin_system.c | 291 size_t page_index; in add_mapping_to_sdma_packet() local 301 page_index = PFN_DOWN(start - cache_entry->rb.addr); in add_mapping_to_sdma_packet() 303 if (page_index >= cache_entry->npages) { in add_mapping_to_sdma_packet() 306 page_index, cache_entry->npages); in add_mapping_to_sdma_packet() 326 cache_entry->pages[page_index], in add_mapping_to_sdma_packet() 338 ret, page_index, page_offset, from_this_page); in add_mapping_to_sdma_packet()
|
| /linux/drivers/android/tests/ |
| H A D | binder_alloc_kunit.c | 147 int page_index; in check_buffer_pages_allocated() local 152 page_index = (page_addr - alloc->vm_start) / PAGE_SIZE; in check_buffer_pages_allocated() 153 if (!alloc->pages[page_index] || in check_buffer_pages_allocated() 154 !list_empty(page_to_lru(alloc->pages[page_index]))) { in check_buffer_pages_allocated() 156 alloc->pages[page_index] ? in check_buffer_pages_allocated() 157 "lru" : "free", page_index); in check_buffer_pages_allocated()
|
| /linux/drivers/md/dm-vdo/ |
| H A D | block-map.c | 50 page_number_t page_index; member 81 page_number_t page_index; member 1374 page_number_t page_index) in get_tree_page_by_index() argument 1382 if (page_index < border) { in get_tree_page_by_index() 1385 return &(tree->segments[segment].levels[height - 1][page_index - offset]); in get_tree_page_by_index() 1400 lock->tree_slots[lock->height].page_index); in get_tree_page() 1809 lock->tree_slots[lock->height - 1].page_index, in continue_with_loaded_page() 1917 .page_index = tree_slot.page_index, in attempt_page_lock() 2240 page_number_t page_index; in vdo_find_block_map_slot() local 2255 page_index = (lock->tree_slots[0].page_index / zone->block_map->root_count); in vdo_find_block_map_slot() [all …]
|
| H A D | data-vio.h | 64 page_number_t page_index; member
|
| /linux/drivers/android/binder/ |
| H A D | page_range.rs | 522 let mut page_index = offset >> PAGE_SHIFT; in iterate() localVariable 527 let page_info = unsafe { pages.add(page_index) }; in iterate() 537 page_index += 1; in iterate() 706 let page_index; in rust_shrink_free_page() localVariable 746 page_index = unsafe { info.offset_from(inner.pages) } as usize; in rust_shrink_free_page() 766 let user_page_addr = vma_addr + (page_index << PAGE_SHIFT); in rust_shrink_free_page()
|
| /linux/drivers/gpu/drm/vc4/ |
| H A D | vc4_bo.c | 200 uint32_t page_index = bo_page_index(size); in vc4_get_cache_list_for_size() local 202 if (vc4->bo_cache.size_list_size <= page_index) { in vc4_get_cache_list_for_size() 204 page_index + 1); in vc4_get_cache_list_for_size() 233 return &vc4->bo_cache.size_list[page_index]; in vc4_get_cache_list_for_size() 364 uint32_t page_index = bo_page_index(size); in vc4_bo_get_from_cache() local 368 if (page_index >= vc4->bo_cache.size_list_size) in vc4_bo_get_from_cache() 371 if (list_empty(&vc4->bo_cache.size_list[page_index])) in vc4_bo_get_from_cache() 374 bo = list_first_entry(&vc4->bo_cache.size_list[page_index], in vc4_bo_get_from_cache()
|
| /linux/fs/ecryptfs/ |
| H A D | read_write.c | 247 pgoff_t page_index, in ecryptfs_read_lower_page_segment() argument 255 offset = (loff_t)page_index * PAGE_SIZE + offset_in_page; in ecryptfs_read_lower_page_segment()
|
| H A D | crypto.c | 301 pgoff_t page_index, in crypt_extent() argument 310 extent_base = (((loff_t)page_index) * (PAGE_SIZE / extent_size)); in crypt_extent() 326 "rc = [%d]\n", __func__, page_index, extent_offset, rc); in crypt_extent()
|
| /linux/tools/testing/selftests/mm/ |
| H A D | gup_test.c | 193 long page_index = strtol(argv[optind], 0, 0) + 1; in main() local 195 gup.which_pages[extra_arg_count] = page_index; in main()
|
| /linux/include/linux/qed/ |
| H A D | qed_chain.h | 286 u32 page_index = 0; in qed_chain_advance_page() local 305 page_index = *(u16 *)page_to_inc; in qed_chain_advance_page() 309 page_index = *(u32 *)page_to_inc; in qed_chain_advance_page() 311 *p_next_elem = p_chain->pbl.pp_addr_tbl[page_index].virt_addr; in qed_chain_advance_page()
|
| /linux/fs/jfs/ |
| H A D | jfs_metapage.c | 692 unsigned long page_index; in __get_metapage() local 700 page_index = lblock >> l2BlocksPerPage; in __get_metapage() 701 page_offset = (lblock - (page_index << l2BlocksPerPage)) << l2bsize; in __get_metapage() 722 folio = filemap_grab_folio(mapping, page_index); in __get_metapage() 729 folio = read_mapping_folio(mapping, page_index, NULL); in __get_metapage()
|
| /linux/drivers/net/ethernet/mellanox/mlxsw/ |
| H A D | pci.c | 291 int page_index) in __mlxsw_pci_queue_page_get() argument 293 return q->mem_item.mapaddr + MLXSW_PCI_PAGE_SIZE * page_index; in __mlxsw_pci_queue_page_get() 401 int page_index = 0; in mlxsw_pci_rdq_build_skb() local 411 page_pool_dma_sync_for_cpu(page_pool, pages[page_index], in mlxsw_pci_rdq_build_skb() 414 data = page_address(pages[page_index]); in mlxsw_pci_rdq_build_skb() 428 page_index++; in mlxsw_pci_rdq_build_skb() 434 page = pages[page_index]; in mlxsw_pci_rdq_build_skb() 440 page_index++; in mlxsw_pci_rdq_build_skb()
|
| /linux/fs/btrfs/ |
| H A D | raid56.c | 364 int page_index = offset >> PAGE_SHIFT; in index_stripe_sectors() local 366 ASSERT(page_index < rbio->nr_pages); in index_stripe_sectors() 367 if (!rbio->stripe_pages[page_index]) in index_stripe_sectors() 370 rbio->stripe_paddrs[i] = page_to_phys(rbio->stripe_pages[page_index]) + in index_stripe_sectors() 2613 const unsigned int page_index = (i * step) >> PAGE_SHIFT; in alloc_rbio_sector_pages() local 2616 if (rbio->stripe_pages[page_index]) in alloc_rbio_sector_pages() 2621 rbio->stripe_pages[page_index] = page; in alloc_rbio_sector_pages()
|
| /linux/drivers/misc/vmw_vmci/ |
| H A D | vmci_queue_pair.c | 340 const u64 page_index = in qp_memcpy_to_queue_iter() local 348 va = kmap_local_page(kernel_if->u.h.page[page_index]); in qp_memcpy_to_queue_iter() 350 va = kernel_if->u.g.vas[page_index + 1]; in qp_memcpy_to_queue_iter() 387 const u64 page_index = in qp_memcpy_from_queue_iter() local 396 va = kmap_local_page(kernel_if->u.h.page[page_index]); in qp_memcpy_from_queue_iter() 398 va = kernel_if->u.g.vas[page_index + 1]; in qp_memcpy_from_queue_iter()
|
| /linux/net/ceph/ |
| H A D | messenger.c | 846 cursor->page_index = 0; in ceph_msg_data_pages_cursor_init() 860 BUG_ON(cursor->page_index >= cursor->page_count); in ceph_msg_data_pages_next() 865 return data->pages[cursor->page_index]; in ceph_msg_data_pages_next() 887 BUG_ON(cursor->page_index >= cursor->page_count); in ceph_msg_data_pages_advance() 888 cursor->page_index++; in ceph_msg_data_pages_advance()
|
| /linux/Documentation/admin-guide/blockdev/ |
| H A D | zram.rst | 375 echo "page_index=1251" > /sys/block/zramX/writeback 388 We also now permit multiple page_index params per call and a mix of 391 echo page_index=42 page_index=99 page_indexes=100-200 \
|
| /linux/fs/nfs/ |
| H A D | nfstrace.h | 352 pgoff_t page_index, 356 TP_ARGS(file, verifier, cookie, page_index, dtsize), 384 __entry->index = page_index; 405 pgoff_t page_index, \ 408 TP_ARGS(file, verifier, cookie, page_index, dtsize))
|
| /linux/mm/kfence/ |
| H A D | core.c | 1265 const int page_index = (addr - (unsigned long)__kfence_pool) / PAGE_SIZE; in kfence_handle_page_fault() local 1279 if (page_index % 2) { in kfence_handle_page_fault()
|
| /linux/include/linux/ceph/ |
| H A D | messenger.h | 247 unsigned short page_index; /* index in array */ member
|