| /linux/arch/mips/mm/ |
| H A D | mmap.c | 23 #define COLOUR_ALIGN(addr, pgoff) \ argument 25 (((pgoff) << PAGE_SHIFT) & shm_align_mask)) 30 unsigned long addr0, unsigned long len, unsigned long pgoff, in arch_get_unmapped_area_common() argument 52 ((addr - (pgoff << PAGE_SHIFT)) & shm_align_mask)) in arch_get_unmapped_area_common() 64 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area_common() 76 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area_common() 101 unsigned long len, unsigned long pgoff, unsigned long flags, in arch_get_unmapped_area() argument 105 addr0, len, pgoff, flags, UP); in arch_get_unmapped_area() 113 unsigned long addr0, unsigned long len, unsigned long pgoff, in arch_get_unmapped_area_topdown() argument 117 addr0, len, pgoff, flags, DOWN); in arch_get_unmapped_area_topdown()
|
| /linux/arch/loongarch/mm/ |
| H A D | mmap.c | 15 #define COLOUR_ALIGN(addr, pgoff) \ argument 17 + (((pgoff) << PAGE_SHIFT) & SHM_ALIGN_MASK)) 22 unsigned long addr0, unsigned long len, unsigned long pgoff, in arch_get_unmapped_area_common() argument 44 ((addr - (pgoff << PAGE_SHIFT)) & SHM_ALIGN_MASK)) in arch_get_unmapped_area_common() 56 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area_common() 67 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area_common() 96 unsigned long len, unsigned long pgoff, unsigned long flags, in arch_get_unmapped_area() argument 100 addr0, len, pgoff, flags, UP); in arch_get_unmapped_area() 108 unsigned long addr0, unsigned long len, unsigned long pgoff, in arch_get_unmapped_area_topdown() argument 112 addr0, len, pgoff, flags, DOWN); in arch_get_unmapped_area_topdown()
|
| /linux/tools/testing/nvdimm/ |
| H A D | dax-dev.c | 9 phys_addr_t dax_pgoff_to_phys(struct dev_dax *dev_dax, pgoff_t pgoff, in dax_pgoff_to_phys() argument 20 pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1; in dax_pgoff_to_phys() 21 if (pgoff < dax_range->pgoff || pgoff > pgoff_end) in dax_pgoff_to_phys() 23 addr = PFN_PHYS(pgoff - dax_range->pgoff) + range->start; in dax_pgoff_to_phys()
|
| H A D | pmem-dax.c | 11 long __pmem_direct_access(struct pmem_device *pmem, pgoff_t pgoff, in __pmem_direct_access() argument 15 resource_size_t offset = PFN_PHYS(pgoff) + pmem->data_offset; in __pmem_direct_access() 17 if (unlikely(is_bad_pmem(&pmem->bb, PFN_PHYS(pgoff) / 512, in __pmem_direct_access() 34 __func__, pmem, pgoff, page_to_pfn(page)); in __pmem_direct_access()
|
| /linux/drivers/dax/ |
| H A D | device.c | 53 __weak phys_addr_t dax_pgoff_to_phys(struct dev_dax *dev_dax, pgoff_t pgoff, in dax_pgoff_to_phys() argument 64 pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1; in dax_pgoff_to_phys() 65 if (pgoff < dax_range->pgoff || pgoff > pgoff_end) in dax_pgoff_to_phys() 67 phys = PFN_PHYS(pgoff - dax_range->pgoff) + range->start; in dax_pgoff_to_phys() 81 pgoff_t pgoff; in dax_set_mapping() local 87 pgoff = linear_page_index(vmf->vma, in dax_set_mapping() 97 folio->index = pgoff + i; in dax_set_mapping() 121 phys = dax_pgoff_to_phys(dev_dax, vmf->pgoff, PAGE_SIZE); in __dev_dax_pte_fault() 123 dev_dbg(dev, "pgoff_to_phys(%#lx) failed\n", vmf->pgoff); in __dev_dax_pte_fault() 141 pgoff_t pgoff; in __dev_dax_pmd_fault() local [all …]
|
| /linux/arch/arm/mm/ |
| H A D | mmap.c | 16 #define COLOUR_ALIGN(addr,pgoff) \ argument 18 (((pgoff)<<PAGE_SHIFT) & (SHMLBA-1))) 31 unsigned long len, unsigned long pgoff, in arch_get_unmapped_area() argument 52 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) in arch_get_unmapped_area() 62 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area() 76 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area() 82 const unsigned long len, const unsigned long pgoff, in arch_get_unmapped_area_topdown() argument 105 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) in arch_get_unmapped_area_topdown() 113 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area_topdown() 127 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area_topdown()
|
| /linux/arch/microblaze/kernel/ |
| H A D | sys_microblaze.c | 38 unsigned long, pgoff) in SYSCALL_DEFINE6() argument 40 if (pgoff & ~PAGE_MASK) in SYSCALL_DEFINE6() 43 return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff >> PAGE_SHIFT); in SYSCALL_DEFINE6() 48 unsigned long, pgoff) in SYSCALL_DEFINE6() argument 50 if (pgoff & (~PAGE_MASK >> 12)) in SYSCALL_DEFINE6() 54 pgoff >> (PAGE_SHIFT - 12)); in SYSCALL_DEFINE6()
|
| /linux/drivers/md/ |
| H A D | dm-linear.c | 162 static struct dax_device *linear_dax_pgoff(struct dm_target *ti, pgoff_t *pgoff) in linear_dax_pgoff() argument 165 sector_t sector = linear_map_sector(ti, *pgoff << PAGE_SECTORS_SHIFT); in linear_dax_pgoff() 167 *pgoff = (get_start_sect(lc->dev->bdev) + sector) >> PAGE_SECTORS_SHIFT; in linear_dax_pgoff() 171 static long linear_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, in linear_dax_direct_access() argument 175 struct dax_device *dax_dev = linear_dax_pgoff(ti, &pgoff); in linear_dax_direct_access() 177 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in linear_dax_direct_access() 180 static int linear_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, in linear_dax_zero_page_range() argument 183 struct dax_device *dax_dev = linear_dax_pgoff(ti, &pgoff); in linear_dax_zero_page_range() 185 return dax_zero_page_range(dax_dev, pgoff, nr_pages); in linear_dax_zero_page_range() 188 static size_t linear_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff, in linear_dax_recovery_write() argument [all …]
|
| H A D | dm-stripe.c | 302 static struct dax_device *stripe_dax_pgoff(struct dm_target *ti, pgoff_t *pgoff) in stripe_dax_pgoff() argument 309 stripe_map_sector(sc, *pgoff * PAGE_SECTORS, &stripe, &dev_sector); in stripe_dax_pgoff() 313 *pgoff = (get_start_sect(bdev) + dev_sector) >> PAGE_SECTORS_SHIFT; in stripe_dax_pgoff() 317 static long stripe_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, in stripe_dax_direct_access() argument 321 struct dax_device *dax_dev = stripe_dax_pgoff(ti, &pgoff); in stripe_dax_direct_access() 323 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in stripe_dax_direct_access() 326 static int stripe_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, in stripe_dax_zero_page_range() argument 329 struct dax_device *dax_dev = stripe_dax_pgoff(ti, &pgoff); in stripe_dax_zero_page_range() 331 return dax_zero_page_range(dax_dev, pgoff, nr_pages); in stripe_dax_zero_page_range() 334 static size_t stripe_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff, in stripe_dax_recovery_write() argument [all …]
|
| /linux/arch/sh/mm/ |
| H A D | mmap.c | 46 unsigned long pgoff) in COLOUR_ALIGN() argument 49 unsigned long off = (pgoff << PAGE_SHIFT) & shm_align_mask; in COLOUR_ALIGN() 55 unsigned long len, unsigned long pgoff, unsigned long flags, in arch_get_unmapped_area() argument 68 ((addr - (pgoff << PAGE_SHIFT)) & shm_align_mask)) in arch_get_unmapped_area() 82 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area() 96 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area() 102 const unsigned long len, const unsigned long pgoff, in arch_get_unmapped_area_topdown() argument 116 ((addr - (pgoff << PAGE_SHIFT)) & shm_align_mask)) in arch_get_unmapped_area_topdown() 131 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area_topdown() 146 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area_topdown()
|
| /linux/tools/testing/radix-tree/ |
| H A D | iteration_check.c | 51 unsigned long pgoff; in add_entries_fn() local 53 for (pgoff = 0; pgoff < MAX_IDX; pgoff++) { in add_entries_fn() 54 my_item_insert(&array, pgoff); in add_entries_fn() 138 int pgoff; in remove_entries_fn() local 141 pgoff = rand_r(&seeds[2]) % MAX_IDX; in remove_entries_fn() 143 item = xa_erase(&array, pgoff); in remove_entries_fn() 145 item_free(item, pgoff); in remove_entries_fn()
|
| /linux/include/trace/events/ |
| H A D | fs_dax.h | 20 __field(pgoff_t, pgoff) 34 __entry->pgoff = vmf->pgoff; 48 __entry->pgoff, 112 __field(pgoff_t, pgoff) 123 __entry->pgoff = vmf->pgoff; 133 __entry->pgoff, 182 TP_PROTO(struct inode *inode, pgoff_t pgoff, pgoff_t pglen), 183 TP_ARGS(inode, pgoff, pglen), 186 __field(pgoff_t, pgoff) 193 __entry->pgoff = pgoff; [all …]
|
| /linux/kernel/bpf/ |
| H A D | arena.c | 285 ret = range_tree_clear(&arena->rt, vmf->pgoff, 1); in arena_vm_fault() 292 range_tree_set(&arena->rt, vmf->pgoff, 1); in arena_vm_fault() 298 range_tree_set(&arena->rt, vmf->pgoff, 1); in arena_vm_fault() 315 unsigned long len, unsigned long pgoff, in arena_get_unmapped_area() argument 322 if (pgoff) in arena_get_unmapped_area() 432 long pgoff = 0; in arena_alloc_pages() local 442 pgoff = compute_pgoff(arena, uaddr); in arena_alloc_pages() 443 if (pgoff > page_cnt_max - page_cnt) in arena_alloc_pages() 456 ret = is_range_tree_set(&arena->rt, pgoff, page_cnt); in arena_alloc_pages() 459 ret = range_tree_clear(&arena->rt, pgoff, page_cnt); in arena_alloc_pages() [all …]
|
| /linux/arch/csky/abiv1/ |
| H A D | mmap.c | 12 #define COLOUR_ALIGN(addr,pgoff) \ argument 14 (((pgoff)<<PAGE_SHIFT) & (SHMLBA-1))) 26 unsigned long len, unsigned long pgoff, in arch_get_unmapped_area() argument 36 .align_offset = pgoff << PAGE_SHIFT in arch_get_unmapped_area() 50 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) in arch_get_unmapped_area() 60 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area()
|
| /linux/arch/parisc/kernel/ |
| H A D | sys_parisc.c | 38 unsigned long pgoff) in shared_align_offset() argument 40 return (filp_pgoff + pgoff) << PAGE_SHIFT; in shared_align_offset() 44 unsigned long filp_pgoff, unsigned long pgoff) in COLOR_ALIGN() argument 48 shared_align_offset(filp_pgoff, pgoff); in COLOR_ALIGN() 100 unsigned long addr, unsigned long len, unsigned long pgoff, in arch_get_unmapped_area_common() argument 125 (addr - shared_align_offset(filp_pgoff, pgoff)) in arch_get_unmapped_area_common() 133 addr = COLOR_ALIGN(addr, filp_pgoff, pgoff); in arch_get_unmapped_area_common() 145 info.align_offset = shared_align_offset(filp_pgoff, pgoff); in arch_get_unmapped_area_common() 170 unsigned long len, unsigned long pgoff, unsigned long flags, in arch_get_unmapped_area() argument 174 addr, len, pgoff, flags, UP); in arch_get_unmapped_area() [all …]
|
| /linux/arch/xtensa/kernel/ |
| H A D | syscall.c | 35 #define COLOUR_ALIGN(addr, pgoff) \ argument 37 (((pgoff) << PAGE_SHIFT) & (SHMLBA - 1))) 58 unsigned long len, unsigned long pgoff, unsigned long flags, in arch_get_unmapped_area() argument 69 ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))) in arch_get_unmapped_area() 80 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area() 92 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area()
|
| /linux/mm/ |
| H A D | mmap.c | 265 unsigned long pgoff, unsigned long len) in file_mmap_ok() argument 272 if (pgoff > maxsize >> PAGE_SHIFT) in file_mmap_ok() 337 unsigned long pgoff, unsigned long *populate, in do_mmap() argument 373 if ((pgoff + (len >> PAGE_SHIFT)) < pgoff) in do_mmap() 406 addr = __get_unmapped_area(file, addr, len, pgoff, flags, vm_flags); in do_mmap() 427 if (!file_mmap_ok(file, inode, pgoff, len)) in do_mmap() 500 pgoff = 0; in do_mmap() 537 pgoff = addr >> PAGE_SHIFT; in do_mmap() 558 addr = mmap_region(file, addr, len, vm_flags, pgoff, uf); in do_mmap() 568 unsigned long fd, unsigned long pgoff) in ksys_mmap_pgoff() argument [all …]
|
| /linux/drivers/w1/slaves/ |
| H A D | w1_ds250x.c | 49 int pgoff = pageno * W1_PAGE_SIZE; in w1_ds2502_read_page() local 63 buf[1] = pgoff & 0xff; in w1_ds2502_read_page() 64 buf[2] = pgoff >> 8; in w1_ds2502_read_page() 71 w1_read_block(sl->master, &data->eprom[pgoff], W1_PAGE_SIZE); in w1_ds2502_read_page() 74 if (w1_calc_crc8(&data->eprom[pgoff], W1_PAGE_SIZE) != crc8) in w1_ds2502_read_page() 88 int pgoff, epoff; in w1_ds2505_read_page() local 97 epoff = pgoff = pageno * W1_PAGE_SIZE; in w1_ds2505_read_page() 105 buf[1] = pgoff & 0xff; in w1_ds2505_read_page() 106 buf[2] = pgoff >> 8; in w1_ds2505_read_page() 121 pgoff = (redir ^ 0xff) * W1_PAGE_SIZE; in w1_ds2505_read_page()
|
| /linux/fs/romfs/ |
| H A D | mmap-nommu.c | 21 unsigned long pgoff, in romfs_get_unmapped_area() argument 35 offset = pgoff << PAGE_SHIFT; in romfs_get_unmapped_area() 38 if ((pgoff >= maxpages) || (maxpages - pgoff < lpages)) in romfs_get_unmapped_area() 44 if (len > mtd->size || pgoff >= (mtd->size >> PAGE_SHIFT)) in romfs_get_unmapped_area()
|
| /linux/arch/sparc/kernel/ |
| H A D | sys_sparc_32.c | 43 …a(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long fla… in arch_get_unmapped_area() argument 56 ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))) in arch_get_unmapped_area() 73 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area() 111 unsigned long, pgoff) in SYSCALL_DEFINE6() argument 116 pgoff >> (PAGE_SHIFT - 12)); in SYSCALL_DEFINE6() 128 unsigned long, prot, unsigned long, pgoff, in SYSCALL_DEFINE5() argument 135 (pgoff >> (PAGE_SHIFT - 12)), flags); in SYSCALL_DEFINE5()
|
| H A D | sys_sparc_64.c | 83 unsigned long pgoff) in COLOR_ALIGN() argument 86 unsigned long off = (pgoff<<PAGE_SHIFT) & (SHMLBA-1); in COLOR_ALIGN() 101 …a(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long fla… in arch_get_unmapped_area() argument 118 ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))) in arch_get_unmapped_area() 134 addr = COLOR_ALIGN(addr, pgoff); in arch_get_unmapped_area() 149 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area() 164 const unsigned long len, const unsigned long pgoff, in arch_get_unmapped_area_topdown() argument 186 ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))) in arch_get_unmapped_area_topdown() 201 addr = COLOR_ALIGN(addr, pgoff); in arch_get_unmapped_area_topdown() 217 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area_topdown() [all …]
|
| /linux/io_uring/ |
| H A D | memmap.c | 259 loff_t pgoff) in io_mmap_get_region() argument 261 loff_t offset = pgoff << PAGE_SHIFT; in io_mmap_get_region() 296 static void *io_uring_validate_mmap_request(struct file *file, loff_t pgoff, in io_uring_validate_mmap_request() argument 302 region = io_mmap_get_region(ctx, pgoff); in io_uring_validate_mmap_request() 348 unsigned long len, unsigned long pgoff, in io_uring_get_unmapped_area() argument 364 ptr = io_uring_validate_mmap_request(filp, pgoff, len); in io_uring_get_unmapped_area() 383 pgoff = 0; /* has been translated to ptr above */ in io_uring_get_unmapped_area() 386 pgoff = addr >> PAGE_SHIFT; in io_uring_get_unmapped_area() 390 return mm_get_unmapped_area(current->mm, filp, addr, len, pgoff, flags); in io_uring_get_unmapped_area() 406 unsigned long len, unsigned long pgoff, in io_uring_get_unmapped_area() argument [all …]
|
| /linux/drivers/dma-buf/ |
| H A D | udmabuf.c | 51 pgoff_t pgoff = vmf->pgoff; in udmabuf_vm_fault() local 55 if (pgoff >= ubuf->pagecount) in udmabuf_vm_fault() 58 pfn = folio_pfn(ubuf->folios[pgoff]); in udmabuf_vm_fault() 59 pfn += ubuf->offsets[pgoff] >> PAGE_SHIFT; in udmabuf_vm_fault() 66 pgoff = vma->vm_pgoff; in udmabuf_vm_fault() 69 for (; addr < vma->vm_end; pgoff++, addr += PAGE_SIZE) { in udmabuf_vm_fault() 73 if (WARN_ON(pgoff >= ubuf->pagecount)) in udmabuf_vm_fault() 76 pfn = folio_pfn(ubuf->folios[pgoff]); in udmabuf_vm_fault() 77 pfn += ubuf->offsets[pgoff] >> PAGE_SHIFT; in udmabuf_vm_fault() 331 pgoff_t pgoff, pgcnt; in udmabuf_pin_folios() local [all …]
|
| /linux/arch/sh/kernel/ |
| H A D | sys_sh.c | 43 unsigned long fd, unsigned long pgoff) in sys_mmap2() argument 49 if (pgoff & ((1 << (PAGE_SHIFT - 12)) - 1)) in sys_mmap2() 52 pgoff >>= PAGE_SHIFT - 12; in sys_mmap2() 54 return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); in sys_mmap2()
|
| /linux/virt/kvm/ |
| H A D | guest_memfd.c | 33 gfn_t gfn = slot->base_gfn + index - slot->gmem.pgoff; in __kvm_gmem_prepare_folio() 80 WARN_ON(!IS_ALIGNED(slot->gmem.pgoff, 1 << folio_order(folio))); in kvm_gmem_prepare_folio() 81 index = gfn - slot->base_gfn + slot->gmem.pgoff; in kvm_gmem_prepare_folio() 123 pgoff_t pgoff = slot->gmem.pgoff; in __kvm_gmem_invalidate_begin() local 126 .start = slot->base_gfn + max(pgoff, start) - pgoff, in __kvm_gmem_invalidate_begin() 127 .end = slot->base_gfn + min(pgoff + slot->npages, end) - pgoff, in __kvm_gmem_invalidate_begin() 340 return gfn - slot->base_gfn + slot->gmem.pgoff; in kvm_gmem_get_index() 356 if (((loff_t)vmf->pgoff << PAGE_SHIFT) >= i_size_read(inode)) in kvm_gmem_fault_user_mapping() 362 folio = kvm_gmem_get_folio(inode, vmf->pgoff); in kvm_gmem_fault_user_mapping() 382 vmf->page = folio_file_page(folio, vmf->pgoff); in kvm_gmem_fault_user_mapping() [all …]
|