| /linux/kernel/ |
| H A D | range.c | 12 int add_range(struct range *range, int az, int nr_range, u64 start, u64 end) in add_range() argument 15 return nr_range; in add_range() 18 if (nr_range >= az) in add_range() 19 return nr_range; in add_range() 21 range[nr_range].start = start; in add_range() 22 range[nr_range].end = end; in add_range() 24 nr_range++; in add_range() 26 return nr_range; in add_range() 29 int add_range_with_merge(struct range *range, int az, int nr_range, in add_range_with_merge() argument 35 return nr_range; in add_range_with_merge() [all …]
|
| /linux/arch/x86/mm/ |
| H A D | init.c | 326 static int __meminit save_mr(struct map_range *mr, int nr_range, in save_mr() argument 331 if (nr_range >= NR_RANGE_MR) in save_mr() 333 mr[nr_range].start = start_pfn<<PAGE_SHIFT; in save_mr() 334 mr[nr_range].end = end_pfn<<PAGE_SHIFT; in save_mr() 335 mr[nr_range].page_size_mask = page_size_mask; in save_mr() 336 nr_range++; in save_mr() 339 return nr_range; in save_mr() 347 int nr_range) in adjust_range_page_size_mask() argument 351 for (i = 0; i < nr_range; i++) { in adjust_range_page_size_mask() 401 static int __meminit split_mem_range(struct map_range *mr, int nr_range, in split_mem_range() argument [all …]
|
| /linux/arch/x86/kernel/cpu/mtrr/ |
| H A D | cleanup.c | 41 static int __initdata nr_range; variable 49 x86_get_mtrr_mem_range(struct range *range, int nr_range, in x86_get_mtrr_mem_range() argument 63 nr_range = add_range_with_merge(range, RANGE_NUM, nr_range, in x86_get_mtrr_mem_range() 68 for (i = 0; i < nr_range; i++) in x86_get_mtrr_mem_range() 108 nr_range = clean_sort_range(range, RANGE_NUM); in x86_get_mtrr_mem_range() 111 for (i = 0; i < nr_range; i++) in x86_get_mtrr_mem_range() 115 return nr_range; in x86_get_mtrr_mem_range() 120 static unsigned long __init sum_ranges(struct range *range, int nr_range) in sum_ranges() argument 125 for (i = 0; i < nr_range; i++) in sum_ranges() 447 x86_setup_var_mtrrs(struct range *range, int nr_range, in x86_setup_var_mtrrs() argument [all …]
|
| /linux/mm/ |
| H A D | memremap.c | 62 for (i = 0; i < pgmap->nr_range; i++) { in pgmap_pfn_valid() 119 for (i = 0; i < pgmap->nr_range; i++) in memunmap_pages() 124 for (i = 0; i < pgmap->nr_range; i++) in memunmap_pages() 273 const int nr_range = pgmap->nr_range; in memremap_pages() local 276 if (WARN_ONCE(!nr_range, "nr_range must be specified\n")) in memremap_pages() 335 pgmap->nr_range = 0; in memremap_pages() 337 for (i = 0; i < nr_range; i++) { in memremap_pages() 341 pgmap->nr_range++; in memremap_pages() 344 if (i < nr_range) { in memremap_pages() 346 pgmap->nr_range = nr_range; in memremap_pages()
|
| H A D | sparse-vmemmap.c | 484 PHYS_PFN(pgmap->ranges[pgmap->nr_range].start); in reuse_compound_section()
|
| /linux/include/linux/ |
| H A D | range.h | 30 int add_range(struct range *range, int az, int nr_range, 34 int add_range_with_merge(struct range *range, int az, int nr_range, 41 void sort_range(struct range *range, int nr_range);
|
| H A D | memremap.h | 142 int nr_range; member
|
| /linux/drivers/dax/ |
| H A D | kmem.c | 97 for (i = 0; i < dev_dax->nr_range; i++) { in dev_dax_kmem_probe() 124 data = kzalloc_flex(*data, res, dev_dax->nr_range); in dev_dax_kmem_probe() 137 for (i = 0; i < dev_dax->nr_range; i++) { in dev_dax_kmem_probe() 222 for (i = 0; i < dev_dax->nr_range; i++) { in dev_dax_kmem_remove() 244 if (success >= dev_dax->nr_range) { in dev_dax_kmem_remove()
|
| H A D | bus.c | 194 for (i = 0; i < dev_dax->nr_range; i++) in dev_dax_size() 443 int i = dev_dax->nr_range - 1; in trim_dev_dax_range() 453 if (--dev_dax->nr_range == 0) { in trim_dev_dax_range() 461 while (dev_dax->nr_range) in free_dev_dax_ranges() 854 if (dev_WARN_ONCE(dev, dev_dax->nr_range, in alloc_dev_dax_range() 866 * (dev_dax->nr_range + 1), GFP_KERNEL); in alloc_dev_dax_range() 872 for (i = 0; i < dev_dax->nr_range; i++) in alloc_dev_dax_range() 875 ranges[dev_dax->nr_range++] = (struct dev_dax_range) { in alloc_dev_dax_range() 883 dev_dbg(dev, "alloc range[%d]: %pa:%pa\n", dev_dax->nr_range - 1, in alloc_dev_dax_range() 893 rc = devm_register_dax_mapping(dev_dax, dev_dax->nr_range - 1); in alloc_dev_dax_range() [all …]
|
| H A D | device.c | 66 for (i = 0; i < dev_dax->nr_range; i++) { in dax_pgoff_to_phys() 414 if (dev_dax->nr_range > 1) { in dev_dax_probe() 429 struct_size(pgmap, ranges, dev_dax->nr_range - 1), in dev_dax_probe() 434 pgmap->nr_range = dev_dax->nr_range; in dev_dax_probe() 437 for (i = 0; i < dev_dax->nr_range; i++) { in dev_dax_probe() 443 for (i = 0; i < dev_dax->nr_range; i++) { in dev_dax_probe()
|
| H A D | dax-private.h | 94 int nr_range; member
|
| /linux/tools/testing/nvdimm/ |
| H A D | dax-dev.c | 14 for (i = 0; i < dev_dax->nr_range; i++) { in dax_pgoff_to_phys()
|
| /linux/drivers/xen/ |
| H A D | unpopulated-alloc.c | 98 pgmap->nr_range = 1; in fill_list()
|
| /linux/Documentation/translations/zh_CN/mm/ |
| H A D | hmm.rst | 235 pagemap.nr_range = 1;
|
| /linux/drivers/nvdimm/ |
| H A D | pmem.c | 528 pmem->pgmap.nr_range = 1; in pmem_attach_disk()
|
| H A D | pfn_devs.c | 695 pgmap->nr_range = 1; in __nvdimm_setup_pfn()
|
| /linux/drivers/pci/ |
| H A D | p2pdma.c | 428 pgmap->nr_range = 1; in pci_p2pdma_add_resource()
|
| /linux/drivers/s390/block/ |
| H A D | dcssblk.c | 696 dev_info->pgmap.nr_range = 1; in dcssblk_add_store()
|
| /linux/arch/powerpc/kvm/ |
| H A D | book3s_hv_uvmem.c | 1186 kvmppc_uvmem_pgmap.nr_range = 1; in kvmppc_uvmem_init()
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_dmem.c | 324 chunk->pagemap.nr_range = 1; in nouveau_dmem_chunk_alloc()
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_migrate.c | 1060 pgmap->nr_range = 1; in kgd2kfd_init_zone_device()
|
| /linux/drivers/hv/ |
| H A D | mshv_vtl_main.c | 402 pgmap->nr_range = 1; in mshv_vtl_ioctl_add_vtl0_mem()
|
| /linux/fs/fuse/ |
| H A D | virtio_fs.c | 1101 pgmap->nr_range = 1; in virtio_fs_setup_dax()
|
| /linux/lib/ |
| H A D | test_hmm.c | 534 devmem->pagemap.nr_range = 1; in dmirror_allocate_chunk()
|
| /linux/Documentation/mm/ |
| H A D | hmm.rst | 284 pagemap.nr_range = 1;
|