| /linux/mm/ |
| H A D | page_ext.c | 294 table_size = page_ext_size * PAGES_PER_SECTION; in init_section_page_ext() 324 table_size = page_ext_size * PAGES_PER_SECTION; in free_page_ext() 381 for (pfn = start; !fail && pfn < end; pfn += PAGES_PER_SECTION) in online_page_ext() 387 end = pfn - PAGES_PER_SECTION; in online_page_ext() 388 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) in online_page_ext() 411 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) in offline_page_ext() 416 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) in offline_page_ext() 467 pfn = ALIGN(pfn + 1, PAGES_PER_SECTION)) { in page_ext_init()
|
| H A D | sparse.c | 197 pfns = min(nr_pages, PAGES_PER_SECTION in subsection_map_init() 223 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { in memory_present() 408 return ALIGN(sizeof(struct page) * PAGES_PER_SECTION, PMD_SIZE); in section_map_size() 414 return PAGE_ALIGN(sizeof(struct page) * PAGES_PER_SECTION); in section_map_size() 557 map = __populate_section_memmap(pfn, PAGES_PER_SECTION, in sparse_init_nid() 567 memmap_boot_pages_add(DIV_ROUND_UP(PAGES_PER_SECTION * sizeof(struct page), in sparse_init_nid() 634 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in online_mem_sections() 652 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in offline_mem_sections() 687 unsigned long end = (unsigned long)(memmap + PAGES_PER_SECTION); in free_map_bootmem() 746 PAGES_PER_SECTION), GFP_KERNEL, nid); in populate_section_memmap() [all …]
|
| H A D | bootmem_info.c | 62 mapsize = sizeof(struct page) * PAGES_PER_SECTION; in register_page_bootmem_info_section() 93 PAGES_PER_SECTION); in register_page_bootmem_info_section() 121 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node()
|
| H A D | memory_hotplug.c | 335 min_align = PAGES_PER_SECTION; in check_pfn_span() 770 if (!IS_ALIGNED(start_pfn, PAGES_PER_SECTION)) in move_pfn_range_to_zone() 772 if (!IS_ALIGNED(start_pfn + nr_pages, PAGES_PER_SECTION)) in move_pfn_range_to_zone() 994 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in auto_movable_zone_for_pfn() 1001 online_pages += PAGES_PER_SECTION; in auto_movable_zone_for_pfn() 1124 if (nr_pages >= PAGES_PER_SECTION) in mhp_init_memmap_on_memory() 1125 online_mem_sections(pfn, ALIGN_DOWN(end_pfn, PAGES_PER_SECTION)); in mhp_init_memmap_on_memory() 1139 if (nr_pages >= PAGES_PER_SECTION) in mhp_deinit_memmap_on_memory() 1140 offline_mem_sections(pfn, ALIGN_DOWN(end_pfn, PAGES_PER_SECTION)); in mhp_deinit_memmap_on_memory() 1176 !IS_ALIGNED(pfn + nr_pages, PAGES_PER_SECTION))) in online_pages() [all …]
|
| H A D | numa_memblks.c | 429 if (pfn_align && pfn_align < PAGES_PER_SECTION) { in numa_register_meminfo() 432 unsigned long sect_align_mb = PFN_PHYS(PAGES_PER_SECTION) / SZ_1M; in numa_register_meminfo()
|
| H A D | mm_init.c | 740 if ((nr_initialised > PAGES_PER_SECTION) && in defer_init() 741 (pfn & (PAGES_PER_SECTION - 1)) == 0) { in defer_init() 998 end_pfn = round_up(end_pfn, PAGES_PER_SECTION); in memmap_init() 2156 .align = PAGES_PER_SECTION, in deferred_init_memmap() 2157 .min_chunk = PAGES_PER_SECTION, in deferred_init_memmap() 2218 spfn = epfn, epfn += PAGES_PER_SECTION) { in deferred_grow_zone()
|
| H A D | util.c | 1339 for (pfn = ALIGN(start_pfn, PAGES_PER_SECTION); in page_range_contiguous() 1340 pfn < end_pfn; pfn += PAGES_PER_SECTION) in page_range_contiguous()
|
| H A D | memblock.c | 2166 start = min(start, ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap() 2191 if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION)) { in free_unused_memmap() 2193 free_memmap(prev_end, ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap()
|
| H A D | compaction.c | 240 return section_nr_to_pfn(start_nr) + PAGES_PER_SECTION; in skip_offline_sections_reverse()
|
| H A D | page_alloc.c | 349 pfn &= (PAGES_PER_SECTION-1); in pfn_to_bitidx()
|
| /linux/drivers/xen/ |
| H A D | unpopulated-alloc.c | 39 unsigned int i, alloc_pages = round_up(nr_pages, PAGES_PER_SECTION); in fill_list() 54 PAGES_PER_SECTION * PAGE_SIZE, NULL, NULL); in fill_list()
|
| H A D | balloon.c | 254 PAGES_PER_SECTION * PAGE_SIZE, NULL, NULL); in additional_memory_resource() 281 balloon_hotplug = round_up(credit, PAGES_PER_SECTION); in reserve_additional_memory()
|
| /linux/drivers/base/ |
| H A D | memory.c | 227 unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; in memory_block_online() 298 unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; in memory_block_offline() 492 unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; in valid_zones_show() 589 unsigned long pages_per_block = PAGES_PER_SECTION * sections_per_block; in probe_store() 740 const unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; in early_node_zone_for_memory_block()
|
| /linux/include/linux/ |
| H A D | page_ext.h | 80 return next_pfn % PAGES_PER_SECTION; in page_ext_iter_next_fast_possible()
|
| H A D | mmzone.h | 1856 #define PAGES_PER_SECTION (1UL << PFN_SECTION_SHIFT) macro 1857 #define PAGE_SECTION_MASK (~(PAGES_PER_SECTION-1)) 1875 #define SECTION_ALIGN_UP(pfn) (((pfn) + PAGES_PER_SECTION - 1) & PAGE_SECTION_MASK)
|
| /linux/scripts/gdb/linux/ |
| H A D | mm.py | 72 self.PAGES_PER_SECTION = 1 << self.PFN_SECTION_SHIFT 73 self.PAGE_SECTION_MASK = (~(self.PAGES_PER_SECTION - 1)) & ((1 << 64) - 1)
|
| /linux/drivers/virtio/ |
| H A D | virtio_mem.c | 2181 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in virtio_mem_bbm_offline_remove_and_unplug_bb() 2186 rc = virtio_mem_fake_offline(vm, pfn, PAGES_PER_SECTION); in virtio_mem_bbm_offline_remove_and_unplug_bb() 2210 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in virtio_mem_bbm_offline_remove_and_unplug_bb() 2214 virtio_mem_fake_online(pfn, PAGES_PER_SECTION); in virtio_mem_bbm_offline_remove_and_unplug_bb() 2232 pfn += PAGES_PER_SECTION) { in virtio_mem_bbm_bb_is_offline() 2252 pfn += PAGES_PER_SECTION) { in virtio_mem_bbm_bb_is_movable() 2760 chunk_size = max_t(uint64_t, PFN_PHYS(PAGES_PER_SECTION), in virtio_mem_vmcore_get_device_ram()
|
| /linux/fs/proc/ |
| H A D | page.c | 40 return round_up(max_pfn, PAGES_PER_SECTION); in get_max_dump_pfn()
|
| /linux/arch/arm64/mm/ |
| H A D | mmu.c | 1724 WARN_ON_ONCE(end - start > PAGES_PER_SECTION * sizeof(struct page)); in vmemmap_populate() 1727 (end - start < PAGES_PER_SECTION * sizeof(struct page))) in vmemmap_populate() 1978 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in prevent_bootmem_remove_notifier()
|
| /linux/arch/x86/mm/ |
| H A D | init_64.c | 1573 if (end - start < PAGES_PER_SECTION * sizeof(struct page)) in vmemmap_populate()
|