Searched refs:zone_start_pfn (Results 1 – 14 of 14) sorted by relevance
| /linux/mm/ |
| H A D | mm_init.c | 949 unsigned long zone_start_pfn = zone->zone_start_pfn; in memmap_init_zone_range() local 950 unsigned long zone_end_pfn = zone_start_pfn + zone->spanned_pages; in memmap_init_zone_range() 953 start_pfn = clamp(start_pfn, zone_start_pfn, zone_end_pfn); in memmap_init_zone_range() 954 end_pfn = clamp(end_pfn, zone_start_pfn, zone_end_pfn); in memmap_init_zone_range() 1167 unsigned long *zone_start_pfn, in adjust_zone_range_for_zone_movable() argument 1174 *zone_start_pfn = zone_movable_pfn[nid]; in adjust_zone_range_for_zone_movable() 1180 *zone_start_pfn < zone_movable_pfn[nid] && in adjust_zone_range_for_zone_movable() 1185 } else if (*zone_start_pfn >= zone_movable_pfn[nid]) in adjust_zone_range_for_zone_movable() 1186 *zone_start_pfn = *zone_end_pfn; in adjust_zone_range_for_zone_movable() 1226 unsigned long zone_start_pfn, in zone_absent_pages_in_node() argument [all …]
|
| H A D | compaction.c | 199 zone->compact_cached_migrate_pfn[0] = zone->zone_start_pfn; in reset_cached_positions() 200 zone->compact_cached_migrate_pfn[1] = zone->zone_start_pfn; in reset_cached_positions() 308 block_pfn = max(block_pfn, zone->zone_start_pfn); in __reset_isolation_pfn() 351 unsigned long migrate_pfn = zone->zone_start_pfn; in __reset_isolation_suitable() 706 if (block_start_pfn < cc->zone->zone_start_pfn) in isolate_freepages_range() 707 block_start_pfn = cc->zone->zone_start_pfn; in isolate_freepages_range() 941 low_pfn == cc->zone->zone_start_pfn)) { in isolate_migratepages_block() 1329 if (block_start_pfn < cc->zone->zone_start_pfn) in isolate_migratepages_range() 1330 block_start_pfn = cc->zone->zone_start_pfn; in isolate_migratepages_range() 1472 start_pfn = max(pageblock_start_pfn(pfn), cc->zone->zone_start_pfn); in fast_isolate_around() [all …]
|
| H A D | shuffle.c | 83 unsigned long start_pfn = z->zone_start_pfn; in __shuffle_zone() 112 j = z->zone_start_pfn + in __shuffle_zone()
|
| H A D | memory_hotplug.c | 485 if (zone->zone_start_pfn == start_pfn) { in shrink_zone_span() 496 zone->zone_start_pfn = pfn; in shrink_zone_span() 498 zone->zone_start_pfn = 0; in shrink_zone_span() 508 pfn = find_biggest_section_pfn(nid, zone, zone->zone_start_pfn, in shrink_zone_span() 511 zone->spanned_pages = pfn - zone->zone_start_pfn + 1; in shrink_zone_span() 513 zone->zone_start_pfn = 0; in shrink_zone_span() 532 node_start_pfn = zone->zone_start_pfn; in update_pgdat_span() 539 if (zone->zone_start_pfn < node_start_pfn) in update_pgdat_span() 540 node_start_pfn = zone->zone_start_pfn; in update_pgdat_span() 708 if (zone_is_empty(zone) || start_pfn < zone->zone_start_pfn) in resize_zone_range() [all …]
|
| H A D | page_owner.c | 429 pfn = zone->zone_start_pfn; in pagetypeinfo_showmixedcount_print() 765 unsigned long pfn = zone->zone_start_pfn; in init_pages_in_zone()
|
| H A D | page_isolation.c | 338 zone->zone_start_pfn); in isolate_single_pageblock()
|
| H A D | vmstat.c | 1644 unsigned long start_pfn = zone->zone_start_pfn; in pagetypeinfo_showblockcount_print() 1852 zone->zone_start_pfn); in zoneinfo_show_print()
|
| H A D | kmemleak.c | 1755 unsigned long start_pfn = zone->zone_start_pfn; in kmemleak_scan()
|
| H A D | page_alloc.c | 351 pfn = pfn - pageblock_start_pfn(page_zone(page)->zone_start_pfn); in pfn_to_bitidx() 586 start_pfn = zone->zone_start_pfn; in page_outside_zone_boundaries() 7101 pfn = ALIGN(zone->zone_start_pfn, nr_pages); in alloc_contig_pages_noprof()
|
| H A D | huge_memory.c | 4260 for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) { in split_huge_pages_all()
|
| /linux/include/linux/ |
| H A D | mmzone.h | 923 unsigned long zone_start_pfn; member 1123 return zone->zone_start_pfn + zone->spanned_pages; in zone_end_pfn() 1128 return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone); in zone_spans_pfn() 1282 start_pfn + nr_pages <= zone->zone_start_pfn) in zone_intersects()
|
| /linux/arch/arm64/kernel/ |
| H A D | hibernate.c | 267 for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) { in swsusp_mte_save_tags()
|
| /linux/kernel/power/ |
| H A D | snapshot.c | 638 zone_start = zone->zone_start_pfn; in create_mem_extents() 1257 for_each_valid_pfn(pfn, zone->zone_start_pfn, max_zone_pfn) { in mark_free_pages() 1357 for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) in count_highmem_pages() 1419 for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) in count_data_pages() 1532 for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) in copy_data_pages()
|
| /linux/Documentation/mm/ |
| H A D | physical_memory.rst | 458 ``zone_start_pfn`` 473 ``spanned_pages`` = ``zone_end_pfn`` - ``zone_start_pfn``. It is initialized 506 The seqlock to protect ``zone_start_pfn`` and ``spanned_pages``. It is a
|