/linux/include/linux/ |
H A D | compaction.h | 96 extern void reset_isolation_suitable(pg_data_t *pgdat); 108 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx); 111 static inline void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() 129 static inline void wakeup_kcompactd(pg_data_t *pgdat, in wakeup_kcompactd()
|
H A D | memory_hotplug.h | 180 static inline void pgdat_kswapd_lock(pg_data_t *pgdat) in pgdat_kswapd_lock() 185 static inline void pgdat_kswapd_unlock(pg_data_t *pgdat) in pgdat_kswapd_unlock() 190 static inline void pgdat_kswapd_lock_init(pg_data_t *pgdat) in pgdat_kswapd_lock_init() 237 static inline void pgdat_kswapd_lock(pg_data_t *pgdat) {} in pgdat_kswapd_lock() 238 static inline void pgdat_kswapd_unlock(pg_data_t *pgdat) {} in pgdat_kswapd_unlock() 239 static inline void pgdat_kswapd_lock_init(pg_data_t *pgdat) {} in pgdat_kswapd_lock_init()
|
H A D | memory-tiers.h | 57 void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets); 65 static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets() 109 static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets()
|
H A D | page_owner.h | 20 pg_data_t *pgdat, struct zone *zone);
|
H A D | mmzone.h | 1353 * On NUMA machines, each NUMA node would have a pg_data_t to describe 1500 } pg_data_t; typedef 1508 static inline unsigned long pgdat_end_pfn(pg_data_t *pgdat) in pgdat_end_pfn() 1515 void build_all_zonelists(pg_data_t *pgdat); 1660 * @pgdat: pointer to a pg_data_t variable
|
/linux/mm/ |
H A D | shuffle.h | 11 extern void __shuffle_free_memory(pg_data_t *pgdat); 13 static inline void __meminit shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() 40 static inline void shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory()
|
H A D | numa.c | 14 const size_t nd_size = roundup(sizeof(pg_data_t), SMP_CACHE_BYTES); in alloc_node_data() 32 memset(NODE_DATA(nid), 0, sizeof(pg_data_t)); in alloc_node_data() 37 pg_data_t *pgdat; in alloc_offline_node_data()
|
H A D | vmstat.c | 320 void set_pgdat_percpu_threshold(pg_data_t *pgdat, in set_pgdat_percpu_threshold() 1506 pg_data_t *pgdat; in frag_start() 1519 pg_data_t *pgdat = (pg_data_t *)arg; in frag_next() 1533 static void walk_zones_in_node(struct seq_file *m, pg_data_t *pgdat, in walk_zones_in_node() argument 1535 void (*print)(struct seq_file *m, pg_data_t *, struct zone *)) in walk_zones_in_node() 1555 static void frag_show_print(struct seq_file *m, pg_data_t *pgdat, in frag_show_print() 1575 pg_data_t *pgdat = (pg_data_t *)arg; in frag_show() 1581 pg_data_t *pgda in pagetypeinfo_showfree_print() [all...] |
H A D | compaction.c | 402 void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() 763 pg_data_t *pgdat = cc->zone->zone_pgdat; in too_many_isolated() 840 pg_data_t *pgdat = cc->zone->zone_pgdat; in isolate_migratepages_block() 2152 static bool kswapd_is_running(pg_data_t *pgdat) in kswapd_is_running() 2197 static unsigned int fragmentation_score_node(pg_data_t *pgdat) in fragmentation_score_node() 2223 static bool should_proactive_compact_node(pg_data_t *pgdat) in should_proactive_compact_node() 2262 pg_data_t *pgdat; in __compact_finished() 2893 static int compact_node(pg_data_t *pgdat, bool proactive) in compact_node() 2957 pg_data_t *pgdat = NODE_DATA(nid); in compaction_proactiveness_sysctl_handler() 3023 static inline bool kcompactd_work_requested(pg_data_t *pgda [all...] |
H A D | memory-tiers.c | 262 pg_data_t *pgdat; in __node_get_memory_tier() 280 pg_data_t *pgdat; in node_is_toptier() 302 void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets() 307 * pg_data_t.memtier updates includes a synchronize_rcu() in node_get_allowed_targets() 537 pg_data_t *pgdat = NODE_DATA(node); in set_node_memory_tier() 573 pg_data_t *pgdat; in clear_node_memory_tier()
|
H A D | show_mem.c | 95 pg_data_t *pgdat = NODE_DATA(nid); in si_meminfo_node() 162 static bool node_has_managed_zones(pg_data_t *pgdat, int max_zone_idx) in node_has_managed_zones() 185 pg_data_t *pgdat; in show_free_areas()
|
H A D | vmscan.c | 519 static bool skip_throttle_noprogress(pg_data_t *pgdat) in skip_throttle_noprogress() 547 void reclaim_throttle(pg_data_t *pgdat, enum vmscan_throttle_state reason) in reclaim_throttle() 621 void __acct_reclaim_writeback(pg_data_t *pgdat, struct folio *folio, in __acct_reclaim_writeback() 2344 static void prepare_scan_control(pg_data_t *pgdat, struct scan_control *sc) in prepare_scan_control() 5991 static void shrink_node_memcgs(pg_data_t *pgdat, struct scan_control *sc) in shrink_node_memcgs() 6070 static void shrink_node(pg_data_t *pgdat, struct scan_control *sc) in shrink_node() 6219 static void consider_reclaim_throttle(pg_data_t *pgdat, struct scan_control *sc) in consider_reclaim_throttle() 6264 pg_data_t *last_pgdat = NULL; in shrink_zones() 6265 pg_data_t *first_pgdat = NULL; in shrink_zones() 6349 static void snapshot_refaults(struct mem_cgroup *target_memcg, pg_data_t *pgda [all...] |
H A D | internal.h | 409 void __acct_reclaim_writeback(pg_data_t *pgdat, struct folio *folio, 413 pg_data_t *pgdat = folio_pgdat(folio); in acct_reclaim_writeback() 420 static inline void wake_throttle_isolated(pg_data_t *pgdat) in wake_throttle_isolated() 540 extern void reclaim_throttle(pg_data_t *pgdat, enum vmscan_throttle_state reason); 543 struct mem_cgroup *memcg, pg_data_t *pgdat); 546 struct mem_cgroup *memcg, pg_data_t *pgdat) in user_proactive_reclaim()
|
H A D | mmzone.c | 32 pg_data_t *pgdat = zone->zone_pgdat; in next_zone()
|
H A D | mm_init.c | 68 pg_data_t *pgdat = NODE_DATA(nid); in mminit_verify_zonelist() 674 pg_data_t *pgdat; in __init_page_from_nid() 693 static inline void pgdat_set_deferred_range(pg_data_t *pgdat) in pgdat_set_deferred_range() 756 static inline void pgdat_set_deferred_range(pg_data_t *pgdat) {} in pgdat_set_deferred_range() 1720 pg_data_t *pgdat = NODE_DATA(nid); in free_area_init_node() 1724 /* pg_data_t should be reset to zero when it's allocated */ in free_area_init_node() 1753 static void __init check_for_memory(pg_data_t *pgdat) in check_for_memory() 1813 * free_area_init - Initialise all pg_data_t and zone data 1904 pg_data_t *pgdat; in free_area_init() 2166 pg_data_t *pgda in deferred_init_memmap() [all...] |
H A D | shuffle.c | 153 void __meminit __shuffle_free_memory(pg_data_t *pgdat) in __shuffle_free_memory()
|
H A D | workingset.c | 199 static void *pack_shadow(int memcgid, pg_data_t *pgdat, unsigned long eviction, in pack_shadow() 210 static void unpack_shadow(void *shadow, int *memcgidp, pg_data_t **pgdat, in unpack_shadow()
|
H A D | page_alloc.c | 4361 pg_data_t *last_pgdat = NULL; in wake_all_kswapds() 5435 static int build_zonerefs_node(pg_data_t *pgdat, struct zoneref *zonerefs) in build_zonerefs_node() 5553 static void build_zonelists_in_node_order(pg_data_t *pgdat, int *node_order, in build_zonelists_in_node_order() 5564 pg_data_t *node = NODE_DATA(node_order[i]); in build_zonelists_in_node_order() 5576 static void build_thisnode_zonelists(pg_data_t *pgdat) in build_thisnode_zonelists() 5588 static void build_zonelists(pg_data_t *pgdat) in build_zonelists() 5644 static void build_zonelists(pg_data_t *pgdat) in build_zonelists() 5685 pg_data_t *self = data; in __build_all_zonelists() 5717 pg_data_t *pgdat = NODE_DATA(nid); in __build_all_zonelists() 5773 void __ref build_all_zonelists(pg_data_t *pgda [all...] |
/linux/Documentation/translations/zh_CN/mm/ |
H A D | physical_memory.rst | 26 即使是在UMA架构中也是这样表示。该结构总是通过 ``pg_data_t`` 来引用。特 27 定节点的 ``pg_data_t`` 结构体可以通过NODE_DATA(nid)引用,其中nid被称 32 架构,只使用一个静态的 ``pg_data_t`` 结构体,称为 ``contig_page_data``。 133 正如我们所提到的,内存中的每个节点由 ``pg_data_t`` 描述,通过 188 pg_data_t *pgdat = NODE_DATA(nid);
|
/linux/arch/s390/kernel/ |
H A D | numa.c | 25 NODE_DATA(nid) = memblock_alloc_or_panic(sizeof(pg_data_t), 8); in numa_setup()
|
/linux/tools/testing/memblock/linux/ |
H A D | mmzone.h | 36 } pg_data_t; typedef
|
/linux/tools/testing/vma/linux/ |
H A D | mmzone.h | 36 } pg_data_t; typedef
|
/linux/arch/x86/kernel/ |
H A D | setup_percpu.c | 62 pg_data_t *last = NULL; in pcpu_need_numa()
|
/linux/mm/kmsan/ |
H A D | init.c | 77 const size_t nd_size = sizeof(pg_data_t); in kmsan_init_shadow()
|
/linux/Documentation/mm/ |
H A D | physical_memory.rst | 22 always referenced by its typedef ``pg_data_t``. A ``pg_data_t`` structure 29 static ``pg_data_t`` structure called ``contig_page_data`` is used. Nodes will 141 As we have mentioned, each node in memory is described by a ``pg_data_t`` which 196 pg_data_t *pgdat = NODE_DATA(nid);
|