Home
last modified time | relevance | path

Searched refs:nr_node_ids (Results 1 – 25 of 45) sorted by relevance

12

/linux/arch/x86/mm/
H A Dnuma.c115 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map()
119 for (node = 0; node < nr_node_ids; node++) in setup_node_to_cpumask_map()
123 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map()
419 if ((unsigned)node >= nr_node_ids) { in cpumask_of_node()
422 node, nr_node_ids); in cpumask_of_node()
/linux/scripts/gdb/linux/
H A Dslab.py183 nr_node_ids = int(gdb.parse_and_eval("nr_node_ids"))
185 nr_node_ids = 1
198 for i in range(0, nr_node_ids):
278 nr_node_ids = None
294 nr_node_ids = int(gdb.parse_and_eval("nr_node_ids"))
296 nr_node_ids = 1
302 for i in range(0, nr_node_ids):
/linux/lib/
H A Dgroup_cpus.c50 masks = kzalloc_objs(cpumask_var_t, nr_node_ids); in alloc_node_to_cpumask()
54 for (node = 0; node < nr_node_ids; node++) { in alloc_node_to_cpumask()
72 for (node = 0; node < nr_node_ids; node++) in free_node_to_cpumask()
237 for (n = 0; n < nr_node_ids; n++) { in alloc_nodes_groups()
255 alloc_groups_to_nodes(numgrps, numcpus, node_groups, nr_node_ids); in alloc_nodes_groups()
435 node_groups = kzalloc_objs(struct node_groups, nr_node_ids); in __group_cpus_evenly()
442 for (i = 0; i < nr_node_ids; i++) { in __group_cpus_evenly()
/linux/drivers/base/
H A Darch_numa.c50 if (WARN_ON(node < 0 || node >= nr_node_ids)) in cpumask_of_node()
105 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map()
109 for (node = 0; node < nr_node_ids; node++) { in setup_node_to_cpumask_map()
115 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map()
/linux/mm/
H A Dmempolicy.c228 new_bw = kcalloc(nr_node_ids, sizeof(unsigned int), GFP_KERNEL); in mempolicy_set_node_perf()
232 new_wi_state = kmalloc_flex(*new_wi_state, iw_table, nr_node_ids); in mempolicy_set_node_perf()
238 for (i = 0; i < nr_node_ids; i++) in mempolicy_set_node_perf()
250 memcpy(new_bw, old_bw, nr_node_ids * sizeof(*old_bw)); in mempolicy_set_node_perf()
1698 unsigned int nbytes = BITS_TO_LONGS(nr_node_ids) * sizeof(long); in copy_nodes_to_user()
1702 nbytes = BITS_TO_COMPAT_LONGS(nr_node_ids) * sizeof(compat_long_t); in copy_nodes_to_user()
1710 maxnode = nr_node_ids; in copy_nodes_to_user()
1964 if (nmask != NULL && maxnode < nr_node_ids) in kernel_get_mempolicy()
2686 weights = kzalloc(nr_node_ids, GFP_KERNEL); in alloc_pages_bulk_weighted_interleave()
2693 memcpy(weights, state->iw_table, nr_node_ids * sizeof(u8)); in alloc_pages_bulk_weighted_interleave()
[all …]
H A Dshrinker_debug.c56 count_per_node = kcalloc(nr_node_ids, sizeof(unsigned long), GFP_KERNEL); in shrinker_debugfs_count_show()
126 if (nid < 0 || nid >= nr_node_ids) in shrinker_debugfs_scan_write()
H A Dlist_lru.c410 mlru = kmalloc_flex(*mlru, node, nr_node_ids, gfp); in memcg_init_list_lru_one()
588 lru->node = kzalloc_objs(*lru->node, nr_node_ids); in __list_lru_init()
H A Dhugetlb_sysfs.c377 for (nid = 0; nid < nr_node_ids; nid++) { in kobj_to_node_hstate()
H A Dksm.c3589 buf = kzalloc_objs(*buf, nr_node_ids + nr_node_ids); in merge_across_nodes_store()
3595 root_unstable_tree = buf + nr_node_ids; in merge_across_nodes_store()
3602 ksm_nr_node_ids = knob ? 1 : nr_node_ids; in merge_across_nodes_store()
H A Dshrinker.c722 size *= nr_node_ids; in shrinker_alloc()
H A Dmemory-tiers.c915 node_demotion = kzalloc_objs(struct demotion_nodes, nr_node_ids); in memory_tier_init()
/linux/arch/loongarch/kernel/
H A Dnuma.c72 if (nr_node_ids >= 8) in setup_per_cpu_areas()
255 loongson_sysconf.nr_nodes = nr_node_ids; in init_numa_memory()
/linux/tools/workqueue/
H A Dwq_dump.py223 nr_node_ids = prog['nr_node_ids'].value_() variable
250 nna = wq.node_nr_active[nr_node_ids]
/linux/arch/x86/kernel/
H A Dsetup_percpu.c118 NR_CPUS, nr_cpumask_bits, nr_cpu_ids, nr_node_ids); in setup_per_cpu_areas()
/linux/include/linux/
H A Dnodemask.h441 extern unsigned int nr_node_ids;
483 #define nr_node_ids 1U macro
/linux/arch/powerpc/mm/
H A Dnuma.c78 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map()
86 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map()
183 if (nid == 0xffff || nid >= nr_node_ids) in __associativity_to_nid()
/linux/include/net/
H A Dudp.h297 up->udp_prod_queue = kzalloc_objs(*up->udp_prod_queue, nr_node_ids); in udp_lib_init_sock()
300 for (int i = 0; i < nr_node_ids; i++) in udp_lib_init_sock()
/linux/kernel/sched/
H A Dtopology.c1844 for (i = 0; i < nr_node_ids; i++) { in sched_numa_warn()
1846 for (j = 0; j < nr_node_ids; j++) { in sched_numa_warn()
2072 masks[i] = kzalloc(nr_node_ids * sizeof(void *), GFP_KERNEL); in sched_init_numa()
2201 for (j = 0; j < nr_node_ids; j++) { in sched_domains_numa_masks_set()
2218 for (j = 0; j < nr_node_ids; j++) { in sched_domains_numa_masks_clear()
2350 if (node >= nr_node_ids || hops >= sched_domains_numa_levels) in sched_numa_hop_mask()
H A Dext_idle.c667 scx_idle_node_masks = kzalloc_objs(*scx_idle_node_masks, nr_node_ids); in scx_idle_init_masks()
833 if (node < 0 || node >= nr_node_ids) { in validate_node()
/linux/kernel/futex/
H A Dfutex.h103 if (nr_node_ids >= max) in futex_flags_valid()
H A Dcore.c442 node = (hash >> futex_hashshift) % nr_node_ids; in __futex_hash()
445 nr_node_ids, node); in __futex_hash()
/linux/drivers/infiniband/sw/siw/
H A Dsiw_main.c124 int i, num_nodes = nr_node_ids; in siw_init_cpulist()
/linux/drivers/hv/
H A Dhv.c186 hv_context.hv_numa_map = kzalloc_objs(struct cpumask, nr_node_ids); in hv_synic_alloc()
/linux/arch/powerpc/sysdev/xive/
H A Dcommon.c1142 ipi_domain = irq_domain_create_linear(fwnode, nr_node_ids, in xive_init_ipis()
1147 xive_ipis = kzalloc_objs(*xive_ipis, nr_node_ids, in xive_init_ipis()
/linux/arch/powerpc/platforms/pseries/
H A Dhotplug-cpu.c244 if (rc && nr_node_ids > 1) { in pseries_add_processor()

12