Home
last modified time | relevance | path

Searched refs:NUMA_NO_NODE (Results 1 – 25 of 349) sorted by relevance

12345678910>>...14

/linux/tools/testing/selftests/bpf/progs/
H A Dverifier_arena.c33 page1 = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in basic_alloc1()
37 page2 = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in basic_alloc1()
41 no_page = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in basic_alloc1()
64 page1 = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0);
68 page2 = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in basic_alloc2()
72 no_page = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in basic_alloc2()
84 page3 = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in basic_alloc2()
103 page1 = bpf_arena_alloc_pages(&arena, NULL, 2, NUMA_NO_NODE, 0);
141 page1 = bpf_arena_alloc_pages(&arena, NULL, 2, NUMA_NO_NODE, 0); in basic_reserve1()
183 pages = bpf_arena_alloc_pages(&ar->map, NULL, ar->map.max_entries, NUMA_NO_NODE, in reserve_twice()
[all...]
H A Dverifier_arena_large.c30 page1 = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in big_alloc1()
39 1, NUMA_NO_NODE, 0); in big_alloc1()
46 1, NUMA_NO_NODE, 0); in big_alloc1()
51 1, NUMA_NO_NODE, 0); in big_alloc1()
63 page3 = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in big_alloc1()
139 page = bpf_arena_alloc_pages(&arena, base, 5, NUMA_NO_NODE, 0); in free_reserved()
158 page = bpf_arena_alloc_pages(&arena, addr, 2, NUMA_NO_NODE, 0); in free_reserved()
174 page = bpf_arena_alloc_pages(&arena, addr + __PAGE_SIZE, 2, NUMA_NO_NODE, 0);
198 NUMA_NO_NODE, 0); in alloc_pages()
225 base = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, in big_alloc2()
[all...]
H A Dverifier_ldsx.c310 __imm_const(numa_no_node, NUMA_NO_NODE) in arena_ldsx_disasm()
362 __imm_const(numa_no_node, NUMA_NO_NODE) in arena_ldsx_s8()
393 __imm_const(numa_no_node, NUMA_NO_NODE) in arena_ldsx_s16()
424 __imm_const(numa_no_node, NUMA_NO_NODE) in arena_ldsx_s32()
/linux/arch/x86/mm/
H A Dnuma.c48 [0 ... MAX_LOCAL_APIC-1] = NUMA_NO_NODE
59 return NUMA_NO_NODE; in numa_cpu_node()
73 DEFINE_EARLY_PER_CPU(int, x86_cpu_to_node_map, NUMA_NO_NODE);
100 numa_set_node(cpu, NUMA_NO_NODE); in numa_clear_node()
167 if (early_cpu_to_node(i) != NUMA_NO_NODE) in numa_init_array()
180 set_apicid_to_node(i, NUMA_NO_NODE); in numa_init()
193 if (nid == NUMA_NO_NODE) in numa_init()
306 if (node == NUMA_NO_NODE) in init_cpu_to_node()
366 return NUMA_NO_NODE; in early_cpu_to_node()
375 if (node == NUMA_NO_NODE) { in debug_cpumask_set_cpu()
[all …]
/linux/mm/
H A Dnuma_emulation.c66 if (emu_nid_to_phys[nid] == NUMA_NO_NODE) in emu_setup_memblk()
335 *dfl_phys_nid = NUMA_NO_NODE; in setup_emu2phys_nid()
337 if (emu_nid_to_phys[i] != NUMA_NO_NODE) { in setup_emu2phys_nid()
339 if (*dfl_phys_nid == NUMA_NO_NODE) in setup_emu2phys_nid()
392 emu_nid_to_phys[i] = NUMA_NO_NODE; in numa_emulation()
473 ei.blk[i].nid != NUMA_NO_NODE) in numa_emulation()
491 if (emu_nid_to_phys[i] == NUMA_NO_NODE) in numa_emulation()
519 if (emu_nid_to_phys[i] != NUMA_NO_NODE && in numa_emulation()
520 emu_nid_to_phys[j] != NUMA_NO_NODE) in numa_emulation()
522 if (emu_nid_to_phys[i] != NUMA_NO_NODE) in numa_emulation()
[all …]
H A Dnuma_memblks.c30 mi->blk[i].nid != NUMA_NO_NODE) in numa_nodemask_from_meminfo()
325 mi->blk[i].nid = NUMA_NO_NODE; in numa_cleanup_meminfo()
455 WARN_ON(memblock_set_node(0, max_addr, &memblock.memory, NUMA_NO_NODE)); in numa_memblks_init()
457 NUMA_NO_NODE)); in numa_memblks_init()
567 return NUMA_NO_NODE; in meminfo_to_nid()
579 if (nid != NUMA_NO_NODE && reserved_nid == NUMA_NO_NODE) in phys_to_target_node()
590 if (nid == NUMA_NO_NODE) in memory_add_physaddr_to_nid()
H A Dhugetlb_sysfs.c35 *nidp = NUMA_NO_NODE; in kobj_to_hstate()
50 if (nid == NUMA_NO_NODE) in nr_hugepages_show_common()
147 if (nid == NUMA_NO_NODE) in free_hugepages_show()
172 if (nid == NUMA_NO_NODE) in surplus_hugepages_show()
196 if (nid != NUMA_NO_NODE) { in demote_store()
214 if (nid != NUMA_NO_NODE) in demote_store()
/linux/drivers/acpi/numa/
H A Dsrat.c28 = { [0 ... MAX_PXM_DOMAINS - 1] = NUMA_NO_NODE };
45 return NUMA_NO_NODE; in pxm_to_node()
60 if (pxm_to_node_map[pxm] == NUMA_NO_NODE || node < pxm_to_node_map[pxm]) in __acpi_map_pxm_to_node()
71 return NUMA_NO_NODE; in acpi_map_pxm_to_node()
75 if (node == NUMA_NO_NODE) { in acpi_map_pxm_to_node()
78 return NUMA_NO_NODE; in acpi_map_pxm_to_node()
95 = { [0 ... MAX_PXM_DOMAINS - 1] = NUMA_NO_NODE }; in fix_pxm_node_maps()
165 pxm_to_node_map_copy[node_to_pxm_map_copy[i]] == NUMA_NO_NODE) in fix_pxm_node_maps()
339 if (from_node == NUMA_NO_NODE) in acpi_parse_slit()
345 if (to_node == NUMA_NO_NODE) in acpi_parse_slit()
[all …]
/linux/arch/powerpc/mm/
H A Dnuma.c65 static int numa_id_index_table[MAX_NUMNODES] = { [0 ... MAX_NUMNODES - 1] = NUMA_NO_NODE };
178 return NUMA_NO_NODE; in __associativity_to_nid()
184 nid = NUMA_NO_NODE; in __associativity_to_nid()
274 int nid = NUMA_NO_NODE; in of_node_to_nid_single()
286 int nid = NUMA_NO_NODE; in of_node_to_nid()
311 if (nid != NUMA_NO_NODE) { in __initialize_form1_numa_distance()
359 if (nid == NUMA_NO_NODE) in update_numa_distance()
587 int default_nid = NUMA_NO_NODE; in get_nid_and_numa_distance()
623 int default_nid = NUMA_NO_NODE; in of_drconf_to_nid_single()
682 return NUMA_NO_NODE; in vphn_get_nid()
[all …]
/linux/drivers/base/
H A Darch_numa.c19 static int cpu_to_node_map[NR_CPUS] = { [0 ... NR_CPUS-1] = NUMA_NO_NODE };
47 if (node == NUMA_NO_NODE) in cpumask_of_node()
67 if (nid == NUMA_NO_NODE) in numa_update_cpu()
90 set_cpu_numa_node(cpu, NUMA_NO_NODE); in numa_clear_node()
342 if (cpu_to_node_map[i] == NUMA_NO_NODE) in numa_emu_update_cpu_to_node()
360 if (node == NUMA_NO_NODE) in debug_cpumask_set_cpu()
/linux/include/linux/
H A Dmemblock.h135 return __memblock_reserve(base, size, NUMA_NO_NODE, 0); in memblock_reserve()
140 return __memblock_reserve(base, size, NUMA_NO_NODE, MEMBLOCK_RSRV_KERN); in memblock_reserve_kern()
184 __next_mem_range(idx, NUMA_NO_NODE, MEMBLOCK_NONE, &physmem, type, in __next_physmem_range()
249 __for_each_mem_range(i, &memblock.memory, NULL, NUMA_NO_NODE, \
261 __for_each_mem_range_rev(i, &memblock.memory, NULL, NUMA_NO_NODE, \
275 __for_each_mem_range(i, &memblock.reserved, NULL, NUMA_NO_NODE, \
428 MEMBLOCK_ALLOC_ACCESSIBLE, NUMA_NO_NODE); in memblock_alloc()
442 NUMA_NO_NODE); in memblock_alloc_raw()
450 MEMBLOCK_ALLOC_ACCESSIBLE, NUMA_NO_NODE); in memblock_alloc_from()
457 ARCH_LOW_ADDRESS_LIMIT, NUMA_NO_NODE); in memblock_alloc_low()
H A Dslab.h508 #define krealloc_noprof(_o, _s, _f) krealloc_node_align_noprof(_o, _s, 1, _f, NUMA_NO_NODE)
511 #define krealloc(...) krealloc_node(__VA_ARGS__, NUMA_NO_NODE)
1063 alloc_hooks(__kmalloc_node_noprof(PASS_BUCKET_PARAMS(_size, _b), _flags, NUMA_NO_NODE))
1066 …alloc_hooks(__kmalloc_node_track_caller_noprof(PASS_BUCKET_PARAMS(_size, _b), _flags, NUMA_NO_NODE
1155 #define kmalloc_track_caller(...) kmalloc_node_track_caller(__VA_ARGS__, NUMA_NO_NODE)
1158 kmalloc_node_track_caller_noprof(__VA_ARGS__, NUMA_NO_NODE, _RET_IP_)
1200 #define kvmalloc(...) kvmalloc_node(__VA_ARGS__, NUMA_NO_NODE)
1206 alloc_hooks(__kvmalloc_node_noprof(PASS_BUCKET_PARAMS(_size, _b), 1, _flags, NUMA_NO_NODE))
1219 #define kvmalloc_array_noprof(...) kvmalloc_array_node_noprof(__VA_ARGS__, NUMA_NO_NODE)
1221 #define kvcalloc_noprof(...) kvcalloc_node_noprof(__VA_ARGS__, NUMA_NO_NODE)
[all …]
H A Dnuma.h49 return NUMA_NO_NODE; in numa_nearest_node()
54 return NUMA_NO_NODE; in nearest_node_nodemask()
H A Dasync.h54 return async_schedule_node(func, data, NUMA_NO_NODE); in async_schedule()
72 return async_schedule_node_domain(func, data, NUMA_NO_NODE, domain); in async_schedule_domain()
H A Dmemory-tiers.h62 return NUMA_NO_NODE; in next_demotion_node()
106 return NUMA_NO_NODE; in next_demotion_node()
/linux/kernel/sched/
H A Dext_idle.c62 return node == NUMA_NO_NODE ? &scx_idle_global_masks : scx_idle_node_masks[node]; in idle_cpumask()
72 return NUMA_NO_NODE; in scx_cpu_node_if_enabled()
218 if (node == NUMA_NO_NODE || flags & SCX_PICK_IDLE_IN_NODE) in scx_pick_idle_cpu()
783 cpumask_copy(idle_cpumask(NUMA_NO_NODE)->cpu, cpu_online_mask); in reset_idle_masks()
784 cpumask_copy(idle_cpumask(NUMA_NO_NODE)->smt, cpu_online_mask); in reset_idle_masks()
829 if (node == NUMA_NO_NODE) in validate_node()
964 return NUMA_NO_NODE; in scx_bpf_cpu_node()
1120 return idle_cpumask(NUMA_NO_NODE)->cpu; in scx_bpf_get_idle_cpumask()
1180 return idle_cpumask(NUMA_NO_NODE)->smt; in scx_bpf_get_idle_smtmask()
1182 return idle_cpumask(NUMA_NO_NODE)->cpu; in scx_bpf_get_idle_smtmask()
[all …]
/linux/tools/testing/memblock/tests/
H A Dalloc_nid_api.c81 NUMA_NO_NODE); in alloc_nid_top_down_simple_check()
134 NUMA_NO_NODE); in alloc_nid_top_down_end_misaligned_check()
184 NUMA_NO_NODE); in alloc_nid_exact_address_generic_check()
235 NUMA_NO_NODE); in alloc_nid_top_down_narrow_range_check()
286 NUMA_NO_NODE); in alloc_nid_low_max_generic_check()
331 NUMA_NO_NODE); in alloc_nid_min_reserved_generic_check()
381 NUMA_NO_NODE); in alloc_nid_max_reserved_generic_check()
444 NUMA_NO_NODE); in alloc_nid_top_down_reserved_with_space_check()
507 NUMA_NO_NODE); in alloc_nid_reserved_full_merge_generic_check()
571 NUMA_NO_NODE); in alloc_nid_top_down_reserved_no_space_check()
[all …]
/linux/arch/arm64/kernel/
H A Dacpi_numa.c30 static int acpi_early_node_map[NR_CPUS] __initdata = { [0 ... NR_CPUS - 1] = NUMA_NO_NODE };
101 if (node == NUMA_NO_NODE) { in acpi_numa_gicc_affinity_init()
/linux/arch/loongarch/kernel/
H A Dnuma.c39 [0 ... CONFIG_NR_CPUS - 1] = NUMA_NO_NODE
115 return NUMA_NO_NODE; in early_cpu_to_node()
232 set_cpuid_to_node(i, NUMA_NO_NODE); in init_numa_memory()
/linux/drivers/gpu/drm/ttm/tests/
H A Dttm_pool_test.c90 ttm_pool_init(pool, devs->dev, NUMA_NO_NODE, TTM_ALLOCATION_POOL_USE_DMA_ALLOC); in ttm_pool_pre_populated()
154 ttm_pool_init(pool, devs->dev, NUMA_NO_NODE, params->alloc_flags); in ttm_pool_alloc_basic()
157 KUNIT_ASSERT_EQ(test, pool->nid, NUMA_NO_NODE); in ttm_pool_alloc_basic()
221 ttm_pool_init(pool, devs->dev, NUMA_NO_NODE, TTM_ALLOCATION_POOL_USE_DMA_ALLOC); in ttm_pool_alloc_basic_dma_addr()
351 ttm_pool_init(pool, devs->dev, NUMA_NO_NODE, TTM_ALLOCATION_POOL_USE_DMA_ALLOC); in ttm_pool_free_dma_alloc()
382 ttm_pool_init(pool, devs->dev, NUMA_NO_NODE, 0); in ttm_pool_free_no_dma_alloc()
/linux/tools/testing/selftests/bpf/
H A Dbpf_arena_common.h9 #ifndef NUMA_NO_NODE
10 #define NUMA_NO_NODE (-1) macro
/linux/arch/riscv/kernel/
H A Dacpi_numa.c31 static int acpi_early_node_map[NR_CPUS] __initdata = { [0 ... NR_CPUS - 1] = NUMA_NO_NODE };
124 if (node == NUMA_NO_NODE) { in acpi_numa_rintc_affinity_init()
/linux/drivers/base/test/
H A Dtest_async_driver_probe.c89 if (nid != NUMA_NO_NODE) in test_platform_device_register_node()
180 NUMA_NO_NODE); in test_async_probe_init()
213 NUMA_NO_NODE); in test_async_probe_init()
/linux/tools/testing/selftests/bpf/prog_tests/
H A Dbloom_filter_map.c9 #ifndef NUMA_NO_NODE
10 #define NUMA_NO_NODE (-1) macro
76 opts.numa_node = NUMA_NO_NODE; in test_success_cases()
/linux/arch/x86/pci/
H A Dacpi.c457 if (node == NUMA_NO_NODE) { in pci_acpi_root_get_node()
459 if (node != 0 && node != NUMA_NO_NODE) in pci_acpi_root_get_node()
463 if (node != NUMA_NO_NODE && !node_online(node)) in pci_acpi_root_get_node()
464 node = NUMA_NO_NODE; in pci_acpi_root_get_node()

12345678910>>...14