Lines Matching +full:memory +full:- +full:region

1 /* SPDX-License-Identifier: GPL-2.0-or-later */
7 * Logical memory blocks.
29 * enum memblock_flags - definition of memory region attributes
31 * @MEMBLOCK_HOTPLUG: hotpluggable region
32 * @MEMBLOCK_MIRROR: mirrored region
37 MEMBLOCK_HOTPLUG = 0x1, /* hotpluggable region */
38 MEMBLOCK_MIRROR = 0x2, /* mirrored region */
43 * struct memblock_region - represents a memory region
44 * @base: base address of the region
45 * @size: size of the region
46 * @flags: memory region attributes
59 * struct memblock_type - collection of memory regions of certain type
64 * @name: the memory type symbolic name
75 * struct memblock - memblock allocator metadata
78 * @memory: usable memory regions
79 * @reserved: reserved memory regions
84 struct memblock_type memory; member
149 * for_each_physmem_range - iterate through physmem areas not included in type.
162 * __for_each_mem_range - iterate through memblock areas from type_a and not
168 * @flags: pick from blocks based on memory attributes
182 * __for_each_mem_range_rev - reverse iterate through memblock areas from
188 * @flags: pick from blocks based on memory attributes
203 * for_each_mem_range - iterate through memory areas.
209 __for_each_mem_range(i, &memblock.memory, NULL, NUMA_NO_NODE, \
213 * for_each_mem_range_rev - reverse iterate through memblock areas from
220 __for_each_mem_range_rev(i, &memblock.memory, NULL, NUMA_NO_NODE, \
224 * for_each_reserved_mem_range - iterate over all reserved memblock areas
238 return m->flags & MEMBLOCK_HOTPLUG; in memblock_is_hotpluggable()
243 return m->flags & MEMBLOCK_MIRROR; in memblock_is_mirror()
248 return m->flags & MEMBLOCK_NOMAP; in memblock_is_nomap()
257 * for_each_mem_pfn_range - early memory pfn range iterator
264 * Walks over configured memory ranges.
267 for (i = -1, __next_mem_pfn_range(&i, nid, p_start, p_end, p_nid); \
275 * for_each_free_mem_range_in_zone - iterate through zone specific free
278 * @zone: zone in which all of the memory blocks reside
282 * Walks over free (memory && !reserved) areas of memblock in a specific
295 * for_each_free_mem_range_in_zone_from - iterate through zone specific
298 * @zone: zone in which all of the memory blocks reside
302 * Walks over free (memory && !reserved) areas of memblock in a specific
315 * for_each_free_mem_range - iterate through free memblock areas
318 * @flags: pick from blocks based on memory attributes
323 * Walks over free (memory && !reserved) areas of memblock. Available as
327 __for_each_mem_range(i, &memblock.memory, &memblock.reserved, \
331 * for_each_free_mem_range_reverse - rev-iterate through free memblock areas
334 * @flags: pick from blocks based on memory attributes
339 * Walks over free (memory && !reserved) areas of memblock in reverse
344 __for_each_mem_range_rev(i, &memblock.memory, &memblock.reserved, \
353 r->nid = nid; in memblock_set_region_node()
358 return r->nid; in memblock_get_region_node()
461 * Set the allocation direction to bottom-up or top-down.
469 * Check if the allocation direction is bottom-up or not.
470 * if this is true, that said, memblock will allocate memory
471 * in bottom-up direction.
494 * memblock_set_current_limit - Set the current allocation limit to allow
507 * While the memory MEMBLOCKs should always be page aligned, the reserved
513 * memblock_region_memory_base_pfn - get the lowest pfn of the memory region
516 * Return: the lowest pfn intersecting with the memory region
520 return PFN_UP(reg->base); in memblock_region_memory_base_pfn()
524 * memblock_region_memory_end_pfn - get the end pfn of the memory region
527 * Return: the end_pfn of the reserved region
531 return PFN_DOWN(reg->base + reg->size); in memblock_region_memory_end_pfn()
535 * memblock_region_reserved_base_pfn - get the lowest pfn of the reserved region
538 * Return: the lowest pfn intersecting with the reserved region
542 return PFN_DOWN(reg->base); in memblock_region_reserved_base_pfn()
546 * memblock_region_reserved_end_pfn - get the end pfn of the reserved region
549 * Return: the end_pfn of the reserved region
553 return PFN_UP(reg->base + reg->size); in memblock_region_reserved_end_pfn()
557 * for_each_mem_region - itereate over memory regions
558 * @region: loop variable
560 #define for_each_mem_region(region) \ argument
561 for (region = memblock.memory.regions; \
562 region < (memblock.memory.regions + memblock.memory.cnt); \
563 region++)
566 * for_each_reserved_mem_region - itereate over reserved memory regions
567 * @region: loop variable
569 #define for_each_reserved_mem_region(region) \ argument
570 for (region = memblock.reserved.regions; \
571 region < (memblock.reserved.regions + memblock.reserved.cnt); \
572 region++)
585 #define HASH_SMALL 0x00000002 /* sub-page allocation allowed, min