| /linux/tools/testing/vma/linux/ |
| H A D | mmzone.h | 21 #define MAX_PAGE_ORDER 10 macro 22 #define MAX_ORDER_NR_PAGES (1 << MAX_PAGE_ORDER) 24 #define pageblock_order MAX_PAGE_ORDER
|
| /linux/tools/testing/memblock/linux/ |
| H A D | mmzone.h | 21 #define MAX_PAGE_ORDER 10 macro 22 #define MAX_ORDER_NR_PAGES (1 << MAX_PAGE_ORDER) 24 #define pageblock_order MAX_PAGE_ORDER
|
| /linux/mm/kmsan/ |
| H A D | init.c | 146 .order = MAX_PAGE_ORDER, 216 collect.order = MAX_PAGE_ORDER; in kmsan_memblock_discard() 217 for (int i = MAX_PAGE_ORDER; i >= 0; i--) { in kmsan_memblock_discard()
|
| /linux/include/linux/ |
| H A D | mmzone.h | 30 #define MAX_PAGE_ORDER 10 macro 32 #define MAX_PAGE_ORDER CONFIG_ARCH_FORCE_MAX_ORDER macro 34 #define MAX_ORDER_NR_PAGES (1 << MAX_PAGE_ORDER) 38 #define NR_PAGE_ORDERS (MAX_PAGE_ORDER + 1) 42 #define PAGE_BLOCK_MAX_ORDER MAX_PAGE_ORDER 52 #if (PAGE_BLOCK_MAX_ORDER > MAX_PAGE_ORDER) 53 #error MAX_PAGE_ORDER must be >= PAGE_BLOCK_MAX_ORDER 1862 #if (MAX_PAGE_ORDER + PAGE_SHIFT) > SECTION_SIZE_BITS 1863 #error Allocator MAX_PAGE_ORDER exceeds SECTION_SIZE
|
| /linux/drivers/gpu/drm/ttm/tests/ |
| H A D | ttm_pool_test.c | 111 .order = MAX_PAGE_ORDER + 1, 120 .order = MAX_PAGE_ORDER + 1, 167 if (params->order <= MAX_PAGE_ORDER) { in ttm_pool_alloc_basic() 184 min_t(unsigned int, MAX_PAGE_ORDER, in ttm_pool_alloc_basic()
|
| /linux/mm/ |
| H A D | shuffle.h | 7 #define SHUFFLE_ORDER MAX_PAGE_ORDER
|
| H A D | debug_page_alloc.c | 25 if (kstrtoul(buf, 10, &res) < 0 || res > MAX_PAGE_ORDER / 2) { in debug_guardpage_minorder_setup()
|
| H A D | page_reporting.c | 23 return param_set_uint_minmax(val, kp, 0, MAX_PAGE_ORDER); in page_order_update_notify() 373 if (prdev->order > 0 && prdev->order <= MAX_PAGE_ORDER) in page_reporting_register()
|
| H A D | page_owner.c | 459 if (freepage_order <= MAX_PAGE_ORDER) in pagetypeinfo_showmixedcount_print() 694 if (freepage_order <= MAX_PAGE_ORDER) in read_page_owner() 802 if (order > 0 && order <= MAX_PAGE_ORDER) in init_pages_in_zone()
|
| H A D | hugetlb_cma.c | 149 VM_WARN_ON(order <= MAX_PAGE_ORDER); in hugetlb_cma_reserve()
|
| H A D | page_alloc.c | 906 if (order >= MAX_PAGE_ORDER - 1) in buddy_merge_likely() 960 while (order < MAX_PAGE_ORDER) { in __free_one_page() 1609 if (order == MAX_PAGE_ORDER && __free_unaccepted(page)) in __free_pages_core() 2042 int order = start_pfn ? __ffs(start_pfn) : MAX_PAGE_ORDER; in find_large_buddy() 2048 if (++order > MAX_PAGE_ORDER) in find_large_buddy() 2108 if (pageblock_order == MAX_PAGE_ORDER) in __move_freepages_block_isolate() 2381 for (current_order = MAX_PAGE_ORDER; current_order >= min_order; in __rmqueue_claim() 5159 if (WARN_ON_ONCE_GFP(order > MAX_PAGE_ORDER, gfp)) in __alloc_frozen_pages_noprof() 7265 return order <= MAX_PAGE_ORDER; in is_free_buddy_page() 7419 accept_memory(page_to_phys(page), PAGE_SIZE << MAX_PAGE_ORDER); in __accept_page() [all …]
|
| H A D | page_isolation.c | 235 if (order >= pageblock_order && order < MAX_PAGE_ORDER) { in unset_migratetype_isolate()
|
| /linux/drivers/base/regmap/ |
| H A D | regmap-debugfs.c | 229 if (count > (PAGE_SIZE << MAX_PAGE_ORDER)) in regmap_read_debugfs() 230 count = PAGE_SIZE << MAX_PAGE_ORDER; in regmap_read_debugfs() 376 if (count > (PAGE_SIZE << MAX_PAGE_ORDER)) in regmap_reg_ranges_read_file() 377 count = PAGE_SIZE << MAX_PAGE_ORDER; in regmap_reg_ranges_read_file()
|
| /linux/kernel/events/ |
| H A D | ring_buffer.c | 620 if (order > MAX_PAGE_ORDER) in rb_alloc_aux_page() 621 order = MAX_PAGE_ORDER; in rb_alloc_aux_page() 729 if (get_order((unsigned long)nr_pages * sizeof(void *)) > MAX_PAGE_ORDER) in rb_alloc_aux() 845 if (order_base_2(size) > PAGE_SHIFT+MAX_PAGE_ORDER) in rb_alloc()
|
| /linux/drivers/crypto/hisilicon/ |
| H A D | sgl.c | 76 block_size = 1 << (PAGE_SHIFT + MAX_PAGE_ORDER < 32 ? in hisi_acc_create_sgl_pool() 77 PAGE_SHIFT + MAX_PAGE_ORDER : 31); in hisi_acc_create_sgl_pool()
|
| /linux/drivers/gpu/drm/i915/gem/ |
| H A D | i915_gem_internal.c | 38 int max_order = MAX_PAGE_ORDER; in i915_gem_object_get_pages_internal()
|
| /linux/arch/sparc/mm/ |
| H A D | tsb.c | 405 if (max_tsb_size > PAGE_SIZE << MAX_PAGE_ORDER) in tsb_grow() 406 max_tsb_size = PAGE_SIZE << MAX_PAGE_ORDER; in tsb_grow()
|
| /linux/fs/ramfs/ |
| H A D | file-nommu.c | 73 if (unlikely(order > MAX_PAGE_ORDER)) in ramfs_nommu_expand_for_mapping()
|
| /linux/arch/arm64/kvm/hyp/nvhe/ |
| H A D | page_alloc.c | 231 pool->max_order = min(MAX_PAGE_ORDER, in hyp_pool_init()
|
| /linux/kernel/dma/ |
| H A D | pool.c | 88 order = min(get_order(pool_size), MAX_PAGE_ORDER); in atomic_pool_expand()
|
| /linux/lib/ |
| H A D | test_kho.c | 29 static long max_mem = (PAGE_SIZE << MAX_PAGE_ORDER) * 2;
|
| /linux/Documentation/networking/ |
| H A D | packet_mmap.rst | 267 region allocated by __get_free_pages is determined by the MAX_PAGE_ORDER macro. 270 PAGE_SIZE << MAX_PAGE_ORDER 273 In a 2.4/i386 kernel MAX_PAGE_ORDER is 10 274 In a 2.6/i386 kernel MAX_PAGE_ORDER is 11 280 /usr/include/linux/mmzone.h to get PAGE_SIZE MAX_PAGE_ORDER declarations. 328 <max-order> is the value defined with MAX_PAGE_ORDER
|
| /linux/arch/powerpc/mm/book3s64/ |
| H A D | iommu_api.c | 100 chunk = (1UL << (PAGE_SHIFT + MAX_PAGE_ORDER)) / in mm_iommu_do_alloc()
|
| /linux/drivers/misc/genwqe/ |
| H A D | card_utils.c | 213 if (get_order(size) > MAX_PAGE_ORDER) in __genwqe_alloc_consistent() 311 if (get_order(sgl->sgl_size) > MAX_PAGE_ORDER) { in genwqe_alloc_sync_sgl()
|
| /linux/arch/nios2/ |
| H A D | Kconfig | 55 contiguous allocations. The limit is called MAX_PAGE_ORDER and it
|