Home
last modified time | relevance | path

Searched refs:alloc_flags (Results 1 – 25 of 40) sorted by relevance

12

/linux/mm/
H A Dpage_alloc.c321 int alloc_flags);
1895 unsigned int alloc_flags) in prep_new_page() argument
1908 if (alloc_flags & ALLOC_NO_WATERMARKS) in prep_new_page()
2314 int block_type, unsigned int alloc_flags) in try_to_claim_block() argument
2335 if (boost_watermark(zone) && (alloc_flags & ALLOC_KSWAPD)) in try_to_claim_block()
2388 unsigned int alloc_flags) in __rmqueue_claim() argument
2401 if (order < pageblock_order && alloc_flags & ALLOC_NOFRAGMENT) in __rmqueue_claim()
2426 alloc_flags); in __rmqueue_claim()
2479 unsigned int alloc_flags, enum rmqueue_mode *mode) in __rmqueue() argument
2489 if (alloc_flags & ALLOC_CMA && in __rmqueue()
[all …]
H A Dcompaction.c2299 cc->highest_zoneidx, cc->alloc_flags, in __compact_finished()
2433 int alloc_flags) in compaction_zonelist_suitable() argument
2471 int highest_zoneidx, unsigned int alloc_flags, in compaction_suit_allocation_order() argument
2482 watermark = wmark_pages(zone, alloc_flags & ALLOC_WMARK_MASK); in compaction_suit_allocation_order()
2484 alloc_flags, free_pages)) in compaction_suit_allocation_order()
2497 !(alloc_flags & ALLOC_CMA)) { in compaction_suit_allocation_order()
2539 cc->alloc_flags, in compact_zone()
2751 unsigned int alloc_flags, int highest_zoneidx, in compact_zone_order() argument
2762 .alloc_flags = alloc_flags, in compact_zone_order()
2815 unsigned int alloc_flags, const struct alloc_context *ac, in try_to_compact_pages() argument
[all …]
/linux/drivers/base/regmap/
H A Dregcache-maple.c76 entry = kmalloc_array(last - index + 1, sizeof(*entry), map->alloc_flags); in regcache_maple_write()
94 ret = mas_store_gfp(&mas, entry, map->alloc_flags); in regcache_maple_write()
138 map->alloc_flags); in regcache_maple_drop()
151 map->alloc_flags); in regcache_maple_drop()
165 ret = mas_store_gfp(&mas, lower, map->alloc_flags); in regcache_maple_drop()
173 ret = mas_store_gfp(&mas, upper, map->alloc_flags); in regcache_maple_drop()
207 buf = kmalloc_array(max - min, val_bytes, map->alloc_flags); in regcache_maple_sync_block()
297 mt = kmalloc_obj(*mt, map->alloc_flags); in regcache_maple_init()
340 entry = kmalloc_array(last - first + 1, sizeof(*entry), map->alloc_flags); in regcache_maple_insert_block()
351 ret = mas_store_gfp(&mas, entry, map->alloc_flags); in regcache_maple_insert_block()
H A Dregcache-rbtree.c188 map->cache = kmalloc_obj(*rbtree_ctx, map->alloc_flags); in regcache_rbtree_init()
280 blk = krealloc_array(rbnode->block, blklen, map->cache_word_size, map->alloc_flags); in regcache_rbtree_insert_to_block()
289 map->alloc_flags); in regcache_rbtree_insert_to_block()
323 rbnode = kzalloc_obj(*rbnode, map->alloc_flags); in regcache_rbtree_node_alloc()
349 map->alloc_flags); in regcache_rbtree_node_alloc()
355 map->alloc_flags); in regcache_rbtree_node_alloc()
H A Dregcache-flat.c39 cache = kzalloc_flex(*cache, data, cache_size, map->alloc_flags); in regcache_flat_init()
43 cache->valid = bitmap_zalloc(cache_size, map->alloc_flags); in regcache_flat_init()
/linux/drivers/gpu/drm/ttm/tests/
H A Dttm_device_test.c14 unsigned int alloc_flags; member
126 .alloc_flags = TTM_ALLOCATION_POOL_USE_DMA_ALLOC |
132 .alloc_flags = TTM_ALLOCATION_POOL_USE_DMA32,
137 .alloc_flags = TTM_ALLOCATION_POOL_USE_DMA_ALLOC,
161 err = ttm_device_kunit_init(priv, ttm_dev, params->alloc_flags); in ttm_device_init_pools()
167 KUNIT_EXPECT_EQ(test, pool->alloc_flags, params->alloc_flags); in ttm_device_init_pools()
H A Dttm_kunit_helpers.c120 unsigned int alloc_flags, in ttm_device_kunit_init_with_funcs() argument
129 alloc_flags); in ttm_device_kunit_init_with_funcs()
145 unsigned int alloc_flags) in ttm_device_kunit_init() argument
147 return ttm_device_kunit_init_with_funcs(priv, ttm, alloc_flags, in ttm_device_kunit_init()
H A Dttm_pool_test.c16 unsigned int alloc_flags; member
117 .alloc_flags = TTM_ALLOCATION_POOL_USE_DMA_ALLOC,
122 .alloc_flags = TTM_ALLOCATION_POOL_USE_DMA_ALLOC,
154 ttm_pool_init(pool, devs->dev, NUMA_NO_NODE, params->alloc_flags); in ttm_pool_alloc_basic()
158 KUNIT_ASSERT_EQ(test, pool->alloc_flags, params->alloc_flags); in ttm_pool_alloc_basic()
H A Dttm_kunit_helpers.h31 unsigned int alloc_flags);
/linux/drivers/gpu/drm/ttm/
H A Dttm_pool_internal.h12 return pool->alloc_flags & TTM_ALLOCATION_POOL_USE_DMA_ALLOC; in ttm_pool_uses_dma_alloc()
17 return pool->alloc_flags & TTM_ALLOCATION_POOL_USE_DMA32; in ttm_pool_uses_dma32()
22 return pool->alloc_flags & 0xff; in ttm_pool_beneficial_order()
H A Dttm_pool.c1077 int nid, unsigned int alloc_flags) in ttm_pool_init() argument
1085 pool->alloc_flags = alloc_flags; in ttm_pool_init()
/linux/fs/xfs/libxfs/
H A Dxfs_alloc.c1676 uint32_t alloc_flags) in xfs_alloc_ag_vextent_near() argument
1696 alloc_flags |= XFS_ALLOC_FLAG_TRYFLUSH; in xfs_alloc_ag_vextent_near()
1762 alloc_flags); in xfs_alloc_ag_vextent_near()
1766 alloc_flags &= ~XFS_ALLOC_FLAG_TRYFLUSH; in xfs_alloc_ag_vextent_near()
1792 uint32_t alloc_flags) in xfs_alloc_ag_vextent_size() argument
1807 alloc_flags |= XFS_ALLOC_FLAG_TRYFLUSH; in xfs_alloc_ag_vextent_size()
1879 alloc_flags); in xfs_alloc_ag_vextent_size()
1883 alloc_flags &= ~XFS_ALLOC_FLAG_TRYFLUSH; in xfs_alloc_ag_vextent_size()
1979 alloc_flags); in xfs_alloc_ag_vextent_size()
1983 alloc_flags &= ~XFS_ALLOC_FLAG_TRYFLUSH; in xfs_alloc_ag_vextent_size()
[all …]
H A Dxfs_alloc.h198 int xfs_alloc_fix_freelist(struct xfs_alloc_arg *args, uint32_t alloc_flags);
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_amdkfd_gpuvm.c312 u32 alloc_flags = bo->kfd_bo->alloc_flags; in amdgpu_amdkfd_release_notify() local
315 amdgpu_amdkfd_unreserve_mem_limit(adev, size, alloc_flags, in amdgpu_amdkfd_release_notify()
342 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) in create_dmamap_sg_bo()
519 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE) in get_pte_flags()
521 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_EXECUTABLE) in get_pte_flags()
564 mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmamap_userptr()
660 mmio = (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_MMIO_REMAP); in kfd_mem_dmamap_sg_bo()
666 dir = mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmamap_sg_bo()
725 mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmaunmap_userptr()
787 dir = mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmaunmap_sg_bo()
[all …]
/linux/include/drm/ttm/
H A Dttm_pool.h74 unsigned int alloc_flags; member
86 int nid, unsigned int alloc_flags);
H A Dttm_device.h226 unsigned int alloc_flags; member
301 unsigned int alloc_flags);
/linux/include/linux/
H A Dstackdepot.h159 gfp_t alloc_flags,
178 unsigned int nr_entries, gfp_t alloc_flags);
H A Dcompaction.h93 unsigned int order, unsigned int alloc_flags,
104 int alloc_flags);
H A Dgfp.h447 acr_flags_t alloc_flags, gfp_t gfp_mask);
452 acr_flags_t alloc_flags, gfp_t gfp_mask);
/linux/lib/
H A Dstackdepot.c643 gfp_t alloc_flags, in stack_depot_save_flags() argument
651 bool allow_spin = gfpflags_allow_spinning(alloc_flags); in stack_depot_save_flags()
685 page = alloc_pages(gfp_nested_mask(alloc_flags), in stack_depot_save_flags()
745 gfp_t alloc_flags) in stack_depot_save() argument
747 return stack_depot_save_flags(entries, nr_entries, alloc_flags, in stack_depot_save()
/linux/drivers/md/
H A Ddm-zoned-reclaim.c286 int alloc_flags = DMZ_ALLOC_SEQ; in dmz_reclaim_rnd_data() local
292 alloc_flags | DMZ_ALLOC_RECLAIM); in dmz_reclaim_rnd_data()
293 if (!szone && alloc_flags == DMZ_ALLOC_SEQ && dmz_nr_cache_zones(zmd)) { in dmz_reclaim_rnd_data()
294 alloc_flags = DMZ_ALLOC_RND; in dmz_reclaim_rnd_data()
/linux/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_chardev.c1972 bo_bucket->alloc_flags = (uint32_t)kgd_mem->alloc_flags; in criu_checkpoint_bos()
1975 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) { in criu_checkpoint_bos()
1983 if (bo_bucket->alloc_flags in criu_checkpoint_bos()
1986 bo_bucket->alloc_flags & in criu_checkpoint_bos()
1995 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_DOORBELL) in criu_checkpoint_bos()
1998 else if (bo_bucket->alloc_flags & in criu_checkpoint_bos()
2016 bo_bucket->alloc_flags, in criu_checkpoint_bos()
2329 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_DOORBELL) { in criu_restore_memory_of_gpu()
2337 } else if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_MMIO_REMAP) { in criu_restore_memory_of_gpu()
2348 } else if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) { in criu_restore_memory_of_gpu()
[all …]
/linux/fs/xfs/
H A Dxfs_extent_busy.c613 uint32_t alloc_flags) in xfs_extent_busy_flush() argument
625 if (alloc_flags & XFS_ALLOC_FLAG_TRYFLUSH) in xfs_extent_busy_flush()
631 if (alloc_flags & XFS_ALLOC_FLAG_FREEING) in xfs_extent_busy_flush()
H A Dxfs_extent_busy.h59 unsigned busy_gen, uint32_t alloc_flags);
/linux/fs/btrfs/
H A Dblock-group.c3073 u64 alloc_flags; in btrfs_inc_block_group_ro() local
3126 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); in btrfs_inc_block_group_ro()
3127 if (alloc_flags != cache->flags) { in btrfs_inc_block_group_ro()
3128 ret = btrfs_chunk_alloc(trans, space_info, alloc_flags, in btrfs_inc_block_group_ro()
3156 alloc_flags = btrfs_get_alloc_profile(fs_info, space_info->flags); in btrfs_inc_block_group_ro()
3157 ret = btrfs_chunk_alloc(trans, space_info, alloc_flags, CHUNK_ALLOC_FORCE); in btrfs_inc_block_group_ro()
3173 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); in btrfs_inc_block_group_ro()
3175 check_system_chunk(trans, alloc_flags); in btrfs_inc_block_group_ro()
4054 u64 alloc_flags = btrfs_get_alloc_profile(trans->fs_info, type); in btrfs_force_chunk_alloc() local
4063 return btrfs_chunk_alloc(trans, space_info, alloc_flags, CHUNK_ALLOC_FORCE); in btrfs_force_chunk_alloc()

12