Home
last modified time | relevance | path

Searched refs:alloc_flags (Results 1 – 25 of 28) sorted by relevance

12

/linux/mm/
H A Dpage_alloc.c294 int alloc_flags);
1857 unsigned int alloc_flags) in prep_new_page() argument
1870 if (alloc_flags & ALLOC_NO_WATERMARKS) in prep_new_page()
2290 int block_type, unsigned int alloc_flags) in try_to_claim_block() argument
2311 if (boost_watermark(zone) && (alloc_flags & ALLOC_KSWAPD)) in try_to_claim_block()
2364 unsigned int alloc_flags) in __rmqueue_claim() argument
2377 if (order < pageblock_order && alloc_flags & ALLOC_NOFRAGMENT) in __rmqueue_claim()
2402 alloc_flags); in __rmqueue_claim()
2455 unsigned int alloc_flags, enum rmqueue_mode *mode) in __rmqueue() argument
2465 if (alloc_flags in __rmqueue()
2525 rmqueue_bulk(struct zone * zone,unsigned int order,unsigned long count,struct list_head * list,int migratetype,unsigned int alloc_flags) rmqueue_bulk() argument
3147 rmqueue_buddy(struct zone * preferred_zone,struct zone * zone,unsigned int order,unsigned int alloc_flags,int migratetype) rmqueue_buddy() argument
3245 __rmqueue_pcplist(struct zone * zone,unsigned int order,int migratetype,unsigned int alloc_flags,struct per_cpu_pages * pcp,struct list_head * list) __rmqueue_pcplist() argument
3276 rmqueue_pcplist(struct zone * preferred_zone,struct zone * zone,unsigned int order,int migratetype,unsigned int alloc_flags) rmqueue_pcplist() argument
3323 rmqueue(struct zone * preferred_zone,struct zone * zone,unsigned int order,gfp_t gfp_flags,unsigned int alloc_flags,int migratetype) rmqueue() argument
3487 __zone_watermark_unusable_free(struct zone * z,unsigned int order,unsigned int alloc_flags) __zone_watermark_unusable_free() argument
3514 __zone_watermark_ok(struct zone * z,unsigned int order,unsigned long mark,int highest_zoneidx,unsigned int alloc_flags,long free_pages) __zone_watermark_ok() argument
3592 zone_watermark_ok(struct zone * z,unsigned int order,unsigned long mark,int highest_zoneidx,unsigned int alloc_flags) zone_watermark_ok() argument
3600 zone_watermark_fast(struct zone * z,unsigned int order,unsigned long mark,int highest_zoneidx,unsigned int alloc_flags,gfp_t gfp_mask) zone_watermark_fast() argument
3669 unsigned int alloc_flags; alloc_flags_nofragment() local
3705 gfp_to_alloc_flags_cma(gfp_t gfp_mask,unsigned int alloc_flags) gfp_to_alloc_flags_cma() argument
3719 get_page_from_freelist(gfp_t gfp_mask,unsigned int order,int alloc_flags,const struct alloc_context * ac) get_page_from_freelist() argument
3938 __alloc_pages_cpuset_fallback(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac) __alloc_pages_cpuset_fallback() argument
4052 __alloc_pages_direct_compact(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac,enum compact_priority prio,enum compact_result * compact_result) __alloc_pages_direct_compact() argument
4110 should_compact_retry(struct alloc_context * ac,int order,int alloc_flags,enum compact_result compact_result,enum compact_priority * compact_priority,int * compaction_retries) should_compact_retry() argument
4177 __alloc_pages_direct_compact(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac,enum compact_priority prio,enum compact_result * compact_result) __alloc_pages_direct_compact() argument
4185 should_compact_retry(struct alloc_context * ac,unsigned int order,int alloc_flags,enum compact_result compact_result,enum compact_priority * compact_priority,int * compaction_retries) should_compact_retry() argument
4324 __alloc_pages_direct_reclaim(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac,unsigned long * did_some_progress) __alloc_pages_direct_reclaim() argument
4384 unsigned int alloc_flags = ALLOC_WMARK_MIN | ALLOC_CPUSET; gfp_to_alloc_flags() local
4487 should_reclaim_retry(gfp_t gfp_mask,unsigned order,struct alloc_context * ac,int alloc_flags,bool did_some_progress,int * no_progress_loops) should_reclaim_retry() argument
4604 unsigned int alloc_flags; __alloc_pages_slowpath() local
4897 prepare_alloc_pages(gfp_t gfp_mask,unsigned int order,int preferred_nid,nodemask_t * nodemask,struct alloc_context * ac,gfp_t * alloc_gfp,unsigned int * alloc_flags) prepare_alloc_pages() argument
4970 unsigned int alloc_flags = ALLOC_WMARK_LOW; alloc_pages_bulk_noprof() local
5116 unsigned int alloc_flags = ALLOC_WMARK_LOW; __alloc_frozen_pages_noprof() local
6838 alloc_contig_range_noprof(unsigned long start,unsigned long end,acr_flags_t alloc_flags,gfp_t gfp_mask) alloc_contig_range_noprof() argument
7404 cond_accept_memory(struct zone * zone,unsigned int order,int alloc_flags) cond_accept_memory() argument
7468 cond_accept_memory(struct zone * zone,unsigned int order,int alloc_flags) cond_accept_memory() argument
7520 unsigned int alloc_flags = ALLOC_TRYLOCK; alloc_pages_nolock_noprof() local
[all...]
H A Dcompaction.c2299 cc->highest_zoneidx, cc->alloc_flags, in __compact_finished()
2433 int alloc_flags) in compaction_zonelist_suitable() argument
2471 int highest_zoneidx, unsigned int alloc_flags, in compaction_suit_allocation_order() argument
2482 watermark = wmark_pages(zone, alloc_flags & ALLOC_WMARK_MASK); in compaction_suit_allocation_order()
2484 alloc_flags, free_pages)) in compaction_suit_allocation_order()
2497 !(alloc_flags & ALLOC_CMA)) { in compaction_suit_allocation_order()
2539 cc->alloc_flags, in compact_zone()
2751 unsigned int alloc_flags, int highest_zoneidx, in compact_zone_order() argument
2762 .alloc_flags = alloc_flags, in compact_zone_order()
2815 try_to_compact_pages(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac,enum compact_priority prio,struct page ** capture) try_to_compact_pages() argument
3035 unsigned int alloc_flags = defrag_mode ? kcompactd_node_suitable() local
[all...]
/linux/drivers/base/regmap/
H A Dregcache-maple.c76 entry = kmalloc_array(last - index + 1, sizeof(*entry), map->alloc_flags); in regcache_maple_write()
94 ret = mas_store_gfp(&mas, entry, map->alloc_flags); in regcache_maple_write()
137 map->alloc_flags); in regcache_maple_drop()
150 map->alloc_flags); in regcache_maple_drop()
164 ret = mas_store_gfp(&mas, lower, map->alloc_flags); in regcache_maple_drop()
172 ret = mas_store_gfp(&mas, upper, map->alloc_flags); in regcache_maple_drop()
206 buf = kmalloc_array(max - min, val_bytes, map->alloc_flags); in regcache_maple_sync_block()
322 entry = kmalloc_array(last - first + 1, sizeof(*entry), map->alloc_flags); in regcache_maple_insert_block()
333 ret = mas_store_gfp(&mas, entry, map->alloc_flags); in regcache_maple_insert_block()
350 mt = kmalloc(sizeof(*mt), map->alloc_flags); in regcache_maple_init()
[all...]
H A Dregcache-rbtree.c190 map->cache = kmalloc(sizeof *rbtree_ctx, map->alloc_flags); in regcache_rbtree_init()
278 blk = krealloc_array(rbnode->block, blklen, map->cache_word_size, map->alloc_flags); in regcache_rbtree_insert_to_block()
287 map->alloc_flags); in regcache_rbtree_insert_to_block()
321 rbnode = kzalloc(sizeof(*rbnode), map->alloc_flags); in regcache_rbtree_node_alloc()
347 map->alloc_flags); in regcache_rbtree_node_alloc()
353 map->alloc_flags); in regcache_rbtree_node_alloc()
H A Dregcache-flat.c30 + 1, sizeof(unsigned int), map->alloc_flags); in regcache_flat_init()
H A Dinternal.h66 gfp_t alloc_flags; member
H A Dregmap.c767 map->alloc_flags = GFP_ATOMIC; in __regmap_init()
769 map->alloc_flags = GFP_KERNEL; in __regmap_init()
2365 wval = kmemdup_array(val, val_count, val_bytes, map->alloc_flags); in regmap_bulk_write()
/linux/fs/xfs/libxfs/
H A Dxfs_alloc.c1676 uint32_t alloc_flags) in xfs_alloc_ag_vextent_near() argument
1696 alloc_flags |= XFS_ALLOC_FLAG_TRYFLUSH; in xfs_alloc_ag_vextent_near()
1762 alloc_flags); in xfs_alloc_ag_vextent_near()
1766 alloc_flags &= ~XFS_ALLOC_FLAG_TRYFLUSH; in xfs_alloc_ag_vextent_near()
1792 uint32_t alloc_flags) in xfs_alloc_ag_vextent_size() argument
1807 alloc_flags |= XFS_ALLOC_FLAG_TRYFLUSH; in xfs_alloc_ag_vextent_size()
1879 alloc_flags); in xfs_alloc_ag_vextent_size()
1883 alloc_flags &= ~XFS_ALLOC_FLAG_TRYFLUSH; in xfs_alloc_ag_vextent_size()
1979 alloc_flags); in xfs_alloc_ag_vextent_size()
1983 alloc_flags in xfs_alloc_ag_vextent_size()
2828 xfs_alloc_fix_freelist(struct xfs_alloc_arg * args,uint32_t alloc_flags) xfs_alloc_fix_freelist() argument
3557 xfs_alloc_vextent_prepare_ag(struct xfs_alloc_arg * args,uint32_t alloc_flags) xfs_alloc_vextent_prepare_ag() argument
3687 uint32_t alloc_flags = 0; xfs_alloc_vextent_this_ag() local
3737 xfs_alloc_vextent_iterate_ags(struct xfs_alloc_arg * args,xfs_agnumber_t minimum_agno,xfs_agnumber_t start_agno,xfs_agblock_t target_agbno,uint32_t alloc_flags) xfs_alloc_vextent_iterate_ags() argument
3813 uint32_t alloc_flags = XFS_ALLOC_FLAG_TRYLOCK; xfs_alloc_vextent_start_ag() local
3867 uint32_t alloc_flags = XFS_ALLOC_FLAG_TRYLOCK; xfs_alloc_vextent_first_ag() local
3939 uint32_t alloc_flags = 0; xfs_alloc_vextent_near_bno() local
[all...]
H A Dxfs_alloc.h198 int xfs_alloc_fix_freelist(struct xfs_alloc_arg *args, uint32_t alloc_flags);
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_amdkfd_gpuvm.c296 u32 alloc_flags = bo->kfd_bo->alloc_flags; in amdgpu_amdkfd_release_notify() local
299 amdgpu_amdkfd_unreserve_mem_limit(adev, size, alloc_flags, in amdgpu_amdkfd_release_notify()
326 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) in create_dmamap_sg_bo()
502 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE) in get_pte_flags()
504 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_EXECUTABLE) in get_pte_flags()
547 mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmamap_userptr()
643 mmio = (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_MMIO_REMAP); in kfd_mem_dmamap_sg_bo()
649 dir = mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmamap_sg_bo()
708 mem->alloc_flags in kfd_mem_dmaunmap_userptr()
1674 u64 alloc_flags; amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() local
[all...]
/linux/include/linux/
H A Dcompaction.h93 unsigned int order, unsigned int alloc_flags,
104 int alloc_flags);
H A Dgfp.h433 acr_flags_t alloc_flags, gfp_t gfp_mask);
H A Dmmzone.h1519 int highest_zoneidx, unsigned int alloc_flags,
1523 unsigned int alloc_flags);
/linux/lib/
H A Dstackdepot.c635 gfp_t alloc_flags, in stack_depot_save_flags() argument
643 bool allow_spin = gfpflags_allow_spinning(alloc_flags); in stack_depot_save_flags()
677 page = alloc_pages(gfp_nested_mask(alloc_flags), in stack_depot_save_flags()
737 gfp_t alloc_flags) in stack_depot_save() argument
739 return stack_depot_save_flags(entries, nr_entries, alloc_flags, in stack_depot_save()
/linux/drivers/md/
H A Ddm-zoned-reclaim.c286 int alloc_flags = DMZ_ALLOC_SEQ; in dmz_reclaim_rnd_data() local
292 alloc_flags | DMZ_ALLOC_RECLAIM); in dmz_reclaim_rnd_data()
293 if (!szone && alloc_flags == DMZ_ALLOC_SEQ && dmz_nr_cache_zones(zmd)) { in dmz_reclaim_rnd_data()
294 alloc_flags = DMZ_ALLOC_RND; in dmz_reclaim_rnd_data()
H A Ddm-zoned-metadata.c2051 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; in dmz_get_chunk_mapping() local
2066 dzone = dmz_alloc_zone(zmd, 0, alloc_flags); in dmz_get_chunk_mapping()
2163 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; in dmz_get_chunk_buffer() local
2172 bzone = dmz_alloc_zone(zmd, 0, alloc_flags); in dmz_get_chunk_buffer()
/linux/fs/xfs/
H A Dxfs_extent_busy.c614 uint32_t alloc_flags) in xfs_extent_busy_flush() argument
626 if (alloc_flags & XFS_ALLOC_FLAG_TRYFLUSH) in xfs_extent_busy_flush()
632 if (alloc_flags & XFS_ALLOC_FLAG_FREEING) in xfs_extent_busy_flush()
H A Dxfs_extent_busy.h59 unsigned busy_gen, uint32_t alloc_flags);
/linux/fs/btrfs/
H A Dblock-group.c2982 u64 alloc_flags; in btrfs_inc_block_group_ro() local
3030 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); in btrfs_inc_block_group_ro()
3031 if (alloc_flags != cache->flags) { in btrfs_inc_block_group_ro()
3032 ret = btrfs_chunk_alloc(trans, space_info, alloc_flags, in btrfs_inc_block_group_ro()
3060 alloc_flags = btrfs_get_alloc_profile(fs_info, space_info->flags); in btrfs_inc_block_group_ro()
3061 ret = btrfs_chunk_alloc(trans, space_info, alloc_flags, CHUNK_ALLOC_FORCE); in btrfs_inc_block_group_ro()
3077 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); in btrfs_inc_block_group_ro()
3079 check_system_chunk(trans, alloc_flags); in btrfs_inc_block_group_ro()
3913 u64 alloc_flags = btrfs_get_alloc_profile(trans->fs_info, type); in btrfs_force_chunk_alloc() local
3922 return btrfs_chunk_alloc(trans, space_info, alloc_flags, CHUNK_ALLOC_FORC in btrfs_force_chunk_alloc()
[all...]
H A Dzoned.c2511 u64 alloc_flags; in btrfs_zoned_reserve_data_reloc_bg() local
2527 alloc_flags = btrfs_get_alloc_profile(fs_info, space_info->flags); in btrfs_zoned_reserve_data_reloc_bg()
2528 index = btrfs_bg_flags_to_raid_index(alloc_flags); in btrfs_zoned_reserve_data_reloc_bg()
2555 ret = btrfs_chunk_alloc(trans, space_info, alloc_flags, CHUNK_ALLOC_FORCE); in btrfs_zoned_reserve_data_reloc_bg()
/linux/fs/xfs/scrub/
H A Drepair.h57 int xrep_fix_freelist(struct xfs_scrub *sc, int alloc_flags);
H A Drepair.c454 int alloc_flags) in xrep_fix_freelist() argument
464 return xfs_alloc_fix_freelist(&args, alloc_flags); in xrep_fix_freelist()
/linux/drivers/gpu/drm/i915/gem/
H A Di915_gem_object.h50 unsigned alloc_flags);
/linux/drivers/iommu/
H A Ddma-iommu.c897 gfp_t alloc_flags = gfp; in __iommu_dma_alloc_pages() local
901 alloc_flags |= __GFP_NORETRY; in __iommu_dma_alloc_pages()
902 page = alloc_pages_node(nid, alloc_flags, order); in __iommu_dma_alloc_pages()
/linux/include/uapi/linux/
H A Dkfd_ioctl.h727 __u32 alloc_flags; member

12