| /linux/fs/btrfs/ |
| H A D | extent-io-tree.c | 166 static struct extent_state *alloc_extent_state_atomic(struct extent_state *prealloc) in alloc_extent_state_atomic() argument 168 if (!prealloc) in alloc_extent_state_atomic() 169 prealloc = alloc_extent_state(GFP_ATOMIC); in alloc_extent_state_atomic() 171 return prealloc; in alloc_extent_state_atomic() 504 struct extent_state *prealloc, u64 split) in split_state() argument 512 prealloc->start = orig->start; in split_state() 513 prealloc->end = split - 1; in split_state() 514 prealloc->state = orig->state; in split_state() 525 if (prealloc->end < entry->start) { in split_state() 527 } else if (prealloc->end > entry->end) { in split_state() [all …]
|
| H A D | ulist.c | 53 ulist->prealloc = NULL; in ulist_init() 72 kfree(ulist->prealloc); in ulist_release() 73 ulist->prealloc = NULL; in ulist_release() 113 if (!ulist->prealloc) in ulist_prealloc() 114 ulist->prealloc = kzalloc(sizeof(*ulist->prealloc), gfp_mask); in ulist_prealloc() 218 if (ulist->prealloc) { in ulist_add_merge() 219 node = ulist->prealloc; in ulist_add_merge() 220 ulist->prealloc = NULL; in ulist_add_merge()
|
| H A D | qgroup.c | 202 struct btrfs_qgroup *prealloc, in add_qgroup_rb() argument 208 ASSERT(prealloc); in add_qgroup_rb() 210 prealloc->qgroupid = qgroupid; in add_qgroup_rb() 211 node = rb_find_add(&prealloc->node, &fs_info->qgroup_tree, btrfs_qgroup_qgroupid_cmp); in add_qgroup_rb() 213 kfree(prealloc); in add_qgroup_rb() 217 INIT_LIST_HEAD(&prealloc->groups); in add_qgroup_rb() 218 INIT_LIST_HEAD(&prealloc->members); in add_qgroup_rb() 219 INIT_LIST_HEAD(&prealloc->dirty); in add_qgroup_rb() 220 INIT_LIST_HEAD(&prealloc->iterator); in add_qgroup_rb() 221 INIT_LIST_HEAD(&prealloc->nested_iterator); in add_qgroup_rb() [all …]
|
| H A D | ulist.h | 44 struct ulist_node *prealloc; member
|
| H A D | extent_io.c | 870 struct btrfs_folio_state *prealloc) in attach_extent_buffer_folio() argument 894 btrfs_free_folio_state(prealloc); in attach_extent_buffer_folio() 898 if (prealloc) in attach_extent_buffer_folio() 900 folio_attach_private(folio, prealloc); in attach_extent_buffer_folio() 3286 struct btrfs_folio_state *prealloc, in attach_eb_folio_to_filemap() argument 3347 ret = attach_extent_buffer_folio(eb, eb->folios[i], prealloc); in attach_eb_folio_to_filemap() 3369 struct btrfs_folio_state *prealloc = NULL; in alloc_extent_buffer() local 3414 prealloc = btrfs_alloc_folio_state(fs_info, PAGE_SIZE, BTRFS_SUBPAGE_METADATA); in alloc_extent_buffer() 3415 if (IS_ERR(prealloc)) { in alloc_extent_buffer() 3416 ret = PTR_ERR(prealloc); in alloc_extent_buffer() [all …]
|
| H A D | qgroup.h | 342 struct btrfs_qgroup_list *prealloc);
|
| /linux/lib/ |
| H A D | stackdepot.c | 293 static bool depot_init_pool(void **prealloc) in depot_init_pool() argument 305 if (!new_pool && *prealloc) { in depot_init_pool() 307 WRITE_ONCE(new_pool, *prealloc); in depot_init_pool() 308 *prealloc = NULL; in depot_init_pool() 340 static void depot_keep_new_pool(void **prealloc) in depot_keep_new_pool() argument 351 WRITE_ONCE(new_pool, *prealloc); in depot_keep_new_pool() 352 *prealloc = NULL; in depot_keep_new_pool() 359 static struct stack_record *depot_pop_free_pool(void **prealloc, size_t size) in depot_pop_free_pool() argument 368 if (!depot_init_pool(prealloc)) in depot_pop_free_pool() 430 …ck(unsigned long *entries, unsigned int nr_entries, u32 hash, depot_flags_t flags, void **prealloc) in depot_alloc_stack() argument [all …]
|
| /linux/tools/testing/selftests/net/ |
| H A D | ioam6.sh | 210 encap ioam6 trace prealloc type 0x800000 ns 0 size 4 dev veth0 &>/dev/null 693 encap ioam6 mode $mode trace prealloc type 0x800000 size 4 \ 699 encap ioam6 mode $mode trace prealloc type 0x800000 ns 0 size 4 \ 734 encap ioam6 mode $mode trace prealloc type 0x800000 ns 0 size 4 \ 740 encap ioam6 mode $mode_tunsrc trace prealloc type 0x800000 ns 0 size 4 \ 774 encap ioam6 mode $mode trace prealloc type 0x800000 ns 0 size 4 \ 780 encap ioam6 mode $mode_tundst trace prealloc type 0x800000 ns 0 size 4 \ 811 encap ioam6 mode $mode trace prealloc ns 0 size 4 \ 817 encap ioam6 mode $mode trace prealloc type 0x800000 ns 0 size 4 \ 842 encap ioam6 mode $mode trace prealloc type 0x800000 ns 0 \ [all …]
|
| H A D | lwt_dst_cache_ref_loop.sh | 59 encap ioam6 trace prealloc type 0x800000 ns 0 size 4 \ 168 encap ioam6 trace prealloc type 0x800000 ns 1 size 4 \
|
| /linux/arch/powerpc/platforms/ps3/ |
| H A D | setup.c | 113 static void __init prealloc(struct ps3_prealloc *p) in prealloc() function 132 #define prealloc_ps3fb_videomemory() prealloc(&ps3fb_videomemory) 155 #define prealloc_ps3flash_bounce_buffer() prealloc(&ps3flash_bounce_buffer)
|
| /linux/fs/ext4/ |
| H A D | extents_status.c | 179 struct extent_status *prealloc); 182 struct extent_status *prealloc); 188 struct pending_reservation **prealloc); 819 struct extent_status *prealloc) in __es_insert_extent() argument 859 if (prealloc) in __es_insert_extent() 860 es = prealloc; in __es_insert_extent() 1415 struct extent_status *prealloc) in __es_remove_extent() argument 1462 err = __es_insert_extent(inode, &newes, prealloc); in __es_remove_extent() 1991 struct pending_reservation **prealloc) in __insert_pending() argument 2017 if (likely(*prealloc == NULL)) { in __insert_pending() [all …]
|
| /linux/net/sched/ |
| H A D | sch_gred.c | 484 struct gred_sched_data **prealloc, in gred_change_vq() argument 496 table->tab[dp] = q = *prealloc; in gred_change_vq() 497 *prealloc = NULL; in gred_change_vq() 651 struct gred_sched_data *prealloc; in gred_change() local 703 prealloc = kzalloc(sizeof(*prealloc), GFP_KERNEL); in gred_change() 706 err = gred_change_vq(sch, ctl->DP, ctl, prio, stab, max_P, &prealloc, in gred_change() 721 kfree(prealloc); in gred_change() 728 kfree(prealloc); in gred_change()
|
| /linux/arch/arc/lib/ |
| H A D | memset-archs.S | 21 prealloc [\reg, \off]
|
| /linux/drivers/gpu/drm/msm/ |
| H A D | msm_gem_vma.c | 118 struct msm_mmu_prealloc prealloc; member 690 vm->mmu->prealloc = &job->prealloc; in msm_vma_job_run() 715 vm->mmu->prealloc = NULL; in msm_vma_job_run() 742 vm->mmu->funcs->prealloc_cleanup(vm->mmu, &job->prealloc); in msm_vma_job_free() 744 atomic_sub(job->prealloc.count, &vm->prealloc_throttle.in_flight); in msm_vma_job_free() 1137 mmu->funcs->prealloc_count(mmu, &job->prealloc, start_iova, end_iova - start_iova); in prealloc_count() 1205 atomic_add(job->prealloc.count, &vm->prealloc_throttle.in_flight); in vm_bind_prealloc_count() 1333 ret = mmu->funcs->prealloc_allocate(mmu, &job->prealloc); in vm_bind_job_prepare()
|
| H A D | msm_mmu.h | 66 struct msm_mmu_prealloc *prealloc; member
|
| /linux/drivers/media/platform/renesas/vsp1/ |
| H A D | vsp1_dl.c | 1114 unsigned int prealloc) in vsp1_dlm_create() argument 1149 dlm->pool = vsp1_dl_body_pool_create(vsp1, prealloc + 1, in vsp1_dlm_create() 1154 for (i = 0; i < prealloc; ++i) { in vsp1_dlm_create() 1171 dlm->list_count = prealloc; in vsp1_dlm_create() 1175 VSP1_EXTCMD_AUTOFLD, prealloc); in vsp1_dlm_create()
|
| H A D | vsp1_dl.h | 56 unsigned int prealloc);
|
| /linux/drivers/gpu/drm/i915/display/ |
| H A D | intel_fbdev.c | 220 bool prealloc = false; in intel_fbdev_driver_fbdev_probe() local 245 prealloc = true; in intel_fbdev_driver_fbdev_probe() 290 if (!intel_bo_is_shmem(obj) && !prealloc) in intel_fbdev_driver_fbdev_probe()
|
| /linux/fs/xfs/ |
| H A D | xfs_iomap.c | 1462 xfs_filblks_t prealloc, in xfs_bmapi_reserve_delalloc() argument 1475 whichfork == XFS_COW_FORK && !prealloc; in xfs_bmapi_reserve_delalloc() 1483 alen = XFS_FILBLKS_MIN(len + prealloc, XFS_MAX_BMBT_EXTLEN); in xfs_bmapi_reserve_delalloc() 1486 if (prealloc && alen >= len) in xfs_bmapi_reserve_delalloc() 1487 prealloc = alen - len; in xfs_bmapi_reserve_delalloc() 1555 if (whichfork == XFS_DATA_FORK && prealloc) in xfs_bmapi_reserve_delalloc() 1557 if (whichfork == XFS_COW_FORK && (prealloc || aoff < off || alen > len)) in xfs_bmapi_reserve_delalloc() 1572 if (prealloc || use_cowextszhint) { in xfs_bmapi_reserve_delalloc() 1575 prealloc = 0; in xfs_bmapi_reserve_delalloc()
|
| H A D | xfs_dquot.c | 125 int prealloc = 0; in xfs_qm_adjust_dqlimits() local 132 prealloc = 1; in xfs_qm_adjust_dqlimits() 136 prealloc = 1; in xfs_qm_adjust_dqlimits() 147 if (prealloc) in xfs_qm_adjust_dqlimits()
|
| /linux/fs/sysfs/ |
| H A D | file.c | 238 .prealloc = true, 243 .prealloc = true, 249 .prealloc = true,
|
| /linux/kernel/bpf/ |
| H A D | hashtab.c | 429 bool prealloc = !(attr->map_flags & BPF_F_NO_PREALLOC); in htab_map_alloc_check() local 447 if (lru && !prealloc) in htab_map_alloc_check() 485 bool prealloc = !(attr->map_flags & BPF_F_NO_PREALLOC); in htab_map_alloc() local 568 if (prealloc) { in htab_map_alloc() 998 bool prealloc = htab_is_prealloc(htab); in alloc_htab_elem() local 1002 if (prealloc) { in alloc_htab_elem() 1038 if (prealloc) { in alloc_htab_elem() 1055 if (!prealloc) in alloc_htab_elem() 2220 bool prealloc = htab_is_prealloc(htab); in htab_map_mem_usage() local 2228 if (prealloc) { in htab_map_mem_usage()
|
| /linux/drivers/md/ |
| H A D | dm-cache-target.c | 1403 struct dm_bio_prison_cell_v2 *prealloc; in mg_lock_writes() local 1405 prealloc = alloc_prison_cell(cache); in mg_lock_writes() 1415 prealloc, &mg->cell); in mg_lock_writes() 1417 free_prison_cell(cache, prealloc); in mg_lock_writes() 1422 if (mg->cell != prealloc) in mg_lock_writes() 1423 free_prison_cell(cache, prealloc); in mg_lock_writes() 1533 struct dm_bio_prison_cell_v2 *prealloc; in invalidate_lock() local 1535 prealloc = alloc_prison_cell(cache); in invalidate_lock() 1539 READ_WRITE_LOCK_LEVEL, prealloc, &mg->cell); in invalidate_lock() 1541 free_prison_cell(cache, prealloc); in invalidate_lock() [all …]
|
| /linux/drivers/usb/gadget/function/ |
| H A D | u_ether.c | 313 static int prealloc(struct list_head *list, struct usb_ep *ep, unsigned n) in prealloc() function 357 status = prealloc(&dev->tx_reqs, link->in_ep, n); in alloc_requests() 360 status = prealloc(&dev->rx_reqs, link->out_ep, n); in alloc_requests()
|
| /linux/mm/ |
| H A D | memory.c | 1048 struct folio **prealloc, struct page *page) in copy_present_page() argument 1053 new_folio = *prealloc; in copy_present_page() 1065 *prealloc = NULL; in copy_present_page() 1114 int max_nr, int *rss, struct folio **prealloc) in copy_present_ptes() argument 1132 if (unlikely(!*prealloc && folio_test_large(folio) && max_nr != 1)) { in copy_present_ptes() 1169 addr, rss, prealloc, page); in copy_present_ptes() 1221 struct folio *prealloc = NULL; in copy_pte_range() local 1304 ptent, addr, max_nr, rss, &prealloc); in copy_pte_range() 1312 if (unlikely(prealloc)) { in copy_pte_range() 1319 folio_put(prealloc); in copy_pte_range() [all …]
|