Home
last modified time | relevance | path

Searched full:pools (Results 1 – 25 of 284) sorted by relevance

12345678910>>...12

/linux/tools/testing/selftests/drivers/net/mlxsw/
H A Dsharedbuffer_configuration.py28 def __init__(self, pools): argument
30 for pool in pools:
72 # The threshold type of pools 4, 8, 9 and 10 cannot be changed
110 # The threshold type of pools 4, 8, 9 and 10 cannot be changed
187 pools = PoolList()
190 pools.append(Pool(pooldict))
191 return pools
194 def do_check_pools(dlname, pools, vp): argument
195 for pool in pools:
219 def check_pools(dlname, pools): argument
[all …]
H A Dqos_pfc.sh68 # iPOOL1+ePOOL5 are overflow pools.
69 # iPOOL2+ePOOL6 are PFC pools.
125 # pools
145 # Control traffic pools. Just reduce the size. Keep them dynamic so that
150 # Overflow pools.
154 # PFC pools. As per the writ, the size of egress PFC pool should be
/linux/tools/net/ynl/ynltool/
H A Dpage-pool.c158 printf("page pools: %u (zombies: %u)\n", in print_plain_stats()
199 print_json_pool_list(struct netdev_page_pool_get_list *pools, in print_json_pool_list() argument
205 ynl_dump_foreach(pools, pp) { in print_json_pool_list()
250 print_plain_pool_list(struct netdev_page_pool_get_list *pools, in print_plain_pool_list() argument
254 ynl_dump_foreach(pools, pp) { in print_plain_pool_list()
301 struct netdev_page_pool_get_list *pools, in aggregate_device_stats() argument
304 ynl_dump_foreach(pools, pp) { in aggregate_device_stats()
331 struct netdev_page_pool_get_list *pools; in do_stats() local
377 pools = netdev_page_pool_get_dump(ys); in do_stats()
378 if (!pools) { in do_stats()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/
H A Ddr_arg.c28 struct dr_arg_pool *pools[DR_ARG_CHUNK_SIZE_MAX]; member
201 arg_obj = dr_arg_pool_get_arg_obj(mgr->pools[size]); in mlx5dr_arg_get_obj()
226 dr_arg_pool_put_arg_obj(mgr->pools[arg_obj->log_chunk_size], arg_obj); in mlx5dr_arg_put_obj()
245 pool_mgr->pools[i] = dr_arg_pool_create(dmn, i); in mlx5dr_arg_mgr_create()
246 if (!pool_mgr->pools[i]) in mlx5dr_arg_mgr_create()
254 dr_arg_pool_destroy(pool_mgr->pools[i]); in mlx5dr_arg_mgr_create()
262 struct dr_arg_pool **pools; in mlx5dr_arg_mgr_destroy() local
268 pools = mgr->pools; in mlx5dr_arg_mgr_destroy()
270 dr_arg_pool_destroy(pools[i]); in mlx5dr_arg_mgr_destroy()
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
H A Dpool.c25 if (!xsk->pools) { in mlx5e_xsk_get_pools()
26 xsk->pools = kzalloc_objs(*xsk->pools, MLX5E_MAX_NUM_CHANNELS); in mlx5e_xsk_get_pools()
27 if (unlikely(!xsk->pools)) in mlx5e_xsk_get_pools()
40 kfree(xsk->pools); in mlx5e_xsk_put_pools()
41 xsk->pools = NULL; in mlx5e_xsk_put_pools()
53 xsk->pools[ix] = pool; in mlx5e_xsk_add_pool()
59 xsk->pools[ix] = NULL; in mlx5e_xsk_remove_pool()
H A Dpool.h12 if (!xsk || !xsk->pools) in mlx5e_xsk_get_pool()
18 return xsk->pools[ix]; in mlx5e_xsk_get_pool()
/linux/drivers/usb/core/
H A Dbuffer.c52 * hcd_buffer_create - initialize buffer pools
53 * @hcd: the bus whose buffer pools are to be initialized
58 * memory allocators. It initializes some pools of dma-coherent memory that
61 * Call hcd_buffer_destroy() to clean up after using those pools.
90 * hcd_buffer_destroy - deallocate buffer pools
91 * @hcd: the bus whose buffer pools are to be destroyed
95 * This frees the buffer pools created by hcd_buffer_create().
/linux/drivers/net/ethernet/ti/icssg/
H A Dicssg_config.h29 /* Defines for forwarding path buffer pools:
31 * - 8 total pools per slice
37 /* Defines for local injection path buffer pools:
39 * - 16 total pools per slice
40 * - 8 pools per port per slice and each slice handles both ports
41 * - only 4 out of 8 pools used per port (as only 4 real QoS levels in ICSSG)
42 * - switch mode: 8 total pools used
43 * - mac mode: 4 total pools used
H A Dicssg_config.c308 /* Configure buffer pools for forwarding buffers in prueth_fw_offload_buffer_setup()
310 * - 8 total pools per slice in prueth_fw_offload_buffer_setup()
318 /* Configure buffer pools for Local Injection buffers in prueth_fw_offload_buffer_setup()
320 * - 16 total pools per slice in prueth_fw_offload_buffer_setup()
395 /* Configure buffer pools for forwarding buffers in prueth_emac_buffer_setup()
396 * - in mac mode - no forwarding so initialize all pools to 0 in prueth_emac_buffer_setup()
397 * - 8 total pools per slice in prueth_emac_buffer_setup()
404 /* Configure buffer pools for Local Injection buffers in prueth_emac_buffer_setup()
406 * - 16 total pools per slice in prueth_emac_buffer_setup()
/linux/arch/sparc/kernel/
H A Diommu-common.c82 spin_lock_init(&(iommu->pools[i].lock)); in iommu_tbl_pool_init()
83 iommu->pools[i].start = start; in iommu_tbl_pool_init()
84 iommu->pools[i].hint = start; in iommu_tbl_pool_init()
86 iommu->pools[i].end = start - 1; in iommu_tbl_pool_init()
131 pool = &(iommu->pools[pool_nr]); in iommu_tbl_range_alloc()
161 pool = &(iommu->pools[0]); in iommu_tbl_range_alloc()
193 pool = &(iommu->pools[pool_nr]); in iommu_tbl_range_alloc()
237 p = &tbl->pools[pool_nr]; in get_pool()
/linux/kernel/cgroup/
H A Ddmem.c38 * @pools: List of pools linked to this region.
41 struct list_head pools; member
51 * No new pools should be added to the region afterwards.
59 struct list_head pools; member
198 list_for_each_entry_rcu(pool, &dmemcs->pools, css_node) in dmemcs_offline()
209 list_for_each_entry_safe(pool, next, &dmemcs->pools, css_node) { in dmemcs_free()
229 INIT_LIST_HEAD(&dmemcs->pools); in dmemcs_alloc()
238 list_for_each_entry_rcu(pool, &dmemcs->pools, css_node, spin_is_locked(&dmemcg_lock)) in find_cg_pool_locked()
270 list_for_each_entry_rcu(pool, &dmemcg_iter->pools, css_node) { in dmem_cgroup_calculate_protection()
386 list_add_tail_rcu(&pool->css_node, &dmemcs->pools); in alloc_pool_single()
[all …]
/linux/Documentation/core-api/
H A Dswiotlb.rst142 as one or more "pools". The default pool is allocated during system boot with a
146 CONFIG_SWIOTLB_DYNAMIC is enabled, additional pools may be allocated later in
224 New pools added via dynamic swiotlb are linked together in a linear list.
227 large number of dynamic pools. The data structures could be improved for
232 not wasted, with dynamic pools making more space available if needed (as long
239 which includes the default memory pool and any dynamic or transient pools
307 Restricted pools
309 The swiotlb machinery is also used for "restricted pools", which are pools of
311 use by a particular device. Restricted pools provide a level of DMA memory
318 Restricted pools add swiotlb_alloc() and swiotlb_free() APIs, which are called
H A Dworkqueue.rst60 * Use per-CPU unified worker pools shared by all wq to provide
85 worker-pools.
89 which manages worker-pools and processes the queued work items.
91 There are two worker-pools, one for normal work items and the other
93 worker-pools to serve work items queued on unbound workqueues - the
94 number of these backing pools is dynamic.
140 For unbound workqueues, the number of backing pools is dynamic.
143 backing worker pools matching the attributes. The responsibility of
194 worker-pools which host workers which are not bound to any
197 worker-pools try to start execution of work items as soon as
[all …]
/linux/drivers/soc/ti/
H A Dknav_qmss.h191 * @pools: list of descriptor pools in the region
203 struct list_head pools; member
207 * struct knav_pool: qmss pools
304 struct list_head pools; member
363 list_for_each_entry(pool, &kdev->pools, list)
/linux/include/linux/
H A Dswiotlb.h91 * @nslabs: Total number of IO TLB slabs in all pools.
95 * @can_grow: %true if more pools can be allocated dynamically.
98 * @pools: List of IO TLB memory pool descriptors (if dynamic).
105 * @transient_nslabs: The total number of slots in all transient pools that
118 struct list_head pools; member
157 * dev->dma_uses_io_tlb here and mem->pools in __swiotlb_find_pool(). in swiotlb_find_pool()
/linux/mm/
H A Ddmapool.c61 struct list_head pools; member
81 list_for_each_entry(pool, &dev->dma_pools, pools) { in pools_show()
93 static DEVICE_ATTR_RO(pools);
212 * Given one of these pools, dma_pool_alloc()
270 INIT_LIST_HEAD(&retval->pools); in dma_pool_create_node()
283 list_add(&retval->pools, &dev->dma_pools); in dma_pool_create_node()
291 list_del(&retval->pools); in dma_pool_create_node()
373 list_del(&pool->pools); in dma_pool_destroy()
/linux/drivers/net/ethernet/freescale/fman/
H A Dfman.h74 #define BM_MAX_NUM_OF_POOLS 64 /* Buffers pools */
75 #define FMAN_PORT_MAX_EXT_POOLS_NUM 8 /* External BM pools per Rx port */
169 * buffer pools used by a port or storage-profile.
177 * buffer pools allocated in the BM and used by a port or a
181 u8 num_of_pools_used; /* Number of pools use by this port */
189 * number of pools (all together!) are depleted
192 /* the number of depleted pools that will invoke pause
/linux/Documentation/devicetree/bindings/soc/ti/
H A Dkeystone-navigator-qmss.txt6 processors(PDSP), linking RAM, descriptor pools and infrastructure
49 - queue-pools : child node classifying the queue ranges into pools.
50 Queue ranges are grouped into 3 type of pools:
151 queue-pools {
/linux/lib/
H A Dstackdepot.c70 /* Number of pools in stack_pools. */
213 pr_info("allocating space for %u stack pools via memblock\n", in stack_depot_early_init()
218 pr_err("stack pools allocation failed, disabling\n"); in stack_depot_early_init()
273 pr_info("allocating space for %u stack pools via kvcalloc\n", in stack_depot_init()
277 pr_err("stack pools allocation failed, disabling\n"); in stack_depot_init()
291 * Initializes new stack pool, and updates the list of pools.
324 * pools. in depot_init_pool()
348 * pools is reached, do not use the preallocated memory. in depot_keep_new_pool()
859 seq_printf(seq, "pools: %d\n", data_race(pools_num)); in stats_show()
/linux/drivers/net/ethernet/chelsio/libcxgb/
H A Dlibcxgb_ppm.c348 struct cxgbi_ppm_pool __percpu *pools; in ppm_alloc_cpu_pool() local
350 unsigned int max = (PCPU_MIN_UNIT_SIZE - sizeof(*pools)) << 3; in ppm_alloc_cpu_pool()
367 alloc_sz = sizeof(*pools) + sizeof(unsigned long) * bmap; in ppm_alloc_cpu_pool()
368 pools = __alloc_percpu(alloc_sz, __alignof__(struct cxgbi_ppm_pool)); in ppm_alloc_cpu_pool()
370 if (!pools) in ppm_alloc_cpu_pool()
374 struct cxgbi_ppm_pool *ppool = per_cpu_ptr(pools, cpu); in ppm_alloc_cpu_pool()
384 return pools; in ppm_alloc_cpu_pool()
/linux/include/soc/fsl/
H A Dbman.h47 * hardware pools as a dma_addr_t
77 /* Portal and Buffer Pools */
81 #define BM_POOL_MAX 64 /* max # of buffer pools */
/linux/drivers/net/wireless/ath/
H A Ddfs_pri_detector.c73 * DOC: Singleton Pulse and Sequence Pools
75 * Instances of pri_sequence and pulse_elem are kept in singleton pools to
79 * Memory is freed after all references to the pools are released.
100 /* free singleton pools with no references left */ in pool_deregister_ref()
357 /* free pulse queue and sequences list and give objects back to pools */
/linux/drivers/net/ethernet/freescale/dpaa2/
H A Ddpaa2-xsk.c162 pools_params->pools[curr_bp].priority_mask |= (1 << j); in dpaa2_xsk_set_bp_per_qdbin()
163 if (!pools_params->pools[curr_bp].priority_mask) in dpaa2_xsk_set_bp_per_qdbin()
166 pools_params->pools[curr_bp].dpbp_id = priv->bp[i]->bpid; in dpaa2_xsk_set_bp_per_qdbin()
167 pools_params->pools[curr_bp].buffer_size = priv->rx_buf_size; in dpaa2_xsk_set_bp_per_qdbin()
168 pools_params->pools[curr_bp++].backup_pool = 0; in dpaa2_xsk_set_bp_per_qdbin()
H A Ddpsw.h196 * struct dpsw_ctrl_if_pools_cfg - Control interface buffer pools configuration
198 * @pools: Array of buffer pools parameters; The number of valid entries
200 * @pools.dpbp_id: DPBP object ID
201 * @pools.buffer_size: Buffer size
202 * @pools.backup_pool: Backup pool
210 } pools[DPSW_MAX_DPBP]; member
/linux/Documentation/networking/device_drivers/ethernet/freescale/dpaa2/
H A Dethernet-driver.rst26 - buffer pools
40 object and uses Buffer Pools (DPBPs), I/O Portals (DPIOs) and Concentrators
69 DPBPs represent hardware buffer pools. Packet I/O is performed in the context
124 The role of hardware buffer pools is storage of ingress frame data. Each network

12345678910>>...12