| /linux/tools/testing/selftests/bpf/ |
| H A D | test_lru_map.c | 256 unsigned int batch_size; in test_lru_sanity1() local 269 batch_size = tgt_free / 2; in test_lru_sanity1() 270 assert(batch_size * 2 == tgt_free); in test_lru_sanity1() 272 map_size = __map_size(tgt_free) + batch_size; in test_lru_sanity1() 288 end_key = 1 + batch_size; in test_lru_sanity1() 325 unsigned int batch_size; in test_lru_sanity2() local 338 batch_size = tgt_free / 2; in test_lru_sanity2() 339 assert(batch_size * 2 == tgt_free); in test_lru_sanity2() 341 map_size = __map_size(tgt_free) + batch_size; in test_lru_sanity2() 378 end_key = 1 + batch_size; in test_lru_sanity2() [all …]
|
| /linux/tools/testing/selftests/bpf/map_tests/ |
| H A D | map_in_map_batch_ops.c | 123 __u32 batch_size, bool delete_entries, in fetch_and_validate() argument 128 __u32 batch_key = 0, fetch_count, step_size = batch_size; in fetch_and_validate() 158 step_size += batch_size; in fetch_and_validate() 181 delete_entries ? "LOOKUP_AND_DELETE" : "LOOKUP", batch_size); in fetch_and_validate() 195 int batch_size[2] = {5, 10}; in _map_in_map_batch_ops() local 242 batch_size[op_index], false, in _map_in_map_batch_ops()
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_migrate.c | 909 u32 batch_size = 1; /* MI_BATCH_BUFFER_END */ in xe_migrate_copy() local 931 batch_size += pte_update_size(m, pte_flags, src, &src_it, &src_L0, in xe_migrate_copy() 938 batch_size += pte_update_size(m, pte_flags, dst, in xe_migrate_copy() 947 batch_size += pte_update_size(m, 0, NULL, &ccs_it, &ccs_size, in xe_migrate_copy() 955 batch_size += ((copy_only_ccs) ? 0 : EMIT_COPY_DW) + in xe_migrate_copy() 958 bb = xe_bb_new(gt, batch_size, usm); in xe_migrate_copy() 1106 u32 batch_size, batch_size_allocated; in xe_migrate_ccs_rw_copy() local 1126 batch_size = 0; in xe_migrate_ccs_rw_copy() 1128 batch_size += 10; /* Flush + ggtt addr + 2 NOP */ in xe_migrate_ccs_rw_copy() 1136 batch_size += pte_update_size(m, false, src, &src_it, &src_L0, in xe_migrate_ccs_rw_copy() [all …]
|
| /linux/net/bpf/ |
| H A D | test_run.c | 108 u32 batch_size; member 157 .pool_size = xdp->batch_size, in xdp_test_run_setup() 163 xdp->frames = kvmalloc_array(xdp->batch_size, sizeof(void *), GFP_KERNEL); in xdp_test_run_setup() 167 xdp->skbs = kvmalloc_array(xdp->batch_size, sizeof(void *), GFP_KERNEL); in xdp_test_run_setup() 287 batch_sz = min_t(u32, repeat, xdp->batch_size); in xdp_test_run_batch() 362 u32 repeat, u32 batch_size, u32 *time) in bpf_test_run_xdp_live() argument 365 struct xdp_test_data xdp = { .batch_size = batch_size }; in bpf_test_run_xdp_live() 682 if (kattr->test.flags || kattr->test.cpu || kattr->test.batch_size) in bpf_prog_test_run_tracing() 759 kattr->test.repeat || kattr->test.batch_size) in bpf_prog_test_run_raw_tp() 1004 kattr->test.cpu || kattr->test.batch_size) in bpf_prog_test_run_skb() [all …]
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | test_xsk.c | 149 sock_opt = xsk->batch_size; in enable_busy_poll() 258 ifobj->xsk_arr[j].batch_size = DEFAULT_BATCH_SIZE; in __test_spec_init() 958 static int complete_pkts(struct xsk_socket_info *xsk, int batch_size) in complete_pkts() argument 970 rcvd = xsk_ring_cons__peek(&xsk->umem->cq, batch_size, &idx); in complete_pkts() 1024 rcvd = xsk_ring_cons__peek(&xsk->rx, xsk->batch_size, &idx_rx); in __receive_pkts() 1180 if (pkts_in_flight >= (int)((umem_size(umem) - xsk->batch_size * buffer_len) / in __send_pkts() 1191 while (xsk_ring_prod__reserve(&xsk->tx, xsk->batch_size, &idx) < xsk->batch_size) { in __send_pkts() 1211 complete_pkts(xsk, xsk->batch_size); in __send_pkts() 1214 for (i = 0; i < xsk->batch_size; i++) { in __send_pkts() 1222 if (nb_frags > xsk->batch_size - i) { in __send_pkts() [all …]
|
| H A D | test_xsk.h | 70 u32 batch_size; member
|
| H A D | xdp_do_redirect.c | 114 .batch_size = 64, in test_xdp_do_redirect()
|
| /linux/tools/sched_ext/ |
| H A D | scx_userland.c | 51 static __u32 batch_size = 8; variable 260 for (i = 0; i < batch_size; i++) { in dispatch_batch() 355 batch_size = strtoul(optarg, NULL, 0); in pre_bootstrap()
|
| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | intel_renderstate.c | 92 so->batch_size = rodata->batch_items * sizeof(u32); in render_state_setup() 225 so->batch_offset, so->batch_size, in intel_renderstate_emit()
|
| H A D | intel_renderstate.h | 40 u32 batch_size; member
|
| /linux/drivers/net/ethernet/sfc/falcon/ |
| H A D | rx.c | 334 unsigned int fill_level, batch_size; in ef4_fast_push_rx_descriptors() local 352 batch_size = efx->rx_pages_per_batch * efx->rx_bufs_per_page; in ef4_fast_push_rx_descriptors() 354 EF4_BUG_ON_PARANOID(space < batch_size); in ef4_fast_push_rx_descriptors() 371 } while ((space -= batch_size) >= batch_size); in ef4_fast_push_rx_descriptors()
|
| /linux/drivers/net/ethernet/sfc/ |
| H A D | rx_common.c | 458 unsigned int fill_level, batch_size; in efx_fast_push_rx_descriptors() local 476 batch_size = efx->rx_pages_per_batch * efx->rx_bufs_per_page; in efx_fast_push_rx_descriptors() 478 EFX_WARN_ON_ONCE_PARANOID(space < batch_size); in efx_fast_push_rx_descriptors() 493 } while ((space -= batch_size) >= batch_size); in efx_fast_push_rx_descriptors()
|
| /linux/drivers/net/ethernet/sfc/siena/ |
| H A D | rx_common.c | 456 unsigned int fill_level, batch_size; in efx_siena_fast_push_rx_descriptors() local 474 batch_size = efx->rx_pages_per_batch * efx->rx_bufs_per_page; in efx_siena_fast_push_rx_descriptors() 476 EFX_WARN_ON_ONCE_PARANOID(space < batch_size); in efx_siena_fast_push_rx_descriptors() 491 } while ((space -= batch_size) >= batch_size); in efx_siena_fast_push_rx_descriptors()
|
| /linux/mm/ |
| H A D | shrinker.c | 380 long batch_size = shrinker->batch ? shrinker->batch in do_shrink_slab() local 430 while (total_scan >= batch_size || in do_shrink_slab() 433 unsigned long nr_to_scan = min(batch_size, total_scan); in do_shrink_slab()
|
| H A D | vmscan.c | 7790 unsigned long batch_size = (nr_to_reclaim - nr_reclaimed) / 4; in user_proactive_reclaim() local 7809 batch_size, gfp_mask, in user_proactive_reclaim() 7819 .nr_to_reclaim = max(batch_size, SWAP_CLUSTER_MAX), in user_proactive_reclaim() 7830 batch_size, &sc); in user_proactive_reclaim()
|
| H A D | memory.c | 2444 const int batch_size = min_t(int, pages_to_write_in_pmd, 8); in insert_pages() local 2451 for (pte = start_pte; pte_idx < batch_size; ++pte, ++pte_idx) { in insert_pages() 2464 pages_to_write_in_pmd -= batch_size; in insert_pages() 2465 remaining_pages_total -= batch_size; in insert_pages()
|
| /linux/tools/lib/bpf/ |
| H A D | bpf.h | 692 __u32 batch_size; 694 #define bpf_test_run_opts__last_field batch_size 684 __u32 batch_size; global() member
|
| H A D | bpf.c | 1038 attr.test.batch_size = OPTS_GET(opts, batch_size, 0); in bpf_prog_test_run_opts()
|
| /linux/drivers/dma/idxd/ |
| H A D | sysfs.c | 1049 u64 batch_size; in wq_max_batch_size_store() local 1058 rc = __get_sysfs_u64(buf, &batch_size); in wq_max_batch_size_store() 1062 if (batch_size > idxd->max_batch_size) in wq_max_batch_size_store() 1065 idxd_wq_set_max_batch_size(idxd->data->type, wq, (u32)batch_size); in wq_max_batch_size_store()
|
| /linux/drivers/md/ |
| H A D | raid5.c | 6669 int i, batch_size = 0, hash; in handle_active_stripes() local 6672 while (batch_size < MAX_STRIPE_BATCH && in handle_active_stripes() 6674 batch[batch_size++] = sh; in handle_active_stripes() 6676 if (batch_size == 0) { in handle_active_stripes() 6684 return batch_size; in handle_active_stripes() 6699 for (i = 0; i < batch_size; i++) in handle_active_stripes() 6706 for (i = 0; i < batch_size; i++) { in handle_active_stripes() 6710 return batch_size; in handle_active_stripes() 6729 int batch_size, released; in raid5_do_work() local 6733 batch_size = handle_active_stripes(conf, group_id, worker, in raid5_do_work() [all …]
|
| /linux/drivers/media/pci/tw5864/ |
| H A D | tw5864-video.c | 1321 int batch_size = H264_BUF_CNT; in tw5864_handle_frame_work() local 1324 while (dev->h264_buf_r_index != dev->h264_buf_w_index && batch_size--) { in tw5864_handle_frame_work()
|
| /linux/Documentation/networking/ |
| H A D | af_xdp.rst | 684 int batch_size) 688 xsk_ring_prod__reserve(&xsk->tx, batch_size, &idx); 690 for (i = 0; i < batch_size;) { 711 if (i == batch_size) {
|
| /linux/drivers/net/wireless/mediatek/mt76/ |
| H A D | mt76_connac_mcu.c | 2152 int i, n_chan, batch_size, idx = 0, tx_power, last_ch, err = 0; in mt76_connac_mcu_rate_txpower_band() local 2176 batch_size = DIV_ROUND_UP(n_chan, batch_len); in mt76_connac_mcu_rate_txpower_band() 2185 for (i = 0; i < batch_size; i++) { in mt76_connac_mcu_rate_txpower_band() 2190 num_ch = i == batch_size - 1 ? n_chan - i * batch_len : batch_len; in mt76_connac_mcu_rate_txpower_band()
|
| /linux/drivers/net/wireless/mediatek/mt76/mt7925/ |
| H A D | mcu.c | 3643 int i, sku_len, batch_size, batch_len = 3; in mt7925_mcu_rate_txpower_band() local 3698 batch_size = DIV_ROUND_UP(n_chan, batch_len); in mt7925_mcu_rate_txpower_band() 3710 for (i = 0; i < batch_size; i++) { in mt7925_mcu_rate_txpower_band() 3715 num_ch = i == batch_size - 1 ? n_chan % batch_len : batch_len; in mt7925_mcu_rate_txpower_band()
|
| /linux/fs/btrfs/ |
| H A D | tree-log.c | 4150 int batch_size = 0; in process_dir_items_leaf() local 4230 if (batch_size == 0) in process_dir_items_leaf() 4232 batch_size++; in process_dir_items_leaf() 4235 if (batch_size > 0) { in process_dir_items_leaf() 4237 batch_start, batch_size); in process_dir_items_leaf()
|