| /linux/drivers/net/ethernet/mellanox/mlxbf_gige/ |
| H A D | mlxbf_gige_rx.c | 122 size_t wq_size, cq_size; in mlxbf_gige_rx_init() local 132 wq_size = MLXBF_GIGE_RX_WQE_SZ * priv->rx_q_entries; in mlxbf_gige_rx_init() 133 priv->rx_wqe_base = dma_alloc_coherent(priv->dev, wq_size, in mlxbf_gige_rx_init() 208 dma_free_coherent(priv->dev, wq_size, in mlxbf_gige_rx_init()
|
| /linux/include/uapi/rdma/ |
| H A D | irdma-abi.h | 43 __u32 wq_size; /* size of the WQs (SQ+RQ) in the mmaped area */ member
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
| H A D | params.c | 827 int wq_size = BIT(mlx5e_mpwqe_get_log_rq_size(mdev, params, xsk)); in mlx5e_shampo_get_log_cq_size() local 834 return order_base_2((wqe_size / rsrv_size) * wq_size * (pkt_per_rsrv + 1)); in mlx5e_shampo_get_log_cq_size() 1063 int wq_size = BIT(MLX5_GET(wq, wqc, log_wq_sz)); in mlx5e_shampo_hd_per_wq() local 1067 hd_per_wq = roundup_pow_of_two(hd_per_wqe * wq_size); in mlx5e_shampo_hd_per_wq() 1077 int wq_size = BIT(MLX5_GET(wq, wqc, log_wq_sz)); in mlx5e_shampo_icosq_sz() local 1087 wqebbs *= wq_size; in mlx5e_shampo_icosq_sz()
|
| /linux/drivers/gpu/drm/i915/gt/uc/ |
| H A D | intel_guc_fwif.h | 239 u32 wq_size; member 264 u32 wq_size; member
|
| H A D | intel_guc_submission.c | 2502 action[len++] = info->wq_size; in __guc_action_register_multi_lrc_v70() 2553 info->wq_size, in __guc_action_register_context_v70() 2835 desc->wq_size = WQ_SIZE; in prepare_context_registration_info_v69() 2841 pdesc->wq_size_bytes = desc->wq_size; in prepare_context_registration_info_v69() 2912 info->wq_size = WQ_SIZE; in prepare_context_registration_info_v70()
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_guc_fwif.h | 64 u32 wq_size; member
|
| H A D | xe_guc_submit.c | 556 action[len++] = info->wq_size; in __register_mlrc_exec_queue() 596 info->wq_size, in __register_exec_queue() 643 info.wq_size = WQ_SIZE; in register_exec_queue()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | en.h | 169 static inline u16 mlx5_min_rx_wqes(int wq_type, u32 wq_size) in mlx5_min_rx_wqes() argument 174 wq_size / 2); in mlx5_min_rx_wqes() 177 wq_size / 2); in mlx5_min_rx_wqes()
|
| H A D | en_main.c | 532 u32 wq_size = mlx5_wq_ll_get_size(&rq->mpwqe.wq); in mlx5e_create_rq_umr_mkey() local 540 if (WARN_ON_ONCE(check_mul_overflow(wq_size, (u32)rq->mpwqe.mtts_per_wqe, in mlx5e_create_rq_umr_mkey() 544 __func__, wq_size, rq->mpwqe.mtts_per_wqe, in mlx5e_create_rq_umr_mkey() 797 int wq_size; in mlx5_rq_shampo_alloc() local 820 wq_size = BIT(MLX5_GET(wq, wqc, log_wq_sz)); in mlx5_rq_shampo_alloc() 836 hd_pool_size = (hd_per_wqe * wq_size) >> log_hd_per_page; in mlx5_rq_shampo_alloc()
|
| /linux/drivers/dma/idxd/ |
| H A D | registers.h | 371 u16 wq_size; member
|
| H A D | sysfs.c | 718 int wq_size = 0; in total_claimed_wq_size() local 723 wq_size += wq->size; in total_claimed_wq_size() 726 return wq_size; in total_claimed_wq_size()
|
| H A D | device.c | 934 wq->wqcfg->wq_size = wq->size; in idxd_wq_config_write() 1142 wq->size = wq->wqcfg->wq_size; in idxd_wq_load_config()
|
| /linux/tools/testing/selftests/vfio/lib/drivers/dsa/ |
| H A D | registers.h | 371 u16 wq_size; member
|
| H A D | dsa.c | 151 .wq_size = wq_cap.total_wq_size, in dsa_wq_init()
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | qp.c | 437 int wq_size; in set_rq_size() local 472 wq_size = roundup_pow_of_two(cap->max_recv_wr) * wqe_size; in set_rq_size() 473 wq_size = max_t(int, wq_size, MLX5_SEND_WQE_BB); in set_rq_size() 474 qp->rq.wqe_cnt = wq_size / wqe_size; in set_rq_size() 595 int wq_size; in calc_sq_size() local 615 wq_size = roundup_pow_of_two(attr->cap.max_send_wr * wqe_size); in calc_sq_size() 616 qp->sq.wqe_cnt = wq_size / MLX5_SEND_WQE_BB; in calc_sq_size() 630 qp->sq.max_post = wq_size / wqe_size; in calc_sq_size() 633 return wq_size; in calc_sq_size()
|
| /linux/drivers/net/ethernet/microsoft/mana/ |
| H A D | gdma_main.c | 1215 u32 wq_size = wq->queue_size; in mana_gd_wq_avail_space() local 1217 WARN_ON_ONCE(used_space > wq_size); in mana_gd_wq_avail_space() 1219 return wq_size - used_space; in mana_gd_wq_avail_space()
|
| H A D | mana_en.c | 1397 req.wq_size = wq_spec->queue_size; in mana_create_wq_obj()
|
| /linux/include/net/mana/ |
| H A D | mana.h | 723 u32 wq_size; member
|
| /linux/kernel/ |
| H A D | workqueue.c | 5695 size_t wq_size; in __alloc_workqueue() local 5711 wq_size = struct_size(wq, node_nr_active, nr_node_ids + 1); in __alloc_workqueue() 5713 wq_size = sizeof(*wq); in __alloc_workqueue() 5715 wq = kzalloc_noprof(wq_size, GFP_KERNEL); in __alloc_workqueue()
|
| /linux/drivers/infiniband/hw/irdma/ |
| H A D | verbs.c | 323 uresp.wq_size = iwdev->rf->sc_dev.hw_attrs.max_qp_wr * 2; in irdma_alloc_ucontext()
|