/linux-6.8/drivers/vdpa/ifcvf/ |
D | ifcvf_base.c | 74 u16 queue_size; in ifcvf_get_vq_size() local 77 queue_size = vp_ioread16(&hw->common_cfg->queue_size); in ifcvf_get_vq_size() 79 return queue_size; in ifcvf_get_vq_size() 88 u16 queue_size, max_size, qid; in ifcvf_get_max_vq_size() local 92 queue_size = ifcvf_get_vq_size(hw, qid); in ifcvf_get_max_vq_size() 94 if (!queue_size) in ifcvf_get_max_vq_size() 97 max_size = min(queue_size, max_size); in ifcvf_get_max_vq_size() 353 vp_iowrite16(num, &cfg->queue_size); in ifcvf_set_vq_num()
|
/linux-6.8/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_kernel_queue.c | 42 enum kfd_queue_type type, unsigned int queue_size) in kq_initialize() argument 52 queue_size); in kq_initialize() 85 retval = kfd_gtt_sa_allocate(dev, queue_size, &kq->pq); in kq_initialize() 87 pr_err("Failed to init pq queues size %d\n", queue_size); in kq_initialize() 124 memset(kq->pq_kernel_addr, 0, queue_size); in kq_initialize() 128 prop.queue_size = queue_size; in kq_initialize() 246 queue_size_dwords = kq->queue->properties.queue_size / 4; in kq_acquire_packet_buffer() 317 (kq->queue->properties.queue_size / 4); in kq_rollback_packet()
|
D | kfd_queue.c | 35 pr_debug("Queue Size: %llu\n", q->queue_size); in print_queue_properties() 52 pr_debug("Queue Size: %llu\n", q->properties.queue_size); in print_queue()
|
D | kfd_mqd_manager_cik.c | 165 uint32_t wptr_mask = (uint32_t)((p->queue_size / 4) - 1); in load_mqd() 191 m->cp_hqd_pq_control |= order_base_2(q->queue_size / 4) - 1; in __update_mqd() 230 m->sdma_rlc_rb_cntl = order_base_2(q->queue_size / 4) in update_mqd_sdma() 351 m->cp_hqd_pq_control |= order_base_2(q->queue_size / 4) - 1; in update_mqd_hiq()
|
/linux-6.8/drivers/net/ethernet/microsoft/mana/ |
D | hw_channel.c | 257 enum gdma_queue_type type, u64 queue_size, in mana_hwc_create_gdma_wq() argument 267 spec.queue_size = queue_size; in mana_hwc_create_gdma_wq() 273 u64 queue_size, in mana_hwc_create_gdma_cq() argument 282 spec.queue_size = queue_size; in mana_hwc_create_gdma_cq() 291 u64 queue_size, in mana_hwc_create_gdma_eq() argument 299 spec.queue_size = queue_size; in mana_hwc_create_gdma_eq() 489 u32 queue_size; in mana_hwc_create_wq() local 495 queue_size = roundup_pow_of_two(GDMA_MAX_RQE_SIZE * q_depth); in mana_hwc_create_wq() 497 queue_size = roundup_pow_of_two(GDMA_MAX_SQE_SIZE * q_depth); in mana_hwc_create_wq() 499 if (queue_size < MINIMUM_SUPPORTED_PAGE_SIZE) in mana_hwc_create_wq() [all …]
|
D | gdma_main.c | 224 req.queue_size = queue->queue_size; in mana_gd_create_hw_eq() 336 u32 num_cqe = cq->queue_size / GDMA_CQE_SIZE; in mana_gd_ring_cq() 346 u32 head = eq->head % (eq->queue_size / GDMA_EQE_SIZE); in mana_gd_process_eqe() 409 num_eqe = eq->queue_size / GDMA_EQE_SIZE; in mana_gd_process_eq_events() 589 log2_num_entries = ilog2(queue->queue_size / GDMA_EQE_SIZE); in mana_gd_create_eq() 628 u32 log2_num_entries = ilog2(spec->queue_size / GDMA_CQE_SIZE); in mana_gd_create_cq() 664 err = mana_gd_alloc_memory(gc, spec->queue_size, gmi); in mana_gd_create_hwc_queue() 671 queue->queue_size = spec->queue_size; in mana_gd_create_hwc_queue() 791 err = mana_gd_alloc_memory(gc, spec->queue_size, gmi); in mana_gd_create_mana_eq() 802 queue->queue_size = spec->queue_size; in mana_gd_create_mana_eq() [all …]
|
D | mana_en.c | 1139 req.wq_size = wq_spec->queue_size; in mana_create_wq_obj() 1140 req.cq_size = cq_spec->queue_size; in mana_create_wq_obj() 1241 spec.queue_size = EQ_SIZE; in mana_create_eq() 1916 spec.queue_size = txq_size; in mana_create_txq() 1930 spec.queue_size = cq_size; in mana_create_txq() 1942 wq_spec.queue_size = txq->gdma_sq->queue_size; in mana_create_txq() 1945 cq_spec.queue_size = cq->gdma_cq->queue_size; in mana_create_txq() 2201 spec.queue_size = rq_size; in mana_create_rxq() 2214 spec.queue_size = cq_size; in mana_create_rxq() 2225 wq_spec.queue_size = rxq->gdma_rq->queue_size; in mana_create_rxq() [all …]
|
/linux-6.8/drivers/firmware/tegra/ |
D | ivc.c | 562 unsigned tegra_ivc_total_queue_size(unsigned queue_size) in tegra_ivc_total_queue_size() argument 564 if (!IS_ALIGNED(queue_size, TEGRA_IVC_ALIGN)) { in tegra_ivc_total_queue_size() 565 pr_err("%s: queue_size (%u) must be %u-byte aligned\n", in tegra_ivc_total_queue_size() 566 __func__, queue_size, TEGRA_IVC_ALIGN); in tegra_ivc_total_queue_size() 570 return queue_size + sizeof(struct tegra_ivc_header); in tegra_ivc_total_queue_size() 652 size_t queue_size; in tegra_ivc_init() local 670 queue_size = tegra_ivc_total_queue_size(num_frames * frame_size); in tegra_ivc_init() 673 ivc->rx.phys = dma_map_single(peer, iosys_map_get_vaddr(rx), queue_size, in tegra_ivc_init() 678 ivc->tx.phys = dma_map_single(peer, iosys_map_get_vaddr(tx), queue_size, in tegra_ivc_init() 681 dma_unmap_single(peer, ivc->rx.phys, queue_size, in tegra_ivc_init()
|
D | bpmp-tegra186.c | 112 size_t message_size, queue_size; in tegra186_bpmp_channel_init() local 123 queue_size = tegra_ivc_total_queue_size(message_size); in tegra186_bpmp_channel_init() 124 offset = queue_size * index; in tegra186_bpmp_channel_init()
|
/linux-6.8/sound/firewire/ |
D | amdtp-stream.c | 643 if (++s->packet_index >= s->queue_size) in queue_packet() 899 unsigned int queue_size) in compute_ohci_it_cycle() argument 902 return increment_ohci_cycle_count(cycle, queue_size); in compute_ohci_it_cycle() 912 unsigned int queue_size = s->queue_size; in generate_tx_packet_descs() local 981 packet_index = (packet_index + 1) % queue_size; in generate_tx_packet_descs() 1014 unsigned int index = (s->packet_index + i) % s->queue_size; in generate_rx_packet_descs() 1017 desc->cycle = compute_ohci_it_cycle(*ctx_header, s->queue_size); in generate_rx_packet_descs() 1224 cycle = compute_ohci_it_cycle(ctx_header[packets - 1], s->queue_size); in skip_rx_packets() 1250 const unsigned int queue_size = s->queue_size; in process_rx_packets_intermediately() local 1261 unsigned int cycle = compute_ohci_it_cycle(ctx_header[offset], queue_size); in process_rx_packets_intermediately() [all …]
|
/linux-6.8/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_amdkfd_gc_9_4_3.c | 328 uint32_t queue_size = in kgd_gfx_v9_4_3_hqd_load() local 330 CP_HQD_PQ_CONTROL, QUEUE_SIZE); in kgd_gfx_v9_4_3_hqd_load() 331 uint64_t guessed_wptr = m->cp_hqd_pq_rptr & (queue_size - 1); in kgd_gfx_v9_4_3_hqd_load() 333 if ((m->cp_hqd_pq_wptr_lo & (queue_size - 1)) < guessed_wptr) in kgd_gfx_v9_4_3_hqd_load() 334 guessed_wptr += queue_size; in kgd_gfx_v9_4_3_hqd_load() 335 guessed_wptr += m->cp_hqd_pq_wptr_lo & ~(queue_size - 1); in kgd_gfx_v9_4_3_hqd_load()
|
D | amdgpu_amdkfd_gfx_v10_3.c | 239 uint32_t queue_size = in hqd_load_v10_3() local 241 CP_HQD_PQ_CONTROL, QUEUE_SIZE); in hqd_load_v10_3() 242 uint64_t guessed_wptr = m->cp_hqd_pq_rptr & (queue_size - 1); in hqd_load_v10_3() 244 if ((m->cp_hqd_pq_wptr_lo & (queue_size - 1)) < guessed_wptr) in hqd_load_v10_3() 245 guessed_wptr += queue_size; in hqd_load_v10_3() 246 guessed_wptr += m->cp_hqd_pq_wptr_lo & ~(queue_size - 1); in hqd_load_v10_3()
|
D | amdgpu_amdkfd_gfx_v11.c | 224 uint32_t queue_size = in hqd_load_v11() local 226 CP_HQD_PQ_CONTROL, QUEUE_SIZE); in hqd_load_v11() 227 uint64_t guessed_wptr = m->cp_hqd_pq_rptr & (queue_size - 1); in hqd_load_v11() 229 if ((m->cp_hqd_pq_wptr_lo & (queue_size - 1)) < guessed_wptr) in hqd_load_v11() 230 guessed_wptr += queue_size; in hqd_load_v11() 231 guessed_wptr += m->cp_hqd_pq_wptr_lo & ~(queue_size - 1); in hqd_load_v11()
|
D | amdgpu_amdkfd_gfx_v10.c | 253 uint32_t queue_size = in kgd_hqd_load() local 255 CP_HQD_PQ_CONTROL, QUEUE_SIZE); in kgd_hqd_load() 256 uint64_t guessed_wptr = m->cp_hqd_pq_rptr & (queue_size - 1); in kgd_hqd_load() 258 if ((m->cp_hqd_pq_wptr_lo & (queue_size - 1)) < guessed_wptr) in kgd_hqd_load() 259 guessed_wptr += queue_size; in kgd_hqd_load() 260 guessed_wptr += m->cp_hqd_pq_wptr_lo & ~(queue_size - 1); in kgd_hqd_load()
|
D | amdgpu_amdkfd_gfx_v9.c | 267 uint32_t queue_size = in kgd_gfx_v9_hqd_load() local 269 CP_HQD_PQ_CONTROL, QUEUE_SIZE); in kgd_gfx_v9_hqd_load() 270 uint64_t guessed_wptr = m->cp_hqd_pq_rptr & (queue_size - 1); in kgd_gfx_v9_hqd_load() 272 if ((m->cp_hqd_pq_wptr_lo & (queue_size - 1)) < guessed_wptr) in kgd_gfx_v9_hqd_load() 273 guessed_wptr += queue_size; in kgd_gfx_v9_hqd_load() 274 guessed_wptr += m->cp_hqd_pq_wptr_lo & ~(queue_size - 1); in kgd_gfx_v9_hqd_load()
|
/linux-6.8/drivers/nvme/host/ |
D | rdma.c | 87 int queue_size; member 266 init_attr.cap.max_send_wr = factor * queue->queue_size + 1; in nvme_rdma_create_qp() 268 init_attr.cap.max_recv_wr = queue->queue_size + 1; in nvme_rdma_create_qp() 444 nvme_rdma_free_ring(ibdev, queue->rsp_ring, queue->queue_size, in nvme_rdma_destroy_queue_ib() 505 queue->cq_size = cq_factor * queue->queue_size + 1; in nvme_rdma_create_queue_ib() 515 queue->rsp_ring = nvme_rdma_alloc_ring(ibdev, queue->queue_size, in nvme_rdma_create_queue_ib() 529 queue->queue_size, in nvme_rdma_create_queue_ib() 535 queue->queue_size, nvme_rdma_queue_idx(queue)); in nvme_rdma_create_queue_ib() 541 queue->queue_size, IB_MR_TYPE_INTEGRITY, in nvme_rdma_create_queue_ib() 546 queue->queue_size, nvme_rdma_queue_idx(queue)); in nvme_rdma_create_queue_ib() [all …]
|
D | fabrics.h | 99 * @queue_size: Number of IO queue elements. 128 size_t queue_size; member
|
/linux-6.8/drivers/nvme/target/ |
D | loop.c | 575 if (opts->queue_size > ctrl->ctrl.maxcmd) { in nvme_loop_create_ctrl() 576 /* warn if maxcmd is lower than queue_size */ in nvme_loop_create_ctrl() 578 "queue_size %zu > ctrl maxcmd %u, clamping down\n", in nvme_loop_create_ctrl() 579 opts->queue_size, ctrl->ctrl.maxcmd); in nvme_loop_create_ctrl() 580 opts->queue_size = ctrl->ctrl.maxcmd; in nvme_loop_create_ctrl() 582 ctrl->ctrl.sqsize = opts->queue_size - 1; in nvme_loop_create_ctrl()
|
/linux-6.8/kernel/bpf/ |
D | queue_stack_maps.c | 68 u64 size, queue_size; in queue_stack_map_alloc() local 71 queue_size = sizeof(*qs) + size * attr->value_size; in queue_stack_map_alloc() 73 qs = bpf_map_area_alloc(queue_size, numa_node); in queue_stack_map_alloc()
|
/linux-6.8/sound/core/ |
D | timer.c | 77 int queue_size; member 1334 prev = tu->qtail == 0 ? tu->queue_size - 1 : tu->qtail - 1; in snd_timer_user_interrupt() 1341 if (tu->qused >= tu->queue_size) { in snd_timer_user_interrupt() 1345 tu->qtail %= tu->queue_size; in snd_timer_user_interrupt() 1359 if (tu->qused >= tu->queue_size) { in snd_timer_user_append_to_tqueue() 1363 tu->qtail %= tu->queue_size; in snd_timer_user_append_to_tqueue() 1440 prev = tu->qtail == 0 ? tu->queue_size - 1 : tu->qtail - 1; in snd_timer_user_tinterrupt() 1482 tu->queue_size = size; in realloc_user_queue() 1867 if (params.queue_size > 0 && in snd_timer_user_params() 1868 (params.queue_size < 32 || params.queue_size > 1024)) { in snd_timer_user_params() [all …]
|
/linux-6.8/drivers/net/wireless/intel/iwlwifi/pcie/ |
D | rx.c | 122 WARN_ON(rxq->queue_size & (rxq->queue_size - 1)); in iwl_rxq_space() 130 return (rxq->read - rxq->write - 1) & (rxq->queue_size - 1); in iwl_rxq_space() 272 rxq->write = (rxq->write + 1) & (rxq->queue_size - 1); in iwl_pcie_rxmq_restock() 685 free_size * rxq->queue_size, in iwl_pcie_free_rxq_dma() 696 rxq->queue_size, in iwl_pcie_free_rxq_dma() 724 rxq->queue_size = trans->cfg->num_rbds; in iwl_pcie_alloc_rxq_dma() 726 rxq->queue_size = RX_QUEUE_SIZE; in iwl_pcie_alloc_rxq_dma() 734 rxq->bd = dma_alloc_coherent(dev, free_size * rxq->queue_size, in iwl_pcie_alloc_rxq_dma() 742 rxq->queue_size, in iwl_pcie_alloc_rxq_dma() 1083 int i, err, queue_size, allocator_pool_size, num_alloc; in _iwl_pcie_rx_init() local [all …]
|
/linux-6.8/drivers/i2c/busses/ |
D | i2c-qcom-cci.c | 122 u16 queue_size[NUM_QUEUES]; member 323 if (val == cci->data->queue_size[queue]) in cci_validate_queue() 697 .queue_size = { 64, 16 }, 731 .queue_size = { 64, 16 }, 765 .queue_size = { 64, 16 },
|
/linux-6.8/drivers/infiniband/hw/mana/ |
D | qp.c | 207 wq_spec.queue_size = wq->wq_buf_size; in mana_ib_create_qp_rss() 210 cq_spec.queue_size = cq->cqe * COMP_ENTRY_SIZE; in mana_ib_create_qp_rss() 394 wq_spec.queue_size = ucmd.sq_buf_size; in mana_ib_create_qp_raw() 397 cq_spec.queue_size = send_cq->cqe * COMP_ENTRY_SIZE; in mana_ib_create_qp_raw()
|
/linux-6.8/drivers/usb/gadget/function/ |
D | u_serial.c | 77 /* RX and TX queues can buffer QUEUE_SIZE packets before they hit the 81 #define QUEUE_SIZE 16 macro 250 if (port->write_started >= QUEUE_SIZE) in gs_start_tx() 321 if (port->read_started >= QUEUE_SIZE) in gs_start_rx() 357 * So QUEUE_SIZE packets plus however many the FIFO holds (usually two) 512 int n = allocated ? QUEUE_SIZE - *allocated : QUEUE_SIZE; in gs_alloc_requests() 514 /* Pre-allocate up to QUEUE_SIZE transfers, but if we can't in gs_alloc_requests()
|
/linux-6.8/drivers/infiniband/ulp/srp/ |
D | ib_srp.c | 539 /* queue_size + 1 for ib_drain_rq() */ in srp_create_ch_ib() 540 recv_cq = ib_alloc_cq(dev->dev, ch, target->queue_size + 1, in srp_create_ch_ib() 547 send_cq = ib_alloc_cq(dev->dev, ch, m * target->queue_size, in srp_create_ch_ib() 555 init_attr->cap.max_send_wr = m * target->queue_size; in srp_create_ch_ib() 556 init_attr->cap.max_recv_wr = target->queue_size + 1; in srp_create_ch_ib() 682 for (i = 0; i < target->queue_size; ++i) in srp_free_ch_ib() 688 for (i = 0; i < target->queue_size; ++i) in srp_free_ch_ib() 1372 for (j = 0; j < target->queue_size; ++j) in srp_rport_reconnect() 2258 ch->rx_ring = kcalloc(target->queue_size, sizeof(*ch->rx_ring), in srp_alloc_iu_bufs() 2262 ch->tx_ring = kcalloc(target->queue_size, sizeof(*ch->tx_ring), in srp_alloc_iu_bufs() [all …]
|