| /linux/drivers/infiniband/hw/cxgb4/ |
| H A D | device.c | 344 xa_for_each(&qpd->devp->qps, index, qp) in qp_open() 354 xa_lock_irq(&qpd->devp->qps); in qp_open() 355 xa_for_each(&qpd->devp->qps, index, qp) in qp_open() 357 xa_unlock_irq(&qpd->devp->qps); in qp_open() 938 WARN_ON(!xa_empty(&ctx->dev->qps)); in c4iw_dealloc() 1048 xa_init_flags(&devp->qps, XA_FLAGS_LOCK_IRQ); in c4iw_alloc() 1272 xa_lock_irqsave(&ctx->dev->qps, flags); in stop_queues() 1276 xa_for_each(&ctx->dev->qps, index, qp) in stop_queues() 1281 xa_unlock_irqrestore(&ctx->dev->qps, flags); in stop_queues() 1311 xa_lock_irq(&ctx->dev->qps); in resume_queues() [all …]
|
| H A D | ev.c | 126 xa_lock_irq(&dev->qps); in c4iw_ev_dispatch() 127 qhp = xa_load(&dev->qps, CQE_QPID(err_cqe)); in c4iw_ev_dispatch() 134 xa_unlock_irq(&dev->qps); in c4iw_ev_dispatch() 149 xa_unlock_irq(&dev->qps); in c4iw_ev_dispatch() 155 xa_unlock_irq(&dev->qps); in c4iw_ev_dispatch()
|
| H A D | qp.c | 66 xa_lock_irq(&dev->qps); in alloc_ird() 71 xa_unlock_irq(&dev->qps); in alloc_ird() 82 xa_lock_irq(&dev->qps); in free_ird() 84 xa_unlock_irq(&dev->qps); in free_ird() 913 xa_lock_irqsave(&qhp->rhp->qps, flags); in ring_kernel_sq_db() 922 xa_unlock_irqrestore(&qhp->rhp->qps, flags); in ring_kernel_sq_db() 930 xa_lock_irqsave(&qhp->rhp->qps, flags); in ring_kernel_rq_db() 939 xa_unlock_irqrestore(&qhp->rhp->qps, flags); in ring_kernel_rq_db() 2093 xa_lock_irq(&rhp->qps); in c4iw_destroy_qp() 2094 __xa_erase(&rhp->qps, qhp->wq.sq.qid); in c4iw_destroy_qp() [all …]
|
| H A D | iw_cxgb4.h | 318 struct xarray qps; member 350 return xa_load(&rhp->qps, qpid); in get_qhp()
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_hw_io.c | 129 qp = &func_to_io->qps[i]; in write_sq_ctxts() 173 qp = &func_to_io->qps[i]; in write_rq_ctxts() 381 func_to_io->qps = devm_kcalloc(&pdev->dev, num_qps, in hinic_io_create_qps() 382 sizeof(*func_to_io->qps), GFP_KERNEL); in hinic_io_create_qps() 383 if (!func_to_io->qps) in hinic_io_create_qps() 419 err = init_qp(func_to_io, &func_to_io->qps[i], i, in hinic_io_create_qps() 444 destroy_qp(func_to_io, &func_to_io->qps[j]); in hinic_io_create_qps() 459 devm_kfree(&pdev->dev, func_to_io->qps); in hinic_io_create_qps() 478 destroy_qp(func_to_io, &func_to_io->qps[i]); in hinic_io_destroy_qps() 488 devm_kfree(&pdev->dev, func_to_io->qps); in hinic_io_destroy_qps() [all …]
|
| H A D | hinic_hw_io.h | 70 struct hinic_qp *qps; member
|
| H A D | hinic_hw_dev.c | 1061 struct hinic_qp *qp = &func_to_io->qps[i]; in hinic_hwdev_get_sq() 1079 struct hinic_qp *qp = &func_to_io->qps[i]; in hinic_hwdev_get_rq()
|
| /linux/drivers/crypto/hisilicon/zip/ |
| H A D | zip_crypto.c | 420 struct hisi_qp *qps[HZIP_CTX_Q_NUM] = { NULL }; in hisi_zip_ctx_init() local 430 ret = zip_create_qps(qps, HZIP_CTX_Q_NUM, node, alg_type); in hisi_zip_ctx_init() 436 hisi_zip = container_of(qps[0]->qm, struct hisi_zip, qm); in hisi_zip_ctx_init() 443 qp_ctx->qp = qps[i]; in hisi_zip_ctx_init() 453 struct hisi_qp *qps[HZIP_CTX_Q_NUM] = { NULL }; in hisi_zip_ctx_exit() local 457 qps[i] = hisi_zip_ctx->qp_ctx[i].qp; in hisi_zip_ctx_exit() 459 hisi_qm_free_qps(qps, HZIP_CTX_Q_NUM); in hisi_zip_ctx_exit()
|
| H A D | zip.h | 102 int zip_create_qps(struct hisi_qp **qps, int qp_num, int node, u8 *alg_type);
|
| H A D | zip_main.c | 449 int zip_create_qps(struct hisi_qp **qps, int qp_num, int node, u8 *alg_type) in zip_create_qps() argument 454 return hisi_qm_alloc_qps_node(&zip_devices, qp_num, alg_type, node, qps); in zip_create_qps()
|
| /linux/drivers/crypto/hisilicon/sec2/ |
| H A D | sec_main.c | 409 void sec_destroy_qps(struct hisi_qp **qps, int qp_num) in sec_destroy_qps() argument 411 hisi_qm_free_qps(qps, qp_num); in sec_destroy_qps() 412 kfree(qps); in sec_destroy_qps() 419 struct hisi_qp **qps; in sec_create_qps() local 423 qps = kzalloc_objs(struct hisi_qp *, ctx_num); in sec_create_qps() 424 if (!qps) in sec_create_qps() 430 kfree(qps); in sec_create_qps() 434 ret = hisi_qm_alloc_qps_node(&sec_devices, ctx_num, type, node, qps); in sec_create_qps() 437 kfree(qps); in sec_create_qps() 442 return qps; in sec_create_qps()
|
| H A D | sec.h | 182 struct hisi_qp **qps; member 286 void sec_destroy_qps(struct hisi_qp **qps, int qp_num);
|
| H A D | sec_crypto.c | 628 qp = ctx->qps[qp_ctx_id]; in sec_create_qp_ctx() 665 ctx->qps = sec_create_qps(); in sec_ctx_base_init() 666 if (!ctx->qps) in sec_ctx_base_init() 669 sec = container_of(ctx->qps[0]->qm, struct sec_dev, qm); in sec_ctx_base_init() 694 sec_destroy_qps(ctx->qps, sec->ctx_q_num); in sec_ctx_base_init() 702 if (!ctx->qps) in sec_ctx_base_uninit() 708 sec_destroy_qps(ctx->qps, ctx->sec->ctx_q_num); in sec_ctx_base_uninit() 716 if (!ctx->qps) in sec_cipher_init() 731 if (!ctx->qps) in sec_cipher_uninit() 755 if (!ctx->qps) in sec_auth_uninit() [all …]
|
| /linux/drivers/gpu/drm/amd/display/dc/dml/dsc/ |
| H A D | rc_calc_fpu.c | 64 static void get_qp_set(qp_set qps, enum colour_mode cm, enum bits_per_comp bpc, in get_qp_set() argument 107 memcpy(qps, table[index].qps, sizeof(qp_set)); in get_qp_set()
|
| H A D | rc_calc_fpu.h | 76 const qp_set qps; member
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | en_rx.c | 1197 &rss_map->qps[i]); in mlx4_en_config_rss_steer() 1205 rss_map->indir_qp = &rss_map->qps[0]; in mlx4_en_config_rss_steer() 1281 MLX4_QP_STATE_RST, NULL, 0, 0, &rss_map->qps[i]); in mlx4_en_config_rss_steer() 1282 mlx4_qp_remove(mdev->dev, &rss_map->qps[i]); in mlx4_en_config_rss_steer() 1283 mlx4_qp_free(mdev->dev, &rss_map->qps[i]); in mlx4_en_config_rss_steer() 1307 MLX4_QP_STATE_RST, NULL, 0, 0, &rss_map->qps[i]); in mlx4_en_release_rss_steer() 1308 mlx4_qp_remove(mdev->dev, &rss_map->qps[i]); in mlx4_en_release_rss_steer() 1309 mlx4_qp_free(mdev->dev, &rss_map->qps[i]); in mlx4_en_release_rss_steer()
|
| H A D | en_main.c | 139 &priv->rss_map.qps[i], in mlx4_en_update_loopback_state()
|
| H A D | mlx4_en.h | 436 struct mlx4_qp qps[MAX_RX_RINGS]; member
|
| /linux/drivers/infiniband/hw/qedr/ |
| H A D | qedr_iw_cm.c | 523 xa_lock(&dev->qps); in qedr_iw_load_qp() 524 qp = xa_load(&dev->qps, qpn); in qedr_iw_load_qp() 527 xa_unlock(&dev->qps); in qedr_iw_load_qp() 818 return xa_load(&dev->qps, qpn); in qedr_iw_get_qp()
|
| H A D | qedr.h | 167 struct xarray qps; member
|
| /linux/drivers/crypto/hisilicon/ |
| H A D | qm.c | 3646 void hisi_qm_free_qps(struct hisi_qp **qps, int qp_num) in hisi_qm_free_qps() argument 3650 if (!qps || qp_num <= 0) in hisi_qm_free_qps() 3653 down_write(&qps[0]->qm->qps_lock); in hisi_qm_free_qps() 3656 if (qps[i]->ref_count == 1) in hisi_qm_free_qps() 3657 qm_stop_qp_nolock(qps[i]); in hisi_qm_free_qps() 3659 qm_release_qp_nolock(qps[i]); in hisi_qm_free_qps() 3662 up_write(&qps[0]->qm->qps_lock); in hisi_qm_free_qps() 3663 qm_pm_put_sync(qps[0]->qm); in hisi_qm_free_qps() 3691 static int qm_get_and_start_qp(struct hisi_qm *qm, int qp_num, struct hisi_qp **qps, u8 *alg_type) in qm_get_and_start_qp() argument 3701 qps[i] = qm_create_qp_nolock(qm, alg_type[i], true); in qm_get_and_start_qp() [all …]
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | pio.c | 1562 struct rvt_qp *qps[PIO_WAIT_BATCH_SIZE]; in sc_piobufavail() local 1582 if (n == ARRAY_SIZE(qps)) in sc_piobufavail() 1591 priv = qps[top_idx]->priv; in sc_piobufavail() 1598 qps[n++] = qp; in sc_piobufavail() 1613 hfi1_qp_wakeup(qps[top_idx], in sc_piobufavail() 1617 hfi1_qp_wakeup(qps[i], in sc_piobufavail()
|
| /linux/drivers/infiniband/hw/ocrdma/ |
| H A D | ocrdma_hw.h | 104 enum ib_qp_state get_ibqp_state(enum ocrdma_qp_state qps);
|
| /linux/include/linux/ |
| H A D | hisi_acc_qm.h | 587 u8 *alg_type, int node, struct hisi_qp **qps); 588 void hisi_qm_free_qps(struct hisi_qp **qps, int qp_num);
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_virtchnl_pf.c | 905 u16 vsi_id, qps; in i40e_map_pf_queues_to_vsi() local 913 qps = vf->ch[i].num_qps; in i40e_map_pf_queues_to_vsi() 916 qps = pf->vsi[vf->lan_vsi_idx]->alloc_queue_pairs; in i40e_map_pf_queues_to_vsi() 921 if (j * 2 >= qps) { in i40e_map_pf_queues_to_vsi() 952 u32 qps, num_tc = 1; /* VF has at least one traffic class */ in i40e_map_pf_to_vf_queues() local 961 qps = vf->ch[i].num_qps; in i40e_map_pf_to_vf_queues() 964 qps = pf->vsi[vf->lan_vsi_idx]->alloc_queue_pairs; in i40e_map_pf_to_vf_queues() 968 for (j = 0; j < qps; j++) { in i40e_map_pf_to_vf_queues()
|