/linux/drivers/dma/amd/ptdma/ |
H A D | ptdma-dev.c | 57 void pt_start_queue(struct pt_cmd_queue *cmd_q) in pt_start_queue() argument 60 iowrite32(cmd_q->qcontrol | CMD_Q_RUN, cmd_q->reg_control); in pt_start_queue() 63 void pt_stop_queue(struct pt_cmd_queue *cmd_q) in pt_stop_queue() argument 66 iowrite32(cmd_q->qcontrol & ~CMD_Q_RUN, cmd_q->reg_control); in pt_stop_queue() 69 static int pt_core_execute_cmd(struct ptdma_desc *desc, struct pt_cmd_queue *cmd_q) in pt_core_execute_cmd() argument 72 u8 *q_desc = (u8 *)&cmd_q->qbase[cmd_q->qidx]; in pt_core_execute_cmd() 80 spin_lock_irqsave(&cmd_q in pt_core_execute_cmd() 100 pt_core_perform_passthru(struct pt_cmd_queue * cmd_q,struct pt_passthru_engine * pt_engine) pt_core_perform_passthru() argument 128 struct pt_cmd_queue *cmd_q = &cmd->pt->cmd_q; pt_do_cmd_complete() local 144 pt_check_status_trans(struct pt_device * pt,struct pt_cmd_queue * cmd_q) pt_check_status_trans() argument 167 struct pt_cmd_queue *cmd_q = &pt->cmd_q; pt_core_irq_handler() local 179 struct pt_cmd_queue *cmd_q = &pt->cmd_q; pt_core_init() local 280 struct pt_cmd_queue *cmd_q = &pt->cmd_q; pt_core_destroy() local [all...] |
H A D | ptdma-dmaengine.c | 51 struct pt_cmd_queue *cmd_q = &ae4cmd_q->cmd_q; in ae4_check_status_error() local 55 memcpy(&desc, &cmd_q->qbase[idx], sizeof(struct ae4dma_desc)); in ae4_check_status_error() 58 cmd_q->cmd_error = desc.dw1.err_code; in ae4_check_status_error() 59 if (cmd_q->cmd_error) in ae4_check_status_error() 60 ae4_log_error(cmd_q->pt, cmd_q->cmd_error); in ae4_check_status_error() 100 struct pt_cmd_queue *cmd_q; in pt_get_cmd_queue() local 106 cmd_q = &ae4cmd_q->cmd_q; in pt_get_cmd_queue() 117 struct pt_cmd_queue *cmd_q = &ae4cmd_q->cmd_q; ae4_core_execute_cmd() local 136 pt_core_perform_passthru_ae4(struct pt_cmd_queue * cmd_q,struct pt_passthru_engine * pt_engine) pt_core_perform_passthru_ae4() argument 166 struct pt_cmd_queue *cmd_q; pt_dma_start_desc() local 250 ae4_core_queue_full(struct pt_cmd_queue * cmd_q) ae4_core_queue_full() argument 475 pt_check_status_trans_ae4(struct pt_device * pt,struct pt_cmd_queue * cmd_q) pt_check_status_trans_ae4() argument 490 struct pt_cmd_queue *cmd_q; pt_tx_status() local 506 struct pt_cmd_queue *cmd_q; pt_pause() local 522 struct pt_cmd_queue *cmd_q; pt_resume() local 542 struct pt_cmd_queue *cmd_q; pt_terminate_all() local [all...] |
H A D | ptdma-debugfs.c | 66 struct pt_cmd_queue *cmd_q = s->private; in pt_debugfs_queue_show() local 70 if (!cmd_q) in pt_debugfs_queue_show() 73 seq_printf(s, " Pass-Thru: %ld\n", cmd_q->total_pt_ops); in pt_debugfs_queue_show() 75 pt = cmd_q->pt; in pt_debugfs_queue_show() 77 regval = readl(cmd_q->reg_control + 0x4); in pt_debugfs_queue_show() 80 regval = ioread32(cmd_q->reg_control + 0x000C); in pt_debugfs_queue_show() 105 struct pt_cmd_queue *cmd_q; in ptdma_debugfs_setup() local 124 cmd_q = &ae4cmd_q->cmd_q; in ptdma_debugfs_setup() 132 debugfs_create_file("stats", 0400, debugfs_q_instance, cmd_q, in ptdma_debugfs_setup() [all...] |
H A D | ptdma.h | 252 struct pt_cmd_queue cmd_q; member 322 int pt_core_perform_passthru(struct pt_cmd_queue *cmd_q, 325 void pt_check_status_trans(struct pt_device *pt, struct pt_cmd_queue *cmd_q); 326 void pt_start_queue(struct pt_cmd_queue *cmd_q); 327 void pt_stop_queue(struct pt_cmd_queue *cmd_q); 331 iowrite32(0, pt->cmd_q.reg_control + 0x000C); in pt_core_disable_queue_interrupts() 336 iowrite32(SUPPORTED_INTERRUPTS, pt->cmd_q.reg_control + 0x000C); in pt_core_enable_queue_interrupts()
|
/linux/drivers/crypto/ccp/ |
H A D | ccp-dev-v5.c | 25 static u32 ccp_lsb_alloc(struct ccp_cmd_queue *cmd_q, unsigned int count) in ccp_lsb_alloc() argument 31 if (cmd_q->lsb >= 0) { in ccp_lsb_alloc() 32 start = (u32)bitmap_find_next_zero_area(cmd_q->lsbmap, in ccp_lsb_alloc() 36 bitmap_set(cmd_q->lsbmap, start, count); in ccp_lsb_alloc() 37 return start + cmd_q->lsb * LSB_SIZE; in ccp_lsb_alloc() 42 ccp = cmd_q->ccp; in ccp_lsb_alloc() 70 static void ccp_lsb_free(struct ccp_cmd_queue *cmd_q, unsigned int start, in ccp_lsb_free() argument 76 if (cmd_q->lsb == start) { in ccp_lsb_free() 78 bitmap_clear(cmd_q->lsbmap, start, count); in ccp_lsb_free() 81 struct ccp_device *ccp = cmd_q in ccp_lsb_free() 207 ccp5_get_free_slots(struct ccp_cmd_queue * cmd_q) ccp5_get_free_slots() argument 222 ccp5_do_cmd(struct ccp5_desc * desc,struct ccp_cmd_queue * cmd_q) ccp5_do_cmd() argument 587 ccp_find_lsb_regions(struct ccp_cmd_queue * cmd_q,u64 status) ccp_find_lsb_regions() argument 628 struct ccp_cmd_queue *cmd_q = &ccp->cmd_q[i]; ccp_find_and_assign_lsb_to_q() local 743 struct ccp_cmd_queue *cmd_q = &ccp->cmd_q[i]; ccp5_irq_bh() local 782 struct ccp_cmd_queue *cmd_q; ccp5_init() local 1008 struct ccp_cmd_queue *cmd_q; ccp5_destroy() local [all...] |
H A D | ccp-dev-v3.c | 19 static u32 ccp_alloc_ksb(struct ccp_cmd_queue *cmd_q, unsigned int count) in ccp_alloc_ksb() argument 22 struct ccp_device *ccp = cmd_q->ccp; in ccp_alloc_ksb() 50 static void ccp_free_ksb(struct ccp_cmd_queue *cmd_q, unsigned int start, in ccp_free_ksb() argument 53 struct ccp_device *ccp = cmd_q->ccp; in ccp_free_ksb() 69 static unsigned int ccp_get_free_slots(struct ccp_cmd_queue *cmd_q) in ccp_get_free_slots() argument 71 return CMD_Q_DEPTH(ioread32(cmd_q->reg_status)); in ccp_get_free_slots() 76 struct ccp_cmd_queue *cmd_q = op->cmd_q; in ccp_do_cmd() local 77 struct ccp_device *ccp = cmd_q->ccp; in ccp_do_cmd() 87 cmd_q in ccp_do_cmd() 327 struct ccp_cmd_queue *cmd_q; ccp_irq_bh() local 372 struct ccp_cmd_queue *cmd_q; ccp_init() local 519 struct ccp_cmd_queue *cmd_q; ccp_destroy() local [all...] |
H A D | ccp-debugfs.c | 115 struct ccp_cmd_queue *cmd_q = &ccp->cmd_q[i]; in ccp5_debugfs_stats_read() local 117 total_ops += cmd_q->total_ops; in ccp5_debugfs_stats_read() 118 total_aes_ops += cmd_q->total_aes_ops; in ccp5_debugfs_stats_read() 119 total_xts_aes_ops += cmd_q->total_xts_aes_ops; in ccp5_debugfs_stats_read() 120 total_3des_ops += cmd_q->total_3des_ops; in ccp5_debugfs_stats_read() 121 total_sha_ops += cmd_q->total_sha_ops; in ccp5_debugfs_stats_read() 122 total_rsa_ops += cmd_q->total_rsa_ops; in ccp5_debugfs_stats_read() 123 total_pt_ops += cmd_q->total_pt_ops; in ccp5_debugfs_stats_read() 124 total_ecc_ops += cmd_q in ccp5_debugfs_stats_read() 158 ccp5_debugfs_reset_queue_stats(struct ccp_cmd_queue * cmd_q) ccp5_debugfs_reset_queue_stats() argument 194 struct ccp_cmd_queue *cmd_q = filp->private_data; ccp5_debugfs_queue_read() local 249 struct ccp_cmd_queue *cmd_q = filp->private_data; ccp5_debugfs_queue_write() local 284 struct ccp_cmd_queue *cmd_q; ccp5_debugfs_setup() local [all...] |
H A D | ccp-ops.c | 154 struct ccp_cmd_queue *cmd_q, in ccp_init_dm_workarea() argument 163 wa->dev = cmd_q->ccp->dev; in ccp_init_dm_workarea() 167 wa->dma_pool = cmd_q->dma_pool; in ccp_init_dm_workarea() 266 static void ccp_free_data(struct ccp_data *data, struct ccp_cmd_queue *cmd_q) in ccp_free_data() argument 272 static int ccp_init_data(struct ccp_data *data, struct ccp_cmd_queue *cmd_q, in ccp_init_data() argument 281 ret = ccp_init_sg_workarea(&data->sg_wa, cmd_q->ccp->dev, sg, sg_len, in ccp_init_data() 286 ret = ccp_init_dm_workarea(&data->dm_wa, cmd_q, dm_len, dir); in ccp_init_data() 293 ccp_free_data(data, cmd_q); in ccp_init_data() 430 static int ccp_copy_to_from_sb(struct ccp_cmd_queue *cmd_q, in ccp_copy_to_from_sb() argument 438 op.cmd_q in ccp_copy_to_from_sb() 462 ccp_copy_to_sb(struct ccp_cmd_queue * cmd_q,struct ccp_dm_workarea * wa,u32 jobid,u32 sb,u32 byte_swap) ccp_copy_to_sb() argument 469 ccp_copy_from_sb(struct ccp_cmd_queue * cmd_q,struct ccp_dm_workarea * wa,u32 jobid,u32 sb,u32 byte_swap) ccp_copy_from_sb() argument 477 ccp_run_aes_cmac_cmd(struct ccp_cmd_queue * cmd_q,struct ccp_cmd * cmd) ccp_run_aes_cmac_cmd() argument 633 ccp_run_aes_gcm_cmd(struct ccp_cmd_queue * cmd_q,struct ccp_cmd * cmd) ccp_run_aes_gcm_cmd() argument 926 ccp_run_aes_cmd(struct ccp_cmd_queue * cmd_q,struct ccp_cmd * cmd) ccp_run_aes_cmd() argument 1103 ccp_run_xts_aes_cmd(struct ccp_cmd_queue * cmd_q,struct ccp_cmd * cmd) ccp_run_xts_aes_cmd() argument 1303 ccp_run_des3_cmd(struct ccp_cmd_queue * cmd_q,struct ccp_cmd * cmd) ccp_run_des3_cmd() argument 1499 ccp_run_sha_cmd(struct ccp_cmd_queue * cmd_q,struct ccp_cmd * cmd) ccp_run_sha_cmd() argument 1845 ccp_run_rsa_cmd(struct ccp_cmd_queue * cmd_q,struct ccp_cmd * cmd) ccp_run_rsa_cmd() argument 1977 ccp_run_passthru_cmd(struct ccp_cmd_queue * cmd_q,struct ccp_cmd * cmd) ccp_run_passthru_cmd() argument 2109 ccp_run_passthru_nomap_cmd(struct ccp_cmd_queue * cmd_q,struct ccp_cmd * cmd) ccp_run_passthru_nomap_cmd() argument 2173 ccp_run_ecc_mm_cmd(struct ccp_cmd_queue * cmd_q,struct ccp_cmd * cmd) ccp_run_ecc_mm_cmd() argument 2280 ccp_run_ecc_pm_cmd(struct ccp_cmd_queue * cmd_q,struct ccp_cmd * cmd) ccp_run_ecc_pm_cmd() argument 2450 ccp_run_ecc_cmd(struct ccp_cmd_queue * cmd_q,struct ccp_cmd * cmd) ccp_run_ecc_cmd() argument 2476 ccp_run_cmd(struct ccp_cmd_queue * cmd_q,struct ccp_cmd * cmd) ccp_run_cmd() argument [all...] |
H A D | ccp-dev.c | 327 if (ccp->cmd_q[i].active) in ccp_enqueue_cmd() 339 wake_up_process(ccp->cmd_q[i].kthread); in ccp_enqueue_cmd() 361 if (ccp->cmd_q[i].active) in ccp_do_cmd_backlog() 371 wake_up_process(ccp->cmd_q[i].kthread); in ccp_do_cmd_backlog() 374 static struct ccp_cmd *ccp_dequeue_cmd(struct ccp_cmd_queue *cmd_q) in ccp_dequeue_cmd() argument 376 struct ccp_device *ccp = cmd_q->ccp; in ccp_dequeue_cmd() 383 cmd_q->active = 0; in ccp_dequeue_cmd() 386 cmd_q->suspended = 1; in ccp_dequeue_cmd() 395 cmd_q->active = 1; in ccp_dequeue_cmd() 436 struct ccp_cmd_queue *cmd_q in ccp_cmd_queue_thread() local [all...] |
H A D | ccp-dev.h | 380 struct ccp_cmd_queue cmd_q[MAX_HW_QUEUES]; member 532 struct ccp_cmd_queue *cmd_q; member 641 int ccp_run_cmd(struct ccp_cmd_queue *cmd_q, struct ccp_cmd *cmd);
|
/linux/drivers/dma/amd/ae4dma/ |
H A D | ae4dma-dev.c | 20 struct pt_cmd_queue *cmd_q = &ae4cmd_q->cmd_q; in ae4_pending_work() local 32 cridx = readl(cmd_q->reg_control + AE4_RD_IDX_OFF); in ae4_pending_work() 52 struct pt_cmd_queue *cmd_q; in ae4_core_irq_handler() local 56 cmd_q = &ae4cmd_q->cmd_q; in ae4_core_irq_handler() 57 pt = cmd_q->pt; in ae4_core_irq_handler() 62 status = readl(cmd_q->reg_control + AE4_INTR_STS_OFF); in ae4_core_irq_handler() 65 writel(status, cmd_q->reg_control + AE4_INTR_STS_OFF); in ae4_core_irq_handler() 94 struct pt_cmd_queue *cmd_q; in ae4_core_init() local [all...] |
H A D | ae4dma.h | 49 struct pt_cmd_queue cmd_q; member
|
/linux/drivers/scsi/ibmvscsi_tgt/ |
H A D | ibmvscsi_tgt.c | 278 bytes = vscsi->cmd_q.size * PAGE_SIZE; in ibmvscsis_free_command_q() 279 memset(vscsi->cmd_q.base_addr, 0, bytes); in ibmvscsis_free_command_q() 280 vscsi->cmd_q.index = 0; in ibmvscsis_free_command_q() 363 crq = ibmvscsis_cmd_q_dequeue(vscsi->cmd_q.mask, &vscsi->cmd_q.index, in ibmvscsis_check_init_msg() 364 vscsi->cmd_q.base_addr); in ibmvscsis_check_init_msg() 377 crq = ibmvscsis_cmd_q_dequeue(vscsi->cmd_q.mask, in ibmvscsis_check_init_msg() 378 &vscsi->cmd_q.index, in ibmvscsis_check_init_msg() 379 vscsi->cmd_q.base_addr); in ibmvscsis_check_init_msg() 896 bytes = vscsi->cmd_q in ibmvscsis_reset_queue() [all...] |
H A D | ibmvscsi_tgt.h | 268 struct cmd_queue cmd_q; member
|
/linux/drivers/media/platform/amphion/ |
H A D | vpu_cmds.c | 150 list_for_each_entry_safe(cmd, tmp, &inst->cmd_q, list) { in vpu_process_cmd_request() 184 list_add_tail(&cmd->list, &inst->cmd_q); in vpu_request_cmd() 244 list_for_each_entry_safe(cmd, tmp, &inst->cmd_q, list) { in vpu_clear_request()
|
H A D | vpu.h | 254 struct list_head cmd_q; member
|
/linux/net/nfc/nci/ |
H A D | core.c | 541 skb_queue_purge(&ndev->cmd_q); in nci_open_device() 583 skb_queue_purge(&ndev->cmd_q); in nci_close_device() 1267 skb_queue_head_init(&ndev->cmd_q); in nci_register_device() 1396 skb_queue_tail(&ndev->cmd_q, skb); in nci_send_cmd() 1592 skb = skb_dequeue(&ndev->cmd_q); in nci_cmd_work()
|
H A D | rsp.c | 426 if (!skb_queue_empty(&ndev->cmd_q)) in nci_rsp_packet()
|
/linux/drivers/scsi/ |
H A D | sg.c | 153 char cmd_q; /* 1 -> allow command queuing, 0 -> don't */ member 722 sfp->cmd_q = 1; /* when sg_io_hdr seen, set command queuing on */ in sg_new_write() 1047 sfp->cmd_q = val ? 1 : 0; in sg_ioctl_common() 1050 return put_user((int) sfp->cmd_q, ip); in sg_ioctl_common() 1190 else if (!sfp->cmd_q) { in sg_poll() 2102 if (!sfp->cmd_q) in sg_add_request() 2170 sfp->cmd_q = SG_DEF_COMMAND_Q; in sg_add_sfp() 2539 seq_printf(s, " cmd_q=%d f_packid=%d k_orphan=%d closed=0\n", in sg_proc_debug_helper() 2540 (int) fp->cmd_q, (int) fp->force_packid, in sg_proc_debug_helper()
|
/linux/net/bluetooth/ |
H A D | hci_sync.c | 107 if (skb_queue_empty(&req->cmd_q)) in hci_cmd_sync_add() 112 skb_queue_tail(&req->cmd_q, skb); in hci_cmd_sync_add() 121 bt_dev_dbg(hdev, "length %u", skb_queue_len(&req->cmd_q)); in hci_req_sync_run() 127 skb_queue_purge(&req->cmd_q); in hci_req_sync_run() 132 if (skb_queue_empty(&req->cmd_q)) in hci_req_sync_run() 135 skb = skb_peek_tail(&req->cmd_q); in hci_req_sync_run() 139 spin_lock_irqsave(&hdev->cmd_q.lock, flags); in hci_req_sync_run() 140 skb_queue_splice_tail(&req->cmd_q, &hdev->cmd_q); in hci_req_sync_run() 141 spin_unlock_irqrestore(&hdev->cmd_q in hci_req_sync_run() [all...] |
/linux/include/net/bluetooth/ |
H A D | hci_sync.h | 20 struct sk_buff_head cmd_q; member
|
/linux/drivers/net/ethernet/brocade/bna/ |
H A D | bfa_ioc.h | 109 struct list_head cmd_q; /*!< pending mbox queue */ member
|
/linux/include/net/nfc/ |
H A D | nci_core.h | 225 struct sk_buff_head cmd_q; member
|
/linux/drivers/scsi/bfa/ |
H A D | bfa_ioc.c | 82 (!list_empty(&((__ioc)->mbox_mod.cmd_q)) || \ 1980 INIT_LIST_HEAD(&mod->cmd_q); in bfa_ioc_mbox_attach() 2000 if (list_empty(&mod->cmd_q)) in bfa_ioc_mbox_poll() 2013 bfa_q_deq(&mod->cmd_q, &cmd); in bfa_ioc_mbox_poll() 2026 while (!list_empty(&mod->cmd_q)) in bfa_ioc_mbox_flush() 2027 bfa_q_deq(&mod->cmd_q, &cmd); in bfa_ioc_mbox_flush() 2546 if (!list_empty(&mod->cmd_q)) { in bfa_ioc_mbox_queue() 2547 list_add_tail(&cmd->qe, &mod->cmd_q); in bfa_ioc_mbox_queue() 2556 list_add_tail(&cmd->qe, &mod->cmd_q); in bfa_ioc_mbox_queue()
|
/linux/drivers/net/ethernet/qlogic/qlcnic/ |
H A D | qlcnic_83xx_hw.c | 3897 struct list_head *head = &mbx->cmd_q; in qlcnic_83xx_flush_mbx_queue() 4019 list_add_tail(&cmd->list, &mbx->cmd_q); in qlcnic_83xx_enqueue_mbx_cmd() 4104 struct list_head *head = &mbx->cmd_q; in qlcnic_83xx_mailbox_worker() 4177 INIT_LIST_HEAD(&mbx->cmd_q); in qlcnic_83xx_init_mailbox_work()
|