Home
last modified time | relevance | path

Searched refs:sg_list (Results 1 – 25 of 136) sorted by relevance

123456

/linux/drivers/virt/
H A Dfsl_hypervisor.c153 struct fh_sg_list *sg_list = NULL; in ioctl_memcpy() local
236 * sg_list is the list of fh_sg_list objects that we pass to the in ioctl_memcpy()
246 sg_list = PTR_ALIGN(sg_list_unaligned, sizeof(struct fh_sg_list)); in ioctl_memcpy()
263 sg_list[0].source = page_to_phys(pages[0]) + lb_offset; in ioctl_memcpy()
264 sg_list[0].target = param.remote_paddr; in ioctl_memcpy()
266 sg_list[0].source = param.remote_paddr; in ioctl_memcpy()
267 sg_list[0].target = page_to_phys(pages[0]) + lb_offset; in ioctl_memcpy()
269 sg_list[0].size = min_t(uint64_t, param.count, PAGE_SIZE - lb_offset); in ioctl_memcpy()
271 remote_paddr = param.remote_paddr + sg_list[0].size; in ioctl_memcpy()
272 count = param.count - sg_list[ in ioctl_memcpy()
[all...]
/linux/drivers/ufs/core/
H A Dufs_bsg.c57 sg_copy_to_buffer(job->request_payload.sg_list, in ufs_bsg_alloc_desc_buffer()
73 struct scatterlist *sg_list = NULL; in ufs_bsg_exec_advanced_rpmb_req() local
111 sg_cnt = dma_map_sg(hba->host->dma_dev, payload->sg_list, payload->sg_cnt, dir); in ufs_bsg_exec_advanced_rpmb_req()
114 sg_list = payload->sg_list; in ufs_bsg_exec_advanced_rpmb_req()
120 &rpmb_reply->ehs_rsp, sg_cnt, sg_list, dir); in ufs_bsg_exec_advanced_rpmb_req()
123 dma_unmap_sg(hba->host->dma_dev, payload->sg_list, payload->sg_cnt, dir); in ufs_bsg_exec_advanced_rpmb_req()
166 sg_copy_from_buffer(job->request_payload.sg_list, in ufs_bsg_request()
/linux/drivers/infiniband/sw/rxe/
H A Drxe_odp.c459 u32 pf_flags, struct ib_sge *sg_list, in rxe_ib_prefetch_sg_list() argument
471 sg_list[i].lkey, RXE_LOOKUP_LOCAL); in rxe_ib_prefetch_sg_list()
475 sg_list[i].lkey); in rxe_ib_prefetch_sg_list()
487 mr, sg_list[i].addr, sg_list[i].length, pf_flags); in rxe_ib_prefetch_sg_list()
505 u32 flags, struct ib_sge *sg_list, in rxe_ib_advise_mr_prefetch() argument
522 return rxe_ib_prefetch_sg_list(ibpd, advice, pf_flags, sg_list, in rxe_ib_advise_mr_prefetch()
537 sg_list[i].lkey, RXE_LOOKUP_LOCAL); in rxe_ib_advise_mr_prefetch()
543 work->frags[i].io_virt = sg_list[i].addr; in rxe_ib_advise_mr_prefetch()
544 work->frags[i].length = sg_list[ in rxe_ib_advise_mr_prefetch()
567 rxe_ib_advise_mr(struct ib_pd * ibpd,enum ib_uverbs_advise_mr_advice advice,u32 flags,struct ib_sge * sg_list,u32 num_sge,struct uverbs_attr_bundle * attrs) rxe_ib_advise_mr() argument
[all...]
/linux/drivers/infiniband/hw/mana/
H A Dwr.c27 gdma_sgl[i].address = wr->sg_list[i].addr; in mana_ib_post_recv_ud()
28 gdma_sgl[i].mem_key = wr->sg_list[i].lkey; in mana_ib_post_recv_ud()
29 gdma_sgl[i].size = wr->sg_list[i].length; in mana_ib_post_recv_ud()
107 gdma_sgl[i + 1].address = wr->wr.sg_list[i].addr; in mana_ib_post_send_ud()
108 gdma_sgl[i + 1].mem_key = wr->wr.sg_list[i].lkey; in mana_ib_post_send_ud()
109 gdma_sgl[i + 1].size = wr->wr.sg_list[i].length; in mana_ib_post_send_ud()
/linux/drivers/tee/amdtee/
H A Dcall.c315 cmd->sg_list.count = count; in handle_map_shmem()
320 cmd->sg_list.buf[i].hi_addr = upper_32_bits(paddr); in handle_map_shmem()
321 cmd->sg_list.buf[i].low_addr = lower_32_bits(paddr); in handle_map_shmem()
322 cmd->sg_list.buf[i].size = start[i].size; in handle_map_shmem()
323 cmd->sg_list.size += cmd->sg_list.buf[i].size; in handle_map_shmem()
326 cmd->sg_list.buf[i].hi_addr); in handle_map_shmem()
328 cmd->sg_list.buf[i].low_addr); in handle_map_shmem()
329 pr_debug("buf[%d]:size = 0x%x\n", i, cmd->sg_list.buf[i].size); in handle_map_shmem()
330 pr_debug("list size = 0x%x\n", cmd->sg_list in handle_map_shmem()
[all...]
/linux/block/
H A Dbsg-lib.c163 kfree(job->request_payload.sg_list); in bsg_teardown_job()
164 kfree(job->reply_payload.sg_list); in bsg_teardown_job()
218 buf->sg_list = kmalloc(sz, GFP_KERNEL); in bsg_map_buffer()
219 if (!buf->sg_list) in bsg_map_buffer()
221 sg_init_table(buf->sg_list, req->nr_phys_segments); in bsg_map_buffer()
222 buf->sg_cnt = blk_rq_map_sg(req, buf->sg_list); in bsg_map_buffer()
256 kfree(job->request_payload.sg_list); in bsg_prepare_job()
/linux/drivers/infiniband/sw/rdmavt/
H A Drc.c164 ss->sge = wqe->sg_list[0]; in rvt_restart_sge()
165 ss->sg_list = wqe->sg_list + 1; in rvt_restart_sge()
H A Dqp.c626 struct rvt_sge *sge = &wqe->sg_list[i]; in rvt_swqe_has_lkey()
1078 sz = struct_size(swq, sg_list, init_attr->cap.max_send_sge); in rvt_create_qp()
1828 wqe->sg_list[i].addr = wr->sg_list[i].addr; in rvt_post_recv()
1829 wqe->sg_list[i].length = wr->sg_list[i].length; in rvt_post_recv()
1830 wqe->sg_list[i].lkey = wr->sg_list[i].lkey; in rvt_post_recv()
1877 wr->sg_list[0].length < sizeof(u64) || in rvt_qp_valid_operation()
1878 wr->sg_list[ in rvt_qp_valid_operation()
[all...]
/linux/include/rdma/
H A Drdmavt_mr.h77 struct rvt_sge *sg_list; /* next SGE to be used if any */ member
98 ss->sge = *ss->sg_list++; in rvt_put_ss()
126 *sge = *ss->sg_list++; in rvt_update_sge()
/linux/drivers/dma/
H A Dimx-dma.c160 struct scatterlist *sg_list; member
789 kfree(imxdmac->sg_list); in imxdma_free_chan_resources()
790 imxdmac->sg_list = NULL; in imxdma_free_chan_resources()
866 kfree(imxdmac->sg_list); in imxdma_prep_dma_cyclic()
868 imxdmac->sg_list = kcalloc(periods + 1, in imxdma_prep_dma_cyclic()
870 if (!imxdmac->sg_list) in imxdma_prep_dma_cyclic()
873 sg_init_table(imxdmac->sg_list, periods); in imxdma_prep_dma_cyclic()
876 sg_assign_page(&imxdmac->sg_list[i], NULL); in imxdma_prep_dma_cyclic()
877 imxdmac->sg_list[i].offset = 0; in imxdma_prep_dma_cyclic()
878 imxdmac->sg_list[ in imxdma_prep_dma_cyclic()
[all...]
/linux/drivers/infiniband/hw/qedr/
H A Dqedr_roce_cm.c110 qp->rqe_wr_id[qp->rq.gsi_cons].sg_list[0].length = in qedr_ll2_complete_rx_packet()
407 send_size += swr->sg_list[i].length; in qedr_gsi_build_header()
532 packet->payload[i].baddr = swr->sg_list[i].addr; in qedr_gsi_build_packet()
533 packet->payload[i].len = swr->sg_list[i].length; in qedr_gsi_build_packet()
640 wr->sg_list[0].addr, in qedr_gsi_post_recv()
641 wr->sg_list[0].length, in qedr_gsi_post_recv()
653 qp->rqe_wr_id[qp->rq.prod].sg_list[0] = wr->sg_list[0]; in qedr_gsi_post_recv()
691 wc[i].byte_len = qp->rqe_wr_id[qp->rq.cons].sg_list[0].length; in qedr_gsi_poll_cq()
/linux/drivers/scsi/aacraid/
H A Dcommctrl.c494 void *sg_list[HBA_MAX_SG_EMBEDDED]; in aac_send_raw_srb() local
519 memset(sg_list, 0, sizeof(sg_list)); /* cleanup may take issue */ in aac_send_raw_srb()
553 if (user_srbcmd->sg.count > ARRAY_SIZE(sg_list)) { in aac_send_raw_srb()
681 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
742 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
797 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
854 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
895 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
935 if (copy_to_user(sg_user[i], sg_list[ in aac_send_raw_srb()
[all...]
/linux/drivers/net/ethernet/marvell/octeon_ep/
H A Doctep_ctrl_net.c51 msg->sg_list[0].msg = buf; in init_send_req()
52 msg->sg_list[0].sz = msg->hdr.s.sz; in init_send_req()
309 msg->sg_list[0].msg, in process_mbox_resp()
325 req = (struct octep_ctrl_net_f2h_req *)msg->sg_list[0].msg; in process_mbox_notify()
367 msg.sg_list[0].sz = msg_sz; in octep_ctrl_net_recv_fw_messages()
368 msg.sg_list[0].msg = &data; in octep_ctrl_net_recv_fw_messages()
/linux/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dotx2_txrx.h78 struct sg_list { struct
101 struct sg_list *sg; argument
/linux/drivers/firmware/efi/
H A Druntime-wrappers.c102 unsigned long sg_list; member
283 args->UPDATE_CAPSULE.sg_list); in efi_call_rts()
523 unsigned long sg_list) in virt_efi_update_capsule() argument
532 status = efi_queue_work(UPDATE_CAPSULE, capsules, count, sg_list); in virt_efi_update_capsule()
/linux/drivers/infiniband/hw/cxgb4/
H A Dqp.c424 if ((plen + wr->sg_list[i].length) > max) in build_immd()
426 srcp = (u8 *)(unsigned long)wr->sg_list[i].addr; in build_immd()
427 plen += wr->sg_list[i].length; in build_immd()
428 rem = wr->sg_list[i].length; in build_immd()
454 struct fw_ri_isgl *isglp, struct ib_sge *sg_list, in build_isgl() argument
468 if ((plen + sg_list[i].length) < plen) in build_isgl()
470 plen += sg_list[i].length; in build_isgl()
471 *flitp = cpu_to_be64(((u64)sg_list[i].lkey << 32) | in build_isgl()
472 sg_list[i].length); in build_isgl()
475 *flitp = cpu_to_be64(sg_list[ in build_isgl()
[all...]
/linux/drivers/net/ethernet/ibm/ehea/
H A Dehea_qmr.h106 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member
115 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES-1]; member
132 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member
/linux/arch/powerpc/include/asm/
H A Dfsl_hcalls.h352 * @sg_list: guest physical address of an array of &fh_sg_list structures
353 * @count: the number of entries in @sg_list
358 unsigned int target, phys_addr_t sg_list, unsigned int count) in fh_partition_memcpy() argument
370 r5 = (uint32_t) sg_list; in fh_partition_memcpy()
373 r6 = sg_list >> 32; in fh_partition_memcpy()
/linux/include/uapi/rdma/
H A Drvt-abi.h40 * The size of the sg_list is determined when the QP (or SRQ) is created
47 struct rvt_wqe_sge sg_list[]; member
/linux/fs/smb/server/
H A Dtransport_rdma.c203 struct scatterlist sg_list[]; member
647 wr.sg_list = &recvmsg->sge; in smb_direct_post_recv()
1060 static int get_sg_list(void *buf, int size, struct scatterlist *sg_list, int nentries) in get_sg_list() argument
1079 if (!sg_list) in get_sg_list()
1081 sg_set_page(sg_list, page, len, offset); in get_sg_list()
1082 sg_list = sg_next(sg_list); in get_sg_list()
1093 struct scatterlist *sg_list, int nentries, in get_mapped_sg_list() argument
1098 npages = get_sg_list(buf, size, sg_list, nentries); in get_mapped_sg_list()
1101 return ib_dma_map_sg(device, sg_list, npage in get_mapped_sg_list()
[all...]
/linux/include/linux/
H A Dbsg-lib.h26 struct scatterlist *sg_list; member
/linux/drivers/infiniband/hw/mlx4/
H A Dsrq.c346 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); in mlx4_ib_post_srq_recv()
347 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); in mlx4_ib_post_srq_recv()
348 scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); in mlx4_ib_post_srq_recv()
/linux/drivers/xen/
H A Defi.c220 unsigned long count, unsigned long sg_list) in xen_efi_update_capsule() argument
230 efi_data(op).u.update_capsule.sg_list = sg_list; in xen_efi_update_capsule()
/linux/drivers/scsi/qla4xxx/
H A Dql4_bsg.c62 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_read_flash()
122 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla4xxx_update_flash()
186 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_get_acb_state()
257 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_read_nvram()
321 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla4xxx_update_nvram()
435 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_bsg_get_acb()
/linux/drivers/infiniband/hw/hfi1/
H A Duc.c112 qp->s_sge.sge = wqe->sg_list[0]; in hfi1_make_uc_req()
113 qp->s_sge.sg_list = wqe->sg_list + 1; in hfi1_make_uc_req()
447 qp->r_sge.sg_list = NULL; in hfi1_uc_rcv()

123456