/linux/drivers/bus/mhi/ep/ |
H A D | ring.c | 26 ring->rd_offset = (ring->rd_offset + 1) % ring->ring_size; in mhi_ep_ring_inc_index() 122 if (ring->rd_offset < ring->wr_offset) in mhi_ep_ring_add_element() 123 num_free_elem = (ring->wr_offset - ring->rd_offset) - 1; in mhi_ep_ring_add_element() 125 num_free_elem = ((ring->ring_size - ring->rd_offset) + ring->wr_offset) - 1; in mhi_ep_ring_add_element() 133 old_offset = ring->rd_offset; in mhi_ep_ring_add_element() 135 dev_dbg(dev, "Adding an element to ring at offset (%zu)\n", ring->rd_offset); in mhi_ep_ring_add_element() 147 rp = cpu_to_le64(ring->rd_offset * sizeof(*el) + ring->rbase); in mhi_ep_ring_add_element() 204 ring->rd_offset = mhi_ep_ring_addr2offset(ring, le64_to_cpu(val)); in mhi_ep_ring_start()
|
H A D | main.c | 97 event->ptr = cpu_to_le64(ring->rbase + ring->rd_offset * sizeof(*tre)); in mhi_ep_send_completion_event() 153 event->ptr = cpu_to_le64(ring->rbase + ring->rd_offset * sizeof(struct mhi_ring_element)); in mhi_ep_send_cmd_comp_event() 203 mhi_chan->rd_offset = ch_ring->rd_offset; in mhi_ep_process_cmd_ring() 333 return !!(mhi_chan->rd_offset == ring->wr_offset); in mhi_ep_queue_is_empty() 343 struct mhi_ring_element *el = &ring->ring_cache[ring->rd_offset]; in mhi_ep_read_completion() 424 el = &ring->ring_cache[mhi_chan->rd_offset]; in mhi_ep_read_channel() 469 mhi_chan->rd_offset = (mhi_chan->rd_offset + 1) % ring->ring_size; in mhi_ep_read_channel() 525 struct mhi_ring_element *el = &ring->ring_cache[ring->rd_offset]; in mhi_ep_skb_completion() [all...] |
H A D | internal.h | 131 size_t rd_offset; member 165 size_t rd_offset; member
|
/linux/drivers/target/ |
H A D | target_core_rd.c | 425 u32 rd_offset; in rd_execute_rw() local 438 rd_offset = do_div(tmp, PAGE_SIZE); in rd_execute_rw() 451 cmd->t_task_lba, rd_size, rd_page, rd_offset); in rd_execute_rw() 460 src_len = PAGE_SIZE - rd_offset; in rd_execute_rw() 479 rd_page, rd_offset, rd_size); in rd_execute_rw() 484 rd_addr = sg_virt(rd_sg) + rd_offset; in rd_execute_rw() 497 rd_offset += len; in rd_execute_rw() 503 rd_offset = 0; in rd_execute_rw()
|
/linux/drivers/gpu/drm/xe/ |
H A D | xe_gsc.c | 135 u32 rd_offset; in query_compatibility_version() local 166 &rd_offset); in query_compatibility_version() 172 compat->major = version_query_rd(xe, &bo->vmap, rd_offset, proj_major); in query_compatibility_version() 173 compat->minor = version_query_rd(xe, &bo->vmap, rd_offset, compat_major); in query_compatibility_version() 174 compat->patch = version_query_rd(xe, &bo->vmap, rd_offset, compat_minor); in query_compatibility_version()
|
H A D | xe_huc.c | 159 u32 rd_offset; in huc_auth_via_gsccs() local 195 sizeof(struct pxp43_huc_auth_out), &rd_offset); in huc_auth_via_gsccs() 206 out_status = huc_auth_msg_rd(xe, &pkt->vmap, rd_offset, header.status); in huc_auth_via_gsccs()
|
H A D | xe_pxp_submit.c | 406 u32 rd_offset; in gsccs_send_message() local 468 min_reply_size, &rd_offset); in gsccs_send_message() 475 reply_size = xe_map_rd_field(xe, &gsc_res->msg_out, rd_offset, in gsccs_send_message() 486 rd_offset, reply_size); in gsccs_send_message()
|
/linux/drivers/spi/ |
H A D | spi-mtk-snfi.c | 869 u32 rd_offset = 0; in mtk_snand_read_page_cache() local 892 rd_offset = op_addr & mask; in mtk_snand_read_page_cache() 896 if (rd_offset == 0 && op->data.nbytes >= snf->nfi_cfg.page_size) in mtk_snand_read_page_cache() 1021 } else if (rd_offset < snf->buf_len) { in mtk_snand_read_page_cache() 1022 u32 cap_len = snf->buf_len - rd_offset; in mtk_snand_read_page_cache() 1026 memcpy(op->data.buf.in, snf->buf + rd_offset, cap_len); in mtk_snand_read_page_cache()
|
/linux/drivers/crypto/inside-secure/ |
H A D | safexcel.c | 565 writel(EIP197_xDR_DESC_MODE_64BIT | (priv->config.rd_offset << 14) | in safexcel_hw_setup_rdesc_rings() 571 (rd_fetch_cnt * (priv->config.rd_offset / sizeof(u32))), in safexcel_hw_setup_rdesc_rings() 764 writel((EIP197_DEFAULT_RING_SIZE * priv->config.rd_offset), in safexcel_hw_init() 890 writel((rdesc * priv->config.rd_offset), in safexcel_dequeue() 1053 (tot_descs * priv->config.rd_offset), in safexcel_handle_result_descriptor() 1342 priv->config.rd_offset = (priv->config.rd_size + mask) & ~mask; in safexcel_configure() 1347 priv->config.rd_offset *= sizeof(u32); in safexcel_configure()
|
H A D | safexcel_ring.c | 57 rdr->offset = priv->config.rd_offset; in safexcel_init_ring_descriptors()
|
H A D | safexcel.h | 684 u32 rd_offset; member
|
/linux/drivers/scsi/lpfc/ |
H A D | lpfc_bsg.h | 294 uint32_t rd_offset; member
|
H A D | lpfc_bsg.c | 5458 u32 rd_offset, rd_index, offset; in lpfc_bsg_get_ras_fwlog() local 5491 rd_offset = ras_req->read_offset; in lpfc_bsg_get_ras_fwlog() 5500 rd_index = (rd_offset / LPFC_RAS_MAX_ENTRY_SIZE); in lpfc_bsg_get_ras_fwlog() 5501 offset = (rd_offset % LPFC_RAS_MAX_ENTRY_SIZE); in lpfc_bsg_get_ras_fwlog()
|
/linux/fs/nfsd/ |
H A D | nfs4proc.c | 955 read->rd_offset, read->rd_length); in nfsd4_read() 958 if (read->rd_offset > (u64)OFFSET_MAX) in nfsd4_read() 959 read->rd_offset = (u64)OFFSET_MAX; in nfsd4_read() 960 if (read->rd_offset + read->rd_length > (u64)OFFSET_MAX) in nfsd4_read() 961 read->rd_length = (u64)OFFSET_MAX - read->rd_offset; in nfsd4_read() 994 u->read.rd_offset, u->read.rd_length); in nfsd4_read_release()
|
H A D | xdr4.h | 421 u64 rd_offset; /* request */ member
|
H A D | nfs4xdr.c | 1279 if (xdr_stream_decode_u64(argp->xdr, &read->rd_offset) < 0) in nfsd4_decode_read() 4440 file, read->rd_offset, &maxcount, in nfsd4_encode_splice_read() 4492 read->rd_offset, &maxcount, base, in nfsd4_encode_readv() 5438 wire_offset = cpu_to_be64(read->rd_offset); in nfsd4_encode_read_plus_data() 5468 read->rd_eof = read->rd_offset >= i_size_read(file_inode(file)); in nfsd4_encode_read_plus()
|