Lines Matching full:va
58 static u32 *decode_read_list(u32 *va, u32 *vaend) in decode_read_list() argument
60 struct rpcrdma_read_chunk *ch = (struct rpcrdma_read_chunk *)va; in decode_read_list()
75 va = (u32 *)&ch->rc_target.rs_offset; in decode_read_list()
76 xdr_decode_hyper(va, &ch_offset); in decode_read_list()
77 put_unaligned(ch_offset, (u64 *)va); in decode_read_list()
105 * offset : remove va + <count>
109 static u32 *decode_write_list(u32 *va, u32 *vaend) in decode_write_list() argument
113 (struct rpcrdma_write_array *)va; in decode_write_list()
140 va = (u32 *)&ary->wc_array[ch_no].wc_target.rs_offset; in decode_write_list()
141 xdr_decode_hyper(va, &ch_offset); in decode_write_list()
142 put_unaligned(ch_offset, (u64 *)va); in decode_write_list()
152 static u32 *decode_reply_array(u32 *va, u32 *vaend) in decode_reply_array() argument
156 (struct rpcrdma_write_array *)va; in decode_reply_array()
183 va = (u32 *)&ary->wc_array[ch_no].wc_target.rs_offset; in decode_reply_array()
184 xdr_decode_hyper(va, &ch_offset); in decode_reply_array()
185 put_unaligned(ch_offset, (u64 *)va); in decode_reply_array()
195 u32 *va; in svc_rdma_xdr_decode_req() local
225 va = &rmsgp->rm_body.rm_padded.rm_pempty[4]; in svc_rdma_xdr_decode_req()
226 rqstp->rq_arg.head[0].iov_base = va; in svc_rdma_xdr_decode_req()
227 hdrlen = (u32)((unsigned long)va - (unsigned long)rmsgp); in svc_rdma_xdr_decode_req()
237 va = &rmsgp->rm_body.rm_chunks[0]; in svc_rdma_xdr_decode_req()
239 va = decode_read_list(va, vaend); in svc_rdma_xdr_decode_req()
240 if (!va) in svc_rdma_xdr_decode_req()
242 va = decode_write_list(va, vaend); in svc_rdma_xdr_decode_req()
243 if (!va) in svc_rdma_xdr_decode_req()
245 va = decode_reply_array(va, vaend); in svc_rdma_xdr_decode_req()
246 if (!va) in svc_rdma_xdr_decode_req()
249 rqstp->rq_arg.head[0].iov_base = va; in svc_rdma_xdr_decode_req()
250 hdr_len = (unsigned long)va - (unsigned long)rmsgp; in svc_rdma_xdr_decode_req()
262 u32 *va; in svc_rdma_xdr_decode_deferred_req() local
271 va = &rmsgp->rm_body.rm_padded.rm_pempty[4]; in svc_rdma_xdr_decode_deferred_req()
272 rqstp->rq_arg.head[0].iov_base = va; in svc_rdma_xdr_decode_deferred_req()
273 hdrlen = (u32)((unsigned long)va - (unsigned long)rmsgp); in svc_rdma_xdr_decode_deferred_req()
281 va = &rmsgp->rm_body.rm_chunks[0]; in svc_rdma_xdr_decode_deferred_req()
284 for (ch = (struct rpcrdma_read_chunk *)va; in svc_rdma_xdr_decode_deferred_req()
286 va = (u32 *)&ch->rc_position; in svc_rdma_xdr_decode_deferred_req()
289 ary = (struct rpcrdma_write_array *)va; in svc_rdma_xdr_decode_deferred_req()
291 va = (u32 *)&ary->wc_nchunks; in svc_rdma_xdr_decode_deferred_req()
297 va = (u32 *)&ary->wc_array[ary->wc_nchunks].wc_target.rs_length; in svc_rdma_xdr_decode_deferred_req()
300 ary = (struct rpcrdma_write_array *)va; in svc_rdma_xdr_decode_deferred_req()
302 va = (u32 *)&ary->wc_nchunks; in svc_rdma_xdr_decode_deferred_req()
304 va = (u32 *)&ary->wc_array[ary->wc_nchunks]; in svc_rdma_xdr_decode_deferred_req()
306 rqstp->rq_arg.head[0].iov_base = va; in svc_rdma_xdr_decode_deferred_req()
307 hdrlen = (unsigned long)va - (unsigned long)rmsgp; in svc_rdma_xdr_decode_deferred_req()
315 enum rpcrdma_errcode err, u32 *va) in svc_rdma_xdr_encode_error() argument
317 u32 *startp = va; in svc_rdma_xdr_encode_error()
319 *va++ = htonl(rmsgp->rm_xid); in svc_rdma_xdr_encode_error()
320 *va++ = htonl(rmsgp->rm_vers); in svc_rdma_xdr_encode_error()
321 *va++ = htonl(xprt->sc_max_requests); in svc_rdma_xdr_encode_error()
322 *va++ = htonl(RDMA_ERROR); in svc_rdma_xdr_encode_error()
323 *va++ = htonl(err); in svc_rdma_xdr_encode_error()
325 *va++ = htonl(RPCRDMA_VERSION); in svc_rdma_xdr_encode_error()
326 *va++ = htonl(RPCRDMA_VERSION); in svc_rdma_xdr_encode_error()
329 return (int)((unsigned long)va - (unsigned long)startp); in svc_rdma_xdr_encode_error()