1 /*
2  * Copyright (c) 2005 Ammasso, Inc. All rights reserved.
3  * Copyright (c) 2005 Open Grid Computing, Inc. All rights reserved.
4  *
5  * This software is available to you under a choice of one of two
6  * licenses.  You may choose to be licensed under the terms of the GNU
7  * General Public License (GPL) Version 2, available from the file
8  * COPYING in the main directory of this source tree, or the
9  * OpenIB.org BSD license below:
10  *
11  *     Redistribution and use in source and binary forms, with or
12  *     without modification, are permitted provided that the following
13  *     conditions are met:
14  *
15  *      - Redistributions of source code must retain the above
16  *        copyright notice, this list of conditions and the following
17  *        disclaimer.
18  *
19  *      - Redistributions in binary form must reproduce the above
20  *        copyright notice, this list of conditions and the following
21  *        disclaimer in the documentation and/or other materials
22  *        provided with the distribution.
23  *
24  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
25  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
26  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
27  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
28  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
29  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
30  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
31  * SOFTWARE.
32  */
33 #include <linux/slab.h>
34 
35 #include "c2.h"
36 #include "c2_vq.h"
37 
38 #define PBL_VIRT 1
39 #define PBL_PHYS 2
40 
41 /*
42  * Send all the PBL messages to convey the remainder of the PBL
43  * Wait for the adapter's reply on the last one.
44  * This is indicated by setting the MEM_PBL_COMPLETE in the flags.
45  *
46  * NOTE:  vq_req is _not_ freed by this function.  The VQ Host
47  *	  Reply buffer _is_ freed by this function.
48  */
49 static int
send_pbl_messages(struct c2_dev * c2dev,__be32 stag_index,unsigned long va,u32 pbl_depth,struct c2_vq_req * vq_req,int pbl_type)50 send_pbl_messages(struct c2_dev *c2dev, __be32 stag_index,
51 		  unsigned long va, u32 pbl_depth,
52 		  struct c2_vq_req *vq_req, int pbl_type)
53 {
54 	u32 pbe_count;		/* amt that fits in a PBL msg */
55 	u32 count;		/* amt in this PBL MSG. */
56 	struct c2wr_nsmr_pbl_req *wr;	/* PBL WR ptr */
57 	struct c2wr_nsmr_pbl_rep *reply;	/* reply ptr */
58  	int err, pbl_virt, pbl_index, i;
59 
60 	switch (pbl_type) {
61 	case PBL_VIRT:
62 		pbl_virt = 1;
63 		break;
64 	case PBL_PHYS:
65 		pbl_virt = 0;
66 		break;
67 	default:
68 		return -EINVAL;
69 		break;
70 	}
71 
72 	pbe_count = (c2dev->req_vq.msg_size -
73 		     sizeof(struct c2wr_nsmr_pbl_req)) / sizeof(u64);
74 	wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL);
75 	if (!wr) {
76 		return -ENOMEM;
77 	}
78 	c2_wr_set_id(wr, CCWR_NSMR_PBL);
79 
80 	/*
81 	 * Only the last PBL message will generate a reply from the verbs,
82 	 * so we set the context to 0 indicating there is no kernel verbs
83 	 * handler blocked awaiting this reply.
84 	 */
85 	wr->hdr.context = 0;
86 	wr->rnic_handle = c2dev->adapter_handle;
87 	wr->stag_index = stag_index;	/* already swapped */
88 	wr->flags = 0;
89 	pbl_index = 0;
90 	while (pbl_depth) {
91 		count = min(pbe_count, pbl_depth);
92 		wr->addrs_length = cpu_to_be32(count);
93 
94 		/*
95 		 *  If this is the last message, then reference the
96 		 *  vq request struct cuz we're gonna wait for a reply.
97 		 *  also make this PBL msg as the last one.
98 		 */
99 		if (count == pbl_depth) {
100 			/*
101 			 * reference the request struct.  dereferenced in the
102 			 * int handler.
103 			 */
104 			vq_req_get(c2dev, vq_req);
105 			wr->flags = cpu_to_be32(MEM_PBL_COMPLETE);
106 
107 			/*
108 			 * This is the last PBL message.
109 			 * Set the context to our VQ Request Object so we can
110 			 * wait for the reply.
111 			 */
112 			wr->hdr.context = (unsigned long) vq_req;
113 		}
114 
115 		/*
116 		 * If pbl_virt is set then va is a virtual address
117 		 * that describes a virtually contiguous memory
118 		 * allocation. The wr needs the start of each virtual page
119 		 * to be converted to the corresponding physical address
120 		 * of the page. If pbl_virt is not set then va is an array
121 		 * of physical addresses and there is no conversion to do.
122 		 * Just fill in the wr with what is in the array.
123 		 */
124 		for (i = 0; i < count; i++) {
125 			if (pbl_virt) {
126 				va += PAGE_SIZE;
127 			} else {
128  				wr->paddrs[i] =
129 				    cpu_to_be64(((u64 *)va)[pbl_index + i]);
130 			}
131 		}
132 
133 		/*
134 		 * Send WR to adapter
135 		 */
136 		err = vq_send_wr(c2dev, (union c2wr *) wr);
137 		if (err) {
138 			if (count <= pbe_count) {
139 				vq_req_put(c2dev, vq_req);
140 			}
141 			goto bail0;
142 		}
143 		pbl_depth -= count;
144 		pbl_index += count;
145 	}
146 
147 	/*
148 	 *  Now wait for the reply...
149 	 */
150 	err = vq_wait_for_reply(c2dev, vq_req);
151 	if (err) {
152 		goto bail0;
153 	}
154 
155 	/*
156 	 * Process reply
157 	 */
158 	reply = (struct c2wr_nsmr_pbl_rep *) (unsigned long) vq_req->reply_msg;
159 	if (!reply) {
160 		err = -ENOMEM;
161 		goto bail0;
162 	}
163 
164 	err = c2_errno(reply);
165 
166 	vq_repbuf_free(c2dev, reply);
167       bail0:
168 	kfree(wr);
169 	return err;
170 }
171 
172 #define C2_PBL_MAX_DEPTH 131072
173 int
c2_nsmr_register_phys_kern(struct c2_dev * c2dev,u64 * addr_list,int page_size,int pbl_depth,u32 length,u32 offset,u64 * va,enum c2_acf acf,struct c2_mr * mr)174 c2_nsmr_register_phys_kern(struct c2_dev *c2dev, u64 *addr_list,
175  			   int page_size, int pbl_depth, u32 length,
176  			   u32 offset, u64 *va, enum c2_acf acf,
177 			   struct c2_mr *mr)
178 {
179 	struct c2_vq_req *vq_req;
180 	struct c2wr_nsmr_register_req *wr;
181 	struct c2wr_nsmr_register_rep *reply;
182 	u16 flags;
183 	int i, pbe_count, count;
184 	int err;
185 
186 	if (!va || !length || !addr_list || !pbl_depth)
187 		return -EINTR;
188 
189 	/*
190 	 * Verify PBL depth is within rnic max
191 	 */
192 	if (pbl_depth > C2_PBL_MAX_DEPTH) {
193 		return -EINTR;
194 	}
195 
196 	/*
197 	 * allocate verbs request object
198 	 */
199 	vq_req = vq_req_alloc(c2dev);
200 	if (!vq_req)
201 		return -ENOMEM;
202 
203 	wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL);
204 	if (!wr) {
205 		err = -ENOMEM;
206 		goto bail0;
207 	}
208 
209 	/*
210 	 * build the WR
211 	 */
212 	c2_wr_set_id(wr, CCWR_NSMR_REGISTER);
213 	wr->hdr.context = (unsigned long) vq_req;
214 	wr->rnic_handle = c2dev->adapter_handle;
215 
216 	flags = (acf | MEM_VA_BASED | MEM_REMOTE);
217 
218 	/*
219 	 * compute how many pbes can fit in the message
220 	 */
221 	pbe_count = (c2dev->req_vq.msg_size -
222 		     sizeof(struct c2wr_nsmr_register_req)) / sizeof(u64);
223 
224 	if (pbl_depth <= pbe_count) {
225 		flags |= MEM_PBL_COMPLETE;
226 	}
227 	wr->flags = cpu_to_be16(flags);
228 	wr->stag_key = 0;	//stag_key;
229 	wr->va = cpu_to_be64(*va);
230 	wr->pd_id = mr->pd->pd_id;
231 	wr->pbe_size = cpu_to_be32(page_size);
232 	wr->length = cpu_to_be32(length);
233 	wr->pbl_depth = cpu_to_be32(pbl_depth);
234 	wr->fbo = cpu_to_be32(offset);
235 	count = min(pbl_depth, pbe_count);
236 	wr->addrs_length = cpu_to_be32(count);
237 
238 	/*
239 	 * fill out the PBL for this message
240 	 */
241 	for (i = 0; i < count; i++) {
242 		wr->paddrs[i] = cpu_to_be64(addr_list[i]);
243 	}
244 
245 	/*
246 	 * regerence the request struct
247 	 */
248 	vq_req_get(c2dev, vq_req);
249 
250 	/*
251 	 * send the WR to the adapter
252 	 */
253 	err = vq_send_wr(c2dev, (union c2wr *) wr);
254 	if (err) {
255 		vq_req_put(c2dev, vq_req);
256 		goto bail1;
257 	}
258 
259 	/*
260 	 * wait for reply from adapter
261 	 */
262 	err = vq_wait_for_reply(c2dev, vq_req);
263 	if (err) {
264 		goto bail1;
265 	}
266 
267 	/*
268 	 * process reply
269 	 */
270 	reply =
271 	    (struct c2wr_nsmr_register_rep *) (unsigned long) (vq_req->reply_msg);
272 	if (!reply) {
273 		err = -ENOMEM;
274 		goto bail1;
275 	}
276 	if ((err = c2_errno(reply))) {
277 		goto bail2;
278 	}
279 	//*p_pb_entries = be32_to_cpu(reply->pbl_depth);
280 	mr->ibmr.lkey = mr->ibmr.rkey = be32_to_cpu(reply->stag_index);
281 	vq_repbuf_free(c2dev, reply);
282 
283 	/*
284 	 * if there are still more PBEs we need to send them to
285 	 * the adapter and wait for a reply on the final one.
286 	 * reuse vq_req for this purpose.
287 	 */
288 	pbl_depth -= count;
289 	if (pbl_depth) {
290 
291 		vq_req->reply_msg = (unsigned long) NULL;
292 		atomic_set(&vq_req->reply_ready, 0);
293 		err = send_pbl_messages(c2dev,
294 					cpu_to_be32(mr->ibmr.lkey),
295 					(unsigned long) &addr_list[i],
296 					pbl_depth, vq_req, PBL_PHYS);
297 		if (err) {
298 			goto bail1;
299 		}
300 	}
301 
302 	vq_req_free(c2dev, vq_req);
303 	kfree(wr);
304 
305 	return err;
306 
307       bail2:
308 	vq_repbuf_free(c2dev, reply);
309       bail1:
310 	kfree(wr);
311       bail0:
312 	vq_req_free(c2dev, vq_req);
313 	return err;
314 }
315 
c2_stag_dealloc(struct c2_dev * c2dev,u32 stag_index)316 int c2_stag_dealloc(struct c2_dev *c2dev, u32 stag_index)
317 {
318 	struct c2_vq_req *vq_req;	/* verbs request object */
319 	struct c2wr_stag_dealloc_req wr;	/* work request */
320 	struct c2wr_stag_dealloc_rep *reply;	/* WR reply  */
321 	int err;
322 
323 
324 	/*
325 	 * allocate verbs request object
326 	 */
327 	vq_req = vq_req_alloc(c2dev);
328 	if (!vq_req) {
329 		return -ENOMEM;
330 	}
331 
332 	/*
333 	 * Build the WR
334 	 */
335 	c2_wr_set_id(&wr, CCWR_STAG_DEALLOC);
336 	wr.hdr.context = (u64) (unsigned long) vq_req;
337 	wr.rnic_handle = c2dev->adapter_handle;
338 	wr.stag_index = cpu_to_be32(stag_index);
339 
340 	/*
341 	 * reference the request struct.  dereferenced in the int handler.
342 	 */
343 	vq_req_get(c2dev, vq_req);
344 
345 	/*
346 	 * Send WR to adapter
347 	 */
348 	err = vq_send_wr(c2dev, (union c2wr *) & wr);
349 	if (err) {
350 		vq_req_put(c2dev, vq_req);
351 		goto bail0;
352 	}
353 
354 	/*
355 	 * Wait for reply from adapter
356 	 */
357 	err = vq_wait_for_reply(c2dev, vq_req);
358 	if (err) {
359 		goto bail0;
360 	}
361 
362 	/*
363 	 * Process reply
364 	 */
365 	reply = (struct c2wr_stag_dealloc_rep *) (unsigned long) vq_req->reply_msg;
366 	if (!reply) {
367 		err = -ENOMEM;
368 		goto bail0;
369 	}
370 
371 	err = c2_errno(reply);
372 
373 	vq_repbuf_free(c2dev, reply);
374       bail0:
375 	vq_req_free(c2dev, vq_req);
376 	return err;
377 }
378