1 /*
2 * Copyright (c) 2004, 2005 Topspin Communications. All rights reserved.
3 * Copyright (c) 2006, 2007 Cisco Systems, Inc. All rights reserved.
4 *
5 * This software is available to you under a choice of one of two
6 * licenses. You may choose to be licensed under the terms of the GNU
7 * General Public License (GPL) Version 2, available from the file
8 * COPYING in the main directory of this source tree, or the
9 * OpenIB.org BSD license below:
10 *
11 * Redistribution and use in source and binary forms, with or
12 * without modification, are permitted provided that the following
13 * conditions are met:
14 *
15 * - Redistributions of source code must retain the above
16 * copyright notice, this list of conditions and the following
17 * disclaimer.
18 *
19 * - Redistributions in binary form must reproduce the above
20 * copyright notice, this list of conditions and the following
21 * disclaimer in the documentation and/or other materials
22 * provided with the distribution.
23 *
24 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
25 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
26 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
27 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
28 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
29 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
30 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
31 * SOFTWARE.
32 */
33 #define _GNU_SOURCE
34 #include <config.h>
35
36 #include <infiniband/endian.h>
37 #include <stdio.h>
38 #include <sys/types.h>
39 #include <sys/stat.h>
40 #include <fcntl.h>
41 #include <unistd.h>
42 #include <stdlib.h>
43 #include <alloca.h>
44 #include <errno.h>
45
46 #include "ibverbs.h"
47
48 /* Hack to avoid GCC's -Wmissing-prototypes and the similar error from sparse
49 with these prototypes. Symbol versionining requires the goofy names, the
50 prototype must match the version in verbs.h.
51 */
52 struct ibv_device **__ibv_get_device_list(int *num_devices);
53 void __ibv_free_device_list(struct ibv_device **list);
54 const char *__ibv_get_device_name(struct ibv_device *device);
55 __be64 __ibv_get_device_guid(struct ibv_device *device);
56 struct ibv_context *__ibv_open_device(struct ibv_device *device);
57 int __ibv_close_device(struct ibv_context *context);
58 int __ibv_get_async_event(struct ibv_context *context,
59 struct ibv_async_event *event);
60 void __ibv_ack_async_event(struct ibv_async_event *event);
61
62 static pthread_once_t device_list_once = PTHREAD_ONCE_INIT;
63 static int num_devices;
64 static struct ibv_device **device_list;
65
count_devices(void)66 static void count_devices(void)
67 {
68 num_devices = ibverbs_init(&device_list);
69 }
70
__ibv_get_device_list(int * num)71 struct ibv_device **__ibv_get_device_list(int *num)
72 {
73 struct ibv_device **l;
74 int i;
75
76 if (num)
77 *num = 0;
78
79 pthread_once(&device_list_once, count_devices);
80
81 if (num_devices < 0) {
82 errno = -num_devices;
83 return NULL;
84 }
85
86 l = calloc(num_devices + 1, sizeof (struct ibv_device *));
87 if (!l) {
88 errno = ENOMEM;
89 return NULL;
90 }
91
92 for (i = 0; i < num_devices; ++i)
93 l[i] = device_list[i];
94 if (num)
95 *num = num_devices;
96
97 return l;
98 }
99 default_symver(__ibv_get_device_list, ibv_get_device_list);
100
__ibv_free_device_list(struct ibv_device ** list)101 void __ibv_free_device_list(struct ibv_device **list)
102 {
103 free(list);
104 }
105 default_symver(__ibv_free_device_list, ibv_free_device_list);
106
__ibv_get_device_name(struct ibv_device * device)107 const char *__ibv_get_device_name(struct ibv_device *device)
108 {
109 return device->name;
110 }
111 default_symver(__ibv_get_device_name, ibv_get_device_name);
112
__ibv_get_device_guid(struct ibv_device * device)113 __be64 __ibv_get_device_guid(struct ibv_device *device)
114 {
115 char attr[24];
116 uint64_t guid = 0;
117 uint16_t parts[4];
118 int i;
119
120 if (ibv_read_sysfs_file(device->ibdev_path, "node_guid",
121 attr, sizeof attr) < 0)
122 return 0;
123
124 if (sscanf(attr, "%hx:%hx:%hx:%hx",
125 parts, parts + 1, parts + 2, parts + 3) != 4)
126 return 0;
127
128 for (i = 0; i < 4; ++i)
129 guid = (guid << 16) | parts[i];
130
131 return htobe64(guid);
132 }
133 default_symver(__ibv_get_device_guid, ibv_get_device_guid);
134
verbs_init_cq(struct ibv_cq * cq,struct ibv_context * context,struct ibv_comp_channel * channel,void * cq_context)135 int verbs_init_cq(struct ibv_cq *cq, struct ibv_context *context,
136 struct ibv_comp_channel *channel,
137 void *cq_context)
138 {
139 int err = 0;
140
141 cq->context = context;
142 cq->channel = channel;
143
144 err = pthread_mutex_init(&cq->mutex, NULL);
145 if (err)
146 return err;
147 err = pthread_cond_init(&cq->cond, NULL);
148 if (err)
149 goto err;
150
151 if (cq->channel) {
152 pthread_mutex_lock(&context->mutex);
153 ++cq->channel->refcnt;
154 pthread_mutex_unlock(&context->mutex);
155 }
156
157 cq->cq_context = cq_context;
158 cq->comp_events_completed = 0;
159 cq->async_events_completed = 0;
160
161 return err;
162
163 err:
164 pthread_mutex_destroy(&cq->mutex);
165
166 return err;
167 }
168
verbs_cleanup_cq(struct ibv_cq * cq)169 void verbs_cleanup_cq(struct ibv_cq *cq)
170 {
171 pthread_cond_destroy(&cq->cond);
172 pthread_mutex_destroy(&cq->mutex);
173 }
174
175 static struct ibv_cq_ex *
__lib_ibv_create_cq_ex(struct ibv_context * context,struct ibv_cq_init_attr_ex * cq_attr)176 __lib_ibv_create_cq_ex(struct ibv_context *context,
177 struct ibv_cq_init_attr_ex *cq_attr)
178 {
179 struct verbs_context *vctx = verbs_get_ctx(context);
180 struct ibv_cq_ex *cq;
181 int err = 0;
182
183 if (cq_attr->wc_flags & ~IBV_CREATE_CQ_SUP_WC_FLAGS) {
184 errno = EOPNOTSUPP;
185 return NULL;
186 }
187
188 cq = vctx->priv->create_cq_ex(context, cq_attr);
189 if (!cq)
190 return NULL;
191
192 err = verbs_init_cq(ibv_cq_ex_to_cq(cq), context,
193 cq_attr->channel, cq_attr->cq_context);
194 if (err)
195 goto err;
196
197 return cq;
198
199 err:
200 context->ops.destroy_cq(ibv_cq_ex_to_cq(cq));
201
202 return NULL;
203 }
204
__ibv_open_device(struct ibv_device * device)205 struct ibv_context *__ibv_open_device(struct ibv_device *device)
206 {
207 struct verbs_device *verbs_device = verbs_get_device(device);
208 char *devpath;
209 int cmd_fd, ret;
210 struct ibv_context *context;
211 struct verbs_context *context_ex;
212
213 if (asprintf(&devpath, "/dev/%s", device->dev_name) < 0)
214 return NULL;
215
216 /*
217 * We'll only be doing writes, but we need O_RDWR in case the
218 * provider needs to mmap() the file.
219 */
220 cmd_fd = open(devpath, O_RDWR | O_CLOEXEC);
221 free(devpath);
222
223 if (cmd_fd < 0)
224 return NULL;
225
226 if (!verbs_device->ops->init_context) {
227 context = verbs_device->ops->alloc_context(device, cmd_fd);
228 if (!context)
229 goto err;
230
231 if (pthread_mutex_init(&context->mutex, NULL)) {
232 verbs_device->ops->free_context(context);
233 goto err;
234 }
235 } else {
236 struct verbs_ex_private *priv;
237
238 /* Library now allocates the context */
239 context_ex = calloc(1, sizeof(*context_ex) +
240 verbs_device->size_of_context);
241 if (!context_ex) {
242 errno = ENOMEM;
243 goto err;
244 }
245
246 priv = calloc(1, sizeof(*priv));
247 if (!priv) {
248 errno = ENOMEM;
249 goto err_context;
250 }
251
252 context_ex->priv = priv;
253 context_ex->context.abi_compat = __VERBS_ABI_IS_EXTENDED;
254 context_ex->sz = sizeof(*context_ex);
255
256 context = &context_ex->context;
257 if (pthread_mutex_init(&context->mutex, NULL))
258 goto verbs_err;
259
260 ret = verbs_device->ops->init_context(verbs_device, context, cmd_fd);
261 if (ret)
262 goto err_mutex;
263 /*
264 * In order to maintain backward/forward binary compatibility
265 * with apps compiled against libibverbs-1.1.8 that use the
266 * flow steering addition, we need to set the two
267 * ABI_placeholder entries to match the driver set flow
268 * entries. This is because apps compiled against
269 * libibverbs-1.1.8 use an inline ibv_create_flow and
270 * ibv_destroy_flow function that looks in the placeholder
271 * spots for the proper entry points. For apps compiled
272 * against libibverbs-1.1.9 and later, the inline functions
273 * will be looking in the right place.
274 */
275 context_ex->ABI_placeholder1 = (void (*)(void)) context_ex->ibv_create_flow;
276 context_ex->ABI_placeholder2 = (void (*)(void)) context_ex->ibv_destroy_flow;
277
278 if (context_ex->create_cq_ex) {
279 priv->create_cq_ex = context_ex->create_cq_ex;
280 context_ex->create_cq_ex = __lib_ibv_create_cq_ex;
281 }
282 }
283
284 context->device = device;
285 context->cmd_fd = cmd_fd;
286
287 return context;
288
289 err_mutex:
290 pthread_mutex_destroy(&context->mutex);
291 verbs_err:
292 free(context_ex->priv);
293 err_context:
294 free(context_ex);
295 err:
296 close(cmd_fd);
297 return NULL;
298 }
299 default_symver(__ibv_open_device, ibv_open_device);
300
__ibv_close_device(struct ibv_context * context)301 int __ibv_close_device(struct ibv_context *context)
302 {
303 int async_fd = context->async_fd;
304 int cmd_fd = context->cmd_fd;
305 struct verbs_context *context_ex;
306 struct verbs_device *verbs_device = verbs_get_device(context->device);
307
308 pthread_mutex_destroy(&context->mutex);
309 context_ex = verbs_get_ctx(context);
310 if (context_ex) {
311 verbs_device->ops->uninit_context(verbs_device, context);
312 free(context_ex->priv);
313 free(context_ex);
314 } else {
315 verbs_device->ops->free_context(context);
316 }
317
318 close(async_fd);
319 close(cmd_fd);
320
321 return 0;
322 }
323 default_symver(__ibv_close_device, ibv_close_device);
324
__ibv_get_async_event(struct ibv_context * context,struct ibv_async_event * event)325 int __ibv_get_async_event(struct ibv_context *context,
326 struct ibv_async_event *event)
327 {
328 struct ibv_kern_async_event ev;
329
330 if (read(context->async_fd, &ev, sizeof ev) != sizeof ev)
331 return -1;
332
333 event->event_type = ev.event_type;
334
335 switch (event->event_type) {
336 case IBV_EVENT_CQ_ERR:
337 event->element.cq = (void *) (uintptr_t) ev.element;
338 break;
339
340 case IBV_EVENT_QP_FATAL:
341 case IBV_EVENT_QP_REQ_ERR:
342 case IBV_EVENT_QP_ACCESS_ERR:
343 case IBV_EVENT_COMM_EST:
344 case IBV_EVENT_SQ_DRAINED:
345 case IBV_EVENT_PATH_MIG:
346 case IBV_EVENT_PATH_MIG_ERR:
347 case IBV_EVENT_QP_LAST_WQE_REACHED:
348 event->element.qp = (void *) (uintptr_t) ev.element;
349 break;
350
351 case IBV_EVENT_SRQ_ERR:
352 case IBV_EVENT_SRQ_LIMIT_REACHED:
353 event->element.srq = (void *) (uintptr_t) ev.element;
354 break;
355
356 case IBV_EVENT_WQ_FATAL:
357 event->element.wq = (void *) (uintptr_t) ev.element;
358 break;
359 default:
360 event->element.port_num = ev.element;
361 break;
362 }
363
364 if (context->ops.async_event)
365 context->ops.async_event(event);
366
367 return 0;
368 }
369 default_symver(__ibv_get_async_event, ibv_get_async_event);
370
__ibv_ack_async_event(struct ibv_async_event * event)371 void __ibv_ack_async_event(struct ibv_async_event *event)
372 {
373 switch (event->event_type) {
374 case IBV_EVENT_CQ_ERR:
375 {
376 struct ibv_cq *cq = event->element.cq;
377
378 pthread_mutex_lock(&cq->mutex);
379 ++cq->async_events_completed;
380 pthread_cond_signal(&cq->cond);
381 pthread_mutex_unlock(&cq->mutex);
382
383 return;
384 }
385
386 case IBV_EVENT_QP_FATAL:
387 case IBV_EVENT_QP_REQ_ERR:
388 case IBV_EVENT_QP_ACCESS_ERR:
389 case IBV_EVENT_COMM_EST:
390 case IBV_EVENT_SQ_DRAINED:
391 case IBV_EVENT_PATH_MIG:
392 case IBV_EVENT_PATH_MIG_ERR:
393 case IBV_EVENT_QP_LAST_WQE_REACHED:
394 {
395 struct ibv_qp *qp = event->element.qp;
396
397 pthread_mutex_lock(&qp->mutex);
398 ++qp->events_completed;
399 pthread_cond_signal(&qp->cond);
400 pthread_mutex_unlock(&qp->mutex);
401
402 return;
403 }
404
405 case IBV_EVENT_SRQ_ERR:
406 case IBV_EVENT_SRQ_LIMIT_REACHED:
407 {
408 struct ibv_srq *srq = event->element.srq;
409
410 pthread_mutex_lock(&srq->mutex);
411 ++srq->events_completed;
412 pthread_cond_signal(&srq->cond);
413 pthread_mutex_unlock(&srq->mutex);
414
415 return;
416 }
417
418 case IBV_EVENT_WQ_FATAL:
419 {
420 struct ibv_wq *wq = event->element.wq;
421
422 pthread_mutex_lock(&wq->mutex);
423 ++wq->events_completed;
424 pthread_cond_signal(&wq->cond);
425 pthread_mutex_unlock(&wq->mutex);
426
427 return;
428 }
429
430 default:
431 return;
432 }
433 }
434 default_symver(__ibv_ack_async_event, ibv_ack_async_event);
435
__ibv_init_wq(struct ibv_wq * wq)436 int __ibv_init_wq(struct ibv_wq *wq)
437 {
438 int err = 0;
439 wq->events_completed = 0;
440 err = pthread_mutex_init(&wq->mutex, NULL);
441 if (err)
442 return err;
443
444 err = pthread_cond_init(&wq->cond, NULL);
445 if (err)
446 goto err;
447
448 return err;
449
450 err:
451 pthread_mutex_destroy(&wq->mutex);
452
453 return err;
454 }
455 default_symver(__ibv_init_wq, ibv_init_wq);
456
__ibv_cleanup_wq(struct ibv_wq * wq)457 void __ibv_cleanup_wq(struct ibv_wq *wq)
458 {
459 pthread_cond_destroy(&wq->cond);
460 pthread_mutex_destroy(&wq->mutex);
461 }
462 default_symver(__ibv_cleanup_wq, ibv_cleanup_wq);
463