xref: /linux/drivers/infiniband/core/uverbs_ioctl.c (revision 17cfcb68af3bc7d5e8ae08779b1853310a2949f3)
1 /*
2  * Copyright (c) 2017, Mellanox Technologies inc.  All rights reserved.
3  *
4  * This software is available to you under a choice of one of two
5  * licenses.  You may choose to be licensed under the terms of the GNU
6  * General Public License (GPL) Version 2, available from the file
7  * COPYING in the main directory of this source tree, or the
8  * OpenIB.org BSD license below:
9  *
10  *     Redistribution and use in source and binary forms, with or
11  *     without modification, are permitted provided that the following
12  *     conditions are met:
13  *
14  *      - Redistributions of source code must retain the above
15  *        copyright notice, this list of conditions and the following
16  *        disclaimer.
17  *
18  *      - Redistributions in binary form must reproduce the above
19  *        copyright notice, this list of conditions and the following
20  *        disclaimer in the documentation and/or other materials
21  *        provided with the distribution.
22  *
23  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
24  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
25  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
26  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
27  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
28  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
29  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
30  * SOFTWARE.
31  */
32 
33 #include <rdma/rdma_user_ioctl.h>
34 #include <rdma/uverbs_ioctl.h>
35 #include "rdma_core.h"
36 #include "uverbs.h"
37 
38 struct bundle_alloc_head {
39 	struct bundle_alloc_head *next;
40 	u8 data[];
41 };
42 
43 struct bundle_priv {
44 	/* Must be first */
45 	struct bundle_alloc_head alloc_head;
46 	struct bundle_alloc_head *allocated_mem;
47 	size_t internal_avail;
48 	size_t internal_used;
49 
50 	struct radix_tree_root *radix;
51 	const struct uverbs_api_ioctl_method *method_elm;
52 	void __rcu **radix_slots;
53 	unsigned long radix_slots_len;
54 	u32 method_key;
55 
56 	struct ib_uverbs_attr __user *user_attrs;
57 	struct ib_uverbs_attr *uattrs;
58 
59 	DECLARE_BITMAP(uobj_finalize, UVERBS_API_ATTR_BKEY_LEN);
60 	DECLARE_BITMAP(spec_finalize, UVERBS_API_ATTR_BKEY_LEN);
61 
62 	/*
63 	 * Must be last. bundle ends in a flex array which overlaps
64 	 * internal_buffer.
65 	 */
66 	struct uverbs_attr_bundle bundle;
67 	u64 internal_buffer[32];
68 };
69 
70 /*
71  * Each method has an absolute minimum amount of memory it needs to allocate,
72  * precompute that amount and determine if the onstack memory can be used or
73  * if allocation is need.
74  */
75 void uapi_compute_bundle_size(struct uverbs_api_ioctl_method *method_elm,
76 			      unsigned int num_attrs)
77 {
78 	struct bundle_priv *pbundle;
79 	size_t bundle_size =
80 		offsetof(struct bundle_priv, internal_buffer) +
81 		sizeof(*pbundle->bundle.attrs) * method_elm->key_bitmap_len +
82 		sizeof(*pbundle->uattrs) * num_attrs;
83 
84 	method_elm->use_stack = bundle_size <= sizeof(*pbundle);
85 	method_elm->bundle_size =
86 		ALIGN(bundle_size + 256, sizeof(*pbundle->internal_buffer));
87 
88 	/* Do not want order-2 allocations for this. */
89 	WARN_ON_ONCE(method_elm->bundle_size > PAGE_SIZE);
90 }
91 
92 /**
93  * uverbs_alloc() - Quickly allocate memory for use with a bundle
94  * @bundle: The bundle
95  * @size: Number of bytes to allocate
96  * @flags: Allocator flags
97  *
98  * The bundle allocator is intended for allocations that are connected with
99  * processing the system call related to the bundle. The allocated memory is
100  * always freed once the system call completes, and cannot be freed any other
101  * way.
102  *
103  * This tries to use a small pool of pre-allocated memory for performance.
104  */
105 __malloc void *_uverbs_alloc(struct uverbs_attr_bundle *bundle, size_t size,
106 			     gfp_t flags)
107 {
108 	struct bundle_priv *pbundle =
109 		container_of(bundle, struct bundle_priv, bundle);
110 	size_t new_used;
111 	void *res;
112 
113 	if (check_add_overflow(size, pbundle->internal_used, &new_used))
114 		return ERR_PTR(-EOVERFLOW);
115 
116 	if (new_used > pbundle->internal_avail) {
117 		struct bundle_alloc_head *buf;
118 
119 		buf = kvmalloc(struct_size(buf, data, size), flags);
120 		if (!buf)
121 			return ERR_PTR(-ENOMEM);
122 		buf->next = pbundle->allocated_mem;
123 		pbundle->allocated_mem = buf;
124 		return buf->data;
125 	}
126 
127 	res = (void *)pbundle->internal_buffer + pbundle->internal_used;
128 	pbundle->internal_used =
129 		ALIGN(new_used, sizeof(*pbundle->internal_buffer));
130 	if (want_init_on_alloc(flags))
131 		memset(res, 0, size);
132 	return res;
133 }
134 EXPORT_SYMBOL(_uverbs_alloc);
135 
136 static bool uverbs_is_attr_cleared(const struct ib_uverbs_attr *uattr,
137 				   u16 len)
138 {
139 	if (uattr->len > sizeof(((struct ib_uverbs_attr *)0)->data))
140 		return ib_is_buffer_cleared(u64_to_user_ptr(uattr->data) + len,
141 					    uattr->len - len);
142 
143 	return !memchr_inv((const void *)&uattr->data + len,
144 			   0, uattr->len - len);
145 }
146 
147 static int uverbs_set_output(const struct uverbs_attr_bundle *bundle,
148 			     const struct uverbs_attr *attr)
149 {
150 	struct bundle_priv *pbundle =
151 		container_of(bundle, struct bundle_priv, bundle);
152 	u16 flags;
153 
154 	flags = pbundle->uattrs[attr->ptr_attr.uattr_idx].flags |
155 		UVERBS_ATTR_F_VALID_OUTPUT;
156 	if (put_user(flags,
157 		     &pbundle->user_attrs[attr->ptr_attr.uattr_idx].flags))
158 		return -EFAULT;
159 	return 0;
160 }
161 
162 static int uverbs_process_idrs_array(struct bundle_priv *pbundle,
163 				     const struct uverbs_api_attr *attr_uapi,
164 				     struct uverbs_objs_arr_attr *attr,
165 				     struct ib_uverbs_attr *uattr,
166 				     u32 attr_bkey)
167 {
168 	const struct uverbs_attr_spec *spec = &attr_uapi->spec;
169 	size_t array_len;
170 	u32 *idr_vals;
171 	int ret = 0;
172 	size_t i;
173 
174 	if (uattr->attr_data.reserved)
175 		return -EINVAL;
176 
177 	if (uattr->len % sizeof(u32))
178 		return -EINVAL;
179 
180 	array_len = uattr->len / sizeof(u32);
181 	if (array_len < spec->u2.objs_arr.min_len ||
182 	    array_len > spec->u2.objs_arr.max_len)
183 		return -EINVAL;
184 
185 	attr->uobjects =
186 		uverbs_alloc(&pbundle->bundle,
187 			     array_size(array_len, sizeof(*attr->uobjects)));
188 	if (IS_ERR(attr->uobjects))
189 		return PTR_ERR(attr->uobjects);
190 
191 	/*
192 	 * Since idr is 4B and *uobjects is >= 4B, we can use attr->uobjects
193 	 * to store idrs array and avoid additional memory allocation. The
194 	 * idrs array is offset to the end of the uobjects array so we will be
195 	 * able to read idr and replace with a pointer.
196 	 */
197 	idr_vals = (u32 *)(attr->uobjects + array_len) - array_len;
198 
199 	if (uattr->len > sizeof(uattr->data)) {
200 		ret = copy_from_user(idr_vals, u64_to_user_ptr(uattr->data),
201 				     uattr->len);
202 		if (ret)
203 			return -EFAULT;
204 	} else {
205 		memcpy(idr_vals, &uattr->data, uattr->len);
206 	}
207 
208 	for (i = 0; i != array_len; i++) {
209 		attr->uobjects[i] = uverbs_get_uobject_from_file(
210 			spec->u2.objs_arr.obj_type, spec->u2.objs_arr.access,
211 			idr_vals[i], &pbundle->bundle);
212 		if (IS_ERR(attr->uobjects[i])) {
213 			ret = PTR_ERR(attr->uobjects[i]);
214 			break;
215 		}
216 	}
217 
218 	attr->len = i;
219 	__set_bit(attr_bkey, pbundle->spec_finalize);
220 	return ret;
221 }
222 
223 static int uverbs_free_idrs_array(const struct uverbs_api_attr *attr_uapi,
224 				  struct uverbs_objs_arr_attr *attr,
225 				  bool commit, struct uverbs_attr_bundle *attrs)
226 {
227 	const struct uverbs_attr_spec *spec = &attr_uapi->spec;
228 	int current_ret;
229 	int ret = 0;
230 	size_t i;
231 
232 	for (i = 0; i != attr->len; i++) {
233 		current_ret = uverbs_finalize_object(attr->uobjects[i],
234 						     spec->u2.objs_arr.access,
235 						     commit, attrs);
236 		if (!ret)
237 			ret = current_ret;
238 	}
239 
240 	return ret;
241 }
242 
243 static int uverbs_process_attr(struct bundle_priv *pbundle,
244 			       const struct uverbs_api_attr *attr_uapi,
245 			       struct ib_uverbs_attr *uattr, u32 attr_bkey)
246 {
247 	const struct uverbs_attr_spec *spec = &attr_uapi->spec;
248 	struct uverbs_attr *e = &pbundle->bundle.attrs[attr_bkey];
249 	const struct uverbs_attr_spec *val_spec = spec;
250 	struct uverbs_obj_attr *o_attr;
251 
252 	switch (spec->type) {
253 	case UVERBS_ATTR_TYPE_ENUM_IN:
254 		if (uattr->attr_data.enum_data.elem_id >= spec->u.enum_def.num_elems)
255 			return -EOPNOTSUPP;
256 
257 		if (uattr->attr_data.enum_data.reserved)
258 			return -EINVAL;
259 
260 		val_spec = &spec->u2.enum_def.ids[uattr->attr_data.enum_data.elem_id];
261 
262 		/* Currently we only support PTR_IN based enums */
263 		if (val_spec->type != UVERBS_ATTR_TYPE_PTR_IN)
264 			return -EOPNOTSUPP;
265 
266 		e->ptr_attr.enum_id = uattr->attr_data.enum_data.elem_id;
267 	/* fall through */
268 	case UVERBS_ATTR_TYPE_PTR_IN:
269 		/* Ensure that any data provided by userspace beyond the known
270 		 * struct is zero. Userspace that knows how to use some future
271 		 * longer struct will fail here if used with an old kernel and
272 		 * non-zero content, making ABI compat/discovery simpler.
273 		 */
274 		if (uattr->len > val_spec->u.ptr.len &&
275 		    val_spec->zero_trailing &&
276 		    !uverbs_is_attr_cleared(uattr, val_spec->u.ptr.len))
277 			return -EOPNOTSUPP;
278 
279 	/* fall through */
280 	case UVERBS_ATTR_TYPE_PTR_OUT:
281 		if (uattr->len < val_spec->u.ptr.min_len ||
282 		    (!val_spec->zero_trailing &&
283 		     uattr->len > val_spec->u.ptr.len))
284 			return -EINVAL;
285 
286 		if (spec->type != UVERBS_ATTR_TYPE_ENUM_IN &&
287 		    uattr->attr_data.reserved)
288 			return -EINVAL;
289 
290 		e->ptr_attr.uattr_idx = uattr - pbundle->uattrs;
291 		e->ptr_attr.len = uattr->len;
292 
293 		if (val_spec->alloc_and_copy && !uverbs_attr_ptr_is_inline(e)) {
294 			void *p;
295 
296 			p = uverbs_alloc(&pbundle->bundle, uattr->len);
297 			if (IS_ERR(p))
298 				return PTR_ERR(p);
299 
300 			e->ptr_attr.ptr = p;
301 
302 			if (copy_from_user(p, u64_to_user_ptr(uattr->data),
303 					   uattr->len))
304 				return -EFAULT;
305 		} else {
306 			e->ptr_attr.data = uattr->data;
307 		}
308 		break;
309 
310 	case UVERBS_ATTR_TYPE_IDR:
311 	case UVERBS_ATTR_TYPE_FD:
312 		if (uattr->attr_data.reserved)
313 			return -EINVAL;
314 
315 		if (uattr->len != 0)
316 			return -EINVAL;
317 
318 		o_attr = &e->obj_attr;
319 		o_attr->attr_elm = attr_uapi;
320 
321 		/*
322 		 * The type of uattr->data is u64 for UVERBS_ATTR_TYPE_IDR and
323 		 * s64 for UVERBS_ATTR_TYPE_FD. We can cast the u64 to s64
324 		 * here without caring about truncation as we know that the
325 		 * IDR implementation today rejects negative IDs
326 		 */
327 		o_attr->uobject = uverbs_get_uobject_from_file(
328 			spec->u.obj.obj_type, spec->u.obj.access,
329 			uattr->data_s64, &pbundle->bundle);
330 		if (IS_ERR(o_attr->uobject))
331 			return PTR_ERR(o_attr->uobject);
332 		__set_bit(attr_bkey, pbundle->uobj_finalize);
333 
334 		if (spec->u.obj.access == UVERBS_ACCESS_NEW) {
335 			unsigned int uattr_idx = uattr - pbundle->uattrs;
336 			s64 id = o_attr->uobject->id;
337 
338 			/* Copy the allocated id to the user-space */
339 			if (put_user(id, &pbundle->user_attrs[uattr_idx].data))
340 				return -EFAULT;
341 		}
342 
343 		break;
344 
345 	case UVERBS_ATTR_TYPE_IDRS_ARRAY:
346 		return uverbs_process_idrs_array(pbundle, attr_uapi,
347 						 &e->objs_arr_attr, uattr,
348 						 attr_bkey);
349 	default:
350 		return -EOPNOTSUPP;
351 	}
352 
353 	return 0;
354 }
355 
356 /*
357  * We search the radix tree with the method prefix and now we want to fast
358  * search the suffix bits to get a particular attribute pointer. It is not
359  * totally clear to me if this breaks the radix tree encasulation or not, but
360  * it uses the iter data to determine if the method iter points at the same
361  * chunk that will store the attribute, if so it just derefs it directly. By
362  * construction in most kernel configs the method and attrs will all fit in a
363  * single radix chunk, so in most cases this will have no search. Other cases
364  * this falls back to a full search.
365  */
366 static void __rcu **uapi_get_attr_for_method(struct bundle_priv *pbundle,
367 					     u32 attr_key)
368 {
369 	void __rcu **slot;
370 
371 	if (likely(attr_key < pbundle->radix_slots_len)) {
372 		void *entry;
373 
374 		slot = pbundle->radix_slots + attr_key;
375 		entry = rcu_dereference_raw(*slot);
376 		if (likely(!radix_tree_is_internal_node(entry) && entry))
377 			return slot;
378 	}
379 
380 	return radix_tree_lookup_slot(pbundle->radix,
381 				      pbundle->method_key | attr_key);
382 }
383 
384 static int uverbs_set_attr(struct bundle_priv *pbundle,
385 			   struct ib_uverbs_attr *uattr)
386 {
387 	u32 attr_key = uapi_key_attr(uattr->attr_id);
388 	u32 attr_bkey = uapi_bkey_attr(attr_key);
389 	const struct uverbs_api_attr *attr;
390 	void __rcu **slot;
391 	int ret;
392 
393 	slot = uapi_get_attr_for_method(pbundle, attr_key);
394 	if (!slot) {
395 		/*
396 		 * Kernel does not support the attribute but user-space says it
397 		 * is mandatory
398 		 */
399 		if (uattr->flags & UVERBS_ATTR_F_MANDATORY)
400 			return -EPROTONOSUPPORT;
401 		return 0;
402 	}
403 	attr = rcu_dereference_protected(*slot, true);
404 
405 	/* Reject duplicate attributes from user-space */
406 	if (test_bit(attr_bkey, pbundle->bundle.attr_present))
407 		return -EINVAL;
408 
409 	ret = uverbs_process_attr(pbundle, attr, uattr, attr_bkey);
410 	if (ret)
411 		return ret;
412 
413 	__set_bit(attr_bkey, pbundle->bundle.attr_present);
414 
415 	return 0;
416 }
417 
418 static int ib_uverbs_run_method(struct bundle_priv *pbundle,
419 				unsigned int num_attrs)
420 {
421 	int (*handler)(struct uverbs_attr_bundle *attrs);
422 	size_t uattrs_size = array_size(sizeof(*pbundle->uattrs), num_attrs);
423 	unsigned int destroy_bkey = pbundle->method_elm->destroy_bkey;
424 	unsigned int i;
425 	int ret;
426 
427 	/* See uverbs_disassociate_api() */
428 	handler = srcu_dereference(
429 		pbundle->method_elm->handler,
430 		&pbundle->bundle.ufile->device->disassociate_srcu);
431 	if (!handler)
432 		return -EIO;
433 
434 	pbundle->uattrs = uverbs_alloc(&pbundle->bundle, uattrs_size);
435 	if (IS_ERR(pbundle->uattrs))
436 		return PTR_ERR(pbundle->uattrs);
437 	if (copy_from_user(pbundle->uattrs, pbundle->user_attrs, uattrs_size))
438 		return -EFAULT;
439 
440 	for (i = 0; i != num_attrs; i++) {
441 		ret = uverbs_set_attr(pbundle, &pbundle->uattrs[i]);
442 		if (unlikely(ret))
443 			return ret;
444 	}
445 
446 	/* User space did not provide all the mandatory attributes */
447 	if (unlikely(!bitmap_subset(pbundle->method_elm->attr_mandatory,
448 				    pbundle->bundle.attr_present,
449 				    pbundle->method_elm->key_bitmap_len)))
450 		return -EINVAL;
451 
452 	if (pbundle->method_elm->has_udata)
453 		uverbs_fill_udata(&pbundle->bundle,
454 				  &pbundle->bundle.driver_udata,
455 				  UVERBS_ATTR_UHW_IN, UVERBS_ATTR_UHW_OUT);
456 	else
457 		pbundle->bundle.driver_udata = (struct ib_udata){};
458 
459 	if (destroy_bkey != UVERBS_API_ATTR_BKEY_LEN) {
460 		struct uverbs_obj_attr *destroy_attr =
461 			&pbundle->bundle.attrs[destroy_bkey].obj_attr;
462 
463 		ret = uobj_destroy(destroy_attr->uobject, &pbundle->bundle);
464 		if (ret)
465 			return ret;
466 		__clear_bit(destroy_bkey, pbundle->uobj_finalize);
467 
468 		ret = handler(&pbundle->bundle);
469 		uobj_put_destroy(destroy_attr->uobject);
470 	} else {
471 		ret = handler(&pbundle->bundle);
472 	}
473 
474 	/*
475 	 * Until the drivers are revised to use the bundle directly we have to
476 	 * assume that the driver wrote to its UHW_OUT and flag userspace
477 	 * appropriately.
478 	 */
479 	if (!ret && pbundle->method_elm->has_udata) {
480 		const struct uverbs_attr *attr =
481 			uverbs_attr_get(&pbundle->bundle, UVERBS_ATTR_UHW_OUT);
482 
483 		if (!IS_ERR(attr))
484 			ret = uverbs_set_output(&pbundle->bundle, attr);
485 	}
486 
487 	/*
488 	 * EPROTONOSUPPORT is ONLY to be returned if the ioctl framework can
489 	 * not invoke the method because the request is not supported.  No
490 	 * other cases should return this code.
491 	 */
492 	if (WARN_ON_ONCE(ret == -EPROTONOSUPPORT))
493 		return -EINVAL;
494 
495 	return ret;
496 }
497 
498 static int bundle_destroy(struct bundle_priv *pbundle, bool commit)
499 {
500 	unsigned int key_bitmap_len = pbundle->method_elm->key_bitmap_len;
501 	struct bundle_alloc_head *memblock;
502 	unsigned int i;
503 	int ret = 0;
504 
505 	/* fast path for simple uobjects */
506 	i = -1;
507 	while ((i = find_next_bit(pbundle->uobj_finalize, key_bitmap_len,
508 				  i + 1)) < key_bitmap_len) {
509 		struct uverbs_attr *attr = &pbundle->bundle.attrs[i];
510 		int current_ret;
511 
512 		current_ret = uverbs_finalize_object(
513 			attr->obj_attr.uobject,
514 			attr->obj_attr.attr_elm->spec.u.obj.access, commit,
515 			&pbundle->bundle);
516 		if (!ret)
517 			ret = current_ret;
518 	}
519 
520 	i = -1;
521 	while ((i = find_next_bit(pbundle->spec_finalize, key_bitmap_len,
522 				  i + 1)) < key_bitmap_len) {
523 		struct uverbs_attr *attr = &pbundle->bundle.attrs[i];
524 		const struct uverbs_api_attr *attr_uapi;
525 		void __rcu **slot;
526 		int current_ret;
527 
528 		slot = uapi_get_attr_for_method(
529 			pbundle,
530 			pbundle->method_key | uapi_bkey_to_key_attr(i));
531 		if (WARN_ON(!slot))
532 			continue;
533 
534 		attr_uapi = rcu_dereference_protected(*slot, true);
535 
536 		if (attr_uapi->spec.type == UVERBS_ATTR_TYPE_IDRS_ARRAY) {
537 			current_ret = uverbs_free_idrs_array(
538 				attr_uapi, &attr->objs_arr_attr, commit,
539 				&pbundle->bundle);
540 			if (!ret)
541 				ret = current_ret;
542 		}
543 	}
544 
545 	for (memblock = pbundle->allocated_mem; memblock;) {
546 		struct bundle_alloc_head *tmp = memblock;
547 
548 		memblock = memblock->next;
549 		kvfree(tmp);
550 	}
551 
552 	return ret;
553 }
554 
555 static int ib_uverbs_cmd_verbs(struct ib_uverbs_file *ufile,
556 			       struct ib_uverbs_ioctl_hdr *hdr,
557 			       struct ib_uverbs_attr __user *user_attrs)
558 {
559 	const struct uverbs_api_ioctl_method *method_elm;
560 	struct uverbs_api *uapi = ufile->device->uapi;
561 	struct radix_tree_iter attrs_iter;
562 	struct bundle_priv *pbundle;
563 	struct bundle_priv onstack;
564 	void __rcu **slot;
565 	int destroy_ret;
566 	int ret;
567 
568 	if (unlikely(hdr->driver_id != uapi->driver_id))
569 		return -EINVAL;
570 
571 	slot = radix_tree_iter_lookup(
572 		&uapi->radix, &attrs_iter,
573 		uapi_key_obj(hdr->object_id) |
574 			uapi_key_ioctl_method(hdr->method_id));
575 	if (unlikely(!slot))
576 		return -EPROTONOSUPPORT;
577 	method_elm = rcu_dereference_protected(*slot, true);
578 
579 	if (!method_elm->use_stack) {
580 		pbundle = kmalloc(method_elm->bundle_size, GFP_KERNEL);
581 		if (!pbundle)
582 			return -ENOMEM;
583 		pbundle->internal_avail =
584 			method_elm->bundle_size -
585 			offsetof(struct bundle_priv, internal_buffer);
586 		pbundle->alloc_head.next = NULL;
587 		pbundle->allocated_mem = &pbundle->alloc_head;
588 	} else {
589 		pbundle = &onstack;
590 		pbundle->internal_avail = sizeof(pbundle->internal_buffer);
591 		pbundle->allocated_mem = NULL;
592 	}
593 
594 	/* Space for the pbundle->bundle.attrs flex array */
595 	pbundle->method_elm = method_elm;
596 	pbundle->method_key = attrs_iter.index;
597 	pbundle->bundle.ufile = ufile;
598 	pbundle->bundle.context = NULL; /* only valid if bundle has uobject */
599 	pbundle->radix = &uapi->radix;
600 	pbundle->radix_slots = slot;
601 	pbundle->radix_slots_len = radix_tree_chunk_size(&attrs_iter);
602 	pbundle->user_attrs = user_attrs;
603 
604 	pbundle->internal_used = ALIGN(pbundle->method_elm->key_bitmap_len *
605 					       sizeof(*pbundle->bundle.attrs),
606 				       sizeof(*pbundle->internal_buffer));
607 	memset(pbundle->bundle.attr_present, 0,
608 	       sizeof(pbundle->bundle.attr_present));
609 	memset(pbundle->uobj_finalize, 0, sizeof(pbundle->uobj_finalize));
610 	memset(pbundle->spec_finalize, 0, sizeof(pbundle->spec_finalize));
611 
612 	ret = ib_uverbs_run_method(pbundle, hdr->num_attrs);
613 	destroy_ret = bundle_destroy(pbundle, ret == 0);
614 	if (unlikely(destroy_ret && !ret))
615 		return destroy_ret;
616 
617 	return ret;
618 }
619 
620 long ib_uverbs_ioctl(struct file *filp, unsigned int cmd, unsigned long arg)
621 {
622 	struct ib_uverbs_file *file = filp->private_data;
623 	struct ib_uverbs_ioctl_hdr __user *user_hdr =
624 		(struct ib_uverbs_ioctl_hdr __user *)arg;
625 	struct ib_uverbs_ioctl_hdr hdr;
626 	int srcu_key;
627 	int err;
628 
629 	if (unlikely(cmd != RDMA_VERBS_IOCTL))
630 		return -ENOIOCTLCMD;
631 
632 	err = copy_from_user(&hdr, user_hdr, sizeof(hdr));
633 	if (err)
634 		return -EFAULT;
635 
636 	if (hdr.length > PAGE_SIZE ||
637 	    hdr.length != struct_size(&hdr, attrs, hdr.num_attrs))
638 		return -EINVAL;
639 
640 	if (hdr.reserved1 || hdr.reserved2)
641 		return -EPROTONOSUPPORT;
642 
643 	srcu_key = srcu_read_lock(&file->device->disassociate_srcu);
644 	err = ib_uverbs_cmd_verbs(file, &hdr, user_hdr->attrs);
645 	srcu_read_unlock(&file->device->disassociate_srcu, srcu_key);
646 	return err;
647 }
648 
649 int uverbs_get_flags64(u64 *to, const struct uverbs_attr_bundle *attrs_bundle,
650 		       size_t idx, u64 allowed_bits)
651 {
652 	const struct uverbs_attr *attr;
653 	u64 flags;
654 
655 	attr = uverbs_attr_get(attrs_bundle, idx);
656 	/* Missing attribute means 0 flags */
657 	if (IS_ERR(attr)) {
658 		*to = 0;
659 		return 0;
660 	}
661 
662 	/*
663 	 * New userspace code should use 8 bytes to pass flags, but we
664 	 * transparently support old userspaces that were using 4 bytes as
665 	 * well.
666 	 */
667 	if (attr->ptr_attr.len == 8)
668 		flags = attr->ptr_attr.data;
669 	else if (attr->ptr_attr.len == 4)
670 		flags = *(u32 *)&attr->ptr_attr.data;
671 	else
672 		return -EINVAL;
673 
674 	if (flags & ~allowed_bits)
675 		return -EINVAL;
676 
677 	*to = flags;
678 	return 0;
679 }
680 EXPORT_SYMBOL(uverbs_get_flags64);
681 
682 int uverbs_get_flags32(u32 *to, const struct uverbs_attr_bundle *attrs_bundle,
683 		       size_t idx, u64 allowed_bits)
684 {
685 	u64 flags;
686 	int ret;
687 
688 	ret = uverbs_get_flags64(&flags, attrs_bundle, idx, allowed_bits);
689 	if (ret)
690 		return ret;
691 
692 	if (flags > U32_MAX)
693 		return -EINVAL;
694 	*to = flags;
695 
696 	return 0;
697 }
698 EXPORT_SYMBOL(uverbs_get_flags32);
699 
700 /*
701  * Fill a ib_udata struct (core or uhw) using the given attribute IDs.
702  * This is primarily used to convert the UVERBS_ATTR_UHW() into the
703  * ib_udata format used by the drivers.
704  */
705 void uverbs_fill_udata(struct uverbs_attr_bundle *bundle,
706 		       struct ib_udata *udata, unsigned int attr_in,
707 		       unsigned int attr_out)
708 {
709 	struct bundle_priv *pbundle =
710 		container_of(bundle, struct bundle_priv, bundle);
711 	const struct uverbs_attr *in =
712 		uverbs_attr_get(&pbundle->bundle, attr_in);
713 	const struct uverbs_attr *out =
714 		uverbs_attr_get(&pbundle->bundle, attr_out);
715 
716 	if (!IS_ERR(in)) {
717 		udata->inlen = in->ptr_attr.len;
718 		if (uverbs_attr_ptr_is_inline(in))
719 			udata->inbuf =
720 				&pbundle->user_attrs[in->ptr_attr.uattr_idx]
721 					 .data;
722 		else
723 			udata->inbuf = u64_to_user_ptr(in->ptr_attr.data);
724 	} else {
725 		udata->inbuf = NULL;
726 		udata->inlen = 0;
727 	}
728 
729 	if (!IS_ERR(out)) {
730 		udata->outbuf = u64_to_user_ptr(out->ptr_attr.data);
731 		udata->outlen = out->ptr_attr.len;
732 	} else {
733 		udata->outbuf = NULL;
734 		udata->outlen = 0;
735 	}
736 }
737 
738 int uverbs_copy_to(const struct uverbs_attr_bundle *bundle, size_t idx,
739 		   const void *from, size_t size)
740 {
741 	const struct uverbs_attr *attr = uverbs_attr_get(bundle, idx);
742 	size_t min_size;
743 
744 	if (IS_ERR(attr))
745 		return PTR_ERR(attr);
746 
747 	min_size = min_t(size_t, attr->ptr_attr.len, size);
748 	if (copy_to_user(u64_to_user_ptr(attr->ptr_attr.data), from, min_size))
749 		return -EFAULT;
750 
751 	return uverbs_set_output(bundle, attr);
752 }
753 EXPORT_SYMBOL(uverbs_copy_to);
754 
755 
756 /*
757  * This is only used if the caller has directly used copy_to_use to write the
758  * data.  It signals to user space that the buffer is filled in.
759  */
760 int uverbs_output_written(const struct uverbs_attr_bundle *bundle, size_t idx)
761 {
762 	const struct uverbs_attr *attr = uverbs_attr_get(bundle, idx);
763 
764 	if (IS_ERR(attr))
765 		return PTR_ERR(attr);
766 
767 	return uverbs_set_output(bundle, attr);
768 }
769 
770 int _uverbs_get_const(s64 *to, const struct uverbs_attr_bundle *attrs_bundle,
771 		      size_t idx, s64 lower_bound, u64 upper_bound,
772 		      s64  *def_val)
773 {
774 	const struct uverbs_attr *attr;
775 
776 	attr = uverbs_attr_get(attrs_bundle, idx);
777 	if (IS_ERR(attr)) {
778 		if ((PTR_ERR(attr) != -ENOENT) || !def_val)
779 			return PTR_ERR(attr);
780 
781 		*to = *def_val;
782 	} else {
783 		*to = attr->ptr_attr.data;
784 	}
785 
786 	if (*to < lower_bound || (*to > 0 && (u64)*to > upper_bound))
787 		return -EINVAL;
788 
789 	return 0;
790 }
791 EXPORT_SYMBOL(_uverbs_get_const);
792 
793 int uverbs_copy_to_struct_or_zero(const struct uverbs_attr_bundle *bundle,
794 				  size_t idx, const void *from, size_t size)
795 {
796 	const struct uverbs_attr *attr = uverbs_attr_get(bundle, idx);
797 
798 	if (size < attr->ptr_attr.len) {
799 		if (clear_user(u64_to_user_ptr(attr->ptr_attr.data) + size,
800 			       attr->ptr_attr.len - size))
801 			return -EFAULT;
802 	}
803 	return uverbs_copy_to(bundle, idx, from, size);
804 }
805