Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
1/*
2 * Copyright (c) 2017, Mellanox Technologies inc. All rights reserved.
3 *
4 * This software is available to you under a choice of one of two
5 * licenses. You may choose to be licensed under the terms of the GNU
6 * General Public License (GPL) Version 2, available from the file
7 * COPYING in the main directory of this source tree, or the
8 * OpenIB.org BSD license below:
9 *
10 * Redistribution and use in source and binary forms, with or
11 * without modification, are permitted provided that the following
12 * conditions are met:
13 *
14 * - Redistributions of source code must retain the above
15 * copyright notice, this list of conditions and the following
16 * disclaimer.
17 *
18 * - Redistributions in binary form must reproduce the above
19 * copyright notice, this list of conditions and the following
20 * disclaimer in the documentation and/or other materials
21 * provided with the distribution.
22 *
23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
24 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
25 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
26 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
27 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
28 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
29 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
30 * SOFTWARE.
31 */
32
33#include <rdma/rdma_user_ioctl.h>
34#include <rdma/uverbs_ioctl.h>
35#include "rdma_core.h"
36#include "uverbs.h"
37
38struct bundle_alloc_head {
39 struct bundle_alloc_head *next;
40 u8 data[];
41};
42
43struct bundle_priv {
44 /* Must be first */
45 struct bundle_alloc_head alloc_head;
46 struct bundle_alloc_head *allocated_mem;
47 size_t internal_avail;
48 size_t internal_used;
49
50 struct radix_tree_root *radix;
51 const struct uverbs_api_ioctl_method *method_elm;
52 void __rcu **radix_slots;
53 unsigned long radix_slots_len;
54 u32 method_key;
55
56 struct ib_uverbs_attr __user *user_attrs;
57 struct ib_uverbs_attr *uattrs;
58
59 DECLARE_BITMAP(uobj_finalize, UVERBS_API_ATTR_BKEY_LEN);
60 DECLARE_BITMAP(spec_finalize, UVERBS_API_ATTR_BKEY_LEN);
61
62 /*
63 * Must be last. bundle ends in a flex array which overlaps
64 * internal_buffer.
65 */
66 struct uverbs_attr_bundle bundle;
67 u64 internal_buffer[32];
68};
69
70/*
71 * Each method has an absolute minimum amount of memory it needs to allocate,
72 * precompute that amount and determine if the onstack memory can be used or
73 * if allocation is need.
74 */
75void uapi_compute_bundle_size(struct uverbs_api_ioctl_method *method_elm,
76 unsigned int num_attrs)
77{
78 struct bundle_priv *pbundle;
79 size_t bundle_size =
80 offsetof(struct bundle_priv, internal_buffer) +
81 sizeof(*pbundle->bundle.attrs) * method_elm->key_bitmap_len +
82 sizeof(*pbundle->uattrs) * num_attrs;
83
84 method_elm->use_stack = bundle_size <= sizeof(*pbundle);
85 method_elm->bundle_size =
86 ALIGN(bundle_size + 256, sizeof(*pbundle->internal_buffer));
87
88 /* Do not want order-2 allocations for this. */
89 WARN_ON_ONCE(method_elm->bundle_size > PAGE_SIZE);
90}
91
92/**
93 * uverbs_alloc() - Quickly allocate memory for use with a bundle
94 * @bundle: The bundle
95 * @size: Number of bytes to allocate
96 * @flags: Allocator flags
97 *
98 * The bundle allocator is intended for allocations that are connected with
99 * processing the system call related to the bundle. The allocated memory is
100 * always freed once the system call completes, and cannot be freed any other
101 * way.
102 *
103 * This tries to use a small pool of pre-allocated memory for performance.
104 */
105__malloc void *_uverbs_alloc(struct uverbs_attr_bundle *bundle, size_t size,
106 gfp_t flags)
107{
108 struct bundle_priv *pbundle =
109 container_of(bundle, struct bundle_priv, bundle);
110 size_t new_used;
111 void *res;
112
113 if (check_add_overflow(size, pbundle->internal_used, &new_used))
114 return ERR_PTR(-EOVERFLOW);
115
116 if (new_used > pbundle->internal_avail) {
117 struct bundle_alloc_head *buf;
118
119 buf = kvmalloc(struct_size(buf, data, size), flags);
120 if (!buf)
121 return ERR_PTR(-ENOMEM);
122 buf->next = pbundle->allocated_mem;
123 pbundle->allocated_mem = buf;
124 return buf->data;
125 }
126
127 res = (void *)pbundle->internal_buffer + pbundle->internal_used;
128 pbundle->internal_used =
129 ALIGN(new_used, sizeof(*pbundle->internal_buffer));
130 if (flags & __GFP_ZERO)
131 memset(res, 0, size);
132 return res;
133}
134EXPORT_SYMBOL(_uverbs_alloc);
135
136static bool uverbs_is_attr_cleared(const struct ib_uverbs_attr *uattr,
137 u16 len)
138{
139 if (uattr->len > sizeof(((struct ib_uverbs_attr *)0)->data))
140 return ib_is_buffer_cleared(u64_to_user_ptr(uattr->data) + len,
141 uattr->len - len);
142
143 return !memchr_inv((const void *)&uattr->data + len,
144 0, uattr->len - len);
145}
146
147static int uverbs_set_output(const struct uverbs_attr_bundle *bundle,
148 const struct uverbs_attr *attr)
149{
150 struct bundle_priv *pbundle =
151 container_of(bundle, struct bundle_priv, bundle);
152 u16 flags;
153
154 flags = pbundle->uattrs[attr->ptr_attr.uattr_idx].flags |
155 UVERBS_ATTR_F_VALID_OUTPUT;
156 if (put_user(flags,
157 &pbundle->user_attrs[attr->ptr_attr.uattr_idx].flags))
158 return -EFAULT;
159 return 0;
160}
161
162static int uverbs_process_idrs_array(struct bundle_priv *pbundle,
163 const struct uverbs_api_attr *attr_uapi,
164 struct uverbs_objs_arr_attr *attr,
165 struct ib_uverbs_attr *uattr,
166 u32 attr_bkey)
167{
168 const struct uverbs_attr_spec *spec = &attr_uapi->spec;
169 size_t array_len;
170 u32 *idr_vals;
171 int ret = 0;
172 size_t i;
173
174 if (uattr->attr_data.reserved)
175 return -EINVAL;
176
177 if (uattr->len % sizeof(u32))
178 return -EINVAL;
179
180 array_len = uattr->len / sizeof(u32);
181 if (array_len < spec->u2.objs_arr.min_len ||
182 array_len > spec->u2.objs_arr.max_len)
183 return -EINVAL;
184
185 attr->uobjects =
186 uverbs_alloc(&pbundle->bundle,
187 array_size(array_len, sizeof(*attr->uobjects)));
188 if (IS_ERR(attr->uobjects))
189 return PTR_ERR(attr->uobjects);
190
191 /*
192 * Since idr is 4B and *uobjects is >= 4B, we can use attr->uobjects
193 * to store idrs array and avoid additional memory allocation. The
194 * idrs array is offset to the end of the uobjects array so we will be
195 * able to read idr and replace with a pointer.
196 */
197 idr_vals = (u32 *)(attr->uobjects + array_len) - array_len;
198
199 if (uattr->len > sizeof(uattr->data)) {
200 ret = copy_from_user(idr_vals, u64_to_user_ptr(uattr->data),
201 uattr->len);
202 if (ret)
203 return -EFAULT;
204 } else {
205 memcpy(idr_vals, &uattr->data, uattr->len);
206 }
207
208 for (i = 0; i != array_len; i++) {
209 attr->uobjects[i] = uverbs_get_uobject_from_file(
210 spec->u2.objs_arr.obj_type, pbundle->bundle.ufile,
211 spec->u2.objs_arr.access, idr_vals[i]);
212 if (IS_ERR(attr->uobjects[i])) {
213 ret = PTR_ERR(attr->uobjects[i]);
214 break;
215 }
216 pbundle->bundle.context = attr->uobjects[i]->context;
217 }
218
219 attr->len = i;
220 __set_bit(attr_bkey, pbundle->spec_finalize);
221 return ret;
222}
223
224static int uverbs_free_idrs_array(const struct uverbs_api_attr *attr_uapi,
225 struct uverbs_objs_arr_attr *attr,
226 bool commit)
227{
228 const struct uverbs_attr_spec *spec = &attr_uapi->spec;
229 int current_ret;
230 int ret = 0;
231 size_t i;
232
233 for (i = 0; i != attr->len; i++) {
234 current_ret = uverbs_finalize_object(
235 attr->uobjects[i], spec->u2.objs_arr.access, commit);
236 if (!ret)
237 ret = current_ret;
238 }
239
240 return ret;
241}
242
243static int uverbs_process_attr(struct bundle_priv *pbundle,
244 const struct uverbs_api_attr *attr_uapi,
245 struct ib_uverbs_attr *uattr, u32 attr_bkey)
246{
247 const struct uverbs_attr_spec *spec = &attr_uapi->spec;
248 struct uverbs_attr *e = &pbundle->bundle.attrs[attr_bkey];
249 const struct uverbs_attr_spec *val_spec = spec;
250 struct uverbs_obj_attr *o_attr;
251
252 switch (spec->type) {
253 case UVERBS_ATTR_TYPE_ENUM_IN:
254 if (uattr->attr_data.enum_data.elem_id >= spec->u.enum_def.num_elems)
255 return -EOPNOTSUPP;
256
257 if (uattr->attr_data.enum_data.reserved)
258 return -EINVAL;
259
260 val_spec = &spec->u2.enum_def.ids[uattr->attr_data.enum_data.elem_id];
261
262 /* Currently we only support PTR_IN based enums */
263 if (val_spec->type != UVERBS_ATTR_TYPE_PTR_IN)
264 return -EOPNOTSUPP;
265
266 e->ptr_attr.enum_id = uattr->attr_data.enum_data.elem_id;
267 /* fall through */
268 case UVERBS_ATTR_TYPE_PTR_IN:
269 /* Ensure that any data provided by userspace beyond the known
270 * struct is zero. Userspace that knows how to use some future
271 * longer struct will fail here if used with an old kernel and
272 * non-zero content, making ABI compat/discovery simpler.
273 */
274 if (uattr->len > val_spec->u.ptr.len &&
275 val_spec->zero_trailing &&
276 !uverbs_is_attr_cleared(uattr, val_spec->u.ptr.len))
277 return -EOPNOTSUPP;
278
279 /* fall through */
280 case UVERBS_ATTR_TYPE_PTR_OUT:
281 if (uattr->len < val_spec->u.ptr.min_len ||
282 (!val_spec->zero_trailing &&
283 uattr->len > val_spec->u.ptr.len))
284 return -EINVAL;
285
286 if (spec->type != UVERBS_ATTR_TYPE_ENUM_IN &&
287 uattr->attr_data.reserved)
288 return -EINVAL;
289
290 e->ptr_attr.uattr_idx = uattr - pbundle->uattrs;
291 e->ptr_attr.len = uattr->len;
292
293 if (val_spec->alloc_and_copy && !uverbs_attr_ptr_is_inline(e)) {
294 void *p;
295
296 p = uverbs_alloc(&pbundle->bundle, uattr->len);
297 if (IS_ERR(p))
298 return PTR_ERR(p);
299
300 e->ptr_attr.ptr = p;
301
302 if (copy_from_user(p, u64_to_user_ptr(uattr->data),
303 uattr->len))
304 return -EFAULT;
305 } else {
306 e->ptr_attr.data = uattr->data;
307 }
308 break;
309
310 case UVERBS_ATTR_TYPE_IDR:
311 case UVERBS_ATTR_TYPE_FD:
312 if (uattr->attr_data.reserved)
313 return -EINVAL;
314
315 if (uattr->len != 0)
316 return -EINVAL;
317
318 o_attr = &e->obj_attr;
319 o_attr->attr_elm = attr_uapi;
320
321 /*
322 * The type of uattr->data is u64 for UVERBS_ATTR_TYPE_IDR and
323 * s64 for UVERBS_ATTR_TYPE_FD. We can cast the u64 to s64
324 * here without caring about truncation as we know that the
325 * IDR implementation today rejects negative IDs
326 */
327 o_attr->uobject = uverbs_get_uobject_from_file(
328 spec->u.obj.obj_type,
329 pbundle->bundle.ufile,
330 spec->u.obj.access,
331 uattr->data_s64);
332 if (IS_ERR(o_attr->uobject))
333 return PTR_ERR(o_attr->uobject);
334 pbundle->bundle.context = o_attr->uobject->context;
335 __set_bit(attr_bkey, pbundle->uobj_finalize);
336
337 if (spec->u.obj.access == UVERBS_ACCESS_NEW) {
338 unsigned int uattr_idx = uattr - pbundle->uattrs;
339 s64 id = o_attr->uobject->id;
340
341 /* Copy the allocated id to the user-space */
342 if (put_user(id, &pbundle->user_attrs[uattr_idx].data))
343 return -EFAULT;
344 }
345
346 break;
347
348 case UVERBS_ATTR_TYPE_IDRS_ARRAY:
349 return uverbs_process_idrs_array(pbundle, attr_uapi,
350 &e->objs_arr_attr, uattr,
351 attr_bkey);
352 default:
353 return -EOPNOTSUPP;
354 }
355
356 return 0;
357}
358
359/*
360 * We search the radix tree with the method prefix and now we want to fast
361 * search the suffix bits to get a particular attribute pointer. It is not
362 * totally clear to me if this breaks the radix tree encasulation or not, but
363 * it uses the iter data to determine if the method iter points at the same
364 * chunk that will store the attribute, if so it just derefs it directly. By
365 * construction in most kernel configs the method and attrs will all fit in a
366 * single radix chunk, so in most cases this will have no search. Other cases
367 * this falls back to a full search.
368 */
369static void __rcu **uapi_get_attr_for_method(struct bundle_priv *pbundle,
370 u32 attr_key)
371{
372 void __rcu **slot;
373
374 if (likely(attr_key < pbundle->radix_slots_len)) {
375 void *entry;
376
377 slot = pbundle->radix_slots + attr_key;
378 entry = rcu_dereference_raw(*slot);
379 if (likely(!radix_tree_is_internal_node(entry) && entry))
380 return slot;
381 }
382
383 return radix_tree_lookup_slot(pbundle->radix,
384 pbundle->method_key | attr_key);
385}
386
387static int uverbs_set_attr(struct bundle_priv *pbundle,
388 struct ib_uverbs_attr *uattr)
389{
390 u32 attr_key = uapi_key_attr(uattr->attr_id);
391 u32 attr_bkey = uapi_bkey_attr(attr_key);
392 const struct uverbs_api_attr *attr;
393 void __rcu **slot;
394 int ret;
395
396 slot = uapi_get_attr_for_method(pbundle, attr_key);
397 if (!slot) {
398 /*
399 * Kernel does not support the attribute but user-space says it
400 * is mandatory
401 */
402 if (uattr->flags & UVERBS_ATTR_F_MANDATORY)
403 return -EPROTONOSUPPORT;
404 return 0;
405 }
406 attr = rcu_dereference_protected(*slot, true);
407
408 /* Reject duplicate attributes from user-space */
409 if (test_bit(attr_bkey, pbundle->bundle.attr_present))
410 return -EINVAL;
411
412 ret = uverbs_process_attr(pbundle, attr, uattr, attr_bkey);
413 if (ret)
414 return ret;
415
416 __set_bit(attr_bkey, pbundle->bundle.attr_present);
417
418 return 0;
419}
420
421static int ib_uverbs_run_method(struct bundle_priv *pbundle,
422 unsigned int num_attrs)
423{
424 int (*handler)(struct uverbs_attr_bundle *attrs);
425 size_t uattrs_size = array_size(sizeof(*pbundle->uattrs), num_attrs);
426 unsigned int destroy_bkey = pbundle->method_elm->destroy_bkey;
427 unsigned int i;
428 int ret;
429
430 /* See uverbs_disassociate_api() */
431 handler = srcu_dereference(
432 pbundle->method_elm->handler,
433 &pbundle->bundle.ufile->device->disassociate_srcu);
434 if (!handler)
435 return -EIO;
436
437 pbundle->uattrs = uverbs_alloc(&pbundle->bundle, uattrs_size);
438 if (IS_ERR(pbundle->uattrs))
439 return PTR_ERR(pbundle->uattrs);
440 if (copy_from_user(pbundle->uattrs, pbundle->user_attrs, uattrs_size))
441 return -EFAULT;
442
443 for (i = 0; i != num_attrs; i++) {
444 ret = uverbs_set_attr(pbundle, &pbundle->uattrs[i]);
445 if (unlikely(ret))
446 return ret;
447 }
448
449 /* User space did not provide all the mandatory attributes */
450 if (unlikely(!bitmap_subset(pbundle->method_elm->attr_mandatory,
451 pbundle->bundle.attr_present,
452 pbundle->method_elm->key_bitmap_len)))
453 return -EINVAL;
454
455 if (pbundle->method_elm->has_udata)
456 uverbs_fill_udata(&pbundle->bundle,
457 &pbundle->bundle.driver_udata,
458 UVERBS_ATTR_UHW_IN, UVERBS_ATTR_UHW_OUT);
459
460 if (destroy_bkey != UVERBS_API_ATTR_BKEY_LEN) {
461 struct uverbs_obj_attr *destroy_attr =
462 &pbundle->bundle.attrs[destroy_bkey].obj_attr;
463
464 ret = uobj_destroy(destroy_attr->uobject);
465 if (ret)
466 return ret;
467 __clear_bit(destroy_bkey, pbundle->uobj_finalize);
468
469 ret = handler(&pbundle->bundle);
470 uobj_put_destroy(destroy_attr->uobject);
471 } else {
472 ret = handler(&pbundle->bundle);
473 }
474
475 /*
476 * Until the drivers are revised to use the bundle directly we have to
477 * assume that the driver wrote to its UHW_OUT and flag userspace
478 * appropriately.
479 */
480 if (!ret && pbundle->method_elm->has_udata) {
481 const struct uverbs_attr *attr =
482 uverbs_attr_get(&pbundle->bundle, UVERBS_ATTR_UHW_OUT);
483
484 if (!IS_ERR(attr))
485 ret = uverbs_set_output(&pbundle->bundle, attr);
486 }
487
488 /*
489 * EPROTONOSUPPORT is ONLY to be returned if the ioctl framework can
490 * not invoke the method because the request is not supported. No
491 * other cases should return this code.
492 */
493 if (WARN_ON_ONCE(ret == -EPROTONOSUPPORT))
494 return -EINVAL;
495
496 return ret;
497}
498
499static int bundle_destroy(struct bundle_priv *pbundle, bool commit)
500{
501 unsigned int key_bitmap_len = pbundle->method_elm->key_bitmap_len;
502 struct bundle_alloc_head *memblock;
503 unsigned int i;
504 int ret = 0;
505
506 /* fast path for simple uobjects */
507 i = -1;
508 while ((i = find_next_bit(pbundle->uobj_finalize, key_bitmap_len,
509 i + 1)) < key_bitmap_len) {
510 struct uverbs_attr *attr = &pbundle->bundle.attrs[i];
511 int current_ret;
512
513 current_ret = uverbs_finalize_object(
514 attr->obj_attr.uobject,
515 attr->obj_attr.attr_elm->spec.u.obj.access, commit);
516 if (!ret)
517 ret = current_ret;
518 }
519
520 i = -1;
521 while ((i = find_next_bit(pbundle->spec_finalize, key_bitmap_len,
522 i + 1)) < key_bitmap_len) {
523 struct uverbs_attr *attr = &pbundle->bundle.attrs[i];
524 const struct uverbs_api_attr *attr_uapi;
525 void __rcu **slot;
526 int current_ret;
527
528 slot = uapi_get_attr_for_method(
529 pbundle,
530 pbundle->method_key | uapi_bkey_to_key_attr(i));
531 if (WARN_ON(!slot))
532 continue;
533
534 attr_uapi = rcu_dereference_protected(*slot, true);
535
536 if (attr_uapi->spec.type == UVERBS_ATTR_TYPE_IDRS_ARRAY) {
537 current_ret = uverbs_free_idrs_array(
538 attr_uapi, &attr->objs_arr_attr, commit);
539 if (!ret)
540 ret = current_ret;
541 }
542 }
543
544 for (memblock = pbundle->allocated_mem; memblock;) {
545 struct bundle_alloc_head *tmp = memblock;
546
547 memblock = memblock->next;
548 kvfree(tmp);
549 }
550
551 return ret;
552}
553
554static int ib_uverbs_cmd_verbs(struct ib_uverbs_file *ufile,
555 struct ib_uverbs_ioctl_hdr *hdr,
556 struct ib_uverbs_attr __user *user_attrs)
557{
558 const struct uverbs_api_ioctl_method *method_elm;
559 struct uverbs_api *uapi = ufile->device->uapi;
560 struct radix_tree_iter attrs_iter;
561 struct bundle_priv *pbundle;
562 struct bundle_priv onstack;
563 void __rcu **slot;
564 int destroy_ret;
565 int ret;
566
567 if (unlikely(hdr->driver_id != uapi->driver_id))
568 return -EINVAL;
569
570 slot = radix_tree_iter_lookup(
571 &uapi->radix, &attrs_iter,
572 uapi_key_obj(hdr->object_id) |
573 uapi_key_ioctl_method(hdr->method_id));
574 if (unlikely(!slot))
575 return -EPROTONOSUPPORT;
576 method_elm = rcu_dereference_protected(*slot, true);
577
578 if (!method_elm->use_stack) {
579 pbundle = kmalloc(method_elm->bundle_size, GFP_KERNEL);
580 if (!pbundle)
581 return -ENOMEM;
582 pbundle->internal_avail =
583 method_elm->bundle_size -
584 offsetof(struct bundle_priv, internal_buffer);
585 pbundle->alloc_head.next = NULL;
586 pbundle->allocated_mem = &pbundle->alloc_head;
587 } else {
588 pbundle = &onstack;
589 pbundle->internal_avail = sizeof(pbundle->internal_buffer);
590 pbundle->allocated_mem = NULL;
591 }
592
593 /* Space for the pbundle->bundle.attrs flex array */
594 pbundle->method_elm = method_elm;
595 pbundle->method_key = attrs_iter.index;
596 pbundle->bundle.ufile = ufile;
597 pbundle->bundle.context = NULL; /* only valid if bundle has uobject */
598 pbundle->radix = &uapi->radix;
599 pbundle->radix_slots = slot;
600 pbundle->radix_slots_len = radix_tree_chunk_size(&attrs_iter);
601 pbundle->user_attrs = user_attrs;
602
603 pbundle->internal_used = ALIGN(pbundle->method_elm->key_bitmap_len *
604 sizeof(*pbundle->bundle.attrs),
605 sizeof(*pbundle->internal_buffer));
606 memset(pbundle->bundle.attr_present, 0,
607 sizeof(pbundle->bundle.attr_present));
608 memset(pbundle->uobj_finalize, 0, sizeof(pbundle->uobj_finalize));
609 memset(pbundle->spec_finalize, 0, sizeof(pbundle->spec_finalize));
610
611 ret = ib_uverbs_run_method(pbundle, hdr->num_attrs);
612 destroy_ret = bundle_destroy(pbundle, ret == 0);
613 if (unlikely(destroy_ret && !ret))
614 return destroy_ret;
615
616 return ret;
617}
618
619long ib_uverbs_ioctl(struct file *filp, unsigned int cmd, unsigned long arg)
620{
621 struct ib_uverbs_file *file = filp->private_data;
622 struct ib_uverbs_ioctl_hdr __user *user_hdr =
623 (struct ib_uverbs_ioctl_hdr __user *)arg;
624 struct ib_uverbs_ioctl_hdr hdr;
625 int srcu_key;
626 int err;
627
628 if (unlikely(cmd != RDMA_VERBS_IOCTL))
629 return -ENOIOCTLCMD;
630
631 err = copy_from_user(&hdr, user_hdr, sizeof(hdr));
632 if (err)
633 return -EFAULT;
634
635 if (hdr.length > PAGE_SIZE ||
636 hdr.length != struct_size(&hdr, attrs, hdr.num_attrs))
637 return -EINVAL;
638
639 if (hdr.reserved1 || hdr.reserved2)
640 return -EPROTONOSUPPORT;
641
642 srcu_key = srcu_read_lock(&file->device->disassociate_srcu);
643 err = ib_uverbs_cmd_verbs(file, &hdr, user_hdr->attrs);
644 srcu_read_unlock(&file->device->disassociate_srcu, srcu_key);
645 return err;
646}
647
648int uverbs_get_flags64(u64 *to, const struct uverbs_attr_bundle *attrs_bundle,
649 size_t idx, u64 allowed_bits)
650{
651 const struct uverbs_attr *attr;
652 u64 flags;
653
654 attr = uverbs_attr_get(attrs_bundle, idx);
655 /* Missing attribute means 0 flags */
656 if (IS_ERR(attr)) {
657 *to = 0;
658 return 0;
659 }
660
661 /*
662 * New userspace code should use 8 bytes to pass flags, but we
663 * transparently support old userspaces that were using 4 bytes as
664 * well.
665 */
666 if (attr->ptr_attr.len == 8)
667 flags = attr->ptr_attr.data;
668 else if (attr->ptr_attr.len == 4)
669 flags = *(u32 *)&attr->ptr_attr.data;
670 else
671 return -EINVAL;
672
673 if (flags & ~allowed_bits)
674 return -EINVAL;
675
676 *to = flags;
677 return 0;
678}
679EXPORT_SYMBOL(uverbs_get_flags64);
680
681int uverbs_get_flags32(u32 *to, const struct uverbs_attr_bundle *attrs_bundle,
682 size_t idx, u64 allowed_bits)
683{
684 u64 flags;
685 int ret;
686
687 ret = uverbs_get_flags64(&flags, attrs_bundle, idx, allowed_bits);
688 if (ret)
689 return ret;
690
691 if (flags > U32_MAX)
692 return -EINVAL;
693 *to = flags;
694
695 return 0;
696}
697EXPORT_SYMBOL(uverbs_get_flags32);
698
699/*
700 * Fill a ib_udata struct (core or uhw) using the given attribute IDs.
701 * This is primarily used to convert the UVERBS_ATTR_UHW() into the
702 * ib_udata format used by the drivers.
703 */
704void uverbs_fill_udata(struct uverbs_attr_bundle *bundle,
705 struct ib_udata *udata, unsigned int attr_in,
706 unsigned int attr_out)
707{
708 struct bundle_priv *pbundle =
709 container_of(bundle, struct bundle_priv, bundle);
710 const struct uverbs_attr *in =
711 uverbs_attr_get(&pbundle->bundle, attr_in);
712 const struct uverbs_attr *out =
713 uverbs_attr_get(&pbundle->bundle, attr_out);
714
715 if (!IS_ERR(in)) {
716 udata->inlen = in->ptr_attr.len;
717 if (uverbs_attr_ptr_is_inline(in))
718 udata->inbuf =
719 &pbundle->user_attrs[in->ptr_attr.uattr_idx]
720 .data;
721 else
722 udata->inbuf = u64_to_user_ptr(in->ptr_attr.data);
723 } else {
724 udata->inbuf = NULL;
725 udata->inlen = 0;
726 }
727
728 if (!IS_ERR(out)) {
729 udata->outbuf = u64_to_user_ptr(out->ptr_attr.data);
730 udata->outlen = out->ptr_attr.len;
731 } else {
732 udata->outbuf = NULL;
733 udata->outlen = 0;
734 }
735}
736
737int uverbs_copy_to(const struct uverbs_attr_bundle *bundle, size_t idx,
738 const void *from, size_t size)
739{
740 const struct uverbs_attr *attr = uverbs_attr_get(bundle, idx);
741 size_t min_size;
742
743 if (IS_ERR(attr))
744 return PTR_ERR(attr);
745
746 min_size = min_t(size_t, attr->ptr_attr.len, size);
747 if (copy_to_user(u64_to_user_ptr(attr->ptr_attr.data), from, min_size))
748 return -EFAULT;
749
750 return uverbs_set_output(bundle, attr);
751}
752EXPORT_SYMBOL(uverbs_copy_to);
753
754
755/*
756 * This is only used if the caller has directly used copy_to_use to write the
757 * data. It signals to user space that the buffer is filled in.
758 */
759int uverbs_output_written(const struct uverbs_attr_bundle *bundle, size_t idx)
760{
761 const struct uverbs_attr *attr = uverbs_attr_get(bundle, idx);
762
763 if (IS_ERR(attr))
764 return PTR_ERR(attr);
765
766 return uverbs_set_output(bundle, attr);
767}
768
769int _uverbs_get_const(s64 *to, const struct uverbs_attr_bundle *attrs_bundle,
770 size_t idx, s64 lower_bound, u64 upper_bound,
771 s64 *def_val)
772{
773 const struct uverbs_attr *attr;
774
775 attr = uverbs_attr_get(attrs_bundle, idx);
776 if (IS_ERR(attr)) {
777 if ((PTR_ERR(attr) != -ENOENT) || !def_val)
778 return PTR_ERR(attr);
779
780 *to = *def_val;
781 } else {
782 *to = attr->ptr_attr.data;
783 }
784
785 if (*to < lower_bound || (*to > 0 && (u64)*to > upper_bound))
786 return -EINVAL;
787
788 return 0;
789}
790EXPORT_SYMBOL(_uverbs_get_const);
791
792int uverbs_copy_to_struct_or_zero(const struct uverbs_attr_bundle *bundle,
793 size_t idx, const void *from, size_t size)
794{
795 const struct uverbs_attr *attr = uverbs_attr_get(bundle, idx);
796
797 if (size < attr->ptr_attr.len) {
798 if (clear_user(u64_to_user_ptr(attr->ptr_attr.data) + size,
799 attr->ptr_attr.len - size))
800 return -EFAULT;
801 }
802 return uverbs_copy_to(bundle, idx, from, size);
803}