mirror of
				https://github.com/torvalds/linux.git
				synced 2025-10-31 00:28:52 +02:00 
			
		
		
		
	 155f04366e
			
		
	
	
		155f04366e
		
	
	
	
	
		
			
			-Wflex-array-member-not-at-end is coming in GCC-14, and we are getting
ready to enable it globally.
There are currently a couple of objects (`alloc_head` and `bundle`) in
`struct bundle_priv` that contain a couple of flexible structures:
struct bundle_priv {
        /* Must be first */
        struct bundle_alloc_head alloc_head;
	...
        /*
         * Must be last. bundle ends in a flex array which overlaps
         * internal_buffer.
         */
        struct uverbs_attr_bundle bundle;
        u64 internal_buffer[32];
};
So, in order to avoid ending up with a couple of flexible-array members
in the middle of a struct, we use the `struct_group_tagged()` helper to
separate the flexible array from the rest of the members in the flexible
structures:
struct uverbs_attr_bundle {
        struct_group_tagged(uverbs_attr_bundle_hdr, hdr,
		... the rest of the members
        );
        struct uverbs_attr attrs[];
};
With the change described above, we now declare objects of the type of
the tagged struct without embedding flexible arrays in the middle of
another struct:
struct bundle_priv {
        /* Must be first */
        struct bundle_alloc_head_hdr alloc_head;
        ...
        struct uverbs_attr_bundle_hdr bundle;
        u64 internal_buffer[32];
};
We also use `container_of()` whenever we need to retrieve a pointer
to the flexible structures.
Notice that the `bundle_size` computed in `uapi_compute_bundle_size()`
remains the same.
So, with these changes, fix the following warnings:
drivers/infiniband/core/uverbs_ioctl.c:45:34: warning: structure containing a flexible array member is not at the end of another structure [-Wflex-array-member-not-at-end]
   45 |         struct bundle_alloc_head alloc_head;
      |                                  ^~~~~~~~~~
drivers/infiniband/core/uverbs_ioctl.c:67:35: warning: structure containing a flexible array member is not at the end of another structure [-Wflex-array-member-not-at-end]
   67 |         struct uverbs_attr_bundle bundle;
      |                                   ^~~~~~
Signed-off-by: Gustavo A. R. Silva <gustavoars@kernel.org>
Link: https://lore.kernel.org/r/ZeIgeZ5Sb0IZTOyt@neat
Reviewed-by: Kees Cook <keescook@chromium.org>
Signed-off-by: Leon Romanovsky <leon@kernel.org>
		
	
			
		
			
				
	
	
		
			849 lines
		
	
	
	
		
			24 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			849 lines
		
	
	
	
		
			24 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Copyright (c) 2017, Mellanox Technologies inc.  All rights reserved.
 | |
|  *
 | |
|  * This software is available to you under a choice of one of two
 | |
|  * licenses.  You may choose to be licensed under the terms of the GNU
 | |
|  * General Public License (GPL) Version 2, available from the file
 | |
|  * COPYING in the main directory of this source tree, or the
 | |
|  * OpenIB.org BSD license below:
 | |
|  *
 | |
|  *     Redistribution and use in source and binary forms, with or
 | |
|  *     without modification, are permitted provided that the following
 | |
|  *     conditions are met:
 | |
|  *
 | |
|  *      - Redistributions of source code must retain the above
 | |
|  *        copyright notice, this list of conditions and the following
 | |
|  *        disclaimer.
 | |
|  *
 | |
|  *      - Redistributions in binary form must reproduce the above
 | |
|  *        copyright notice, this list of conditions and the following
 | |
|  *        disclaimer in the documentation and/or other materials
 | |
|  *        provided with the distribution.
 | |
|  *
 | |
|  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
 | |
|  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
 | |
|  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
 | |
|  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
 | |
|  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
 | |
|  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
 | |
|  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
 | |
|  * SOFTWARE.
 | |
|  */
 | |
| 
 | |
| #include <rdma/rdma_user_ioctl.h>
 | |
| #include <rdma/uverbs_ioctl.h>
 | |
| #include "rdma_core.h"
 | |
| #include "uverbs.h"
 | |
| 
 | |
| struct bundle_alloc_head {
 | |
| 	struct_group_tagged(bundle_alloc_head_hdr, hdr,
 | |
| 		struct bundle_alloc_head *next;
 | |
| 	);
 | |
| 	u8 data[];
 | |
| };
 | |
| 
 | |
| struct bundle_priv {
 | |
| 	/* Must be first */
 | |
| 	struct bundle_alloc_head_hdr alloc_head;
 | |
| 	struct bundle_alloc_head *allocated_mem;
 | |
| 	size_t internal_avail;
 | |
| 	size_t internal_used;
 | |
| 
 | |
| 	struct radix_tree_root *radix;
 | |
| 	const struct uverbs_api_ioctl_method *method_elm;
 | |
| 	void __rcu **radix_slots;
 | |
| 	unsigned long radix_slots_len;
 | |
| 	u32 method_key;
 | |
| 
 | |
| 	struct ib_uverbs_attr __user *user_attrs;
 | |
| 	struct ib_uverbs_attr *uattrs;
 | |
| 
 | |
| 	DECLARE_BITMAP(uobj_finalize, UVERBS_API_ATTR_BKEY_LEN);
 | |
| 	DECLARE_BITMAP(spec_finalize, UVERBS_API_ATTR_BKEY_LEN);
 | |
| 	DECLARE_BITMAP(uobj_hw_obj_valid, UVERBS_API_ATTR_BKEY_LEN);
 | |
| 
 | |
| 	/*
 | |
| 	 * Must be last. bundle ends in a flex array which overlaps
 | |
| 	 * internal_buffer.
 | |
| 	 */
 | |
| 	struct uverbs_attr_bundle_hdr bundle;
 | |
| 	u64 internal_buffer[32];
 | |
| };
 | |
| 
 | |
| /*
 | |
|  * Each method has an absolute minimum amount of memory it needs to allocate,
 | |
|  * precompute that amount and determine if the onstack memory can be used or
 | |
|  * if allocation is need.
 | |
|  */
 | |
| void uapi_compute_bundle_size(struct uverbs_api_ioctl_method *method_elm,
 | |
| 			      unsigned int num_attrs)
 | |
| {
 | |
| 	struct bundle_priv *pbundle;
 | |
| 	struct uverbs_attr_bundle *bundle;
 | |
| 	size_t bundle_size =
 | |
| 		offsetof(struct bundle_priv, internal_buffer) +
 | |
| 		sizeof(*bundle->attrs) * method_elm->key_bitmap_len +
 | |
| 		sizeof(*pbundle->uattrs) * num_attrs;
 | |
| 
 | |
| 	method_elm->use_stack = bundle_size <= sizeof(*pbundle);
 | |
| 	method_elm->bundle_size =
 | |
| 		ALIGN(bundle_size + 256, sizeof(*pbundle->internal_buffer));
 | |
| 
 | |
| 	/* Do not want order-2 allocations for this. */
 | |
| 	WARN_ON_ONCE(method_elm->bundle_size > PAGE_SIZE);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * _uverbs_alloc() - Quickly allocate memory for use with a bundle
 | |
|  * @bundle: The bundle
 | |
|  * @size: Number of bytes to allocate
 | |
|  * @flags: Allocator flags
 | |
|  *
 | |
|  * The bundle allocator is intended for allocations that are connected with
 | |
|  * processing the system call related to the bundle. The allocated memory is
 | |
|  * always freed once the system call completes, and cannot be freed any other
 | |
|  * way.
 | |
|  *
 | |
|  * This tries to use a small pool of pre-allocated memory for performance.
 | |
|  */
 | |
| __malloc void *_uverbs_alloc(struct uverbs_attr_bundle *bundle, size_t size,
 | |
| 			     gfp_t flags)
 | |
| {
 | |
| 	struct bundle_priv *pbundle =
 | |
| 		container_of(&bundle->hdr, struct bundle_priv, bundle);
 | |
| 	size_t new_used;
 | |
| 	void *res;
 | |
| 
 | |
| 	if (check_add_overflow(size, pbundle->internal_used, &new_used))
 | |
| 		return ERR_PTR(-EOVERFLOW);
 | |
| 
 | |
| 	if (new_used > pbundle->internal_avail) {
 | |
| 		struct bundle_alloc_head *buf;
 | |
| 
 | |
| 		buf = kvmalloc(struct_size(buf, data, size), flags);
 | |
| 		if (!buf)
 | |
| 			return ERR_PTR(-ENOMEM);
 | |
| 		buf->next = pbundle->allocated_mem;
 | |
| 		pbundle->allocated_mem = buf;
 | |
| 		return buf->data;
 | |
| 	}
 | |
| 
 | |
| 	res = (void *)pbundle->internal_buffer + pbundle->internal_used;
 | |
| 	pbundle->internal_used =
 | |
| 		ALIGN(new_used, sizeof(*pbundle->internal_buffer));
 | |
| 	if (want_init_on_alloc(flags))
 | |
| 		memset(res, 0, size);
 | |
| 	return res;
 | |
| }
 | |
| EXPORT_SYMBOL(_uverbs_alloc);
 | |
| 
 | |
| static bool uverbs_is_attr_cleared(const struct ib_uverbs_attr *uattr,
 | |
| 				   u16 len)
 | |
| {
 | |
| 	if (uattr->len > sizeof_field(struct ib_uverbs_attr, data))
 | |
| 		return ib_is_buffer_cleared(u64_to_user_ptr(uattr->data) + len,
 | |
| 					    uattr->len - len);
 | |
| 
 | |
| 	return !memchr_inv((const void *)&uattr->data + len,
 | |
| 			   0, uattr->len - len);
 | |
| }
 | |
| 
 | |
| static int uverbs_set_output(const struct uverbs_attr_bundle *bundle,
 | |
| 			     const struct uverbs_attr *attr)
 | |
| {
 | |
| 	struct bundle_priv *pbundle =
 | |
| 		container_of(&bundle->hdr, struct bundle_priv, bundle);
 | |
| 	u16 flags;
 | |
| 
 | |
| 	flags = pbundle->uattrs[attr->ptr_attr.uattr_idx].flags |
 | |
| 		UVERBS_ATTR_F_VALID_OUTPUT;
 | |
| 	if (put_user(flags,
 | |
| 		     &pbundle->user_attrs[attr->ptr_attr.uattr_idx].flags))
 | |
| 		return -EFAULT;
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int uverbs_process_idrs_array(struct bundle_priv *pbundle,
 | |
| 				     const struct uverbs_api_attr *attr_uapi,
 | |
| 				     struct uverbs_objs_arr_attr *attr,
 | |
| 				     struct ib_uverbs_attr *uattr,
 | |
| 				     u32 attr_bkey)
 | |
| {
 | |
| 	struct uverbs_attr_bundle *bundle =
 | |
| 		container_of(&pbundle->bundle, struct uverbs_attr_bundle, hdr);
 | |
| 	const struct uverbs_attr_spec *spec = &attr_uapi->spec;
 | |
| 	size_t array_len;
 | |
| 	u32 *idr_vals;
 | |
| 	int ret = 0;
 | |
| 	size_t i;
 | |
| 
 | |
| 	if (uattr->attr_data.reserved)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	if (uattr->len % sizeof(u32))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	array_len = uattr->len / sizeof(u32);
 | |
| 	if (array_len < spec->u2.objs_arr.min_len ||
 | |
| 	    array_len > spec->u2.objs_arr.max_len)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	attr->uobjects =
 | |
| 		uverbs_alloc(bundle,
 | |
| 			     array_size(array_len, sizeof(*attr->uobjects)));
 | |
| 	if (IS_ERR(attr->uobjects))
 | |
| 		return PTR_ERR(attr->uobjects);
 | |
| 
 | |
| 	/*
 | |
| 	 * Since idr is 4B and *uobjects is >= 4B, we can use attr->uobjects
 | |
| 	 * to store idrs array and avoid additional memory allocation. The
 | |
| 	 * idrs array is offset to the end of the uobjects array so we will be
 | |
| 	 * able to read idr and replace with a pointer.
 | |
| 	 */
 | |
| 	idr_vals = (u32 *)(attr->uobjects + array_len) - array_len;
 | |
| 
 | |
| 	if (uattr->len > sizeof(uattr->data)) {
 | |
| 		ret = copy_from_user(idr_vals, u64_to_user_ptr(uattr->data),
 | |
| 				     uattr->len);
 | |
| 		if (ret)
 | |
| 			return -EFAULT;
 | |
| 	} else {
 | |
| 		memcpy(idr_vals, &uattr->data, uattr->len);
 | |
| 	}
 | |
| 
 | |
| 	for (i = 0; i != array_len; i++) {
 | |
| 		attr->uobjects[i] = uverbs_get_uobject_from_file(
 | |
| 			spec->u2.objs_arr.obj_type, spec->u2.objs_arr.access,
 | |
| 			idr_vals[i], bundle);
 | |
| 		if (IS_ERR(attr->uobjects[i])) {
 | |
| 			ret = PTR_ERR(attr->uobjects[i]);
 | |
| 			break;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	attr->len = i;
 | |
| 	__set_bit(attr_bkey, pbundle->spec_finalize);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static void uverbs_free_idrs_array(const struct uverbs_api_attr *attr_uapi,
 | |
| 				   struct uverbs_objs_arr_attr *attr,
 | |
| 				   bool commit,
 | |
| 				   struct uverbs_attr_bundle *attrs)
 | |
| {
 | |
| 	const struct uverbs_attr_spec *spec = &attr_uapi->spec;
 | |
| 	size_t i;
 | |
| 
 | |
| 	for (i = 0; i != attr->len; i++)
 | |
| 		uverbs_finalize_object(attr->uobjects[i],
 | |
| 				       spec->u2.objs_arr.access, false, commit,
 | |
| 				       attrs);
 | |
| }
 | |
| 
 | |
| static int uverbs_process_attr(struct bundle_priv *pbundle,
 | |
| 			       const struct uverbs_api_attr *attr_uapi,
 | |
| 			       struct ib_uverbs_attr *uattr, u32 attr_bkey)
 | |
| {
 | |
| 	const struct uverbs_attr_spec *spec = &attr_uapi->spec;
 | |
| 	struct uverbs_attr_bundle *bundle =
 | |
| 		container_of(&pbundle->bundle, struct uverbs_attr_bundle, hdr);
 | |
| 	struct uverbs_attr *e = &bundle->attrs[attr_bkey];
 | |
| 	const struct uverbs_attr_spec *val_spec = spec;
 | |
| 	struct uverbs_obj_attr *o_attr;
 | |
| 
 | |
| 	switch (spec->type) {
 | |
| 	case UVERBS_ATTR_TYPE_ENUM_IN:
 | |
| 		if (uattr->attr_data.enum_data.elem_id >= spec->u.enum_def.num_elems)
 | |
| 			return -EOPNOTSUPP;
 | |
| 
 | |
| 		if (uattr->attr_data.enum_data.reserved)
 | |
| 			return -EINVAL;
 | |
| 
 | |
| 		val_spec = &spec->u2.enum_def.ids[uattr->attr_data.enum_data.elem_id];
 | |
| 
 | |
| 		/* Currently we only support PTR_IN based enums */
 | |
| 		if (val_spec->type != UVERBS_ATTR_TYPE_PTR_IN)
 | |
| 			return -EOPNOTSUPP;
 | |
| 
 | |
| 		e->ptr_attr.enum_id = uattr->attr_data.enum_data.elem_id;
 | |
| 		fallthrough;
 | |
| 	case UVERBS_ATTR_TYPE_PTR_IN:
 | |
| 		/* Ensure that any data provided by userspace beyond the known
 | |
| 		 * struct is zero. Userspace that knows how to use some future
 | |
| 		 * longer struct will fail here if used with an old kernel and
 | |
| 		 * non-zero content, making ABI compat/discovery simpler.
 | |
| 		 */
 | |
| 		if (uattr->len > val_spec->u.ptr.len &&
 | |
| 		    val_spec->zero_trailing &&
 | |
| 		    !uverbs_is_attr_cleared(uattr, val_spec->u.ptr.len))
 | |
| 			return -EOPNOTSUPP;
 | |
| 
 | |
| 		fallthrough;
 | |
| 	case UVERBS_ATTR_TYPE_PTR_OUT:
 | |
| 		if (uattr->len < val_spec->u.ptr.min_len ||
 | |
| 		    (!val_spec->zero_trailing &&
 | |
| 		     uattr->len > val_spec->u.ptr.len))
 | |
| 			return -EINVAL;
 | |
| 
 | |
| 		if (spec->type != UVERBS_ATTR_TYPE_ENUM_IN &&
 | |
| 		    uattr->attr_data.reserved)
 | |
| 			return -EINVAL;
 | |
| 
 | |
| 		e->ptr_attr.uattr_idx = uattr - pbundle->uattrs;
 | |
| 		e->ptr_attr.len = uattr->len;
 | |
| 
 | |
| 		if (val_spec->alloc_and_copy && !uverbs_attr_ptr_is_inline(e)) {
 | |
| 			void *p;
 | |
| 
 | |
| 			p = uverbs_alloc(bundle, uattr->len);
 | |
| 			if (IS_ERR(p))
 | |
| 				return PTR_ERR(p);
 | |
| 
 | |
| 			e->ptr_attr.ptr = p;
 | |
| 
 | |
| 			if (copy_from_user(p, u64_to_user_ptr(uattr->data),
 | |
| 					   uattr->len))
 | |
| 				return -EFAULT;
 | |
| 		} else {
 | |
| 			e->ptr_attr.data = uattr->data;
 | |
| 		}
 | |
| 		break;
 | |
| 
 | |
| 	case UVERBS_ATTR_TYPE_IDR:
 | |
| 	case UVERBS_ATTR_TYPE_FD:
 | |
| 		if (uattr->attr_data.reserved)
 | |
| 			return -EINVAL;
 | |
| 
 | |
| 		if (uattr->len != 0)
 | |
| 			return -EINVAL;
 | |
| 
 | |
| 		o_attr = &e->obj_attr;
 | |
| 		o_attr->attr_elm = attr_uapi;
 | |
| 
 | |
| 		/*
 | |
| 		 * The type of uattr->data is u64 for UVERBS_ATTR_TYPE_IDR and
 | |
| 		 * s64 for UVERBS_ATTR_TYPE_FD. We can cast the u64 to s64
 | |
| 		 * here without caring about truncation as we know that the
 | |
| 		 * IDR implementation today rejects negative IDs
 | |
| 		 */
 | |
| 		o_attr->uobject = uverbs_get_uobject_from_file(
 | |
| 			spec->u.obj.obj_type, spec->u.obj.access,
 | |
| 			uattr->data_s64, bundle);
 | |
| 		if (IS_ERR(o_attr->uobject))
 | |
| 			return PTR_ERR(o_attr->uobject);
 | |
| 		__set_bit(attr_bkey, pbundle->uobj_finalize);
 | |
| 
 | |
| 		if (spec->u.obj.access == UVERBS_ACCESS_NEW) {
 | |
| 			unsigned int uattr_idx = uattr - pbundle->uattrs;
 | |
| 			s64 id = o_attr->uobject->id;
 | |
| 
 | |
| 			/* Copy the allocated id to the user-space */
 | |
| 			if (put_user(id, &pbundle->user_attrs[uattr_idx].data))
 | |
| 				return -EFAULT;
 | |
| 		}
 | |
| 
 | |
| 		break;
 | |
| 
 | |
| 	case UVERBS_ATTR_TYPE_RAW_FD:
 | |
| 		if (uattr->attr_data.reserved || uattr->len != 0 ||
 | |
| 		    uattr->data_s64 < INT_MIN || uattr->data_s64 > INT_MAX)
 | |
| 			return -EINVAL;
 | |
| 		/* _uverbs_get_const_signed() is the accessor */
 | |
| 		e->ptr_attr.data = uattr->data_s64;
 | |
| 		break;
 | |
| 
 | |
| 	case UVERBS_ATTR_TYPE_IDRS_ARRAY:
 | |
| 		return uverbs_process_idrs_array(pbundle, attr_uapi,
 | |
| 						 &e->objs_arr_attr, uattr,
 | |
| 						 attr_bkey);
 | |
| 	default:
 | |
| 		return -EOPNOTSUPP;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * We search the radix tree with the method prefix and now we want to fast
 | |
|  * search the suffix bits to get a particular attribute pointer. It is not
 | |
|  * totally clear to me if this breaks the radix tree encasulation or not, but
 | |
|  * it uses the iter data to determine if the method iter points at the same
 | |
|  * chunk that will store the attribute, if so it just derefs it directly. By
 | |
|  * construction in most kernel configs the method and attrs will all fit in a
 | |
|  * single radix chunk, so in most cases this will have no search. Other cases
 | |
|  * this falls back to a full search.
 | |
|  */
 | |
| static void __rcu **uapi_get_attr_for_method(struct bundle_priv *pbundle,
 | |
| 					     u32 attr_key)
 | |
| {
 | |
| 	void __rcu **slot;
 | |
| 
 | |
| 	if (likely(attr_key < pbundle->radix_slots_len)) {
 | |
| 		void *entry;
 | |
| 
 | |
| 		slot = pbundle->radix_slots + attr_key;
 | |
| 		entry = rcu_dereference_raw(*slot);
 | |
| 		if (likely(!radix_tree_is_internal_node(entry) && entry))
 | |
| 			return slot;
 | |
| 	}
 | |
| 
 | |
| 	return radix_tree_lookup_slot(pbundle->radix,
 | |
| 				      pbundle->method_key | attr_key);
 | |
| }
 | |
| 
 | |
| static int uverbs_set_attr(struct bundle_priv *pbundle,
 | |
| 			   struct ib_uverbs_attr *uattr)
 | |
| {
 | |
| 	u32 attr_key = uapi_key_attr(uattr->attr_id);
 | |
| 	u32 attr_bkey = uapi_bkey_attr(attr_key);
 | |
| 	const struct uverbs_api_attr *attr;
 | |
| 	void __rcu **slot;
 | |
| 	int ret;
 | |
| 
 | |
| 	slot = uapi_get_attr_for_method(pbundle, attr_key);
 | |
| 	if (!slot) {
 | |
| 		/*
 | |
| 		 * Kernel does not support the attribute but user-space says it
 | |
| 		 * is mandatory
 | |
| 		 */
 | |
| 		if (uattr->flags & UVERBS_ATTR_F_MANDATORY)
 | |
| 			return -EPROTONOSUPPORT;
 | |
| 		return 0;
 | |
| 	}
 | |
| 	attr = rcu_dereference_protected(*slot, true);
 | |
| 
 | |
| 	/* Reject duplicate attributes from user-space */
 | |
| 	if (test_bit(attr_bkey, pbundle->bundle.attr_present))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	ret = uverbs_process_attr(pbundle, attr, uattr, attr_bkey);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	__set_bit(attr_bkey, pbundle->bundle.attr_present);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int ib_uverbs_run_method(struct bundle_priv *pbundle,
 | |
| 				unsigned int num_attrs)
 | |
| {
 | |
| 	int (*handler)(struct uverbs_attr_bundle *attrs);
 | |
| 	struct uverbs_attr_bundle *bundle =
 | |
| 		container_of(&pbundle->bundle, struct uverbs_attr_bundle, hdr);
 | |
| 	size_t uattrs_size = array_size(sizeof(*pbundle->uattrs), num_attrs);
 | |
| 	unsigned int destroy_bkey = pbundle->method_elm->destroy_bkey;
 | |
| 	unsigned int i;
 | |
| 	int ret;
 | |
| 
 | |
| 	/* See uverbs_disassociate_api() */
 | |
| 	handler = srcu_dereference(
 | |
| 		pbundle->method_elm->handler,
 | |
| 		&pbundle->bundle.ufile->device->disassociate_srcu);
 | |
| 	if (!handler)
 | |
| 		return -EIO;
 | |
| 
 | |
| 	pbundle->uattrs = uverbs_alloc(bundle, uattrs_size);
 | |
| 	if (IS_ERR(pbundle->uattrs))
 | |
| 		return PTR_ERR(pbundle->uattrs);
 | |
| 	if (copy_from_user(pbundle->uattrs, pbundle->user_attrs, uattrs_size))
 | |
| 		return -EFAULT;
 | |
| 
 | |
| 	for (i = 0; i != num_attrs; i++) {
 | |
| 		ret = uverbs_set_attr(pbundle, &pbundle->uattrs[i]);
 | |
| 		if (unlikely(ret))
 | |
| 			return ret;
 | |
| 	}
 | |
| 
 | |
| 	/* User space did not provide all the mandatory attributes */
 | |
| 	if (unlikely(!bitmap_subset(pbundle->method_elm->attr_mandatory,
 | |
| 				    pbundle->bundle.attr_present,
 | |
| 				    pbundle->method_elm->key_bitmap_len)))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	if (pbundle->method_elm->has_udata)
 | |
| 		uverbs_fill_udata(bundle, &pbundle->bundle.driver_udata,
 | |
| 				  UVERBS_ATTR_UHW_IN, UVERBS_ATTR_UHW_OUT);
 | |
| 	else
 | |
| 		pbundle->bundle.driver_udata = (struct ib_udata){};
 | |
| 
 | |
| 	if (destroy_bkey != UVERBS_API_ATTR_BKEY_LEN) {
 | |
| 		struct uverbs_obj_attr *destroy_attr = &bundle->attrs[destroy_bkey].obj_attr;
 | |
| 
 | |
| 		ret = uobj_destroy(destroy_attr->uobject, bundle);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 		__clear_bit(destroy_bkey, pbundle->uobj_finalize);
 | |
| 
 | |
| 		ret = handler(bundle);
 | |
| 		uobj_put_destroy(destroy_attr->uobject);
 | |
| 	} else {
 | |
| 		ret = handler(bundle);
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * Until the drivers are revised to use the bundle directly we have to
 | |
| 	 * assume that the driver wrote to its UHW_OUT and flag userspace
 | |
| 	 * appropriately.
 | |
| 	 */
 | |
| 	if (!ret && pbundle->method_elm->has_udata) {
 | |
| 		const struct uverbs_attr *attr =
 | |
| 			uverbs_attr_get(bundle, UVERBS_ATTR_UHW_OUT);
 | |
| 
 | |
| 		if (!IS_ERR(attr))
 | |
| 			ret = uverbs_set_output(bundle, attr);
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * EPROTONOSUPPORT is ONLY to be returned if the ioctl framework can
 | |
| 	 * not invoke the method because the request is not supported.  No
 | |
| 	 * other cases should return this code.
 | |
| 	 */
 | |
| 	if (WARN_ON_ONCE(ret == -EPROTONOSUPPORT))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static void bundle_destroy(struct bundle_priv *pbundle, bool commit)
 | |
| {
 | |
| 	unsigned int key_bitmap_len = pbundle->method_elm->key_bitmap_len;
 | |
| 	struct uverbs_attr_bundle *bundle =
 | |
| 		container_of(&pbundle->bundle, struct uverbs_attr_bundle, hdr);
 | |
| 	struct bundle_alloc_head *memblock;
 | |
| 	unsigned int i;
 | |
| 
 | |
| 	/* fast path for simple uobjects */
 | |
| 	i = -1;
 | |
| 	while ((i = find_next_bit(pbundle->uobj_finalize, key_bitmap_len,
 | |
| 				  i + 1)) < key_bitmap_len) {
 | |
| 		struct uverbs_attr *attr = &bundle->attrs[i];
 | |
| 
 | |
| 		uverbs_finalize_object(
 | |
| 			attr->obj_attr.uobject,
 | |
| 			attr->obj_attr.attr_elm->spec.u.obj.access,
 | |
| 			test_bit(i, pbundle->uobj_hw_obj_valid),
 | |
| 			commit, bundle);
 | |
| 	}
 | |
| 
 | |
| 	i = -1;
 | |
| 	while ((i = find_next_bit(pbundle->spec_finalize, key_bitmap_len,
 | |
| 				  i + 1)) < key_bitmap_len) {
 | |
| 		struct uverbs_attr *attr = &bundle->attrs[i];
 | |
| 		const struct uverbs_api_attr *attr_uapi;
 | |
| 		void __rcu **slot;
 | |
| 
 | |
| 		slot = uapi_get_attr_for_method(
 | |
| 			pbundle,
 | |
| 			pbundle->method_key | uapi_bkey_to_key_attr(i));
 | |
| 		if (WARN_ON(!slot))
 | |
| 			continue;
 | |
| 
 | |
| 		attr_uapi = rcu_dereference_protected(*slot, true);
 | |
| 
 | |
| 		if (attr_uapi->spec.type == UVERBS_ATTR_TYPE_IDRS_ARRAY) {
 | |
| 			uverbs_free_idrs_array(attr_uapi, &attr->objs_arr_attr,
 | |
| 					       commit, bundle);
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	for (memblock = pbundle->allocated_mem; memblock;) {
 | |
| 		struct bundle_alloc_head *tmp = memblock;
 | |
| 
 | |
| 		memblock = memblock->next;
 | |
| 		kvfree(tmp);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static int ib_uverbs_cmd_verbs(struct ib_uverbs_file *ufile,
 | |
| 			       struct ib_uverbs_ioctl_hdr *hdr,
 | |
| 			       struct ib_uverbs_attr __user *user_attrs)
 | |
| {
 | |
| 	const struct uverbs_api_ioctl_method *method_elm;
 | |
| 	struct uverbs_api *uapi = ufile->device->uapi;
 | |
| 	struct radix_tree_iter attrs_iter;
 | |
| 	struct bundle_priv *pbundle;
 | |
| 	struct bundle_priv onstack;
 | |
| 	void __rcu **slot;
 | |
| 	int ret;
 | |
| 
 | |
| 	if (unlikely(hdr->driver_id != uapi->driver_id))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	slot = radix_tree_iter_lookup(
 | |
| 		&uapi->radix, &attrs_iter,
 | |
| 		uapi_key_obj(hdr->object_id) |
 | |
| 			uapi_key_ioctl_method(hdr->method_id));
 | |
| 	if (unlikely(!slot))
 | |
| 		return -EPROTONOSUPPORT;
 | |
| 	method_elm = rcu_dereference_protected(*slot, true);
 | |
| 
 | |
| 	if (!method_elm->use_stack) {
 | |
| 		pbundle = kmalloc(method_elm->bundle_size, GFP_KERNEL);
 | |
| 		if (!pbundle)
 | |
| 			return -ENOMEM;
 | |
| 		pbundle->internal_avail =
 | |
| 			method_elm->bundle_size -
 | |
| 			offsetof(struct bundle_priv, internal_buffer);
 | |
| 		pbundle->alloc_head.next = NULL;
 | |
| 		pbundle->allocated_mem = container_of(&pbundle->alloc_head,
 | |
| 						struct bundle_alloc_head, hdr);
 | |
| 	} else {
 | |
| 		pbundle = &onstack;
 | |
| 		pbundle->internal_avail = sizeof(pbundle->internal_buffer);
 | |
| 		pbundle->allocated_mem = NULL;
 | |
| 	}
 | |
| 
 | |
| 	/* Space for the pbundle->bundle.attrs flex array */
 | |
| 	pbundle->method_elm = method_elm;
 | |
| 	pbundle->method_key = attrs_iter.index;
 | |
| 	pbundle->bundle.ufile = ufile;
 | |
| 	pbundle->bundle.context = NULL; /* only valid if bundle has uobject */
 | |
| 	pbundle->radix = &uapi->radix;
 | |
| 	pbundle->radix_slots = slot;
 | |
| 	pbundle->radix_slots_len = radix_tree_chunk_size(&attrs_iter);
 | |
| 	pbundle->user_attrs = user_attrs;
 | |
| 
 | |
| 	pbundle->internal_used = ALIGN(pbundle->method_elm->key_bitmap_len *
 | |
| 					       sizeof(*container_of(&pbundle->bundle,
 | |
| 							struct uverbs_attr_bundle, hdr)->attrs),
 | |
| 					       sizeof(*pbundle->internal_buffer));
 | |
| 	memset(pbundle->bundle.attr_present, 0,
 | |
| 	       sizeof(pbundle->bundle.attr_present));
 | |
| 	memset(pbundle->uobj_finalize, 0, sizeof(pbundle->uobj_finalize));
 | |
| 	memset(pbundle->spec_finalize, 0, sizeof(pbundle->spec_finalize));
 | |
| 	memset(pbundle->uobj_hw_obj_valid, 0,
 | |
| 	       sizeof(pbundle->uobj_hw_obj_valid));
 | |
| 
 | |
| 	ret = ib_uverbs_run_method(pbundle, hdr->num_attrs);
 | |
| 	bundle_destroy(pbundle, ret == 0);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| long ib_uverbs_ioctl(struct file *filp, unsigned int cmd, unsigned long arg)
 | |
| {
 | |
| 	struct ib_uverbs_file *file = filp->private_data;
 | |
| 	struct ib_uverbs_ioctl_hdr __user *user_hdr =
 | |
| 		(struct ib_uverbs_ioctl_hdr __user *)arg;
 | |
| 	struct ib_uverbs_ioctl_hdr hdr;
 | |
| 	int srcu_key;
 | |
| 	int err;
 | |
| 
 | |
| 	if (unlikely(cmd != RDMA_VERBS_IOCTL))
 | |
| 		return -ENOIOCTLCMD;
 | |
| 
 | |
| 	err = copy_from_user(&hdr, user_hdr, sizeof(hdr));
 | |
| 	if (err)
 | |
| 		return -EFAULT;
 | |
| 
 | |
| 	if (hdr.length > PAGE_SIZE ||
 | |
| 	    hdr.length != struct_size(&hdr, attrs, hdr.num_attrs))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	if (hdr.reserved1 || hdr.reserved2)
 | |
| 		return -EPROTONOSUPPORT;
 | |
| 
 | |
| 	srcu_key = srcu_read_lock(&file->device->disassociate_srcu);
 | |
| 	err = ib_uverbs_cmd_verbs(file, &hdr, user_hdr->attrs);
 | |
| 	srcu_read_unlock(&file->device->disassociate_srcu, srcu_key);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| int uverbs_get_flags64(u64 *to, const struct uverbs_attr_bundle *attrs_bundle,
 | |
| 		       size_t idx, u64 allowed_bits)
 | |
| {
 | |
| 	const struct uverbs_attr *attr;
 | |
| 	u64 flags;
 | |
| 
 | |
| 	attr = uverbs_attr_get(attrs_bundle, idx);
 | |
| 	/* Missing attribute means 0 flags */
 | |
| 	if (IS_ERR(attr)) {
 | |
| 		*to = 0;
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * New userspace code should use 8 bytes to pass flags, but we
 | |
| 	 * transparently support old userspaces that were using 4 bytes as
 | |
| 	 * well.
 | |
| 	 */
 | |
| 	if (attr->ptr_attr.len == 8)
 | |
| 		flags = attr->ptr_attr.data;
 | |
| 	else if (attr->ptr_attr.len == 4)
 | |
| 		flags = *(u32 *)&attr->ptr_attr.data;
 | |
| 	else
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	if (flags & ~allowed_bits)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	*to = flags;
 | |
| 	return 0;
 | |
| }
 | |
| EXPORT_SYMBOL(uverbs_get_flags64);
 | |
| 
 | |
| int uverbs_get_flags32(u32 *to, const struct uverbs_attr_bundle *attrs_bundle,
 | |
| 		       size_t idx, u64 allowed_bits)
 | |
| {
 | |
| 	u64 flags;
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = uverbs_get_flags64(&flags, attrs_bundle, idx, allowed_bits);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	if (flags > U32_MAX)
 | |
| 		return -EINVAL;
 | |
| 	*to = flags;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| EXPORT_SYMBOL(uverbs_get_flags32);
 | |
| 
 | |
| /*
 | |
|  * Fill a ib_udata struct (core or uhw) using the given attribute IDs.
 | |
|  * This is primarily used to convert the UVERBS_ATTR_UHW() into the
 | |
|  * ib_udata format used by the drivers.
 | |
|  */
 | |
| void uverbs_fill_udata(struct uverbs_attr_bundle *bundle,
 | |
| 		       struct ib_udata *udata, unsigned int attr_in,
 | |
| 		       unsigned int attr_out)
 | |
| {
 | |
| 	struct bundle_priv *pbundle =
 | |
| 		container_of(&bundle->hdr, struct bundle_priv, bundle);
 | |
| 	struct uverbs_attr_bundle *bundle_aux =
 | |
| 		container_of(&pbundle->bundle, struct uverbs_attr_bundle, hdr);
 | |
| 	const struct uverbs_attr *in =
 | |
| 		uverbs_attr_get(bundle_aux, attr_in);
 | |
| 	const struct uverbs_attr *out =
 | |
| 		uverbs_attr_get(bundle_aux, attr_out);
 | |
| 
 | |
| 	if (!IS_ERR(in)) {
 | |
| 		udata->inlen = in->ptr_attr.len;
 | |
| 		if (uverbs_attr_ptr_is_inline(in))
 | |
| 			udata->inbuf =
 | |
| 				&pbundle->user_attrs[in->ptr_attr.uattr_idx]
 | |
| 					 .data;
 | |
| 		else
 | |
| 			udata->inbuf = u64_to_user_ptr(in->ptr_attr.data);
 | |
| 	} else {
 | |
| 		udata->inbuf = NULL;
 | |
| 		udata->inlen = 0;
 | |
| 	}
 | |
| 
 | |
| 	if (!IS_ERR(out)) {
 | |
| 		udata->outbuf = u64_to_user_ptr(out->ptr_attr.data);
 | |
| 		udata->outlen = out->ptr_attr.len;
 | |
| 	} else {
 | |
| 		udata->outbuf = NULL;
 | |
| 		udata->outlen = 0;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| int uverbs_copy_to(const struct uverbs_attr_bundle *bundle, size_t idx,
 | |
| 		   const void *from, size_t size)
 | |
| {
 | |
| 	const struct uverbs_attr *attr = uverbs_attr_get(bundle, idx);
 | |
| 	size_t min_size;
 | |
| 
 | |
| 	if (IS_ERR(attr))
 | |
| 		return PTR_ERR(attr);
 | |
| 
 | |
| 	min_size = min_t(size_t, attr->ptr_attr.len, size);
 | |
| 	if (copy_to_user(u64_to_user_ptr(attr->ptr_attr.data), from, min_size))
 | |
| 		return -EFAULT;
 | |
| 
 | |
| 	return uverbs_set_output(bundle, attr);
 | |
| }
 | |
| EXPORT_SYMBOL(uverbs_copy_to);
 | |
| 
 | |
| 
 | |
| /*
 | |
|  * This is only used if the caller has directly used copy_to_use to write the
 | |
|  * data.  It signals to user space that the buffer is filled in.
 | |
|  */
 | |
| int uverbs_output_written(const struct uverbs_attr_bundle *bundle, size_t idx)
 | |
| {
 | |
| 	const struct uverbs_attr *attr = uverbs_attr_get(bundle, idx);
 | |
| 
 | |
| 	if (IS_ERR(attr))
 | |
| 		return PTR_ERR(attr);
 | |
| 
 | |
| 	return uverbs_set_output(bundle, attr);
 | |
| }
 | |
| 
 | |
| int _uverbs_get_const_signed(s64 *to,
 | |
| 			     const struct uverbs_attr_bundle *attrs_bundle,
 | |
| 			     size_t idx, s64 lower_bound, u64 upper_bound,
 | |
| 			     s64  *def_val)
 | |
| {
 | |
| 	const struct uverbs_attr *attr;
 | |
| 
 | |
| 	attr = uverbs_attr_get(attrs_bundle, idx);
 | |
| 	if (IS_ERR(attr)) {
 | |
| 		if ((PTR_ERR(attr) != -ENOENT) || !def_val)
 | |
| 			return PTR_ERR(attr);
 | |
| 
 | |
| 		*to = *def_val;
 | |
| 	} else {
 | |
| 		*to = attr->ptr_attr.data;
 | |
| 	}
 | |
| 
 | |
| 	if (*to < lower_bound || (*to > 0 && (u64)*to > upper_bound))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| EXPORT_SYMBOL(_uverbs_get_const_signed);
 | |
| 
 | |
| int _uverbs_get_const_unsigned(u64 *to,
 | |
| 			       const struct uverbs_attr_bundle *attrs_bundle,
 | |
| 			       size_t idx, u64 upper_bound, u64 *def_val)
 | |
| {
 | |
| 	const struct uverbs_attr *attr;
 | |
| 
 | |
| 	attr = uverbs_attr_get(attrs_bundle, idx);
 | |
| 	if (IS_ERR(attr)) {
 | |
| 		if ((PTR_ERR(attr) != -ENOENT) || !def_val)
 | |
| 			return PTR_ERR(attr);
 | |
| 
 | |
| 		*to = *def_val;
 | |
| 	} else {
 | |
| 		*to = attr->ptr_attr.data;
 | |
| 	}
 | |
| 
 | |
| 	if (*to > upper_bound)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| EXPORT_SYMBOL(_uverbs_get_const_unsigned);
 | |
| 
 | |
| int uverbs_copy_to_struct_or_zero(const struct uverbs_attr_bundle *bundle,
 | |
| 				  size_t idx, const void *from, size_t size)
 | |
| {
 | |
| 	const struct uverbs_attr *attr = uverbs_attr_get(bundle, idx);
 | |
| 
 | |
| 	if (IS_ERR(attr))
 | |
| 		return PTR_ERR(attr);
 | |
| 
 | |
| 	if (size < attr->ptr_attr.len) {
 | |
| 		if (clear_user(u64_to_user_ptr(attr->ptr_attr.data) + size,
 | |
| 			       attr->ptr_attr.len - size))
 | |
| 			return -EFAULT;
 | |
| 	}
 | |
| 	return uverbs_copy_to(bundle, idx, from, size);
 | |
| }
 | |
| EXPORT_SYMBOL(uverbs_copy_to_struct_or_zero);
 | |
| 
 | |
| /* Once called an abort will call through to the type's destroy_hw() */
 | |
| void uverbs_finalize_uobj_create(const struct uverbs_attr_bundle *bundle,
 | |
| 				 u16 idx)
 | |
| {
 | |
| 	struct bundle_priv *pbundle =
 | |
| 		container_of(&bundle->hdr, struct bundle_priv, bundle);
 | |
| 
 | |
| 	__set_bit(uapi_bkey_attr(uapi_key_attr(idx)),
 | |
| 		  pbundle->uobj_hw_obj_valid);
 | |
| }
 | |
| EXPORT_SYMBOL(uverbs_finalize_uobj_create);
 |