mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 02:30:34 +02:00 
			
		
		
		
	Add a NULL check before accessing device memory to prevent a crash if
dev->dm allocation in mlx5_init_once() fails.
Fixes: c9b9dcb430 ("net/mlx5: Move device memory management to mlx5_core")
Signed-off-by: Stav Aviram <saviram@nvidia.com>
Link: https://patch.msgid.link/c88711327f4d74d5cebc730dc629607e989ca187.1751370035.git.leon@kernel.org
Signed-off-by: Leon Romanovsky <leon@kernel.org>
		
	
			
		
			
				
	
	
		
			612 lines
		
	
	
	
		
			17 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			612 lines
		
	
	
	
		
			17 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
// SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
 | 
						|
/*
 | 
						|
 * Copyright (c) 2021, Mellanox Technologies inc. All rights reserved.
 | 
						|
 */
 | 
						|
 | 
						|
#include <rdma/uverbs_std_types.h>
 | 
						|
#include "dm.h"
 | 
						|
 | 
						|
#define UVERBS_MODULE_NAME mlx5_ib
 | 
						|
#include <rdma/uverbs_named_ioctl.h>
 | 
						|
 | 
						|
static int mlx5_cmd_alloc_memic(struct mlx5_dm *dm, phys_addr_t *addr,
 | 
						|
				u64 length, u32 alignment)
 | 
						|
{
 | 
						|
	struct mlx5_core_dev *dev = dm->dev;
 | 
						|
	u64 num_memic_hw_pages = MLX5_CAP_DEV_MEM(dev, memic_bar_size)
 | 
						|
					>> PAGE_SHIFT;
 | 
						|
	u64 hw_start_addr = MLX5_CAP64_DEV_MEM(dev, memic_bar_start_addr);
 | 
						|
	u32 max_alignment = MLX5_CAP_DEV_MEM(dev, log_max_memic_addr_alignment);
 | 
						|
	u32 num_pages = DIV_ROUND_UP(length, PAGE_SIZE);
 | 
						|
	u32 out[MLX5_ST_SZ_DW(alloc_memic_out)] = {};
 | 
						|
	u32 in[MLX5_ST_SZ_DW(alloc_memic_in)] = {};
 | 
						|
	u32 mlx5_alignment;
 | 
						|
	u64 page_idx = 0;
 | 
						|
	int ret = 0;
 | 
						|
 | 
						|
	if (!length || (length & MLX5_MEMIC_ALLOC_SIZE_MASK))
 | 
						|
		return -EINVAL;
 | 
						|
 | 
						|
	/* mlx5 device sets alignment as 64*2^driver_value
 | 
						|
	 * so normalizing is needed.
 | 
						|
	 */
 | 
						|
	mlx5_alignment = (alignment < MLX5_MEMIC_BASE_ALIGN) ? 0 :
 | 
						|
			 alignment - MLX5_MEMIC_BASE_ALIGN;
 | 
						|
	if (mlx5_alignment > max_alignment)
 | 
						|
		return -EINVAL;
 | 
						|
 | 
						|
	MLX5_SET(alloc_memic_in, in, opcode, MLX5_CMD_OP_ALLOC_MEMIC);
 | 
						|
	MLX5_SET(alloc_memic_in, in, range_size, num_pages * PAGE_SIZE);
 | 
						|
	MLX5_SET(alloc_memic_in, in, memic_size, length);
 | 
						|
	MLX5_SET(alloc_memic_in, in, log_memic_addr_alignment,
 | 
						|
		 mlx5_alignment);
 | 
						|
 | 
						|
	while (page_idx < num_memic_hw_pages) {
 | 
						|
		spin_lock(&dm->lock);
 | 
						|
		page_idx = bitmap_find_next_zero_area(dm->memic_alloc_pages,
 | 
						|
						      num_memic_hw_pages,
 | 
						|
						      page_idx,
 | 
						|
						      num_pages, 0);
 | 
						|
 | 
						|
		if (page_idx < num_memic_hw_pages)
 | 
						|
			bitmap_set(dm->memic_alloc_pages,
 | 
						|
				   page_idx, num_pages);
 | 
						|
 | 
						|
		spin_unlock(&dm->lock);
 | 
						|
 | 
						|
		if (page_idx >= num_memic_hw_pages)
 | 
						|
			break;
 | 
						|
 | 
						|
		MLX5_SET64(alloc_memic_in, in, range_start_addr,
 | 
						|
			   hw_start_addr + (page_idx * PAGE_SIZE));
 | 
						|
 | 
						|
		ret = mlx5_cmd_exec_inout(dev, alloc_memic, in, out);
 | 
						|
		if (ret) {
 | 
						|
			spin_lock(&dm->lock);
 | 
						|
			bitmap_clear(dm->memic_alloc_pages,
 | 
						|
				     page_idx, num_pages);
 | 
						|
			spin_unlock(&dm->lock);
 | 
						|
 | 
						|
			if (ret == -EAGAIN) {
 | 
						|
				page_idx++;
 | 
						|
				continue;
 | 
						|
			}
 | 
						|
 | 
						|
			return ret;
 | 
						|
		}
 | 
						|
 | 
						|
		*addr = dev->bar_addr +
 | 
						|
			MLX5_GET64(alloc_memic_out, out, memic_start_addr);
 | 
						|
 | 
						|
		return 0;
 | 
						|
	}
 | 
						|
 | 
						|
	return -ENOMEM;
 | 
						|
}
 | 
						|
 | 
						|
void mlx5_cmd_dealloc_memic(struct mlx5_dm *dm, phys_addr_t addr,
 | 
						|
			    u64 length)
 | 
						|
{
 | 
						|
	struct mlx5_core_dev *dev = dm->dev;
 | 
						|
	u64 hw_start_addr = MLX5_CAP64_DEV_MEM(dev, memic_bar_start_addr);
 | 
						|
	u32 num_pages = DIV_ROUND_UP(length, PAGE_SIZE);
 | 
						|
	u32 in[MLX5_ST_SZ_DW(dealloc_memic_in)] = {};
 | 
						|
	u64 start_page_idx;
 | 
						|
	int err;
 | 
						|
 | 
						|
	addr -= dev->bar_addr;
 | 
						|
	start_page_idx = (addr - hw_start_addr) >> PAGE_SHIFT;
 | 
						|
 | 
						|
	MLX5_SET(dealloc_memic_in, in, opcode, MLX5_CMD_OP_DEALLOC_MEMIC);
 | 
						|
	MLX5_SET64(dealloc_memic_in, in, memic_start_addr, addr);
 | 
						|
	MLX5_SET(dealloc_memic_in, in, memic_size, length);
 | 
						|
 | 
						|
	err =  mlx5_cmd_exec_in(dev, dealloc_memic, in);
 | 
						|
	if (err)
 | 
						|
		return;
 | 
						|
 | 
						|
	spin_lock(&dm->lock);
 | 
						|
	bitmap_clear(dm->memic_alloc_pages,
 | 
						|
		     start_page_idx, num_pages);
 | 
						|
	spin_unlock(&dm->lock);
 | 
						|
}
 | 
						|
 | 
						|
void mlx5_cmd_dealloc_memic_op(struct mlx5_dm *dm, phys_addr_t addr,
 | 
						|
			       u8 operation)
 | 
						|
{
 | 
						|
	u32 in[MLX5_ST_SZ_DW(modify_memic_in)] = {};
 | 
						|
	struct mlx5_core_dev *dev = dm->dev;
 | 
						|
 | 
						|
	MLX5_SET(modify_memic_in, in, opcode, MLX5_CMD_OP_MODIFY_MEMIC);
 | 
						|
	MLX5_SET(modify_memic_in, in, op_mod, MLX5_MODIFY_MEMIC_OP_MOD_DEALLOC);
 | 
						|
	MLX5_SET(modify_memic_in, in, memic_operation_type, operation);
 | 
						|
	MLX5_SET64(modify_memic_in, in, memic_start_addr, addr - dev->bar_addr);
 | 
						|
 | 
						|
	mlx5_cmd_exec_in(dev, modify_memic, in);
 | 
						|
}
 | 
						|
 | 
						|
static int mlx5_cmd_alloc_memic_op(struct mlx5_dm *dm, phys_addr_t addr,
 | 
						|
				   u8 operation, phys_addr_t *op_addr)
 | 
						|
{
 | 
						|
	u32 out[MLX5_ST_SZ_DW(modify_memic_out)] = {};
 | 
						|
	u32 in[MLX5_ST_SZ_DW(modify_memic_in)] = {};
 | 
						|
	struct mlx5_core_dev *dev = dm->dev;
 | 
						|
	int err;
 | 
						|
 | 
						|
	MLX5_SET(modify_memic_in, in, opcode, MLX5_CMD_OP_MODIFY_MEMIC);
 | 
						|
	MLX5_SET(modify_memic_in, in, op_mod, MLX5_MODIFY_MEMIC_OP_MOD_ALLOC);
 | 
						|
	MLX5_SET(modify_memic_in, in, memic_operation_type, operation);
 | 
						|
	MLX5_SET64(modify_memic_in, in, memic_start_addr, addr - dev->bar_addr);
 | 
						|
 | 
						|
	err = mlx5_cmd_exec_inout(dev, modify_memic, in, out);
 | 
						|
	if (err)
 | 
						|
		return err;
 | 
						|
 | 
						|
	*op_addr = dev->bar_addr +
 | 
						|
		   MLX5_GET64(modify_memic_out, out, memic_operation_addr);
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int add_dm_mmap_entry(struct ib_ucontext *context,
 | 
						|
			     struct mlx5_user_mmap_entry *mentry, u8 mmap_flag,
 | 
						|
			     size_t size, u64 address)
 | 
						|
{
 | 
						|
	mentry->mmap_flag = mmap_flag;
 | 
						|
	mentry->address = address;
 | 
						|
 | 
						|
	return rdma_user_mmap_entry_insert_range(
 | 
						|
		context, &mentry->rdma_entry, size,
 | 
						|
		MLX5_IB_MMAP_DEVICE_MEM << 16,
 | 
						|
		(MLX5_IB_MMAP_DEVICE_MEM << 16) + (1UL << 16) - 1);
 | 
						|
}
 | 
						|
 | 
						|
static void mlx5_ib_dm_memic_free(struct kref *kref)
 | 
						|
{
 | 
						|
	struct mlx5_ib_dm_memic *dm =
 | 
						|
		container_of(kref, struct mlx5_ib_dm_memic, ref);
 | 
						|
	struct mlx5_ib_dev *dev = to_mdev(dm->base.ibdm.device);
 | 
						|
 | 
						|
	mlx5_cmd_dealloc_memic(&dev->dm, dm->base.dev_addr, dm->base.size);
 | 
						|
	kfree(dm);
 | 
						|
}
 | 
						|
 | 
						|
static int copy_op_to_user(struct mlx5_ib_dm_op_entry *op_entry,
 | 
						|
			   struct uverbs_attr_bundle *attrs)
 | 
						|
{
 | 
						|
	u64 start_offset;
 | 
						|
	u16 page_idx;
 | 
						|
	int err;
 | 
						|
 | 
						|
	page_idx = op_entry->mentry.rdma_entry.start_pgoff & 0xFFFF;
 | 
						|
	start_offset = op_entry->op_addr & ~PAGE_MASK;
 | 
						|
	err = uverbs_copy_to(attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_PAGE_INDEX,
 | 
						|
			     &page_idx, sizeof(page_idx));
 | 
						|
	if (err)
 | 
						|
		return err;
 | 
						|
 | 
						|
	return uverbs_copy_to(attrs,
 | 
						|
			      MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_START_OFFSET,
 | 
						|
			      &start_offset, sizeof(start_offset));
 | 
						|
}
 | 
						|
 | 
						|
static int map_existing_op(struct mlx5_ib_dm_memic *dm, u8 op,
 | 
						|
			   struct uverbs_attr_bundle *attrs)
 | 
						|
{
 | 
						|
	struct mlx5_ib_dm_op_entry *op_entry;
 | 
						|
 | 
						|
	op_entry = xa_load(&dm->ops, op);
 | 
						|
	if (!op_entry)
 | 
						|
		return -ENOENT;
 | 
						|
 | 
						|
	return copy_op_to_user(op_entry, attrs);
 | 
						|
}
 | 
						|
 | 
						|
static int UVERBS_HANDLER(MLX5_IB_METHOD_DM_MAP_OP_ADDR)(
 | 
						|
	struct uverbs_attr_bundle *attrs)
 | 
						|
{
 | 
						|
	struct ib_uobject *uobj = uverbs_attr_get_uobject(
 | 
						|
		attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_HANDLE);
 | 
						|
	struct mlx5_ib_dev *dev = to_mdev(uobj->context->device);
 | 
						|
	struct ib_dm *ibdm = uobj->object;
 | 
						|
	struct mlx5_ib_dm_memic *dm = to_memic(ibdm);
 | 
						|
	struct mlx5_ib_dm_op_entry *op_entry;
 | 
						|
	int err;
 | 
						|
	u8 op;
 | 
						|
 | 
						|
	err = uverbs_copy_from(&op, attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_OP);
 | 
						|
	if (err)
 | 
						|
		return err;
 | 
						|
 | 
						|
	if (op >= BITS_PER_TYPE(u32))
 | 
						|
		return -EOPNOTSUPP;
 | 
						|
 | 
						|
	if (!(MLX5_CAP_DEV_MEM(dev->mdev, memic_operations) & BIT(op)))
 | 
						|
		return -EOPNOTSUPP;
 | 
						|
 | 
						|
	mutex_lock(&dm->ops_xa_lock);
 | 
						|
	err = map_existing_op(dm, op, attrs);
 | 
						|
	if (!err || err != -ENOENT)
 | 
						|
		goto err_unlock;
 | 
						|
 | 
						|
	op_entry = kzalloc(sizeof(*op_entry), GFP_KERNEL);
 | 
						|
	if (!op_entry)
 | 
						|
		goto err_unlock;
 | 
						|
 | 
						|
	err = mlx5_cmd_alloc_memic_op(&dev->dm, dm->base.dev_addr, op,
 | 
						|
				      &op_entry->op_addr);
 | 
						|
	if (err) {
 | 
						|
		kfree(op_entry);
 | 
						|
		goto err_unlock;
 | 
						|
	}
 | 
						|
	op_entry->op = op;
 | 
						|
	op_entry->dm = dm;
 | 
						|
 | 
						|
	err = add_dm_mmap_entry(uobj->context, &op_entry->mentry,
 | 
						|
				MLX5_IB_MMAP_TYPE_MEMIC_OP, dm->base.size,
 | 
						|
				op_entry->op_addr & PAGE_MASK);
 | 
						|
	if (err) {
 | 
						|
		mlx5_cmd_dealloc_memic_op(&dev->dm, dm->base.dev_addr, op);
 | 
						|
		kfree(op_entry);
 | 
						|
		goto err_unlock;
 | 
						|
	}
 | 
						|
	/* From this point, entry will be freed by mmap_free */
 | 
						|
	kref_get(&dm->ref);
 | 
						|
 | 
						|
	err = copy_op_to_user(op_entry, attrs);
 | 
						|
	if (err)
 | 
						|
		goto err_remove;
 | 
						|
 | 
						|
	err = xa_insert(&dm->ops, op, op_entry, GFP_KERNEL);
 | 
						|
	if (err)
 | 
						|
		goto err_remove;
 | 
						|
	mutex_unlock(&dm->ops_xa_lock);
 | 
						|
 | 
						|
	return 0;
 | 
						|
 | 
						|
err_remove:
 | 
						|
	rdma_user_mmap_entry_remove(&op_entry->mentry.rdma_entry);
 | 
						|
err_unlock:
 | 
						|
	mutex_unlock(&dm->ops_xa_lock);
 | 
						|
 | 
						|
	return err;
 | 
						|
}
 | 
						|
 | 
						|
static struct ib_dm *handle_alloc_dm_memic(struct ib_ucontext *ctx,
 | 
						|
					   struct ib_dm_alloc_attr *attr,
 | 
						|
					   struct uverbs_attr_bundle *attrs)
 | 
						|
{
 | 
						|
	struct mlx5_dm *dm_db = &to_mdev(ctx->device)->dm;
 | 
						|
	struct mlx5_ib_dm_memic *dm;
 | 
						|
	u64 start_offset;
 | 
						|
	u16 page_idx;
 | 
						|
	int err;
 | 
						|
	u64 address;
 | 
						|
 | 
						|
	if (!dm_db || !MLX5_CAP_DEV_MEM(dm_db->dev, memic))
 | 
						|
		return ERR_PTR(-EOPNOTSUPP);
 | 
						|
 | 
						|
	dm = kzalloc(sizeof(*dm), GFP_KERNEL);
 | 
						|
	if (!dm)
 | 
						|
		return ERR_PTR(-ENOMEM);
 | 
						|
 | 
						|
	dm->base.type = MLX5_IB_UAPI_DM_TYPE_MEMIC;
 | 
						|
	dm->base.size = roundup(attr->length, MLX5_MEMIC_BASE_SIZE);
 | 
						|
	dm->base.ibdm.device = ctx->device;
 | 
						|
 | 
						|
	kref_init(&dm->ref);
 | 
						|
	xa_init(&dm->ops);
 | 
						|
	mutex_init(&dm->ops_xa_lock);
 | 
						|
	dm->req_length = attr->length;
 | 
						|
 | 
						|
	err = mlx5_cmd_alloc_memic(dm_db, &dm->base.dev_addr,
 | 
						|
				   dm->base.size, attr->alignment);
 | 
						|
	if (err) {
 | 
						|
		kfree(dm);
 | 
						|
		return ERR_PTR(err);
 | 
						|
	}
 | 
						|
 | 
						|
	address = dm->base.dev_addr & PAGE_MASK;
 | 
						|
	err = add_dm_mmap_entry(ctx, &dm->mentry, MLX5_IB_MMAP_TYPE_MEMIC,
 | 
						|
				dm->base.size, address);
 | 
						|
	if (err) {
 | 
						|
		mlx5_cmd_dealloc_memic(dm_db, dm->base.dev_addr, dm->base.size);
 | 
						|
		kfree(dm);
 | 
						|
		return ERR_PTR(err);
 | 
						|
	}
 | 
						|
 | 
						|
	page_idx = dm->mentry.rdma_entry.start_pgoff & 0xFFFF;
 | 
						|
	err = uverbs_copy_to(attrs, MLX5_IB_ATTR_ALLOC_DM_RESP_PAGE_INDEX,
 | 
						|
			     &page_idx, sizeof(page_idx));
 | 
						|
	if (err)
 | 
						|
		goto err_copy;
 | 
						|
 | 
						|
	start_offset = dm->base.dev_addr & ~PAGE_MASK;
 | 
						|
	err = uverbs_copy_to(attrs,
 | 
						|
			     MLX5_IB_ATTR_ALLOC_DM_RESP_START_OFFSET,
 | 
						|
			     &start_offset, sizeof(start_offset));
 | 
						|
	if (err)
 | 
						|
		goto err_copy;
 | 
						|
 | 
						|
	return &dm->base.ibdm;
 | 
						|
 | 
						|
err_copy:
 | 
						|
	rdma_user_mmap_entry_remove(&dm->mentry.rdma_entry);
 | 
						|
	return ERR_PTR(err);
 | 
						|
}
 | 
						|
 | 
						|
static enum mlx5_sw_icm_type get_icm_type(int uapi_type)
 | 
						|
{
 | 
						|
	switch (uapi_type) {
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM:
 | 
						|
		return MLX5_SW_ICM_TYPE_HEADER_MODIFY;
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM:
 | 
						|
		return MLX5_SW_ICM_TYPE_HEADER_MODIFY_PATTERN;
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_ENCAP_SW_ICM:
 | 
						|
		return MLX5_SW_ICM_TYPE_SW_ENCAP;
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM:
 | 
						|
	default:
 | 
						|
		return MLX5_SW_ICM_TYPE_STEERING;
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static struct ib_dm *handle_alloc_dm_sw_icm(struct ib_ucontext *ctx,
 | 
						|
					    struct ib_dm_alloc_attr *attr,
 | 
						|
					    struct uverbs_attr_bundle *attrs,
 | 
						|
					    int type)
 | 
						|
{
 | 
						|
	struct mlx5_core_dev *dev = to_mdev(ctx->device)->mdev;
 | 
						|
	enum mlx5_sw_icm_type icm_type;
 | 
						|
	struct mlx5_ib_dm_icm *dm;
 | 
						|
	u64 act_size;
 | 
						|
	int err;
 | 
						|
 | 
						|
	if (!capable(CAP_SYS_RAWIO) || !capable(CAP_NET_RAW))
 | 
						|
		return ERR_PTR(-EPERM);
 | 
						|
 | 
						|
	switch (type) {
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM:
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM:
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_ENCAP_SW_ICM:
 | 
						|
		if (!(MLX5_CAP_FLOWTABLE_NIC_RX(dev, sw_owner) ||
 | 
						|
		      MLX5_CAP_FLOWTABLE_NIC_TX(dev, sw_owner) ||
 | 
						|
		      MLX5_CAP_FLOWTABLE_NIC_RX(dev, sw_owner_v2) ||
 | 
						|
		      MLX5_CAP_FLOWTABLE_NIC_TX(dev, sw_owner_v2)))
 | 
						|
			return ERR_PTR(-EOPNOTSUPP);
 | 
						|
		break;
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM:
 | 
						|
		if (!MLX5_CAP_FLOWTABLE_NIC_RX(dev, sw_owner_v2) ||
 | 
						|
		    !MLX5_CAP_FLOWTABLE_NIC_TX(dev, sw_owner_v2))
 | 
						|
			return ERR_PTR(-EOPNOTSUPP);
 | 
						|
		break;
 | 
						|
	default:
 | 
						|
		return ERR_PTR(-EOPNOTSUPP);
 | 
						|
	}
 | 
						|
 | 
						|
	dm = kzalloc(sizeof(*dm), GFP_KERNEL);
 | 
						|
	if (!dm)
 | 
						|
		return ERR_PTR(-ENOMEM);
 | 
						|
 | 
						|
	dm->base.type = type;
 | 
						|
	dm->base.ibdm.device = ctx->device;
 | 
						|
 | 
						|
	/* Allocation size must a multiple of the basic block size
 | 
						|
	 * and a power of 2.
 | 
						|
	 */
 | 
						|
	act_size = round_up(attr->length, MLX5_SW_ICM_BLOCK_SIZE(dev));
 | 
						|
	act_size = roundup_pow_of_two(act_size);
 | 
						|
 | 
						|
	dm->base.size = act_size;
 | 
						|
	icm_type = get_icm_type(type);
 | 
						|
 | 
						|
	err = mlx5_dm_sw_icm_alloc(dev, icm_type, act_size, attr->alignment,
 | 
						|
				   to_mucontext(ctx)->devx_uid,
 | 
						|
				   &dm->base.dev_addr, &dm->obj_id);
 | 
						|
	if (err)
 | 
						|
		goto free;
 | 
						|
 | 
						|
	err = uverbs_copy_to(attrs, MLX5_IB_ATTR_ALLOC_DM_RESP_START_OFFSET,
 | 
						|
			     &dm->base.dev_addr, sizeof(dm->base.dev_addr));
 | 
						|
	if (err) {
 | 
						|
		mlx5_dm_sw_icm_dealloc(dev, icm_type, dm->base.size,
 | 
						|
				       to_mucontext(ctx)->devx_uid,
 | 
						|
				       dm->base.dev_addr, dm->obj_id);
 | 
						|
		goto free;
 | 
						|
	}
 | 
						|
	return &dm->base.ibdm;
 | 
						|
free:
 | 
						|
	kfree(dm);
 | 
						|
	return ERR_PTR(err);
 | 
						|
}
 | 
						|
 | 
						|
struct ib_dm *mlx5_ib_alloc_dm(struct ib_device *ibdev,
 | 
						|
			       struct ib_ucontext *context,
 | 
						|
			       struct ib_dm_alloc_attr *attr,
 | 
						|
			       struct uverbs_attr_bundle *attrs)
 | 
						|
{
 | 
						|
	enum mlx5_ib_uapi_dm_type type;
 | 
						|
	int err;
 | 
						|
 | 
						|
	err = uverbs_get_const_default(&type, attrs,
 | 
						|
				       MLX5_IB_ATTR_ALLOC_DM_REQ_TYPE,
 | 
						|
				       MLX5_IB_UAPI_DM_TYPE_MEMIC);
 | 
						|
	if (err)
 | 
						|
		return ERR_PTR(err);
 | 
						|
 | 
						|
	mlx5_ib_dbg(to_mdev(ibdev), "alloc_dm req: dm_type=%d user_length=0x%llx log_alignment=%d\n",
 | 
						|
		    type, attr->length, attr->alignment);
 | 
						|
 | 
						|
	switch (type) {
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_MEMIC:
 | 
						|
		return handle_alloc_dm_memic(context, attr, attrs);
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM:
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM:
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM:
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_ENCAP_SW_ICM:
 | 
						|
		return handle_alloc_dm_sw_icm(context, attr, attrs, type);
 | 
						|
	default:
 | 
						|
		return ERR_PTR(-EOPNOTSUPP);
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static void dm_memic_remove_ops(struct mlx5_ib_dm_memic *dm)
 | 
						|
{
 | 
						|
	struct mlx5_ib_dm_op_entry *entry;
 | 
						|
	unsigned long idx;
 | 
						|
 | 
						|
	mutex_lock(&dm->ops_xa_lock);
 | 
						|
	xa_for_each(&dm->ops, idx, entry) {
 | 
						|
		xa_erase(&dm->ops, idx);
 | 
						|
		rdma_user_mmap_entry_remove(&entry->mentry.rdma_entry);
 | 
						|
	}
 | 
						|
	mutex_unlock(&dm->ops_xa_lock);
 | 
						|
}
 | 
						|
 | 
						|
static void mlx5_dm_memic_dealloc(struct mlx5_ib_dm_memic *dm)
 | 
						|
{
 | 
						|
	dm_memic_remove_ops(dm);
 | 
						|
	rdma_user_mmap_entry_remove(&dm->mentry.rdma_entry);
 | 
						|
}
 | 
						|
 | 
						|
static int mlx5_dm_icm_dealloc(struct mlx5_ib_ucontext *ctx,
 | 
						|
			       struct mlx5_ib_dm_icm *dm)
 | 
						|
{
 | 
						|
	enum mlx5_sw_icm_type type = get_icm_type(dm->base.type);
 | 
						|
	struct mlx5_core_dev *dev = to_mdev(dm->base.ibdm.device)->mdev;
 | 
						|
	int err;
 | 
						|
 | 
						|
	err = mlx5_dm_sw_icm_dealloc(dev, type, dm->base.size, ctx->devx_uid,
 | 
						|
				     dm->base.dev_addr, dm->obj_id);
 | 
						|
	if (!err)
 | 
						|
		kfree(dm);
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int mlx5_ib_dealloc_dm(struct ib_dm *ibdm,
 | 
						|
			      struct uverbs_attr_bundle *attrs)
 | 
						|
{
 | 
						|
	struct mlx5_ib_ucontext *ctx = rdma_udata_to_drv_context(
 | 
						|
		&attrs->driver_udata, struct mlx5_ib_ucontext, ibucontext);
 | 
						|
	struct mlx5_ib_dm *dm = to_mdm(ibdm);
 | 
						|
 | 
						|
	switch (dm->type) {
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_MEMIC:
 | 
						|
		mlx5_dm_memic_dealloc(to_memic(ibdm));
 | 
						|
		return 0;
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM:
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM:
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM:
 | 
						|
	case MLX5_IB_UAPI_DM_TYPE_ENCAP_SW_ICM:
 | 
						|
		return mlx5_dm_icm_dealloc(ctx, to_icm(ibdm));
 | 
						|
	default:
 | 
						|
		return -EOPNOTSUPP;
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static int UVERBS_HANDLER(MLX5_IB_METHOD_DM_QUERY)(
 | 
						|
	struct uverbs_attr_bundle *attrs)
 | 
						|
{
 | 
						|
	struct ib_dm *ibdm =
 | 
						|
		uverbs_attr_get_obj(attrs, MLX5_IB_ATTR_QUERY_DM_REQ_HANDLE);
 | 
						|
	struct mlx5_ib_dm *dm = to_mdm(ibdm);
 | 
						|
	struct mlx5_ib_dm_memic *memic;
 | 
						|
	u64 start_offset;
 | 
						|
	u16 page_idx;
 | 
						|
	int err;
 | 
						|
 | 
						|
	if (dm->type != MLX5_IB_UAPI_DM_TYPE_MEMIC)
 | 
						|
		return -EOPNOTSUPP;
 | 
						|
 | 
						|
	memic = to_memic(ibdm);
 | 
						|
	page_idx = memic->mentry.rdma_entry.start_pgoff & 0xFFFF;
 | 
						|
	err = uverbs_copy_to(attrs, MLX5_IB_ATTR_QUERY_DM_RESP_PAGE_INDEX,
 | 
						|
			     &page_idx, sizeof(page_idx));
 | 
						|
	if (err)
 | 
						|
		return err;
 | 
						|
 | 
						|
	start_offset = memic->base.dev_addr & ~PAGE_MASK;
 | 
						|
	err =  uverbs_copy_to(attrs, MLX5_IB_ATTR_QUERY_DM_RESP_START_OFFSET,
 | 
						|
			      &start_offset, sizeof(start_offset));
 | 
						|
	if (err)
 | 
						|
		return err;
 | 
						|
 | 
						|
	return uverbs_copy_to(attrs, MLX5_IB_ATTR_QUERY_DM_RESP_LENGTH,
 | 
						|
			      &memic->req_length,
 | 
						|
			      sizeof(memic->req_length));
 | 
						|
}
 | 
						|
 | 
						|
void mlx5_ib_dm_mmap_free(struct mlx5_ib_dev *dev,
 | 
						|
			  struct mlx5_user_mmap_entry *mentry)
 | 
						|
{
 | 
						|
	struct mlx5_ib_dm_op_entry *op_entry;
 | 
						|
	struct mlx5_ib_dm_memic *mdm;
 | 
						|
 | 
						|
	switch (mentry->mmap_flag) {
 | 
						|
	case MLX5_IB_MMAP_TYPE_MEMIC:
 | 
						|
		mdm = container_of(mentry, struct mlx5_ib_dm_memic, mentry);
 | 
						|
		kref_put(&mdm->ref, mlx5_ib_dm_memic_free);
 | 
						|
		break;
 | 
						|
	case MLX5_IB_MMAP_TYPE_MEMIC_OP:
 | 
						|
		op_entry = container_of(mentry, struct mlx5_ib_dm_op_entry,
 | 
						|
					mentry);
 | 
						|
		mdm = op_entry->dm;
 | 
						|
		mlx5_cmd_dealloc_memic_op(&dev->dm, mdm->base.dev_addr,
 | 
						|
					  op_entry->op);
 | 
						|
		kfree(op_entry);
 | 
						|
		kref_put(&mdm->ref, mlx5_ib_dm_memic_free);
 | 
						|
		break;
 | 
						|
	default:
 | 
						|
		WARN_ON(true);
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
DECLARE_UVERBS_NAMED_METHOD(
 | 
						|
	MLX5_IB_METHOD_DM_QUERY,
 | 
						|
	UVERBS_ATTR_IDR(MLX5_IB_ATTR_QUERY_DM_REQ_HANDLE, UVERBS_OBJECT_DM,
 | 
						|
			UVERBS_ACCESS_READ, UA_MANDATORY),
 | 
						|
	UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_QUERY_DM_RESP_START_OFFSET,
 | 
						|
			    UVERBS_ATTR_TYPE(u64), UA_MANDATORY),
 | 
						|
	UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_QUERY_DM_RESP_PAGE_INDEX,
 | 
						|
			    UVERBS_ATTR_TYPE(u16), UA_MANDATORY),
 | 
						|
	UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_QUERY_DM_RESP_LENGTH,
 | 
						|
			    UVERBS_ATTR_TYPE(u64), UA_MANDATORY));
 | 
						|
 | 
						|
ADD_UVERBS_ATTRIBUTES_SIMPLE(
 | 
						|
	mlx5_ib_dm, UVERBS_OBJECT_DM, UVERBS_METHOD_DM_ALLOC,
 | 
						|
	UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_ALLOC_DM_RESP_START_OFFSET,
 | 
						|
			    UVERBS_ATTR_TYPE(u64), UA_MANDATORY),
 | 
						|
	UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_ALLOC_DM_RESP_PAGE_INDEX,
 | 
						|
			    UVERBS_ATTR_TYPE(u16), UA_OPTIONAL),
 | 
						|
	UVERBS_ATTR_CONST_IN(MLX5_IB_ATTR_ALLOC_DM_REQ_TYPE,
 | 
						|
			     enum mlx5_ib_uapi_dm_type, UA_OPTIONAL));
 | 
						|
 | 
						|
DECLARE_UVERBS_NAMED_METHOD(
 | 
						|
	MLX5_IB_METHOD_DM_MAP_OP_ADDR,
 | 
						|
	UVERBS_ATTR_IDR(MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_HANDLE,
 | 
						|
			UVERBS_OBJECT_DM,
 | 
						|
			UVERBS_ACCESS_READ,
 | 
						|
			UA_MANDATORY),
 | 
						|
	UVERBS_ATTR_PTR_IN(MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_OP,
 | 
						|
			   UVERBS_ATTR_TYPE(u8),
 | 
						|
			   UA_MANDATORY),
 | 
						|
	UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_START_OFFSET,
 | 
						|
			    UVERBS_ATTR_TYPE(u64),
 | 
						|
			    UA_MANDATORY),
 | 
						|
	UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_PAGE_INDEX,
 | 
						|
			    UVERBS_ATTR_TYPE(u16),
 | 
						|
			    UA_OPTIONAL));
 | 
						|
 | 
						|
DECLARE_UVERBS_GLOBAL_METHODS(UVERBS_OBJECT_DM,
 | 
						|
			      &UVERBS_METHOD(MLX5_IB_METHOD_DM_MAP_OP_ADDR),
 | 
						|
			      &UVERBS_METHOD(MLX5_IB_METHOD_DM_QUERY));
 | 
						|
 | 
						|
const struct uapi_definition mlx5_ib_dm_defs[] = {
 | 
						|
	UAPI_DEF_CHAIN_OBJ_TREE(UVERBS_OBJECT_DM, &mlx5_ib_dm),
 | 
						|
	UAPI_DEF_CHAIN_OBJ_TREE_NAMED(UVERBS_OBJECT_DM),
 | 
						|
	{},
 | 
						|
};
 | 
						|
 | 
						|
const struct ib_device_ops mlx5_ib_dev_dm_ops = {
 | 
						|
	.alloc_dm = mlx5_ib_alloc_dm,
 | 
						|
	.dealloc_dm = mlx5_ib_dealloc_dm,
 | 
						|
	.reg_dm_mr = mlx5_ib_reg_dm_mr,
 | 
						|
};
 |