1 // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
3 * Copyright (c) 2021, Mellanox Technologies inc. All rights reserved.
6 #include <rdma/uverbs_std_types.h>
9 #define UVERBS_MODULE_NAME mlx5_ib
10 #include <rdma/uverbs_named_ioctl.h>
12 static int mlx5_cmd_alloc_memic(struct mlx5_dm *dm, phys_addr_t *addr,
13 u64 length, u32 alignment)
15 struct mlx5_core_dev *dev = dm->dev;
16 u64 num_memic_hw_pages = MLX5_CAP_DEV_MEM(dev, memic_bar_size)
18 u64 hw_start_addr = MLX5_CAP64_DEV_MEM(dev, memic_bar_start_addr);
19 u32 max_alignment = MLX5_CAP_DEV_MEM(dev, log_max_memic_addr_alignment);
20 u32 num_pages = DIV_ROUND_UP(length, PAGE_SIZE);
21 u32 out[MLX5_ST_SZ_DW(alloc_memic_out)] = {};
22 u32 in[MLX5_ST_SZ_DW(alloc_memic_in)] = {};
27 if (!length || (length & MLX5_MEMIC_ALLOC_SIZE_MASK))
30 /* mlx5 device sets alignment as 64*2^driver_value
31 * so normalizing is needed.
33 mlx5_alignment = (alignment < MLX5_MEMIC_BASE_ALIGN) ? 0 :
34 alignment - MLX5_MEMIC_BASE_ALIGN;
35 if (mlx5_alignment > max_alignment)
38 MLX5_SET(alloc_memic_in, in, opcode, MLX5_CMD_OP_ALLOC_MEMIC);
39 MLX5_SET(alloc_memic_in, in, range_size, num_pages * PAGE_SIZE);
40 MLX5_SET(alloc_memic_in, in, memic_size, length);
41 MLX5_SET(alloc_memic_in, in, log_memic_addr_alignment,
44 while (page_idx < num_memic_hw_pages) {
46 page_idx = bitmap_find_next_zero_area(dm->memic_alloc_pages,
51 if (page_idx < num_memic_hw_pages)
52 bitmap_set(dm->memic_alloc_pages,
55 spin_unlock(&dm->lock);
57 if (page_idx >= num_memic_hw_pages)
60 MLX5_SET64(alloc_memic_in, in, range_start_addr,
61 hw_start_addr + (page_idx * PAGE_SIZE));
63 ret = mlx5_cmd_exec_inout(dev, alloc_memic, in, out);
66 bitmap_clear(dm->memic_alloc_pages,
68 spin_unlock(&dm->lock);
78 *addr = dev->bar_addr +
79 MLX5_GET64(alloc_memic_out, out, memic_start_addr);
87 void mlx5_cmd_dealloc_memic(struct mlx5_dm *dm, phys_addr_t addr,
90 struct mlx5_core_dev *dev = dm->dev;
91 u64 hw_start_addr = MLX5_CAP64_DEV_MEM(dev, memic_bar_start_addr);
92 u32 num_pages = DIV_ROUND_UP(length, PAGE_SIZE);
93 u32 in[MLX5_ST_SZ_DW(dealloc_memic_in)] = {};
97 addr -= dev->bar_addr;
98 start_page_idx = (addr - hw_start_addr) >> PAGE_SHIFT;
100 MLX5_SET(dealloc_memic_in, in, opcode, MLX5_CMD_OP_DEALLOC_MEMIC);
101 MLX5_SET64(dealloc_memic_in, in, memic_start_addr, addr);
102 MLX5_SET(dealloc_memic_in, in, memic_size, length);
104 err = mlx5_cmd_exec_in(dev, dealloc_memic, in);
108 spin_lock(&dm->lock);
109 bitmap_clear(dm->memic_alloc_pages,
110 start_page_idx, num_pages);
111 spin_unlock(&dm->lock);
114 void mlx5_cmd_dealloc_memic_op(struct mlx5_dm *dm, phys_addr_t addr,
117 u32 in[MLX5_ST_SZ_DW(modify_memic_in)] = {};
118 struct mlx5_core_dev *dev = dm->dev;
120 MLX5_SET(modify_memic_in, in, opcode, MLX5_CMD_OP_MODIFY_MEMIC);
121 MLX5_SET(modify_memic_in, in, op_mod, MLX5_MODIFY_MEMIC_OP_MOD_DEALLOC);
122 MLX5_SET(modify_memic_in, in, memic_operation_type, operation);
123 MLX5_SET64(modify_memic_in, in, memic_start_addr, addr - dev->bar_addr);
125 mlx5_cmd_exec_in(dev, modify_memic, in);
128 static int mlx5_cmd_alloc_memic_op(struct mlx5_dm *dm, phys_addr_t addr,
129 u8 operation, phys_addr_t *op_addr)
131 u32 out[MLX5_ST_SZ_DW(modify_memic_out)] = {};
132 u32 in[MLX5_ST_SZ_DW(modify_memic_in)] = {};
133 struct mlx5_core_dev *dev = dm->dev;
136 MLX5_SET(modify_memic_in, in, opcode, MLX5_CMD_OP_MODIFY_MEMIC);
137 MLX5_SET(modify_memic_in, in, op_mod, MLX5_MODIFY_MEMIC_OP_MOD_ALLOC);
138 MLX5_SET(modify_memic_in, in, memic_operation_type, operation);
139 MLX5_SET64(modify_memic_in, in, memic_start_addr, addr - dev->bar_addr);
141 err = mlx5_cmd_exec_inout(dev, modify_memic, in, out);
145 *op_addr = dev->bar_addr +
146 MLX5_GET64(modify_memic_out, out, memic_operation_addr);
150 static int add_dm_mmap_entry(struct ib_ucontext *context,
151 struct mlx5_user_mmap_entry *mentry, u8 mmap_flag,
152 size_t size, u64 address)
154 mentry->mmap_flag = mmap_flag;
155 mentry->address = address;
157 return rdma_user_mmap_entry_insert_range(
158 context, &mentry->rdma_entry, size,
159 MLX5_IB_MMAP_DEVICE_MEM << 16,
160 (MLX5_IB_MMAP_DEVICE_MEM << 16) + (1UL << 16) - 1);
163 static void mlx5_ib_dm_memic_free(struct kref *kref)
165 struct mlx5_ib_dm_memic *dm =
166 container_of(kref, struct mlx5_ib_dm_memic, ref);
167 struct mlx5_ib_dev *dev = to_mdev(dm->base.ibdm.device);
169 mlx5_cmd_dealloc_memic(&dev->dm, dm->base.dev_addr, dm->base.size);
173 static int copy_op_to_user(struct mlx5_ib_dm_op_entry *op_entry,
174 struct uverbs_attr_bundle *attrs)
180 page_idx = op_entry->mentry.rdma_entry.start_pgoff & 0xFFFF;
181 start_offset = op_entry->op_addr & ~PAGE_MASK;
182 err = uverbs_copy_to(attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_PAGE_INDEX,
183 &page_idx, sizeof(page_idx));
187 return uverbs_copy_to(attrs,
188 MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_START_OFFSET,
189 &start_offset, sizeof(start_offset));
192 static int map_existing_op(struct mlx5_ib_dm_memic *dm, u8 op,
193 struct uverbs_attr_bundle *attrs)
195 struct mlx5_ib_dm_op_entry *op_entry;
197 op_entry = xa_load(&dm->ops, op);
201 return copy_op_to_user(op_entry, attrs);
204 static int UVERBS_HANDLER(MLX5_IB_METHOD_DM_MAP_OP_ADDR)(
205 struct uverbs_attr_bundle *attrs)
207 struct ib_uobject *uobj = uverbs_attr_get_uobject(
208 attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_HANDLE);
209 struct mlx5_ib_dev *dev = to_mdev(uobj->context->device);
210 struct ib_dm *ibdm = uobj->object;
211 struct mlx5_ib_dm_memic *dm = to_memic(ibdm);
212 struct mlx5_ib_dm_op_entry *op_entry;
216 err = uverbs_copy_from(&op, attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_OP);
220 if (!(MLX5_CAP_DEV_MEM(dev->mdev, memic_operations) & BIT(op)))
223 mutex_lock(&dm->ops_xa_lock);
224 err = map_existing_op(dm, op, attrs);
225 if (!err || err != -ENOENT)
228 op_entry = kzalloc(sizeof(*op_entry), GFP_KERNEL);
232 err = mlx5_cmd_alloc_memic_op(&dev->dm, dm->base.dev_addr, op,
241 err = add_dm_mmap_entry(uobj->context, &op_entry->mentry,
242 MLX5_IB_MMAP_TYPE_MEMIC_OP, dm->base.size,
243 op_entry->op_addr & PAGE_MASK);
245 mlx5_cmd_dealloc_memic_op(&dev->dm, dm->base.dev_addr, op);
249 /* From this point, entry will be freed by mmap_free */
252 err = copy_op_to_user(op_entry, attrs);
256 err = xa_insert(&dm->ops, op, op_entry, GFP_KERNEL);
259 mutex_unlock(&dm->ops_xa_lock);
264 rdma_user_mmap_entry_remove(&op_entry->mentry.rdma_entry);
266 mutex_unlock(&dm->ops_xa_lock);
271 static struct ib_dm *handle_alloc_dm_memic(struct ib_ucontext *ctx,
272 struct ib_dm_alloc_attr *attr,
273 struct uverbs_attr_bundle *attrs)
275 struct mlx5_dm *dm_db = &to_mdev(ctx->device)->dm;
276 struct mlx5_ib_dm_memic *dm;
282 if (!MLX5_CAP_DEV_MEM(dm_db->dev, memic))
283 return ERR_PTR(-EOPNOTSUPP);
285 dm = kzalloc(sizeof(*dm), GFP_KERNEL);
287 return ERR_PTR(-ENOMEM);
289 dm->base.type = MLX5_IB_UAPI_DM_TYPE_MEMIC;
290 dm->base.size = roundup(attr->length, MLX5_MEMIC_BASE_SIZE);
291 dm->base.ibdm.device = ctx->device;
295 mutex_init(&dm->ops_xa_lock);
296 dm->req_length = attr->length;
298 err = mlx5_cmd_alloc_memic(dm_db, &dm->base.dev_addr,
299 dm->base.size, attr->alignment);
305 address = dm->base.dev_addr & PAGE_MASK;
306 err = add_dm_mmap_entry(ctx, &dm->mentry, MLX5_IB_MMAP_TYPE_MEMIC,
307 dm->base.size, address);
309 mlx5_cmd_dealloc_memic(dm_db, dm->base.dev_addr, dm->base.size);
314 page_idx = dm->mentry.rdma_entry.start_pgoff & 0xFFFF;
315 err = uverbs_copy_to(attrs, MLX5_IB_ATTR_ALLOC_DM_RESP_PAGE_INDEX,
316 &page_idx, sizeof(page_idx));
320 start_offset = dm->base.dev_addr & ~PAGE_MASK;
321 err = uverbs_copy_to(attrs,
322 MLX5_IB_ATTR_ALLOC_DM_RESP_START_OFFSET,
323 &start_offset, sizeof(start_offset));
327 return &dm->base.ibdm;
330 rdma_user_mmap_entry_remove(&dm->mentry.rdma_entry);
334 static enum mlx5_sw_icm_type get_icm_type(int uapi_type)
336 return uapi_type == MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM ?
337 MLX5_SW_ICM_TYPE_STEERING :
338 MLX5_SW_ICM_TYPE_HEADER_MODIFY;
341 static struct ib_dm *handle_alloc_dm_sw_icm(struct ib_ucontext *ctx,
342 struct ib_dm_alloc_attr *attr,
343 struct uverbs_attr_bundle *attrs,
346 struct mlx5_core_dev *dev = to_mdev(ctx->device)->mdev;
347 enum mlx5_sw_icm_type icm_type = get_icm_type(type);
348 struct mlx5_ib_dm_icm *dm;
352 dm = kzalloc(sizeof(*dm), GFP_KERNEL);
354 return ERR_PTR(-ENOMEM);
356 dm->base.type = type;
357 dm->base.ibdm.device = ctx->device;
359 if (!capable(CAP_SYS_RAWIO) || !capable(CAP_NET_RAW)) {
364 if (!(MLX5_CAP_FLOWTABLE_NIC_RX(dev, sw_owner) ||
365 MLX5_CAP_FLOWTABLE_NIC_TX(dev, sw_owner) ||
366 MLX5_CAP_FLOWTABLE_NIC_RX(dev, sw_owner_v2) ||
367 MLX5_CAP_FLOWTABLE_NIC_TX(dev, sw_owner_v2))) {
372 /* Allocation size must a multiple of the basic block size
375 act_size = round_up(attr->length, MLX5_SW_ICM_BLOCK_SIZE(dev));
376 act_size = roundup_pow_of_two(act_size);
378 dm->base.size = act_size;
379 err = mlx5_dm_sw_icm_alloc(dev, icm_type, act_size, attr->alignment,
380 to_mucontext(ctx)->devx_uid,
381 &dm->base.dev_addr, &dm->obj_id);
385 err = uverbs_copy_to(attrs, MLX5_IB_ATTR_ALLOC_DM_RESP_START_OFFSET,
386 &dm->base.dev_addr, sizeof(dm->base.dev_addr));
388 mlx5_dm_sw_icm_dealloc(dev, icm_type, dm->base.size,
389 to_mucontext(ctx)->devx_uid,
390 dm->base.dev_addr, dm->obj_id);
393 return &dm->base.ibdm;
399 struct ib_dm *mlx5_ib_alloc_dm(struct ib_device *ibdev,
400 struct ib_ucontext *context,
401 struct ib_dm_alloc_attr *attr,
402 struct uverbs_attr_bundle *attrs)
404 enum mlx5_ib_uapi_dm_type type;
407 err = uverbs_get_const_default(&type, attrs,
408 MLX5_IB_ATTR_ALLOC_DM_REQ_TYPE,
409 MLX5_IB_UAPI_DM_TYPE_MEMIC);
413 mlx5_ib_dbg(to_mdev(ibdev), "alloc_dm req: dm_type=%d user_length=0x%llx log_alignment=%d\n",
414 type, attr->length, attr->alignment);
417 case MLX5_IB_UAPI_DM_TYPE_MEMIC:
418 return handle_alloc_dm_memic(context, attr, attrs);
419 case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM:
420 return handle_alloc_dm_sw_icm(context, attr, attrs, type);
421 case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM:
422 return handle_alloc_dm_sw_icm(context, attr, attrs, type);
424 return ERR_PTR(-EOPNOTSUPP);
428 static void dm_memic_remove_ops(struct mlx5_ib_dm_memic *dm)
430 struct mlx5_ib_dm_op_entry *entry;
433 mutex_lock(&dm->ops_xa_lock);
434 xa_for_each(&dm->ops, idx, entry) {
435 xa_erase(&dm->ops, idx);
436 rdma_user_mmap_entry_remove(&entry->mentry.rdma_entry);
438 mutex_unlock(&dm->ops_xa_lock);
441 static void mlx5_dm_memic_dealloc(struct mlx5_ib_dm_memic *dm)
443 dm_memic_remove_ops(dm);
444 rdma_user_mmap_entry_remove(&dm->mentry.rdma_entry);
447 static int mlx5_dm_icm_dealloc(struct mlx5_ib_ucontext *ctx,
448 struct mlx5_ib_dm_icm *dm)
450 enum mlx5_sw_icm_type type = get_icm_type(dm->base.type);
451 struct mlx5_core_dev *dev = to_mdev(dm->base.ibdm.device)->mdev;
454 err = mlx5_dm_sw_icm_dealloc(dev, type, dm->base.size, ctx->devx_uid,
455 dm->base.dev_addr, dm->obj_id);
461 static int mlx5_ib_dealloc_dm(struct ib_dm *ibdm,
462 struct uverbs_attr_bundle *attrs)
464 struct mlx5_ib_ucontext *ctx = rdma_udata_to_drv_context(
465 &attrs->driver_udata, struct mlx5_ib_ucontext, ibucontext);
466 struct mlx5_ib_dm *dm = to_mdm(ibdm);
469 case MLX5_IB_UAPI_DM_TYPE_MEMIC:
470 mlx5_dm_memic_dealloc(to_memic(ibdm));
472 case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM:
473 case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM:
474 return mlx5_dm_icm_dealloc(ctx, to_icm(ibdm));
480 static int UVERBS_HANDLER(MLX5_IB_METHOD_DM_QUERY)(
481 struct uverbs_attr_bundle *attrs)
484 uverbs_attr_get_obj(attrs, MLX5_IB_ATTR_QUERY_DM_REQ_HANDLE);
485 struct mlx5_ib_dm *dm = to_mdm(ibdm);
486 struct mlx5_ib_dm_memic *memic;
491 if (dm->type != MLX5_IB_UAPI_DM_TYPE_MEMIC)
494 memic = to_memic(ibdm);
495 page_idx = memic->mentry.rdma_entry.start_pgoff & 0xFFFF;
496 err = uverbs_copy_to(attrs, MLX5_IB_ATTR_QUERY_DM_RESP_PAGE_INDEX,
497 &page_idx, sizeof(page_idx));
501 start_offset = memic->base.dev_addr & ~PAGE_MASK;
502 err = uverbs_copy_to(attrs, MLX5_IB_ATTR_QUERY_DM_RESP_START_OFFSET,
503 &start_offset, sizeof(start_offset));
507 return uverbs_copy_to(attrs, MLX5_IB_ATTR_QUERY_DM_RESP_LENGTH,
509 sizeof(memic->req_length));
512 void mlx5_ib_dm_mmap_free(struct mlx5_ib_dev *dev,
513 struct mlx5_user_mmap_entry *mentry)
515 struct mlx5_ib_dm_op_entry *op_entry;
516 struct mlx5_ib_dm_memic *mdm;
518 switch (mentry->mmap_flag) {
519 case MLX5_IB_MMAP_TYPE_MEMIC:
520 mdm = container_of(mentry, struct mlx5_ib_dm_memic, mentry);
521 kref_put(&mdm->ref, mlx5_ib_dm_memic_free);
523 case MLX5_IB_MMAP_TYPE_MEMIC_OP:
524 op_entry = container_of(mentry, struct mlx5_ib_dm_op_entry,
527 mlx5_cmd_dealloc_memic_op(&dev->dm, mdm->base.dev_addr,
530 kref_put(&mdm->ref, mlx5_ib_dm_memic_free);
537 DECLARE_UVERBS_NAMED_METHOD(
538 MLX5_IB_METHOD_DM_QUERY,
539 UVERBS_ATTR_IDR(MLX5_IB_ATTR_QUERY_DM_REQ_HANDLE, UVERBS_OBJECT_DM,
540 UVERBS_ACCESS_READ, UA_MANDATORY),
541 UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_QUERY_DM_RESP_START_OFFSET,
542 UVERBS_ATTR_TYPE(u64), UA_MANDATORY),
543 UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_QUERY_DM_RESP_PAGE_INDEX,
544 UVERBS_ATTR_TYPE(u16), UA_MANDATORY),
545 UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_QUERY_DM_RESP_LENGTH,
546 UVERBS_ATTR_TYPE(u64), UA_MANDATORY));
548 ADD_UVERBS_ATTRIBUTES_SIMPLE(
549 mlx5_ib_dm, UVERBS_OBJECT_DM, UVERBS_METHOD_DM_ALLOC,
550 UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_ALLOC_DM_RESP_START_OFFSET,
551 UVERBS_ATTR_TYPE(u64), UA_MANDATORY),
552 UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_ALLOC_DM_RESP_PAGE_INDEX,
553 UVERBS_ATTR_TYPE(u16), UA_OPTIONAL),
554 UVERBS_ATTR_CONST_IN(MLX5_IB_ATTR_ALLOC_DM_REQ_TYPE,
555 enum mlx5_ib_uapi_dm_type, UA_OPTIONAL));
557 DECLARE_UVERBS_NAMED_METHOD(
558 MLX5_IB_METHOD_DM_MAP_OP_ADDR,
559 UVERBS_ATTR_IDR(MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_HANDLE,
563 UVERBS_ATTR_PTR_IN(MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_OP,
564 UVERBS_ATTR_TYPE(u8),
566 UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_START_OFFSET,
567 UVERBS_ATTR_TYPE(u64),
569 UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_PAGE_INDEX,
570 UVERBS_ATTR_TYPE(u16),
573 DECLARE_UVERBS_GLOBAL_METHODS(UVERBS_OBJECT_DM,
574 &UVERBS_METHOD(MLX5_IB_METHOD_DM_MAP_OP_ADDR),
575 &UVERBS_METHOD(MLX5_IB_METHOD_DM_QUERY));
577 const struct uapi_definition mlx5_ib_dm_defs[] = {
578 UAPI_DEF_CHAIN_OBJ_TREE(UVERBS_OBJECT_DM, &mlx5_ib_dm),
579 UAPI_DEF_CHAIN_OBJ_TREE_NAMED(UVERBS_OBJECT_DM),
583 const struct ib_device_ops mlx5_ib_dev_dm_ops = {
584 .alloc_dm = mlx5_ib_alloc_dm,
585 .dealloc_dm = mlx5_ib_dealloc_dm,
586 .reg_dm_mr = mlx5_ib_reg_dm_mr,