1 // SPDX-License-Identifier: GPL-2.0
4 * Copyright 2016-2019 HabanaLabs, Ltd.
8 #include <uapi/misc/habanalabs.h>
9 #include "habanalabs.h"
11 #include <linux/kernel.h>
13 #include <linux/uaccess.h>
14 #include <linux/slab.h>
16 static u32 hl_debug_struct_size[HL_DEBUG_OP_TIMESTAMP + 1] = {
17 [HL_DEBUG_OP_ETR] = sizeof(struct hl_debug_params_etr),
18 [HL_DEBUG_OP_ETF] = sizeof(struct hl_debug_params_etf),
19 [HL_DEBUG_OP_STM] = sizeof(struct hl_debug_params_stm),
20 [HL_DEBUG_OP_FUNNEL] = 0,
21 [HL_DEBUG_OP_BMON] = sizeof(struct hl_debug_params_bmon),
22 [HL_DEBUG_OP_SPMU] = sizeof(struct hl_debug_params_spmu),
23 [HL_DEBUG_OP_TIMESTAMP] = 0
27 static int device_status_info(struct hl_device *hdev, struct hl_info_args *args)
29 struct hl_info_device_status dev_stat = {0};
30 u32 size = args->return_size;
31 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
33 if ((!size) || (!out))
36 dev_stat.status = hl_device_status(hdev);
38 return copy_to_user(out, &dev_stat,
39 min((size_t)size, sizeof(dev_stat))) ? -EFAULT : 0;
42 static int hw_ip_info(struct hl_device *hdev, struct hl_info_args *args)
44 struct hl_info_hw_ip_info hw_ip = {0};
45 u32 size = args->return_size;
46 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
47 struct asic_fixed_properties *prop = &hdev->asic_prop;
48 u64 sram_kmd_size, dram_kmd_size;
50 if ((!size) || (!out))
53 sram_kmd_size = (prop->sram_user_base_address -
54 prop->sram_base_address);
55 dram_kmd_size = (prop->dram_user_base_address -
56 prop->dram_base_address);
58 hw_ip.device_id = hdev->asic_funcs->get_pci_id(hdev);
59 hw_ip.sram_base_address = prop->sram_user_base_address;
60 hw_ip.dram_base_address = prop->dram_user_base_address;
61 hw_ip.tpc_enabled_mask = prop->tpc_enabled_mask;
62 hw_ip.sram_size = prop->sram_size - sram_kmd_size;
63 hw_ip.dram_size = prop->dram_size - dram_kmd_size;
64 if (hw_ip.dram_size > PAGE_SIZE)
65 hw_ip.dram_enabled = 1;
66 hw_ip.num_of_events = prop->num_of_events;
68 memcpy(hw_ip.cpucp_version, prop->cpucp_info.cpucp_version,
69 min(VERSION_MAX_LEN, HL_INFO_VERSION_MAX_LEN));
71 memcpy(hw_ip.card_name, prop->cpucp_info.card_name,
72 min(CARD_NAME_MAX_LEN, HL_INFO_CARD_NAME_MAX_LEN));
74 hw_ip.cpld_version = le32_to_cpu(prop->cpucp_info.cpld_version);
75 hw_ip.module_id = le32_to_cpu(prop->cpucp_info.card_location);
77 hw_ip.psoc_pci_pll_nr = prop->psoc_pci_pll_nr;
78 hw_ip.psoc_pci_pll_nf = prop->psoc_pci_pll_nf;
79 hw_ip.psoc_pci_pll_od = prop->psoc_pci_pll_od;
80 hw_ip.psoc_pci_pll_div_factor = prop->psoc_pci_pll_div_factor;
82 return copy_to_user(out, &hw_ip,
83 min((size_t)size, sizeof(hw_ip))) ? -EFAULT : 0;
86 static int hw_events_info(struct hl_device *hdev, bool aggregate,
87 struct hl_info_args *args)
89 u32 size, max_size = args->return_size;
90 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
93 if ((!max_size) || (!out))
96 arr = hdev->asic_funcs->get_events_stat(hdev, aggregate, &size);
98 return copy_to_user(out, arr, min(max_size, size)) ? -EFAULT : 0;
101 static int dram_usage_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
103 struct hl_device *hdev = hpriv->hdev;
104 struct hl_info_dram_usage dram_usage = {0};
105 u32 max_size = args->return_size;
106 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
107 struct asic_fixed_properties *prop = &hdev->asic_prop;
110 if ((!max_size) || (!out))
113 dram_kmd_size = (prop->dram_user_base_address -
114 prop->dram_base_address);
115 dram_usage.dram_free_mem = (prop->dram_size - dram_kmd_size) -
116 atomic64_read(&hdev->dram_used_mem);
118 dram_usage.ctx_dram_mem =
119 atomic64_read(&hpriv->ctx->dram_phys_mem);
121 return copy_to_user(out, &dram_usage,
122 min((size_t) max_size, sizeof(dram_usage))) ? -EFAULT : 0;
125 static int hw_idle(struct hl_device *hdev, struct hl_info_args *args)
127 struct hl_info_hw_idle hw_idle = {0};
128 u32 max_size = args->return_size;
129 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
131 if ((!max_size) || (!out))
134 hw_idle.is_idle = hdev->asic_funcs->is_device_idle(hdev,
135 &hw_idle.busy_engines_mask_ext, NULL);
137 return copy_to_user(out, &hw_idle,
138 min((size_t) max_size, sizeof(hw_idle))) ? -EFAULT : 0;
141 static int debug_coresight(struct hl_device *hdev, struct hl_debug_args *args)
143 struct hl_debug_params *params;
144 void *input = NULL, *output = NULL;
147 params = kzalloc(sizeof(*params), GFP_KERNEL);
151 params->reg_idx = args->reg_idx;
152 params->enable = args->enable;
153 params->op = args->op;
155 if (args->input_ptr && args->input_size) {
156 input = kzalloc(hl_debug_struct_size[args->op], GFP_KERNEL);
162 if (copy_from_user(input, u64_to_user_ptr(args->input_ptr),
165 dev_err(hdev->dev, "failed to copy input debug data\n");
169 params->input = input;
172 if (args->output_ptr && args->output_size) {
173 output = kzalloc(args->output_size, GFP_KERNEL);
179 params->output = output;
180 params->output_size = args->output_size;
183 rc = hdev->asic_funcs->debug_coresight(hdev, params);
186 "debug coresight operation failed %d\n", rc);
190 if (output && copy_to_user((void __user *) (uintptr_t) args->output_ptr,
191 output, args->output_size)) {
192 dev_err(hdev->dev, "copy to user failed in debug ioctl\n");
206 static int device_utilization(struct hl_device *hdev, struct hl_info_args *args)
208 struct hl_info_device_utilization device_util = {0};
209 u32 max_size = args->return_size;
210 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
212 if ((!max_size) || (!out))
215 if ((args->period_ms < 100) || (args->period_ms > 1000) ||
216 (args->period_ms % 100)) {
218 "period %u must be between 100 - 1000 and must be divisible by 100\n",
223 device_util.utilization = hl_device_utilization(hdev, args->period_ms);
225 return copy_to_user(out, &device_util,
226 min((size_t) max_size, sizeof(device_util))) ? -EFAULT : 0;
229 static int get_clk_rate(struct hl_device *hdev, struct hl_info_args *args)
231 struct hl_info_clk_rate clk_rate = {0};
232 u32 max_size = args->return_size;
233 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
236 if ((!max_size) || (!out))
239 rc = hdev->asic_funcs->get_clk_rate(hdev, &clk_rate.cur_clk_rate_mhz,
240 &clk_rate.max_clk_rate_mhz);
244 return copy_to_user(out, &clk_rate,
245 min((size_t) max_size, sizeof(clk_rate))) ? -EFAULT : 0;
248 static int get_reset_count(struct hl_device *hdev, struct hl_info_args *args)
250 struct hl_info_reset_count reset_count = {0};
251 u32 max_size = args->return_size;
252 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
254 if ((!max_size) || (!out))
257 reset_count.hard_reset_cnt = hdev->hard_reset_cnt;
258 reset_count.soft_reset_cnt = hdev->soft_reset_cnt;
260 return copy_to_user(out, &reset_count,
261 min((size_t) max_size, sizeof(reset_count))) ? -EFAULT : 0;
264 static int time_sync_info(struct hl_device *hdev, struct hl_info_args *args)
266 struct hl_info_time_sync time_sync = {0};
267 u32 max_size = args->return_size;
268 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
270 if ((!max_size) || (!out))
273 time_sync.device_time = hdev->asic_funcs->get_device_time(hdev);
274 time_sync.host_time = ktime_get_raw_ns();
276 return copy_to_user(out, &time_sync,
277 min((size_t) max_size, sizeof(time_sync))) ? -EFAULT : 0;
280 static int pci_counters_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
282 struct hl_device *hdev = hpriv->hdev;
283 struct hl_info_pci_counters pci_counters = {0};
284 u32 max_size = args->return_size;
285 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
288 if ((!max_size) || (!out))
291 rc = hl_fw_cpucp_pci_counters_get(hdev, &pci_counters);
295 return copy_to_user(out, &pci_counters,
296 min((size_t) max_size, sizeof(pci_counters))) ? -EFAULT : 0;
299 static int clk_throttle_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
301 struct hl_device *hdev = hpriv->hdev;
302 struct hl_info_clk_throttle clk_throttle = {0};
303 u32 max_size = args->return_size;
304 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
306 if ((!max_size) || (!out))
309 clk_throttle.clk_throttling_reason = hdev->clk_throttling_reason;
311 return copy_to_user(out, &clk_throttle,
312 min((size_t) max_size, sizeof(clk_throttle))) ? -EFAULT : 0;
315 static int cs_counters_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
317 struct hl_device *hdev = hpriv->hdev;
318 struct hl_info_cs_counters cs_counters = { {0} };
319 u32 max_size = args->return_size;
320 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
322 if ((!max_size) || (!out))
325 memcpy(&cs_counters.cs_counters, &hdev->aggregated_cs_counters,
326 sizeof(struct hl_cs_counters));
329 memcpy(&cs_counters.ctx_cs_counters, &hpriv->ctx->cs_counters,
330 sizeof(struct hl_cs_counters));
332 return copy_to_user(out, &cs_counters,
333 min((size_t) max_size, sizeof(cs_counters))) ? -EFAULT : 0;
336 static int sync_manager_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
338 struct hl_device *hdev = hpriv->hdev;
339 struct asic_fixed_properties *prop = &hdev->asic_prop;
340 struct hl_info_sync_manager sm_info = {0};
341 u32 max_size = args->return_size;
342 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
344 if ((!max_size) || (!out))
347 if (args->dcore_id >= HL_MAX_DCORES)
350 sm_info.first_available_sync_object =
351 prop->first_available_user_sob[args->dcore_id];
352 sm_info.first_available_monitor =
353 prop->first_available_user_mon[args->dcore_id];
356 return copy_to_user(out, &sm_info, min_t(size_t, (size_t) max_size,
357 sizeof(sm_info))) ? -EFAULT : 0;
360 static int total_energy_consumption_info(struct hl_fpriv *hpriv,
361 struct hl_info_args *args)
363 struct hl_device *hdev = hpriv->hdev;
364 struct hl_info_energy total_energy = {0};
365 u32 max_size = args->return_size;
366 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
369 if ((!max_size) || (!out))
372 rc = hl_fw_cpucp_total_energy_get(hdev,
373 &total_energy.total_energy_consumption);
377 return copy_to_user(out, &total_energy,
378 min((size_t) max_size, sizeof(total_energy))) ? -EFAULT : 0;
381 static int _hl_info_ioctl(struct hl_fpriv *hpriv, void *data,
384 struct hl_info_args *args = data;
385 struct hl_device *hdev = hpriv->hdev;
389 * Information is returned for the following opcodes even if the device
390 * is disabled or in reset.
393 case HL_INFO_HW_IP_INFO:
394 return hw_ip_info(hdev, args);
396 case HL_INFO_DEVICE_STATUS:
397 return device_status_info(hdev, args);
399 case HL_INFO_RESET_COUNT:
400 return get_reset_count(hdev, args);
406 if (hl_device_disabled_or_in_reset(hdev)) {
407 dev_warn_ratelimited(dev,
408 "Device is %s. Can't execute INFO IOCTL\n",
409 atomic_read(&hdev->in_reset) ? "in_reset" : "disabled");
414 case HL_INFO_HW_EVENTS:
415 rc = hw_events_info(hdev, false, args);
418 case HL_INFO_DRAM_USAGE:
419 rc = dram_usage_info(hpriv, args);
422 case HL_INFO_HW_IDLE:
423 rc = hw_idle(hdev, args);
426 case HL_INFO_DEVICE_UTILIZATION:
427 rc = device_utilization(hdev, args);
430 case HL_INFO_HW_EVENTS_AGGREGATE:
431 rc = hw_events_info(hdev, true, args);
434 case HL_INFO_CLK_RATE:
435 rc = get_clk_rate(hdev, args);
438 case HL_INFO_TIME_SYNC:
439 return time_sync_info(hdev, args);
441 case HL_INFO_CS_COUNTERS:
442 return cs_counters_info(hpriv, args);
444 case HL_INFO_PCI_COUNTERS:
445 return pci_counters_info(hpriv, args);
447 case HL_INFO_CLK_THROTTLE_REASON:
448 return clk_throttle_info(hpriv, args);
450 case HL_INFO_SYNC_MANAGER:
451 return sync_manager_info(hpriv, args);
453 case HL_INFO_TOTAL_ENERGY:
454 return total_energy_consumption_info(hpriv, args);
457 dev_err(dev, "Invalid request %d\n", args->op);
465 static int hl_info_ioctl(struct hl_fpriv *hpriv, void *data)
467 return _hl_info_ioctl(hpriv, data, hpriv->hdev->dev);
470 static int hl_info_ioctl_control(struct hl_fpriv *hpriv, void *data)
472 return _hl_info_ioctl(hpriv, data, hpriv->hdev->dev_ctrl);
475 static int hl_debug_ioctl(struct hl_fpriv *hpriv, void *data)
477 struct hl_debug_args *args = data;
478 struct hl_device *hdev = hpriv->hdev;
481 if (hl_device_disabled_or_in_reset(hdev)) {
482 dev_warn_ratelimited(hdev->dev,
483 "Device is %s. Can't execute DEBUG IOCTL\n",
484 atomic_read(&hdev->in_reset) ? "in_reset" : "disabled");
489 case HL_DEBUG_OP_ETR:
490 case HL_DEBUG_OP_ETF:
491 case HL_DEBUG_OP_STM:
492 case HL_DEBUG_OP_FUNNEL:
493 case HL_DEBUG_OP_BMON:
494 case HL_DEBUG_OP_SPMU:
495 case HL_DEBUG_OP_TIMESTAMP:
496 if (!hdev->in_debug) {
497 dev_err_ratelimited(hdev->dev,
498 "Rejecting debug configuration request because device not in debug mode\n");
502 min(args->input_size, hl_debug_struct_size[args->op]);
503 rc = debug_coresight(hdev, args);
505 case HL_DEBUG_OP_SET_MODE:
506 rc = hl_device_set_debug_mode(hdev, (bool) args->enable);
509 dev_err(hdev->dev, "Invalid request %d\n", args->op);
517 #define HL_IOCTL_DEF(ioctl, _func) \
518 [_IOC_NR(ioctl)] = {.cmd = ioctl, .func = _func}
520 static const struct hl_ioctl_desc hl_ioctls[] = {
521 HL_IOCTL_DEF(HL_IOCTL_INFO, hl_info_ioctl),
522 HL_IOCTL_DEF(HL_IOCTL_CB, hl_cb_ioctl),
523 HL_IOCTL_DEF(HL_IOCTL_CS, hl_cs_ioctl),
524 HL_IOCTL_DEF(HL_IOCTL_WAIT_CS, hl_cs_wait_ioctl),
525 HL_IOCTL_DEF(HL_IOCTL_MEMORY, hl_mem_ioctl),
526 HL_IOCTL_DEF(HL_IOCTL_DEBUG, hl_debug_ioctl)
529 static const struct hl_ioctl_desc hl_ioctls_control[] = {
530 HL_IOCTL_DEF(HL_IOCTL_INFO, hl_info_ioctl_control)
533 static long _hl_ioctl(struct file *filep, unsigned int cmd, unsigned long arg,
534 const struct hl_ioctl_desc *ioctl, struct device *dev)
536 struct hl_fpriv *hpriv = filep->private_data;
537 struct hl_device *hdev = hpriv->hdev;
538 unsigned int nr = _IOC_NR(cmd);
539 char stack_kdata[128] = {0};
541 unsigned int usize, asize;
546 if (hdev->hard_reset_pending) {
547 dev_crit_ratelimited(hdev->dev_ctrl,
548 "Device HARD reset pending! Please close FD\n");
552 /* Do not trust userspace, use our own definition */
555 if (unlikely(!func)) {
556 dev_dbg(dev, "no function\n");
561 hl_size = _IOC_SIZE(ioctl->cmd);
562 usize = asize = _IOC_SIZE(cmd);
568 if (cmd & (IOC_IN | IOC_OUT)) {
569 if (asize <= sizeof(stack_kdata)) {
572 kdata = kzalloc(asize, GFP_KERNEL);
581 if (copy_from_user(kdata, (void __user *)arg, usize)) {
585 } else if (cmd & IOC_OUT) {
586 memset(kdata, 0, usize);
589 retcode = func(hpriv, kdata);
591 if ((cmd & IOC_OUT) && copy_to_user((void __user *)arg, kdata, usize))
596 dev_dbg(dev, "error in ioctl: pid=%d, cmd=0x%02x, nr=0x%02x\n",
597 task_pid_nr(current), cmd, nr);
599 if (kdata != stack_kdata)
605 long hl_ioctl(struct file *filep, unsigned int cmd, unsigned long arg)
607 struct hl_fpriv *hpriv = filep->private_data;
608 struct hl_device *hdev = hpriv->hdev;
609 const struct hl_ioctl_desc *ioctl = NULL;
610 unsigned int nr = _IOC_NR(cmd);
612 if ((nr >= HL_COMMAND_START) && (nr < HL_COMMAND_END)) {
613 ioctl = &hl_ioctls[nr];
615 dev_err(hdev->dev, "invalid ioctl: pid=%d, nr=0x%02x\n",
616 task_pid_nr(current), nr);
620 return _hl_ioctl(filep, cmd, arg, ioctl, hdev->dev);
623 long hl_ioctl_control(struct file *filep, unsigned int cmd, unsigned long arg)
625 struct hl_fpriv *hpriv = filep->private_data;
626 struct hl_device *hdev = hpriv->hdev;
627 const struct hl_ioctl_desc *ioctl = NULL;
628 unsigned int nr = _IOC_NR(cmd);
630 if (nr == _IOC_NR(HL_IOCTL_INFO)) {
631 ioctl = &hl_ioctls_control[nr];
633 dev_err(hdev->dev_ctrl, "invalid ioctl: pid=%d, nr=0x%02x\n",
634 task_pid_nr(current), nr);
638 return _hl_ioctl(filep, cmd, arg, ioctl, hdev->dev_ctrl);