Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Lucas Segarra Fernandez | 1205 | 87.07% | 3 | 17.65% |
Tadeusz Struk | 81 | 5.85% | 4 | 23.53% |
Zeng Xin | 38 | 2.75% | 1 | 5.88% |
Giovanni Cabiddu | 20 | 1.45% | 4 | 23.53% |
Bruce W Allan | 16 | 1.16% | 1 | 5.88% |
Damian Muszynski | 11 | 0.79% | 1 | 5.88% |
Tomasz Kowalik | 11 | 0.79% | 1 | 5.88% |
Marco Chiappero | 1 | 0.07% | 1 | 5.88% |
Shashank Gupta | 1 | 0.07% | 1 | 5.88% |
Total | 1384 | 17 |
// SPDX-License-Identifier: GPL-2.0-only /* Copyright (c) 2023 Intel Corporation. */ #define dev_fmt(fmt) "Telemetry: " fmt #include <asm/errno.h> #include <linux/atomic.h> #include <linux/device.h> #include <linux/dev_printk.h> #include <linux/dma-mapping.h> #include <linux/jiffies.h> #include <linux/kernel.h> #include <linux/mutex.h> #include <linux/slab.h> #include <linux/string.h> #include <linux/workqueue.h> #include "adf_admin.h" #include "adf_accel_devices.h" #include "adf_common_drv.h" #include "adf_telemetry.h" #define TL_IS_ZERO(input) ((input) == 0) static bool is_tl_supported(struct adf_accel_dev *accel_dev) { u16 fw_caps = GET_HW_DATA(accel_dev)->fw_capabilities; return fw_caps & TL_CAPABILITY_BIT; } static int validate_tl_data(struct adf_tl_hw_data *tl_data) { if (!tl_data->dev_counters || TL_IS_ZERO(tl_data->num_dev_counters) || !tl_data->sl_util_counters || !tl_data->sl_exec_counters || !tl_data->rp_counters || TL_IS_ZERO(tl_data->num_rp_counters)) return -EOPNOTSUPP; return 0; } static int validate_tl_slice_counters(struct icp_qat_fw_init_admin_slice_cnt *slice_count, u8 max_slices_per_type) { u8 *sl_counter = (u8 *)slice_count; int i; for (i = 0; i < ADF_TL_SL_CNT_COUNT; i++) { if (sl_counter[i] > max_slices_per_type) return -EINVAL; } return 0; } static int adf_tl_alloc_mem(struct adf_accel_dev *accel_dev) { struct adf_tl_hw_data *tl_data = &GET_TL_DATA(accel_dev); struct device *dev = &GET_DEV(accel_dev); size_t regs_sz = tl_data->layout_sz; struct adf_telemetry *telemetry; int node = dev_to_node(dev); void *tl_data_regs; unsigned int i; telemetry = kzalloc_node(sizeof(*telemetry), GFP_KERNEL, node); if (!telemetry) return -ENOMEM; telemetry->rp_num_indexes = kmalloc_array(tl_data->max_rp, sizeof(*telemetry->rp_num_indexes), GFP_KERNEL); if (!telemetry->rp_num_indexes) goto err_free_tl; telemetry->regs_hist_buff = kmalloc_array(tl_data->num_hbuff, sizeof(*telemetry->regs_hist_buff), GFP_KERNEL); if (!telemetry->regs_hist_buff) goto err_free_rp_indexes; telemetry->regs_data = dma_alloc_coherent(dev, regs_sz, &telemetry->regs_data_p, GFP_KERNEL); if (!telemetry->regs_data) goto err_free_regs_hist_buff; for (i = 0; i < tl_data->num_hbuff; i++) { tl_data_regs = kzalloc_node(regs_sz, GFP_KERNEL, node); if (!tl_data_regs) goto err_free_dma; telemetry->regs_hist_buff[i] = tl_data_regs; } accel_dev->telemetry = telemetry; return 0; err_free_dma: dma_free_coherent(dev, regs_sz, telemetry->regs_data, telemetry->regs_data_p); while (i--) kfree(telemetry->regs_hist_buff[i]); err_free_regs_hist_buff: kfree(telemetry->regs_hist_buff); err_free_rp_indexes: kfree(telemetry->rp_num_indexes); err_free_tl: kfree(telemetry); return -ENOMEM; } static void adf_tl_free_mem(struct adf_accel_dev *accel_dev) { struct adf_tl_hw_data *tl_data = &GET_TL_DATA(accel_dev); struct adf_telemetry *telemetry = accel_dev->telemetry; struct device *dev = &GET_DEV(accel_dev); size_t regs_sz = tl_data->layout_sz; unsigned int i; for (i = 0; i < tl_data->num_hbuff; i++) kfree(telemetry->regs_hist_buff[i]); dma_free_coherent(dev, regs_sz, telemetry->regs_data, telemetry->regs_data_p); kfree(telemetry->regs_hist_buff); kfree(telemetry->rp_num_indexes); kfree(telemetry); accel_dev->telemetry = NULL; } static unsigned long get_next_timeout(void) { return msecs_to_jiffies(ADF_TL_TIMER_INT_MS); } static void snapshot_regs(struct adf_telemetry *telemetry, size_t size) { void *dst = telemetry->regs_hist_buff[telemetry->hb_num]; void *src = telemetry->regs_data; memcpy(dst, src, size); } static void tl_work_handler(struct work_struct *work) { struct delayed_work *delayed_work; struct adf_telemetry *telemetry; struct adf_tl_hw_data *tl_data; u32 msg_cnt, old_msg_cnt; size_t layout_sz; u32 *regs_data; size_t id; delayed_work = to_delayed_work(work); telemetry = container_of(delayed_work, struct adf_telemetry, work_ctx); tl_data = &GET_TL_DATA(telemetry->accel_dev); regs_data = telemetry->regs_data; id = tl_data->msg_cnt_off / sizeof(*regs_data); layout_sz = tl_data->layout_sz; if (!atomic_read(&telemetry->state)) { cancel_delayed_work_sync(&telemetry->work_ctx); return; } msg_cnt = regs_data[id]; old_msg_cnt = msg_cnt; if (msg_cnt == telemetry->msg_cnt) goto out; mutex_lock(&telemetry->regs_hist_lock); snapshot_regs(telemetry, layout_sz); /* Check if data changed while updating it */ msg_cnt = regs_data[id]; if (old_msg_cnt != msg_cnt) snapshot_regs(telemetry, layout_sz); telemetry->msg_cnt = msg_cnt; telemetry->hb_num++; telemetry->hb_num %= telemetry->hbuffs; mutex_unlock(&telemetry->regs_hist_lock); out: adf_misc_wq_queue_delayed_work(&telemetry->work_ctx, get_next_timeout()); } int adf_tl_halt(struct adf_accel_dev *accel_dev) { struct adf_telemetry *telemetry = accel_dev->telemetry; struct device *dev = &GET_DEV(accel_dev); int ret; cancel_delayed_work_sync(&telemetry->work_ctx); atomic_set(&telemetry->state, 0); ret = adf_send_admin_tl_stop(accel_dev); if (ret) dev_err(dev, "failed to stop telemetry\n"); return ret; } int adf_tl_run(struct adf_accel_dev *accel_dev, int state) { struct adf_tl_hw_data *tl_data = &GET_TL_DATA(accel_dev); struct adf_telemetry *telemetry = accel_dev->telemetry; struct device *dev = &GET_DEV(accel_dev); size_t layout_sz = tl_data->layout_sz; int ret; ret = adf_send_admin_tl_start(accel_dev, telemetry->regs_data_p, layout_sz, telemetry->rp_num_indexes, &telemetry->slice_cnt); if (ret) { dev_err(dev, "failed to start telemetry\n"); return ret; } ret = validate_tl_slice_counters(&telemetry->slice_cnt, tl_data->max_sl_cnt); if (ret) { dev_err(dev, "invalid value returned by FW\n"); adf_send_admin_tl_stop(accel_dev); return ret; } telemetry->hbuffs = state; atomic_set(&telemetry->state, state); adf_misc_wq_queue_delayed_work(&telemetry->work_ctx, get_next_timeout()); return 0; } int adf_tl_init(struct adf_accel_dev *accel_dev) { struct adf_tl_hw_data *tl_data = &GET_TL_DATA(accel_dev); u8 max_rp = GET_TL_DATA(accel_dev).max_rp; struct device *dev = &GET_DEV(accel_dev); struct adf_telemetry *telemetry; unsigned int i; int ret; ret = validate_tl_data(tl_data); if (ret) return ret; ret = adf_tl_alloc_mem(accel_dev); if (ret) { dev_err(dev, "failed to initialize: %d\n", ret); return ret; } telemetry = accel_dev->telemetry; telemetry->accel_dev = accel_dev; mutex_init(&telemetry->wr_lock); mutex_init(&telemetry->regs_hist_lock); INIT_DELAYED_WORK(&telemetry->work_ctx, tl_work_handler); for (i = 0; i < max_rp; i++) telemetry->rp_num_indexes[i] = ADF_TL_RP_REGS_DISABLED; return 0; } int adf_tl_start(struct adf_accel_dev *accel_dev) { struct device *dev = &GET_DEV(accel_dev); if (!accel_dev->telemetry) return -EOPNOTSUPP; if (!is_tl_supported(accel_dev)) { dev_info(dev, "feature not supported by FW\n"); adf_tl_free_mem(accel_dev); return -EOPNOTSUPP; } return 0; } void adf_tl_stop(struct adf_accel_dev *accel_dev) { if (!accel_dev->telemetry) return; if (atomic_read(&accel_dev->telemetry->state)) adf_tl_halt(accel_dev); } void adf_tl_shutdown(struct adf_accel_dev *accel_dev) { if (!accel_dev->telemetry) return; adf_tl_free_mem(accel_dev); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1