Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Tadeusz Struk | 718 | 68.77% | 3 | 27.27% |
Giovanni Cabiddu | 222 | 21.26% | 3 | 27.27% |
Bruce W Allan | 53 | 5.08% | 1 | 9.09% |
Tomasz Kowalik | 28 | 2.68% | 1 | 9.09% |
Pingchao Yang | 20 | 1.92% | 1 | 9.09% |
Wojciech Ziemba | 2 | 0.19% | 1 | 9.09% |
Jack Xu | 1 | 0.10% | 1 | 9.09% |
Total | 1044 | 11 |
// SPDX-License-Identifier: (BSD-3-Clause OR GPL-2.0-only) /* Copyright(c) 2014 - 2020 Intel Corporation */ #include <linux/firmware.h> #include <linux/pci.h> #include "adf_cfg.h" #include "adf_accel_devices.h" #include "adf_common_drv.h" #include "icp_qat_uclo.h" static int adf_ae_fw_load_images(struct adf_accel_dev *accel_dev, void *fw_addr, u32 fw_size) { struct adf_fw_loader_data *loader_data = accel_dev->fw_loader; struct adf_hw_device_data *hw_device = accel_dev->hw_device; struct icp_qat_fw_loader_handle *loader; char *obj_name; u32 num_objs; u32 ae_mask; int i; loader = loader_data->fw_loader; num_objs = hw_device->uof_get_num_objs(); for (i = 0; i < num_objs; i++) { obj_name = hw_device->uof_get_name(accel_dev, i); ae_mask = hw_device->uof_get_ae_mask(accel_dev, i); if (!obj_name || !ae_mask) { dev_err(&GET_DEV(accel_dev), "Invalid UOF image\n"); goto out_err; } if (qat_uclo_set_cfg_ae_mask(loader, ae_mask)) { dev_err(&GET_DEV(accel_dev), "Invalid mask for UOF image\n"); goto out_err; } if (qat_uclo_map_obj(loader, fw_addr, fw_size, obj_name)) { dev_err(&GET_DEV(accel_dev), "Failed to map UOF firmware\n"); goto out_err; } if (qat_uclo_wr_all_uimage(loader)) { dev_err(&GET_DEV(accel_dev), "Failed to load UOF firmware\n"); goto out_err; } qat_uclo_del_obj(loader); } return 0; out_err: adf_ae_fw_release(accel_dev); return -EFAULT; } int adf_ae_fw_load(struct adf_accel_dev *accel_dev) { struct adf_fw_loader_data *loader_data = accel_dev->fw_loader; struct adf_hw_device_data *hw_device = accel_dev->hw_device; void *fw_addr, *mmp_addr; u32 fw_size, mmp_size; if (!hw_device->fw_name) return 0; if (request_firmware(&loader_data->mmp_fw, hw_device->fw_mmp_name, &accel_dev->accel_pci_dev.pci_dev->dev)) { dev_err(&GET_DEV(accel_dev), "Failed to load MMP firmware %s\n", hw_device->fw_mmp_name); return -EFAULT; } if (request_firmware(&loader_data->uof_fw, hw_device->fw_name, &accel_dev->accel_pci_dev.pci_dev->dev)) { dev_err(&GET_DEV(accel_dev), "Failed to load UOF firmware %s\n", hw_device->fw_name); goto out_err; } fw_size = loader_data->uof_fw->size; fw_addr = (void *)loader_data->uof_fw->data; mmp_size = loader_data->mmp_fw->size; mmp_addr = (void *)loader_data->mmp_fw->data; if (qat_uclo_wr_mimage(loader_data->fw_loader, mmp_addr, mmp_size)) { dev_err(&GET_DEV(accel_dev), "Failed to load MMP\n"); goto out_err; } if (hw_device->uof_get_num_objs) return adf_ae_fw_load_images(accel_dev, fw_addr, fw_size); if (qat_uclo_map_obj(loader_data->fw_loader, fw_addr, fw_size, NULL)) { dev_err(&GET_DEV(accel_dev), "Failed to map FW\n"); goto out_err; } if (qat_uclo_wr_all_uimage(loader_data->fw_loader)) { dev_err(&GET_DEV(accel_dev), "Failed to load UOF\n"); goto out_err; } return 0; out_err: adf_ae_fw_release(accel_dev); return -EFAULT; } void adf_ae_fw_release(struct adf_accel_dev *accel_dev) { struct adf_fw_loader_data *loader_data = accel_dev->fw_loader; struct adf_hw_device_data *hw_device = accel_dev->hw_device; if (!hw_device->fw_name) return; qat_uclo_del_obj(loader_data->fw_loader); qat_hal_deinit(loader_data->fw_loader); release_firmware(loader_data->uof_fw); release_firmware(loader_data->mmp_fw); loader_data->uof_fw = NULL; loader_data->mmp_fw = NULL; loader_data->fw_loader = NULL; } int adf_ae_start(struct adf_accel_dev *accel_dev) { struct adf_fw_loader_data *loader_data = accel_dev->fw_loader; struct adf_hw_device_data *hw_data = accel_dev->hw_device; u32 ae_ctr; if (!hw_data->fw_name) return 0; ae_ctr = qat_hal_start(loader_data->fw_loader); dev_info(&GET_DEV(accel_dev), "qat_dev%d started %d acceleration engines\n", accel_dev->accel_id, ae_ctr); return 0; } int adf_ae_stop(struct adf_accel_dev *accel_dev) { struct adf_fw_loader_data *loader_data = accel_dev->fw_loader; struct adf_hw_device_data *hw_data = accel_dev->hw_device; u32 ae_ctr, ae, max_aes = GET_MAX_ACCELENGINES(accel_dev); if (!hw_data->fw_name) return 0; for (ae = 0, ae_ctr = 0; ae < max_aes; ae++) { if (hw_data->ae_mask & (1 << ae)) { qat_hal_stop(loader_data->fw_loader, ae, 0xFF); ae_ctr++; } } dev_info(&GET_DEV(accel_dev), "qat_dev%d stopped %d acceleration engines\n", accel_dev->accel_id, ae_ctr); return 0; } static int adf_ae_reset(struct adf_accel_dev *accel_dev, int ae) { struct adf_fw_loader_data *loader_data = accel_dev->fw_loader; qat_hal_reset(loader_data->fw_loader); if (qat_hal_clr_reset(loader_data->fw_loader)) return -EFAULT; return 0; } int adf_ae_init(struct adf_accel_dev *accel_dev) { struct adf_fw_loader_data *loader_data; struct adf_hw_device_data *hw_device = accel_dev->hw_device; if (!hw_device->fw_name) return 0; loader_data = kzalloc(sizeof(*loader_data), GFP_KERNEL); if (!loader_data) return -ENOMEM; accel_dev->fw_loader = loader_data; if (qat_hal_init(accel_dev)) { dev_err(&GET_DEV(accel_dev), "Failed to init the AEs\n"); kfree(loader_data); return -EFAULT; } if (adf_ae_reset(accel_dev, 0)) { dev_err(&GET_DEV(accel_dev), "Failed to reset the AEs\n"); qat_hal_deinit(loader_data->fw_loader); kfree(loader_data); return -EFAULT; } return 0; } int adf_ae_shutdown(struct adf_accel_dev *accel_dev) { struct adf_fw_loader_data *loader_data = accel_dev->fw_loader; struct adf_hw_device_data *hw_device = accel_dev->hw_device; if (!hw_device->fw_name) return 0; qat_hal_deinit(loader_data->fw_loader); kfree(accel_dev->fw_loader); accel_dev->fw_loader = NULL; return 0; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1