Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Zhi Wang | 746 | 42.19% | 13 | 34.21% |
fred gao | 638 | 36.09% | 1 | 2.63% |
Jike Song | 99 | 5.60% | 4 | 10.53% |
Ping Gao | 84 | 4.75% | 3 | 7.89% |
Zhenyu Wang | 79 | 4.47% | 4 | 10.53% |
Changbin Du | 57 | 3.22% | 6 | 15.79% |
Hang Yuan | 45 | 2.55% | 3 | 7.89% |
Tina Zhang | 10 | 0.57% | 1 | 2.63% |
Colin Xu | 8 | 0.45% | 1 | 2.63% |
Xiong Zhang | 1 | 0.06% | 1 | 2.63% |
Hao Peng | 1 | 0.06% | 1 | 2.63% |
Total | 1768 | 38 |
/* * Copyright(c) 2011-2016 Intel Corporation. All rights reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. * * Authors: * Kevin Tian <kevin.tian@intel.com> * Eddie Dong <eddie.dong@intel.com> * * Contributors: * Niu Bing <bing.niu@intel.com> * Zhi Wang <zhi.a.wang@intel.com> * */ #include <linux/types.h> #include <xen/xen.h> #include <linux/kthread.h> #include "i915_drv.h" #include "gvt.h" #include <linux/vfio.h> #include <linux/mdev.h> struct intel_gvt_host intel_gvt_host; static const char * const supported_hypervisors[] = { [INTEL_GVT_HYPERVISOR_XEN] = "XEN", [INTEL_GVT_HYPERVISOR_KVM] = "KVM", }; static struct intel_vgpu_type *intel_gvt_find_vgpu_type(struct intel_gvt *gvt, const char *name) { int i; struct intel_vgpu_type *t; const char *driver_name = dev_driver_string( &gvt->dev_priv->drm.pdev->dev); for (i = 0; i < gvt->num_types; i++) { t = &gvt->types[i]; if (!strncmp(t->name, name + strlen(driver_name) + 1, sizeof(t->name))) return t; } return NULL; } static ssize_t available_instances_show(struct kobject *kobj, struct device *dev, char *buf) { struct intel_vgpu_type *type; unsigned int num = 0; void *gvt = kdev_to_i915(dev)->gvt; type = intel_gvt_find_vgpu_type(gvt, kobject_name(kobj)); if (!type) num = 0; else num = type->avail_instance; return sprintf(buf, "%u\n", num); } static ssize_t device_api_show(struct kobject *kobj, struct device *dev, char *buf) { return sprintf(buf, "%s\n", VFIO_DEVICE_API_PCI_STRING); } static ssize_t description_show(struct kobject *kobj, struct device *dev, char *buf) { struct intel_vgpu_type *type; void *gvt = kdev_to_i915(dev)->gvt; type = intel_gvt_find_vgpu_type(gvt, kobject_name(kobj)); if (!type) return 0; return sprintf(buf, "low_gm_size: %dMB\nhigh_gm_size: %dMB\n" "fence: %d\nresolution: %s\n" "weight: %d\n", BYTES_TO_MB(type->low_gm_size), BYTES_TO_MB(type->high_gm_size), type->fence, vgpu_edid_str(type->resolution), type->weight); } static MDEV_TYPE_ATTR_RO(available_instances); static MDEV_TYPE_ATTR_RO(device_api); static MDEV_TYPE_ATTR_RO(description); static struct attribute *gvt_type_attrs[] = { &mdev_type_attr_available_instances.attr, &mdev_type_attr_device_api.attr, &mdev_type_attr_description.attr, NULL, }; static struct attribute_group *gvt_vgpu_type_groups[] = { [0 ... NR_MAX_INTEL_VGPU_TYPES - 1] = NULL, }; static bool intel_get_gvt_attrs(struct attribute ***type_attrs, struct attribute_group ***intel_vgpu_type_groups) { *type_attrs = gvt_type_attrs; *intel_vgpu_type_groups = gvt_vgpu_type_groups; return true; } static bool intel_gvt_init_vgpu_type_groups(struct intel_gvt *gvt) { int i, j; struct intel_vgpu_type *type; struct attribute_group *group; for (i = 0; i < gvt->num_types; i++) { type = &gvt->types[i]; group = kzalloc(sizeof(struct attribute_group), GFP_KERNEL); if (WARN_ON(!group)) goto unwind; group->name = type->name; group->attrs = gvt_type_attrs; gvt_vgpu_type_groups[i] = group; } return true; unwind: for (j = 0; j < i; j++) { group = gvt_vgpu_type_groups[j]; kfree(group); } return false; } static void intel_gvt_cleanup_vgpu_type_groups(struct intel_gvt *gvt) { int i; struct attribute_group *group; for (i = 0; i < gvt->num_types; i++) { group = gvt_vgpu_type_groups[i]; gvt_vgpu_type_groups[i] = NULL; kfree(group); } } static const struct intel_gvt_ops intel_gvt_ops = { .emulate_cfg_read = intel_vgpu_emulate_cfg_read, .emulate_cfg_write = intel_vgpu_emulate_cfg_write, .emulate_mmio_read = intel_vgpu_emulate_mmio_read, .emulate_mmio_write = intel_vgpu_emulate_mmio_write, .vgpu_create = intel_gvt_create_vgpu, .vgpu_destroy = intel_gvt_destroy_vgpu, .vgpu_release = intel_gvt_release_vgpu, .vgpu_reset = intel_gvt_reset_vgpu, .vgpu_activate = intel_gvt_activate_vgpu, .vgpu_deactivate = intel_gvt_deactivate_vgpu, .gvt_find_vgpu_type = intel_gvt_find_vgpu_type, .get_gvt_attrs = intel_get_gvt_attrs, .vgpu_query_plane = intel_vgpu_query_plane, .vgpu_get_dmabuf = intel_vgpu_get_dmabuf, .write_protect_handler = intel_vgpu_page_track_handler, }; /** * intel_gvt_init_host - Load MPT modules and detect if we're running in host * * This function is called at the driver loading stage. If failed to find a * loadable MPT module or detect currently we're running in a VM, then GVT-g * will be disabled * * Returns: * Zero on success, negative error code if failed. * */ int intel_gvt_init_host(void) { if (intel_gvt_host.initialized) return 0; /* Xen DOM U */ if (xen_domain() && !xen_initial_domain()) return -ENODEV; /* Try to load MPT modules for hypervisors */ if (xen_initial_domain()) { /* In Xen dom0 */ intel_gvt_host.mpt = try_then_request_module( symbol_get(xengt_mpt), "xengt"); intel_gvt_host.hypervisor_type = INTEL_GVT_HYPERVISOR_XEN; } else { #if IS_ENABLED(CONFIG_DRM_I915_GVT_KVMGT) /* not in Xen. Try KVMGT */ intel_gvt_host.mpt = try_then_request_module( symbol_get(kvmgt_mpt), "kvmgt"); intel_gvt_host.hypervisor_type = INTEL_GVT_HYPERVISOR_KVM; #endif } /* Fail to load MPT modules - bail out */ if (!intel_gvt_host.mpt) return -EINVAL; gvt_dbg_core("Running with hypervisor %s in host mode\n", supported_hypervisors[intel_gvt_host.hypervisor_type]); intel_gvt_host.initialized = true; return 0; } static void init_device_info(struct intel_gvt *gvt) { struct intel_gvt_device_info *info = &gvt->device_info; struct pci_dev *pdev = gvt->dev_priv->drm.pdev; info->max_support_vgpus = 8; info->cfg_space_size = PCI_CFG_SPACE_EXP_SIZE; info->mmio_size = 2 * 1024 * 1024; info->mmio_bar = 0; info->gtt_start_offset = 8 * 1024 * 1024; info->gtt_entry_size = 8; info->gtt_entry_size_shift = 3; info->gmadr_bytes_in_cmd = 8; info->max_surface_size = 36 * 1024 * 1024; info->msi_cap_offset = pdev->msi_cap; } static int gvt_service_thread(void *data) { struct intel_gvt *gvt = (struct intel_gvt *)data; int ret; gvt_dbg_core("service thread start\n"); while (!kthread_should_stop()) { ret = wait_event_interruptible(gvt->service_thread_wq, kthread_should_stop() || gvt->service_request); if (kthread_should_stop()) break; if (WARN_ONCE(ret, "service thread is waken up by signal.\n")) continue; if (test_and_clear_bit(INTEL_GVT_REQUEST_EMULATE_VBLANK, (void *)&gvt->service_request)) intel_gvt_emulate_vblank(gvt); if (test_bit(INTEL_GVT_REQUEST_SCHED, (void *)&gvt->service_request) || test_bit(INTEL_GVT_REQUEST_EVENT_SCHED, (void *)&gvt->service_request)) { intel_gvt_schedule(gvt); } } return 0; } static void clean_service_thread(struct intel_gvt *gvt) { kthread_stop(gvt->service_thread); } static int init_service_thread(struct intel_gvt *gvt) { init_waitqueue_head(&gvt->service_thread_wq); gvt->service_thread = kthread_run(gvt_service_thread, gvt, "gvt_service_thread"); if (IS_ERR(gvt->service_thread)) { gvt_err("fail to start service thread.\n"); return PTR_ERR(gvt->service_thread); } return 0; } /** * intel_gvt_clean_device - clean a GVT device * @dev_priv: i915 private * * This function is called at the driver unloading stage, to free the * resources owned by a GVT device. * */ void intel_gvt_clean_device(struct drm_i915_private *dev_priv) { struct intel_gvt *gvt = to_gvt(dev_priv); if (WARN_ON(!gvt)) return; intel_gvt_destroy_idle_vgpu(gvt->idle_vgpu); intel_gvt_hypervisor_host_exit(&dev_priv->drm.pdev->dev, gvt); intel_gvt_cleanup_vgpu_type_groups(gvt); intel_gvt_clean_vgpu_types(gvt); intel_gvt_debugfs_clean(gvt); clean_service_thread(gvt); intel_gvt_clean_cmd_parser(gvt); intel_gvt_clean_sched_policy(gvt); intel_gvt_clean_workload_scheduler(gvt); intel_gvt_clean_gtt(gvt); intel_gvt_clean_irq(gvt); intel_gvt_free_firmware(gvt); intel_gvt_clean_mmio_info(gvt); idr_destroy(&gvt->vgpu_idr); kfree(dev_priv->gvt); dev_priv->gvt = NULL; } /** * intel_gvt_init_device - initialize a GVT device * @dev_priv: drm i915 private data * * This function is called at the initialization stage, to initialize * necessary GVT components. * * Returns: * Zero on success, negative error code if failed. * */ int intel_gvt_init_device(struct drm_i915_private *dev_priv) { struct intel_gvt *gvt; struct intel_vgpu *vgpu; int ret; /* * Cannot initialize GVT device without intel_gvt_host gets * initialized first. */ if (WARN_ON(!intel_gvt_host.initialized)) return -EINVAL; if (WARN_ON(dev_priv->gvt)) return -EEXIST; gvt = kzalloc(sizeof(struct intel_gvt), GFP_KERNEL); if (!gvt) return -ENOMEM; gvt_dbg_core("init gvt device\n"); idr_init(&gvt->vgpu_idr); spin_lock_init(&gvt->scheduler.mmio_context_lock); mutex_init(&gvt->lock); mutex_init(&gvt->sched_lock); gvt->dev_priv = dev_priv; init_device_info(gvt); ret = intel_gvt_setup_mmio_info(gvt); if (ret) goto out_clean_idr; intel_gvt_init_engine_mmio_context(gvt); ret = intel_gvt_load_firmware(gvt); if (ret) goto out_clean_mmio_info; ret = intel_gvt_init_irq(gvt); if (ret) goto out_free_firmware; ret = intel_gvt_init_gtt(gvt); if (ret) goto out_clean_irq; ret = intel_gvt_init_workload_scheduler(gvt); if (ret) goto out_clean_gtt; ret = intel_gvt_init_sched_policy(gvt); if (ret) goto out_clean_workload_scheduler; ret = intel_gvt_init_cmd_parser(gvt); if (ret) goto out_clean_sched_policy; ret = init_service_thread(gvt); if (ret) goto out_clean_cmd_parser; ret = intel_gvt_init_vgpu_types(gvt); if (ret) goto out_clean_thread; ret = intel_gvt_init_vgpu_type_groups(gvt); if (ret == false) { gvt_err("failed to init vgpu type groups: %d\n", ret); goto out_clean_types; } ret = intel_gvt_hypervisor_host_init(&dev_priv->drm.pdev->dev, gvt, &intel_gvt_ops); if (ret) { gvt_err("failed to register gvt-g host device: %d\n", ret); goto out_clean_types; } vgpu = intel_gvt_create_idle_vgpu(gvt); if (IS_ERR(vgpu)) { ret = PTR_ERR(vgpu); gvt_err("failed to create idle vgpu\n"); goto out_clean_types; } gvt->idle_vgpu = vgpu; ret = intel_gvt_debugfs_init(gvt); if (ret) gvt_err("debugfs registration failed, go on.\n"); gvt_dbg_core("gvt device initialization is done\n"); dev_priv->gvt = gvt; return 0; out_clean_types: intel_gvt_clean_vgpu_types(gvt); out_clean_thread: clean_service_thread(gvt); out_clean_cmd_parser: intel_gvt_clean_cmd_parser(gvt); out_clean_sched_policy: intel_gvt_clean_sched_policy(gvt); out_clean_workload_scheduler: intel_gvt_clean_workload_scheduler(gvt); out_clean_gtt: intel_gvt_clean_gtt(gvt); out_clean_irq: intel_gvt_clean_irq(gvt); out_free_firmware: intel_gvt_free_firmware(gvt); out_clean_mmio_info: intel_gvt_clean_mmio_info(gvt); out_clean_idr: idr_destroy(&gvt->vgpu_idr); kfree(gvt); return ret; } #if IS_ENABLED(CONFIG_DRM_I915_GVT_KVMGT) MODULE_SOFTDEP("pre: kvmgt"); #endif
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1