Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Ohad Ben-Cohen | 904 | 35.13% | 8 | 20.00% |
Arnaud Pouliquen | 807 | 31.36% | 4 | 10.00% |
Sjur Brændeland | 380 | 14.77% | 1 | 2.50% |
Loic Pallardy | 289 | 11.23% | 4 | 10.00% |
Michael S. Tsirkin | 47 | 1.83% | 5 | 12.50% |
Nikita Shubin | 30 | 1.17% | 1 | 2.50% |
Tero Kristo | 26 | 1.01% | 1 | 2.50% |
Tiwei Bie | 25 | 0.97% | 1 | 2.50% |
Wei Wang | 24 | 0.93% | 1 | 2.50% |
Xuan Zhuo | 12 | 0.47% | 2 | 5.00% |
Suman Anna | 9 | 0.35% | 2 | 5.00% |
Christoph Hellwig | 6 | 0.23% | 2 | 5.00% |
Heinz Graalfs | 4 | 0.16% | 1 | 2.50% |
Thomas Gleixner | 2 | 0.08% | 1 | 2.50% |
Björn Andersson | 2 | 0.08% | 1 | 2.50% |
Jason (Hui) Wang | 2 | 0.08% | 1 | 2.50% |
Christophe Jaillet | 1 | 0.04% | 1 | 2.50% |
weiping zhang | 1 | 0.04% | 1 | 2.50% |
Stephen Hemminger | 1 | 0.04% | 1 | 2.50% |
Stefan Hajnoczi | 1 | 0.04% | 1 | 2.50% |
Total | 2573 | 40 |
// SPDX-License-Identifier: GPL-2.0-only /* * Remote processor messaging transport (OMAP platform-specific bits) * * Copyright (C) 2011 Texas Instruments, Inc. * Copyright (C) 2011 Google, Inc. * * Ohad Ben-Cohen <ohad@wizery.com> * Brian Swetland <swetland@google.com> */ #include <linux/dma-direct.h> #include <linux/dma-map-ops.h> #include <linux/dma-mapping.h> #include <linux/export.h> #include <linux/of_reserved_mem.h> #include <linux/platform_device.h> #include <linux/remoteproc.h> #include <linux/virtio.h> #include <linux/virtio_config.h> #include <linux/virtio_ids.h> #include <linux/virtio_ring.h> #include <linux/err.h> #include <linux/kref.h> #include <linux/slab.h> #include "remoteproc_internal.h" static int copy_dma_range_map(struct device *to, struct device *from) { const struct bus_dma_region *map = from->dma_range_map, *new_map, *r; int num_ranges = 0; if (!map) return 0; for (r = map; r->size; r++) num_ranges++; new_map = kmemdup(map, array_size(num_ranges + 1, sizeof(*map)), GFP_KERNEL); if (!new_map) return -ENOMEM; to->dma_range_map = new_map; return 0; } static struct rproc_vdev *vdev_to_rvdev(struct virtio_device *vdev) { struct platform_device *pdev; pdev = container_of(vdev->dev.parent, struct platform_device, dev); return platform_get_drvdata(pdev); } static struct rproc *vdev_to_rproc(struct virtio_device *vdev) { struct rproc_vdev *rvdev = vdev_to_rvdev(vdev); return rvdev->rproc; } /* kick the remote processor, and let it know which virtqueue to poke at */ static bool rproc_virtio_notify(struct virtqueue *vq) { struct rproc_vring *rvring = vq->priv; struct rproc *rproc = rvring->rvdev->rproc; int notifyid = rvring->notifyid; dev_dbg(&rproc->dev, "kicking vq index: %d\n", notifyid); rproc->ops->kick(rproc, notifyid); return true; } /** * rproc_vq_interrupt() - tell remoteproc that a virtqueue is interrupted * @rproc: handle to the remote processor * @notifyid: index of the signalled virtqueue (unique per this @rproc) * * This function should be called by the platform-specific rproc driver, * when the remote processor signals that a specific virtqueue has pending * messages available. * * Return: IRQ_NONE if no message was found in the @notifyid virtqueue, * and otherwise returns IRQ_HANDLED. */ irqreturn_t rproc_vq_interrupt(struct rproc *rproc, int notifyid) { struct rproc_vring *rvring; dev_dbg(&rproc->dev, "vq index %d is interrupted\n", notifyid); rvring = idr_find(&rproc->notifyids, notifyid); if (!rvring || !rvring->vq) return IRQ_NONE; return vring_interrupt(0, rvring->vq); } EXPORT_SYMBOL(rproc_vq_interrupt); static struct virtqueue *rp_find_vq(struct virtio_device *vdev, unsigned int id, void (*callback)(struct virtqueue *vq), const char *name, bool ctx) { struct rproc_vdev *rvdev = vdev_to_rvdev(vdev); struct rproc *rproc = vdev_to_rproc(vdev); struct device *dev = &rproc->dev; struct rproc_mem_entry *mem; struct rproc_vring *rvring; struct fw_rsc_vdev *rsc; struct virtqueue *vq; void *addr; int num, size; /* we're temporarily limited to two virtqueues per rvdev */ if (id >= ARRAY_SIZE(rvdev->vring)) return ERR_PTR(-EINVAL); if (!name) return NULL; /* Search allocated memory region by name */ mem = rproc_find_carveout_by_name(rproc, "vdev%dvring%d", rvdev->index, id); if (!mem || !mem->va) return ERR_PTR(-ENOMEM); rvring = &rvdev->vring[id]; addr = mem->va; num = rvring->num; /* zero vring */ size = vring_size(num, rvring->align); memset(addr, 0, size); dev_dbg(dev, "vring%d: va %pK qsz %d notifyid %d\n", id, addr, num, rvring->notifyid); /* * Create the new vq, and tell virtio we're not interested in * the 'weak' smp barriers, since we're talking with a real device. */ vq = vring_new_virtqueue(id, num, rvring->align, vdev, false, ctx, addr, rproc_virtio_notify, callback, name); if (!vq) { dev_err(dev, "vring_new_virtqueue %s failed\n", name); rproc_free_vring(rvring); return ERR_PTR(-ENOMEM); } vq->num_max = num; rvring->vq = vq; vq->priv = rvring; /* Update vring in resource table */ rsc = (void *)rproc->table_ptr + rvdev->rsc_offset; rsc->vring[id].da = mem->da; return vq; } static void __rproc_virtio_del_vqs(struct virtio_device *vdev) { struct virtqueue *vq, *n; struct rproc_vring *rvring; list_for_each_entry_safe(vq, n, &vdev->vqs, list) { rvring = vq->priv; rvring->vq = NULL; vring_del_virtqueue(vq); } } static void rproc_virtio_del_vqs(struct virtio_device *vdev) { __rproc_virtio_del_vqs(vdev); } static int rproc_virtio_find_vqs(struct virtio_device *vdev, unsigned int nvqs, struct virtqueue *vqs[], vq_callback_t *callbacks[], const char * const names[], const bool * ctx, struct irq_affinity *desc) { int i, ret, queue_idx = 0; for (i = 0; i < nvqs; ++i) { if (!names[i]) { vqs[i] = NULL; continue; } vqs[i] = rp_find_vq(vdev, queue_idx++, callbacks[i], names[i], ctx ? ctx[i] : false); if (IS_ERR(vqs[i])) { ret = PTR_ERR(vqs[i]); goto error; } } return 0; error: __rproc_virtio_del_vqs(vdev); return ret; } static u8 rproc_virtio_get_status(struct virtio_device *vdev) { struct rproc_vdev *rvdev = vdev_to_rvdev(vdev); struct fw_rsc_vdev *rsc; rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset; return rsc->status; } static void rproc_virtio_set_status(struct virtio_device *vdev, u8 status) { struct rproc_vdev *rvdev = vdev_to_rvdev(vdev); struct fw_rsc_vdev *rsc; rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset; rsc->status = status; dev_dbg(&vdev->dev, "status: %d\n", status); } static void rproc_virtio_reset(struct virtio_device *vdev) { struct rproc_vdev *rvdev = vdev_to_rvdev(vdev); struct fw_rsc_vdev *rsc; rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset; rsc->status = 0; dev_dbg(&vdev->dev, "reset !\n"); } /* provide the vdev features as retrieved from the firmware */ static u64 rproc_virtio_get_features(struct virtio_device *vdev) { struct rproc_vdev *rvdev = vdev_to_rvdev(vdev); struct fw_rsc_vdev *rsc; rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset; return rsc->dfeatures; } static void rproc_transport_features(struct virtio_device *vdev) { /* * Packed ring isn't enabled on remoteproc for now, * because remoteproc uses vring_new_virtqueue() which * creates virtio rings on preallocated memory. */ __virtio_clear_bit(vdev, VIRTIO_F_RING_PACKED); } static int rproc_virtio_finalize_features(struct virtio_device *vdev) { struct rproc_vdev *rvdev = vdev_to_rvdev(vdev); struct fw_rsc_vdev *rsc; rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset; /* Give virtio_ring a chance to accept features */ vring_transport_features(vdev); /* Give virtio_rproc a chance to accept features. */ rproc_transport_features(vdev); /* Make sure we don't have any features > 32 bits! */ BUG_ON((u32)vdev->features != vdev->features); /* * Remember the finalized features of our vdev, and provide it * to the remote processor once it is powered on. */ rsc->gfeatures = vdev->features; return 0; } static void rproc_virtio_get(struct virtio_device *vdev, unsigned int offset, void *buf, unsigned int len) { struct rproc_vdev *rvdev = vdev_to_rvdev(vdev); struct fw_rsc_vdev *rsc; void *cfg; rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset; cfg = &rsc->vring[rsc->num_of_vrings]; if (offset + len > rsc->config_len || offset + len < len) { dev_err(&vdev->dev, "rproc_virtio_get: access out of bounds\n"); return; } memcpy(buf, cfg + offset, len); } static void rproc_virtio_set(struct virtio_device *vdev, unsigned int offset, const void *buf, unsigned int len) { struct rproc_vdev *rvdev = vdev_to_rvdev(vdev); struct fw_rsc_vdev *rsc; void *cfg; rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset; cfg = &rsc->vring[rsc->num_of_vrings]; if (offset + len > rsc->config_len || offset + len < len) { dev_err(&vdev->dev, "rproc_virtio_set: access out of bounds\n"); return; } memcpy(cfg + offset, buf, len); } static const struct virtio_config_ops rproc_virtio_config_ops = { .get_features = rproc_virtio_get_features, .finalize_features = rproc_virtio_finalize_features, .find_vqs = rproc_virtio_find_vqs, .del_vqs = rproc_virtio_del_vqs, .reset = rproc_virtio_reset, .set_status = rproc_virtio_set_status, .get_status = rproc_virtio_get_status, .get = rproc_virtio_get, .set = rproc_virtio_set, }; /* * This function is called whenever vdev is released, and is responsible * to decrement the remote processor's refcount which was taken when vdev was * added. * * Never call this function directly; it will be called by the driver * core when needed. */ static void rproc_virtio_dev_release(struct device *dev) { struct virtio_device *vdev = dev_to_virtio(dev); struct rproc_vdev *rvdev = vdev_to_rvdev(vdev); kfree(vdev); put_device(&rvdev->pdev->dev); } /** * rproc_add_virtio_dev() - register an rproc-induced virtio device * @rvdev: the remote vdev * @id: the device type identification (used to match it with a driver). * * This function registers a virtio device. This vdev's partent is * the rproc device. * * Return: 0 on success or an appropriate error value otherwise */ static int rproc_add_virtio_dev(struct rproc_vdev *rvdev, int id) { struct rproc *rproc = rvdev->rproc; struct device *dev = &rvdev->pdev->dev; struct virtio_device *vdev; struct rproc_mem_entry *mem; int ret; if (rproc->ops->kick == NULL) { ret = -EINVAL; dev_err(dev, ".kick method not defined for %s\n", rproc->name); goto out; } /* Try to find dedicated vdev buffer carveout */ mem = rproc_find_carveout_by_name(rproc, "vdev%dbuffer", rvdev->index); if (mem) { phys_addr_t pa; if (mem->of_resm_idx != -1) { struct device_node *np = rproc->dev.parent->of_node; /* Associate reserved memory to vdev device */ ret = of_reserved_mem_device_init_by_idx(dev, np, mem->of_resm_idx); if (ret) { dev_err(dev, "Can't associate reserved memory\n"); goto out; } } else { if (mem->va) { dev_warn(dev, "vdev %d buffer already mapped\n", rvdev->index); pa = rproc_va_to_pa(mem->va); } else { /* Use dma address as carveout no memmapped yet */ pa = (phys_addr_t)mem->dma; } /* Associate vdev buffer memory pool to vdev subdev */ ret = dma_declare_coherent_memory(dev, pa, mem->da, mem->len); if (ret < 0) { dev_err(dev, "Failed to associate buffer\n"); goto out; } } } else { struct device_node *np = rproc->dev.parent->of_node; /* * If we don't have dedicated buffer, just attempt to re-assign * the reserved memory from our parent. A default memory-region * at index 0 from the parent's memory-regions is assigned for * the rvdev dev to allocate from. Failure is non-critical and * the allocations will fall back to global pools, so don't * check return value either. */ of_reserved_mem_device_init_by_idx(dev, np, 0); } /* Allocate virtio device */ vdev = kzalloc(sizeof(*vdev), GFP_KERNEL); if (!vdev) { ret = -ENOMEM; goto out; } vdev->id.device = id, vdev->config = &rproc_virtio_config_ops, vdev->dev.parent = dev; vdev->dev.release = rproc_virtio_dev_release; /* Reference the vdev and vring allocations */ get_device(dev); ret = register_virtio_device(vdev); if (ret) { put_device(&vdev->dev); dev_err(dev, "failed to register vdev: %d\n", ret); goto out; } dev_info(dev, "registered %s (type %d)\n", dev_name(&vdev->dev), id); out: return ret; } /** * rproc_remove_virtio_dev() - remove an rproc-induced virtio device * @dev: the virtio device * @data: must be null * * This function unregisters an existing virtio device. * * Return: 0 */ static int rproc_remove_virtio_dev(struct device *dev, void *data) { struct virtio_device *vdev = dev_to_virtio(dev); unregister_virtio_device(vdev); return 0; } static int rproc_vdev_do_start(struct rproc_subdev *subdev) { struct rproc_vdev *rvdev = container_of(subdev, struct rproc_vdev, subdev); return rproc_add_virtio_dev(rvdev, rvdev->id); } static void rproc_vdev_do_stop(struct rproc_subdev *subdev, bool crashed) { struct rproc_vdev *rvdev = container_of(subdev, struct rproc_vdev, subdev); struct device *dev = &rvdev->pdev->dev; int ret; ret = device_for_each_child(dev, NULL, rproc_remove_virtio_dev); if (ret) dev_warn(dev, "can't remove vdev child device: %d\n", ret); } static int rproc_virtio_probe(struct platform_device *pdev) { struct device *dev = &pdev->dev; struct rproc_vdev_data *rvdev_data = dev->platform_data; struct rproc_vdev *rvdev; struct rproc *rproc = container_of(dev->parent, struct rproc, dev); struct fw_rsc_vdev *rsc; int i, ret; if (!rvdev_data) return -EINVAL; rvdev = devm_kzalloc(dev, sizeof(*rvdev), GFP_KERNEL); if (!rvdev) return -ENOMEM; rvdev->id = rvdev_data->id; rvdev->rproc = rproc; rvdev->index = rvdev_data->index; ret = copy_dma_range_map(dev, rproc->dev.parent); if (ret) return ret; /* Make device dma capable by inheriting from parent's capabilities */ set_dma_ops(dev, get_dma_ops(rproc->dev.parent)); ret = dma_coerce_mask_and_coherent(dev, dma_get_mask(rproc->dev.parent)); if (ret) { dev_warn(dev, "Failed to set DMA mask %llx. Trying to continue... (%pe)\n", dma_get_mask(rproc->dev.parent), ERR_PTR(ret)); } platform_set_drvdata(pdev, rvdev); rvdev->pdev = pdev; rsc = rvdev_data->rsc; /* parse the vrings */ for (i = 0; i < rsc->num_of_vrings; i++) { ret = rproc_parse_vring(rvdev, rsc, i); if (ret) return ret; } /* remember the resource offset*/ rvdev->rsc_offset = rvdev_data->rsc_offset; /* allocate the vring resources */ for (i = 0; i < rsc->num_of_vrings; i++) { ret = rproc_alloc_vring(rvdev, i); if (ret) goto unwind_vring_allocations; } rproc_add_rvdev(rproc, rvdev); rvdev->subdev.start = rproc_vdev_do_start; rvdev->subdev.stop = rproc_vdev_do_stop; rproc_add_subdev(rproc, &rvdev->subdev); /* * We're indirectly making a non-temporary copy of the rproc pointer * here, because the platform device or the vdev device will indirectly * access the wrapping rproc. * * Therefore we must increment the rproc refcount here, and decrement * it _only_ on platform remove. */ get_device(&rproc->dev); return 0; unwind_vring_allocations: for (i--; i >= 0; i--) rproc_free_vring(&rvdev->vring[i]); return ret; } static int rproc_virtio_remove(struct platform_device *pdev) { struct rproc_vdev *rvdev = dev_get_drvdata(&pdev->dev); struct rproc *rproc = rvdev->rproc; struct rproc_vring *rvring; int id; for (id = 0; id < ARRAY_SIZE(rvdev->vring); id++) { rvring = &rvdev->vring[id]; rproc_free_vring(rvring); } rproc_remove_subdev(rproc, &rvdev->subdev); rproc_remove_rvdev(rvdev); of_reserved_mem_device_release(&pdev->dev); dma_release_coherent_memory(&pdev->dev); put_device(&rproc->dev); return 0; } /* Platform driver */ static struct platform_driver rproc_virtio_driver = { .probe = rproc_virtio_probe, .remove = rproc_virtio_remove, .driver = { .name = "rproc-virtio", }, }; builtin_platform_driver(rproc_virtio_driver);
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1