Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Terje Bergstrom | 1403 | 82.87% | 5 | 14.71% |
Mikko Perttunen | 155 | 9.16% | 11 | 32.35% |
Thierry Reding | 104 | 6.14% | 12 | 35.29% |
Arto Merilainen | 16 | 0.95% | 3 | 8.82% |
Dmitry Osipenko | 9 | 0.53% | 1 | 2.94% |
Alexandre Courbot | 4 | 0.24% | 1 | 2.94% |
Thomas Gleixner | 2 | 0.12% | 1 | 2.94% |
Total | 1693 | 34 |
/* SPDX-License-Identifier: GPL-2.0-only */ /* * Copyright (c) 2012-2015, NVIDIA Corporation. */ #ifndef HOST1X_DEV_H #define HOST1X_DEV_H #include <linux/device.h> #include <linux/iommu.h> #include <linux/iova.h> #include <linux/platform_device.h> #include <linux/reset.h> #include "cdma.h" #include "channel.h" #include "context.h" #include "intr.h" #include "job.h" #include "syncpt.h" struct host1x_syncpt; struct host1x_syncpt_base; struct host1x_channel; struct host1x_cdma; struct host1x_job; struct push_buffer; struct output; struct dentry; struct host1x_channel_ops { int (*init)(struct host1x_channel *channel, struct host1x *host, unsigned int id); int (*submit)(struct host1x_job *job); }; struct host1x_cdma_ops { void (*start)(struct host1x_cdma *cdma); void (*stop)(struct host1x_cdma *cdma); void (*flush)(struct host1x_cdma *cdma); int (*timeout_init)(struct host1x_cdma *cdma); void (*timeout_destroy)(struct host1x_cdma *cdma); void (*freeze)(struct host1x_cdma *cdma); void (*resume)(struct host1x_cdma *cdma, u32 getptr); void (*timeout_cpu_incr)(struct host1x_cdma *cdma, u32 getptr, u32 syncpt_incrs, u32 syncval, u32 nr_slots); }; struct host1x_pushbuffer_ops { void (*init)(struct push_buffer *pb); }; struct host1x_debug_ops { void (*debug_init)(struct dentry *de); void (*show_channel_cdma)(struct host1x *host, struct host1x_channel *ch, struct output *o); void (*show_channel_fifo)(struct host1x *host, struct host1x_channel *ch, struct output *o); void (*show_mlocks)(struct host1x *host, struct output *output); }; struct host1x_syncpt_ops { void (*restore)(struct host1x_syncpt *syncpt); void (*restore_wait_base)(struct host1x_syncpt *syncpt); void (*load_wait_base)(struct host1x_syncpt *syncpt); u32 (*load)(struct host1x_syncpt *syncpt); int (*cpu_incr)(struct host1x_syncpt *syncpt); void (*assign_to_channel)(struct host1x_syncpt *syncpt, struct host1x_channel *channel); void (*enable_protection)(struct host1x *host); }; struct host1x_intr_ops { int (*init_host_sync)(struct host1x *host, u32 cpm); void (*set_syncpt_threshold)( struct host1x *host, unsigned int id, u32 thresh); void (*enable_syncpt_intr)(struct host1x *host, unsigned int id); void (*disable_syncpt_intr)(struct host1x *host, unsigned int id); void (*disable_all_syncpt_intrs)(struct host1x *host); int (*free_syncpt_irq)(struct host1x *host); }; struct host1x_sid_entry { unsigned int base; unsigned int offset; unsigned int limit; }; struct host1x_table_desc { unsigned int base; unsigned int count; }; struct host1x_info { unsigned int nb_channels; /* host1x: number of channels supported */ unsigned int nb_pts; /* host1x: number of syncpoints supported */ unsigned int nb_bases; /* host1x: number of syncpoint bases supported */ unsigned int nb_mlocks; /* host1x: number of mlocks supported */ int (*init)(struct host1x *host1x); /* initialize per SoC ops */ unsigned int sync_offset; /* offset of syncpoint registers */ u64 dma_mask; /* mask of addressable memory */ bool has_wide_gather; /* supports GATHER_W opcode */ bool has_hypervisor; /* has hypervisor registers */ bool has_common; /* has common registers separate from hypervisor */ unsigned int num_sid_entries; const struct host1x_sid_entry *sid_table; struct host1x_table_desc streamid_vm_table; struct host1x_table_desc classid_vm_table; struct host1x_table_desc mmio_vm_table; /* * On T20-T148, the boot chain may setup DC to increment syncpoints * 26/27 on VBLANK. As such we cannot use these syncpoints until * the display driver disables VBLANK increments. */ bool reserve_vblank_syncpts; /* * On Tegra186, secure world applications may require access to * host1x during suspend/resume. To allow this, we need to leave * host1x not in reset. */ bool skip_reset_assert; }; struct host1x { const struct host1x_info *info; void __iomem *regs; void __iomem *hv_regs; /* hypervisor region */ void __iomem *common_regs; int syncpt_irqs[8]; int num_syncpt_irqs; struct host1x_syncpt *syncpt; struct host1x_syncpt_base *bases; struct device *dev; struct clk *clk; struct reset_control_bulk_data resets[2]; unsigned int nresets; struct iommu_group *group; struct iommu_domain *domain; struct iova_domain iova; dma_addr_t iova_end; struct mutex intr_mutex; const struct host1x_syncpt_ops *syncpt_op; const struct host1x_intr_ops *intr_op; const struct host1x_channel_ops *channel_op; const struct host1x_cdma_ops *cdma_op; const struct host1x_pushbuffer_ops *cdma_pb_op; const struct host1x_debug_ops *debug_op; struct host1x_syncpt *nop_sp; struct mutex syncpt_mutex; struct host1x_channel_list channel_list; struct host1x_memory_context_list context_list; struct dentry *debugfs; struct mutex devices_lock; struct list_head devices; struct list_head list; struct device_dma_parameters dma_parms; struct host1x_bo_cache cache; }; void host1x_common_writel(struct host1x *host1x, u32 v, u32 r); void host1x_hypervisor_writel(struct host1x *host1x, u32 r, u32 v); u32 host1x_hypervisor_readl(struct host1x *host1x, u32 r); void host1x_sync_writel(struct host1x *host1x, u32 r, u32 v); u32 host1x_sync_readl(struct host1x *host1x, u32 r); void host1x_ch_writel(struct host1x_channel *ch, u32 r, u32 v); u32 host1x_ch_readl(struct host1x_channel *ch, u32 r); static inline void host1x_hw_syncpt_restore(struct host1x *host, struct host1x_syncpt *sp) { host->syncpt_op->restore(sp); } static inline void host1x_hw_syncpt_restore_wait_base(struct host1x *host, struct host1x_syncpt *sp) { host->syncpt_op->restore_wait_base(sp); } static inline void host1x_hw_syncpt_load_wait_base(struct host1x *host, struct host1x_syncpt *sp) { host->syncpt_op->load_wait_base(sp); } static inline u32 host1x_hw_syncpt_load(struct host1x *host, struct host1x_syncpt *sp) { return host->syncpt_op->load(sp); } static inline int host1x_hw_syncpt_cpu_incr(struct host1x *host, struct host1x_syncpt *sp) { return host->syncpt_op->cpu_incr(sp); } static inline void host1x_hw_syncpt_assign_to_channel( struct host1x *host, struct host1x_syncpt *sp, struct host1x_channel *ch) { return host->syncpt_op->assign_to_channel(sp, ch); } static inline void host1x_hw_syncpt_enable_protection(struct host1x *host) { return host->syncpt_op->enable_protection(host); } static inline int host1x_hw_intr_init_host_sync(struct host1x *host, u32 cpm) { return host->intr_op->init_host_sync(host, cpm); } static inline void host1x_hw_intr_set_syncpt_threshold(struct host1x *host, unsigned int id, u32 thresh) { host->intr_op->set_syncpt_threshold(host, id, thresh); } static inline void host1x_hw_intr_enable_syncpt_intr(struct host1x *host, unsigned int id) { host->intr_op->enable_syncpt_intr(host, id); } static inline void host1x_hw_intr_disable_syncpt_intr(struct host1x *host, unsigned int id) { host->intr_op->disable_syncpt_intr(host, id); } static inline void host1x_hw_intr_disable_all_syncpt_intrs(struct host1x *host) { host->intr_op->disable_all_syncpt_intrs(host); } static inline int host1x_hw_intr_free_syncpt_irq(struct host1x *host) { return host->intr_op->free_syncpt_irq(host); } static inline int host1x_hw_channel_init(struct host1x *host, struct host1x_channel *channel, unsigned int id) { return host->channel_op->init(channel, host, id); } static inline int host1x_hw_channel_submit(struct host1x *host, struct host1x_job *job) { return host->channel_op->submit(job); } static inline void host1x_hw_cdma_start(struct host1x *host, struct host1x_cdma *cdma) { host->cdma_op->start(cdma); } static inline void host1x_hw_cdma_stop(struct host1x *host, struct host1x_cdma *cdma) { host->cdma_op->stop(cdma); } static inline void host1x_hw_cdma_flush(struct host1x *host, struct host1x_cdma *cdma) { host->cdma_op->flush(cdma); } static inline int host1x_hw_cdma_timeout_init(struct host1x *host, struct host1x_cdma *cdma) { return host->cdma_op->timeout_init(cdma); } static inline void host1x_hw_cdma_timeout_destroy(struct host1x *host, struct host1x_cdma *cdma) { host->cdma_op->timeout_destroy(cdma); } static inline void host1x_hw_cdma_freeze(struct host1x *host, struct host1x_cdma *cdma) { host->cdma_op->freeze(cdma); } static inline void host1x_hw_cdma_resume(struct host1x *host, struct host1x_cdma *cdma, u32 getptr) { host->cdma_op->resume(cdma, getptr); } static inline void host1x_hw_cdma_timeout_cpu_incr(struct host1x *host, struct host1x_cdma *cdma, u32 getptr, u32 syncpt_incrs, u32 syncval, u32 nr_slots) { host->cdma_op->timeout_cpu_incr(cdma, getptr, syncpt_incrs, syncval, nr_slots); } static inline void host1x_hw_pushbuffer_init(struct host1x *host, struct push_buffer *pb) { host->cdma_pb_op->init(pb); } static inline void host1x_hw_debug_init(struct host1x *host, struct dentry *de) { if (host->debug_op && host->debug_op->debug_init) host->debug_op->debug_init(de); } static inline void host1x_hw_show_channel_cdma(struct host1x *host, struct host1x_channel *channel, struct output *o) { host->debug_op->show_channel_cdma(host, channel, o); } static inline void host1x_hw_show_channel_fifo(struct host1x *host, struct host1x_channel *channel, struct output *o) { host->debug_op->show_channel_fifo(host, channel, o); } static inline void host1x_hw_show_mlocks(struct host1x *host, struct output *o) { host->debug_op->show_mlocks(host, o); } extern struct platform_driver tegra_mipi_driver; #endif
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1