Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
CK Hu | 2531 | 44.09% | 6 | 6.32% |
Bibby Hsieh | 747 | 13.01% | 10 | 10.53% |
Chun-Kuang Hu | 538 | 9.37% | 11 | 11.58% |
Yongqiang Niu | 377 | 6.57% | 7 | 7.37% |
jason-jh.lin | 365 | 6.36% | 5 | 5.26% |
Sean Paul | 339 | 5.91% | 5 | 5.26% |
Nancy.Lin | 244 | 4.25% | 3 | 3.16% |
Hsiao Chien Sung | 145 | 2.53% | 8 | 8.42% |
Evan Benn | 66 | 1.15% | 1 | 1.05% |
Yt Shen | 53 | 0.92% | 1 | 1.05% |
stu.hsieh at mediatek.com | 53 | 0.92% | 3 | 3.16% |
Maxime Ripard | 32 | 0.56% | 5 | 5.26% |
Rex-BC Chen | 31 | 0.54% | 1 | 1.05% |
Angelo G. Del Regno | 23 | 0.40% | 2 | 2.11% |
Gustavo Padovan | 22 | 0.38% | 1 | 1.05% |
Stuart Lee | 21 | 0.37% | 1 | 1.05% |
Laurent Pinchart | 18 | 0.31% | 2 | 2.11% |
Dafna Hirschfeld | 16 | 0.28% | 1 | 1.05% |
Peter Shih | 15 | 0.26% | 1 | 1.05% |
Enric Balletbò i Serra | 12 | 0.21% | 1 | 1.05% |
Christophe Jaillet | 11 | 0.19% | 1 | 1.05% |
Daniel Vetter | 11 | 0.19% | 1 | 1.05% |
ruanjinjie | 11 | 0.19% | 1 | 1.05% |
Justin Green | 10 | 0.17% | 1 | 1.05% |
Hsin-Yi, Wang | 8 | 0.14% | 1 | 1.05% |
Ville Syrjälä | 8 | 0.14% | 1 | 1.05% |
Dennis YC Hsieh | 7 | 0.12% | 3 | 3.16% |
Sam Ravnborg | 7 | 0.12% | 2 | 2.11% |
Shawn Guo | 6 | 0.10% | 1 | 1.05% |
Jyri Sarha | 3 | 0.05% | 1 | 1.05% |
Thomas Gleixner | 2 | 0.03% | 1 | 1.05% |
Thierry Reding | 2 | 0.03% | 1 | 1.05% |
Kees Cook | 2 | 0.03% | 1 | 1.05% |
Doug Anderson | 1 | 0.02% | 1 | 1.05% |
Rob Herring | 1 | 0.02% | 1 | 1.05% |
Wang Li | 1 | 0.02% | 1 | 1.05% |
Daniel Kurtz | 1 | 0.02% | 1 | 1.05% |
Total | 5740 | 95 |
// SPDX-License-Identifier: GPL-2.0-only /* * Copyright (c) 2015 MediaTek Inc. */ #include <linux/clk.h> #include <linux/dma-mapping.h> #include <linux/mailbox_controller.h> #include <linux/of.h> #include <linux/pm_runtime.h> #include <linux/soc/mediatek/mtk-cmdq.h> #include <linux/soc/mediatek/mtk-mmsys.h> #include <linux/soc/mediatek/mtk-mutex.h> #include <asm/barrier.h> #include <drm/drm_atomic.h> #include <drm/drm_atomic_helper.h> #include <drm/drm_probe_helper.h> #include <drm/drm_vblank.h> #include "mtk_crtc.h" #include "mtk_ddp_comp.h" #include "mtk_drm_drv.h" #include "mtk_gem.h" #include "mtk_plane.h" /* * struct mtk_crtc - MediaTek specific crtc structure. * @base: crtc object. * @enabled: records whether crtc_enable succeeded * @planes: array of 4 drm_plane structures, one for each overlay plane * @pending_planes: whether any plane has pending changes to be applied * @mmsys_dev: pointer to the mmsys device for configuration registers * @mutex: handle to one of the ten disp_mutex streams * @ddp_comp_nr: number of components in ddp_comp * @ddp_comp: array of pointers the mtk_ddp_comp structures used by this crtc * * TODO: Needs update: this header is missing a bunch of member descriptions. */ struct mtk_crtc { struct drm_crtc base; bool enabled; bool pending_needs_vblank; struct drm_pending_vblank_event *event; struct drm_plane *planes; unsigned int layer_nr; bool pending_planes; bool pending_async_planes; #if IS_REACHABLE(CONFIG_MTK_CMDQ) struct cmdq_client cmdq_client; struct cmdq_pkt cmdq_handle; u32 cmdq_event; u32 cmdq_vblank_cnt; wait_queue_head_t cb_blocking_queue; #endif struct device *mmsys_dev; struct device *dma_dev; struct mtk_mutex *mutex; unsigned int ddp_comp_nr; struct mtk_ddp_comp **ddp_comp; unsigned int num_conn_routes; const struct mtk_drm_route *conn_routes; /* lock for display hardware access */ struct mutex hw_lock; bool config_updating; }; struct mtk_crtc_state { struct drm_crtc_state base; bool pending_config; unsigned int pending_width; unsigned int pending_height; unsigned int pending_vrefresh; }; static inline struct mtk_crtc *to_mtk_crtc(struct drm_crtc *c) { return container_of(c, struct mtk_crtc, base); } static inline struct mtk_crtc_state *to_mtk_crtc_state(struct drm_crtc_state *s) { return container_of(s, struct mtk_crtc_state, base); } static void mtk_crtc_finish_page_flip(struct mtk_crtc *mtk_crtc) { struct drm_crtc *crtc = &mtk_crtc->base; unsigned long flags; if (mtk_crtc->event) { spin_lock_irqsave(&crtc->dev->event_lock, flags); drm_crtc_send_vblank_event(crtc, mtk_crtc->event); drm_crtc_vblank_put(crtc); mtk_crtc->event = NULL; spin_unlock_irqrestore(&crtc->dev->event_lock, flags); } } static void mtk_drm_finish_page_flip(struct mtk_crtc *mtk_crtc) { drm_crtc_handle_vblank(&mtk_crtc->base); if (!mtk_crtc->config_updating && mtk_crtc->pending_needs_vblank) { mtk_crtc_finish_page_flip(mtk_crtc); mtk_crtc->pending_needs_vblank = false; } } #if IS_REACHABLE(CONFIG_MTK_CMDQ) static int mtk_drm_cmdq_pkt_create(struct cmdq_client *client, struct cmdq_pkt *pkt, size_t size) { struct device *dev; dma_addr_t dma_addr; pkt->va_base = kzalloc(size, GFP_KERNEL); if (!pkt->va_base) return -ENOMEM; pkt->buf_size = size; pkt->cl = (void *)client; dev = client->chan->mbox->dev; dma_addr = dma_map_single(dev, pkt->va_base, pkt->buf_size, DMA_TO_DEVICE); if (dma_mapping_error(dev, dma_addr)) { dev_err(dev, "dma map failed, size=%u\n", (u32)(u64)size); kfree(pkt->va_base); return -ENOMEM; } pkt->pa_base = dma_addr; return 0; } static void mtk_drm_cmdq_pkt_destroy(struct cmdq_pkt *pkt) { struct cmdq_client *client = (struct cmdq_client *)pkt->cl; dma_unmap_single(client->chan->mbox->dev, pkt->pa_base, pkt->buf_size, DMA_TO_DEVICE); kfree(pkt->va_base); } #endif static void mtk_crtc_destroy(struct drm_crtc *crtc) { struct mtk_crtc *mtk_crtc = to_mtk_crtc(crtc); int i; mtk_mutex_put(mtk_crtc->mutex); #if IS_REACHABLE(CONFIG_MTK_CMDQ) mtk_drm_cmdq_pkt_destroy(&mtk_crtc->cmdq_handle); if (mtk_crtc->cmdq_client.chan) { mbox_free_channel(mtk_crtc->cmdq_client.chan); mtk_crtc->cmdq_client.chan = NULL; } #endif for (i = 0; i < mtk_crtc->ddp_comp_nr; i++) { struct mtk_ddp_comp *comp; comp = mtk_crtc->ddp_comp[i]; mtk_ddp_comp_unregister_vblank_cb(comp); } drm_crtc_cleanup(crtc); } static void mtk_crtc_reset(struct drm_crtc *crtc) { struct mtk_crtc_state *state; if (crtc->state) __drm_atomic_helper_crtc_destroy_state(crtc->state); kfree(to_mtk_crtc_state(crtc->state)); crtc->state = NULL; state = kzalloc(sizeof(*state), GFP_KERNEL); if (state) __drm_atomic_helper_crtc_reset(crtc, &state->base); } static struct drm_crtc_state *mtk_crtc_duplicate_state(struct drm_crtc *crtc) { struct mtk_crtc_state *state; state = kmalloc(sizeof(*state), GFP_KERNEL); if (!state) return NULL; __drm_atomic_helper_crtc_duplicate_state(crtc, &state->base); WARN_ON(state->base.crtc != crtc); state->base.crtc = crtc; state->pending_config = false; return &state->base; } static void mtk_crtc_destroy_state(struct drm_crtc *crtc, struct drm_crtc_state *state) { __drm_atomic_helper_crtc_destroy_state(state); kfree(to_mtk_crtc_state(state)); } static enum drm_mode_status mtk_crtc_mode_valid(struct drm_crtc *crtc, const struct drm_display_mode *mode) { struct mtk_crtc *mtk_crtc = to_mtk_crtc(crtc); enum drm_mode_status status = MODE_OK; int i; for (i = 0; i < mtk_crtc->ddp_comp_nr; i++) { status = mtk_ddp_comp_mode_valid(mtk_crtc->ddp_comp[i], mode); if (status != MODE_OK) break; } return status; } static bool mtk_crtc_mode_fixup(struct drm_crtc *crtc, const struct drm_display_mode *mode, struct drm_display_mode *adjusted_mode) { /* Nothing to do here, but this callback is mandatory. */ return true; } static void mtk_crtc_mode_set_nofb(struct drm_crtc *crtc) { struct mtk_crtc_state *state = to_mtk_crtc_state(crtc->state); state->pending_width = crtc->mode.hdisplay; state->pending_height = crtc->mode.vdisplay; state->pending_vrefresh = drm_mode_vrefresh(&crtc->mode); wmb(); /* Make sure the above parameters are set before update */ state->pending_config = true; } static int mtk_crtc_ddp_clk_enable(struct mtk_crtc *mtk_crtc) { int ret; int i; for (i = 0; i < mtk_crtc->ddp_comp_nr; i++) { ret = mtk_ddp_comp_clk_enable(mtk_crtc->ddp_comp[i]); if (ret) { DRM_ERROR("Failed to enable clock %d: %d\n", i, ret); goto err; } } return 0; err: while (--i >= 0) mtk_ddp_comp_clk_disable(mtk_crtc->ddp_comp[i]); return ret; } static void mtk_crtc_ddp_clk_disable(struct mtk_crtc *mtk_crtc) { int i; for (i = 0; i < mtk_crtc->ddp_comp_nr; i++) mtk_ddp_comp_clk_disable(mtk_crtc->ddp_comp[i]); } static struct mtk_ddp_comp *mtk_ddp_comp_for_plane(struct drm_crtc *crtc, struct drm_plane *plane, unsigned int *local_layer) { struct mtk_crtc *mtk_crtc = to_mtk_crtc(crtc); struct mtk_ddp_comp *comp; int i, count = 0; unsigned int local_index = plane - mtk_crtc->planes; for (i = 0; i < mtk_crtc->ddp_comp_nr; i++) { comp = mtk_crtc->ddp_comp[i]; if (local_index < (count + mtk_ddp_comp_layer_nr(comp))) { *local_layer = local_index - count; return comp; } count += mtk_ddp_comp_layer_nr(comp); } WARN(1, "Failed to find component for plane %d\n", plane->index); return NULL; } #if IS_REACHABLE(CONFIG_MTK_CMDQ) static void ddp_cmdq_cb(struct mbox_client *cl, void *mssg) { struct cmdq_cb_data *data = mssg; struct cmdq_client *cmdq_cl = container_of(cl, struct cmdq_client, client); struct mtk_crtc *mtk_crtc = container_of(cmdq_cl, struct mtk_crtc, cmdq_client); struct mtk_crtc_state *state; unsigned int i; if (data->sta < 0) return; state = to_mtk_crtc_state(mtk_crtc->base.state); state->pending_config = false; if (mtk_crtc->pending_planes) { for (i = 0; i < mtk_crtc->layer_nr; i++) { struct drm_plane *plane = &mtk_crtc->planes[i]; struct mtk_plane_state *plane_state; plane_state = to_mtk_plane_state(plane->state); plane_state->pending.config = false; } mtk_crtc->pending_planes = false; } if (mtk_crtc->pending_async_planes) { for (i = 0; i < mtk_crtc->layer_nr; i++) { struct drm_plane *plane = &mtk_crtc->planes[i]; struct mtk_plane_state *plane_state; plane_state = to_mtk_plane_state(plane->state); plane_state->pending.async_config = false; } mtk_crtc->pending_async_planes = false; } mtk_crtc->cmdq_vblank_cnt = 0; wake_up(&mtk_crtc->cb_blocking_queue); } #endif static int mtk_crtc_ddp_hw_init(struct mtk_crtc *mtk_crtc) { struct drm_crtc *crtc = &mtk_crtc->base; struct drm_connector *connector; struct drm_encoder *encoder; struct drm_connector_list_iter conn_iter; unsigned int width, height, vrefresh, bpc = MTK_MAX_BPC; int ret; int i; if (WARN_ON(!crtc->state)) return -EINVAL; width = crtc->state->adjusted_mode.hdisplay; height = crtc->state->adjusted_mode.vdisplay; vrefresh = drm_mode_vrefresh(&crtc->state->adjusted_mode); drm_for_each_encoder(encoder, crtc->dev) { if (encoder->crtc != crtc) continue; drm_connector_list_iter_begin(crtc->dev, &conn_iter); drm_for_each_connector_iter(connector, &conn_iter) { if (connector->encoder != encoder) continue; if (connector->display_info.bpc != 0 && bpc > connector->display_info.bpc) bpc = connector->display_info.bpc; } drm_connector_list_iter_end(&conn_iter); } ret = pm_runtime_resume_and_get(crtc->dev->dev); if (ret < 0) { DRM_ERROR("Failed to enable power domain: %d\n", ret); return ret; } ret = mtk_mutex_prepare(mtk_crtc->mutex); if (ret < 0) { DRM_ERROR("Failed to enable mutex clock: %d\n", ret); goto err_pm_runtime_put; } ret = mtk_crtc_ddp_clk_enable(mtk_crtc); if (ret < 0) { DRM_ERROR("Failed to enable component clocks: %d\n", ret); goto err_mutex_unprepare; } for (i = 0; i < mtk_crtc->ddp_comp_nr - 1; i++) { if (!mtk_ddp_comp_connect(mtk_crtc->ddp_comp[i], mtk_crtc->mmsys_dev, mtk_crtc->ddp_comp[i + 1]->id)) mtk_mmsys_ddp_connect(mtk_crtc->mmsys_dev, mtk_crtc->ddp_comp[i]->id, mtk_crtc->ddp_comp[i + 1]->id); if (!mtk_ddp_comp_add(mtk_crtc->ddp_comp[i], mtk_crtc->mutex)) mtk_mutex_add_comp(mtk_crtc->mutex, mtk_crtc->ddp_comp[i]->id); } if (!mtk_ddp_comp_add(mtk_crtc->ddp_comp[i], mtk_crtc->mutex)) mtk_mutex_add_comp(mtk_crtc->mutex, mtk_crtc->ddp_comp[i]->id); mtk_mutex_enable(mtk_crtc->mutex); for (i = 0; i < mtk_crtc->ddp_comp_nr; i++) { struct mtk_ddp_comp *comp = mtk_crtc->ddp_comp[i]; if (i == 1) mtk_ddp_comp_bgclr_in_on(comp); mtk_ddp_comp_config(comp, width, height, vrefresh, bpc, NULL); mtk_ddp_comp_start(comp); } /* Initially configure all planes */ for (i = 0; i < mtk_crtc->layer_nr; i++) { struct drm_plane *plane = &mtk_crtc->planes[i]; struct mtk_plane_state *plane_state; struct mtk_ddp_comp *comp; unsigned int local_layer; plane_state = to_mtk_plane_state(plane->state); /* should not enable layer before crtc enabled */ plane_state->pending.enable = false; comp = mtk_ddp_comp_for_plane(crtc, plane, &local_layer); if (comp) mtk_ddp_comp_layer_config(comp, local_layer, plane_state, NULL); } return 0; err_mutex_unprepare: mtk_mutex_unprepare(mtk_crtc->mutex); err_pm_runtime_put: pm_runtime_put(crtc->dev->dev); return ret; } static void mtk_crtc_ddp_hw_fini(struct mtk_crtc *mtk_crtc) { struct drm_device *drm = mtk_crtc->base.dev; struct drm_crtc *crtc = &mtk_crtc->base; int i; for (i = 0; i < mtk_crtc->ddp_comp_nr; i++) { mtk_ddp_comp_stop(mtk_crtc->ddp_comp[i]); if (i == 1) mtk_ddp_comp_bgclr_in_off(mtk_crtc->ddp_comp[i]); } for (i = 0; i < mtk_crtc->ddp_comp_nr; i++) if (!mtk_ddp_comp_remove(mtk_crtc->ddp_comp[i], mtk_crtc->mutex)) mtk_mutex_remove_comp(mtk_crtc->mutex, mtk_crtc->ddp_comp[i]->id); mtk_mutex_disable(mtk_crtc->mutex); for (i = 0; i < mtk_crtc->ddp_comp_nr - 1; i++) { if (!mtk_ddp_comp_disconnect(mtk_crtc->ddp_comp[i], mtk_crtc->mmsys_dev, mtk_crtc->ddp_comp[i + 1]->id)) mtk_mmsys_ddp_disconnect(mtk_crtc->mmsys_dev, mtk_crtc->ddp_comp[i]->id, mtk_crtc->ddp_comp[i + 1]->id); if (!mtk_ddp_comp_remove(mtk_crtc->ddp_comp[i], mtk_crtc->mutex)) mtk_mutex_remove_comp(mtk_crtc->mutex, mtk_crtc->ddp_comp[i]->id); } if (!mtk_ddp_comp_remove(mtk_crtc->ddp_comp[i], mtk_crtc->mutex)) mtk_mutex_remove_comp(mtk_crtc->mutex, mtk_crtc->ddp_comp[i]->id); mtk_crtc_ddp_clk_disable(mtk_crtc); mtk_mutex_unprepare(mtk_crtc->mutex); pm_runtime_put(drm->dev); if (crtc->state->event && !crtc->state->active) { spin_lock_irq(&crtc->dev->event_lock); drm_crtc_send_vblank_event(crtc, crtc->state->event); crtc->state->event = NULL; spin_unlock_irq(&crtc->dev->event_lock); } } static void mtk_crtc_ddp_config(struct drm_crtc *crtc, struct cmdq_pkt *cmdq_handle) { struct mtk_crtc *mtk_crtc = to_mtk_crtc(crtc); struct mtk_crtc_state *state = to_mtk_crtc_state(mtk_crtc->base.state); struct mtk_ddp_comp *comp = mtk_crtc->ddp_comp[0]; unsigned int i; unsigned int local_layer; /* * TODO: instead of updating the registers here, we should prepare * working registers in atomic_commit and let the hardware command * queue update module registers on vblank. */ if (state->pending_config) { mtk_ddp_comp_config(comp, state->pending_width, state->pending_height, state->pending_vrefresh, 0, cmdq_handle); if (!cmdq_handle) state->pending_config = false; } if (mtk_crtc->pending_planes) { for (i = 0; i < mtk_crtc->layer_nr; i++) { struct drm_plane *plane = &mtk_crtc->planes[i]; struct mtk_plane_state *plane_state; plane_state = to_mtk_plane_state(plane->state); if (!plane_state->pending.config) continue; comp = mtk_ddp_comp_for_plane(crtc, plane, &local_layer); if (comp) mtk_ddp_comp_layer_config(comp, local_layer, plane_state, cmdq_handle); if (!cmdq_handle) plane_state->pending.config = false; } if (!cmdq_handle) mtk_crtc->pending_planes = false; } if (mtk_crtc->pending_async_planes) { for (i = 0; i < mtk_crtc->layer_nr; i++) { struct drm_plane *plane = &mtk_crtc->planes[i]; struct mtk_plane_state *plane_state; plane_state = to_mtk_plane_state(plane->state); if (!plane_state->pending.async_config) continue; comp = mtk_ddp_comp_for_plane(crtc, plane, &local_layer); if (comp) mtk_ddp_comp_layer_config(comp, local_layer, plane_state, cmdq_handle); if (!cmdq_handle) plane_state->pending.async_config = false; } if (!cmdq_handle) mtk_crtc->pending_async_planes = false; } } static void mtk_crtc_update_config(struct mtk_crtc *mtk_crtc, bool needs_vblank) { #if IS_REACHABLE(CONFIG_MTK_CMDQ) struct cmdq_pkt *cmdq_handle = &mtk_crtc->cmdq_handle; #endif struct drm_crtc *crtc = &mtk_crtc->base; struct mtk_drm_private *priv = crtc->dev->dev_private; unsigned int pending_planes = 0, pending_async_planes = 0; int i; mutex_lock(&mtk_crtc->hw_lock); mtk_crtc->config_updating = true; if (needs_vblank) mtk_crtc->pending_needs_vblank = true; for (i = 0; i < mtk_crtc->layer_nr; i++) { struct drm_plane *plane = &mtk_crtc->planes[i]; struct mtk_plane_state *plane_state; plane_state = to_mtk_plane_state(plane->state); if (plane_state->pending.dirty) { plane_state->pending.config = true; plane_state->pending.dirty = false; pending_planes |= BIT(i); } else if (plane_state->pending.async_dirty) { plane_state->pending.async_config = true; plane_state->pending.async_dirty = false; pending_async_planes |= BIT(i); } } if (pending_planes) mtk_crtc->pending_planes = true; if (pending_async_planes) mtk_crtc->pending_async_planes = true; if (priv->data->shadow_register) { mtk_mutex_acquire(mtk_crtc->mutex); mtk_crtc_ddp_config(crtc, NULL); mtk_mutex_release(mtk_crtc->mutex); } #if IS_REACHABLE(CONFIG_MTK_CMDQ) if (mtk_crtc->cmdq_client.chan) { mbox_flush(mtk_crtc->cmdq_client.chan, 2000); cmdq_handle->cmd_buf_size = 0; cmdq_pkt_clear_event(cmdq_handle, mtk_crtc->cmdq_event); cmdq_pkt_wfe(cmdq_handle, mtk_crtc->cmdq_event, false); mtk_crtc_ddp_config(crtc, cmdq_handle); cmdq_pkt_finalize(cmdq_handle); dma_sync_single_for_device(mtk_crtc->cmdq_client.chan->mbox->dev, cmdq_handle->pa_base, cmdq_handle->cmd_buf_size, DMA_TO_DEVICE); /* * CMDQ command should execute in next 3 vblank. * One vblank interrupt before send message (occasionally) * and one vblank interrupt after cmdq done, * so it's timeout after 3 vblank interrupt. * If it fail to execute in next 3 vblank, timeout happen. */ mtk_crtc->cmdq_vblank_cnt = 3; mbox_send_message(mtk_crtc->cmdq_client.chan, cmdq_handle); mbox_client_txdone(mtk_crtc->cmdq_client.chan, 0); } #endif mtk_crtc->config_updating = false; mutex_unlock(&mtk_crtc->hw_lock); } static void mtk_crtc_ddp_irq(void *data) { struct drm_crtc *crtc = data; struct mtk_crtc *mtk_crtc = to_mtk_crtc(crtc); struct mtk_drm_private *priv = crtc->dev->dev_private; #if IS_REACHABLE(CONFIG_MTK_CMDQ) if (!priv->data->shadow_register && !mtk_crtc->cmdq_client.chan) mtk_crtc_ddp_config(crtc, NULL); else if (mtk_crtc->cmdq_vblank_cnt > 0 && --mtk_crtc->cmdq_vblank_cnt == 0) DRM_ERROR("mtk_crtc %d CMDQ execute command timeout!\n", drm_crtc_index(&mtk_crtc->base)); #else if (!priv->data->shadow_register) mtk_crtc_ddp_config(crtc, NULL); #endif mtk_drm_finish_page_flip(mtk_crtc); } static int mtk_crtc_enable_vblank(struct drm_crtc *crtc) { struct mtk_crtc *mtk_crtc = to_mtk_crtc(crtc); struct mtk_ddp_comp *comp = mtk_crtc->ddp_comp[0]; mtk_ddp_comp_enable_vblank(comp); return 0; } static void mtk_crtc_disable_vblank(struct drm_crtc *crtc) { struct mtk_crtc *mtk_crtc = to_mtk_crtc(crtc); struct mtk_ddp_comp *comp = mtk_crtc->ddp_comp[0]; mtk_ddp_comp_disable_vblank(comp); } static void mtk_crtc_update_output(struct drm_crtc *crtc, struct drm_atomic_state *state) { int crtc_index = drm_crtc_index(crtc); int i; struct device *dev; struct drm_crtc_state *crtc_state = state->crtcs[crtc_index].new_state; struct mtk_crtc *mtk_crtc = to_mtk_crtc(crtc); struct mtk_drm_private *priv; unsigned int encoder_mask = crtc_state->encoder_mask; if (!crtc_state->connectors_changed) return; if (!mtk_crtc->num_conn_routes) return; priv = ((struct mtk_drm_private *)crtc->dev->dev_private)->all_drm_private[crtc_index]; dev = priv->dev; dev_dbg(dev, "connector change:%d, encoder mask:0x%x for crtc:%d\n", crtc_state->connectors_changed, encoder_mask, crtc_index); for (i = 0; i < mtk_crtc->num_conn_routes; i++) { unsigned int comp_id = mtk_crtc->conn_routes[i].route_ddp; struct mtk_ddp_comp *comp = &priv->ddp_comp[comp_id]; if (comp->encoder_index >= 0 && (encoder_mask & BIT(comp->encoder_index))) { mtk_crtc->ddp_comp[mtk_crtc->ddp_comp_nr - 1] = comp; dev_dbg(dev, "Add comp_id: %d at path index %d\n", comp->id, mtk_crtc->ddp_comp_nr - 1); break; } } } int mtk_crtc_plane_check(struct drm_crtc *crtc, struct drm_plane *plane, struct mtk_plane_state *state) { unsigned int local_layer; struct mtk_ddp_comp *comp; comp = mtk_ddp_comp_for_plane(crtc, plane, &local_layer); if (comp) return mtk_ddp_comp_layer_check(comp, local_layer, state); return 0; } void mtk_crtc_async_update(struct drm_crtc *crtc, struct drm_plane *plane, struct drm_atomic_state *state) { struct mtk_crtc *mtk_crtc = to_mtk_crtc(crtc); if (!mtk_crtc->enabled) return; mtk_crtc_update_config(mtk_crtc, false); } static void mtk_crtc_atomic_enable(struct drm_crtc *crtc, struct drm_atomic_state *state) { struct mtk_crtc *mtk_crtc = to_mtk_crtc(crtc); struct mtk_ddp_comp *comp = mtk_crtc->ddp_comp[0]; int ret; DRM_DEBUG_DRIVER("%s %d\n", __func__, crtc->base.id); ret = mtk_ddp_comp_power_on(comp); if (ret < 0) { DRM_DEV_ERROR(comp->dev, "Failed to enable power domain: %d\n", ret); return; } mtk_crtc_update_output(crtc, state); ret = mtk_crtc_ddp_hw_init(mtk_crtc); if (ret) { mtk_ddp_comp_power_off(comp); return; } drm_crtc_vblank_on(crtc); mtk_crtc->enabled = true; } static void mtk_crtc_atomic_disable(struct drm_crtc *crtc, struct drm_atomic_state *state) { struct mtk_crtc *mtk_crtc = to_mtk_crtc(crtc); struct mtk_ddp_comp *comp = mtk_crtc->ddp_comp[0]; int i; DRM_DEBUG_DRIVER("%s %d\n", __func__, crtc->base.id); if (!mtk_crtc->enabled) return; /* Set all pending plane state to disabled */ for (i = 0; i < mtk_crtc->layer_nr; i++) { struct drm_plane *plane = &mtk_crtc->planes[i]; struct mtk_plane_state *plane_state; plane_state = to_mtk_plane_state(plane->state); plane_state->pending.enable = false; plane_state->pending.config = true; } mtk_crtc->pending_planes = true; mtk_crtc_update_config(mtk_crtc, false); #if IS_REACHABLE(CONFIG_MTK_CMDQ) /* Wait for planes to be disabled by cmdq */ if (mtk_crtc->cmdq_client.chan) wait_event_timeout(mtk_crtc->cb_blocking_queue, mtk_crtc->cmdq_vblank_cnt == 0, msecs_to_jiffies(500)); #endif /* Wait for planes to be disabled */ drm_crtc_wait_one_vblank(crtc); drm_crtc_vblank_off(crtc); mtk_crtc_ddp_hw_fini(mtk_crtc); mtk_ddp_comp_power_off(comp); mtk_crtc->enabled = false; } static void mtk_crtc_atomic_begin(struct drm_crtc *crtc, struct drm_atomic_state *state) { struct drm_crtc_state *crtc_state = drm_atomic_get_new_crtc_state(state, crtc); struct mtk_crtc_state *mtk_crtc_state = to_mtk_crtc_state(crtc_state); struct mtk_crtc *mtk_crtc = to_mtk_crtc(crtc); unsigned long flags; if (mtk_crtc->event && mtk_crtc_state->base.event) DRM_ERROR("new event while there is still a pending event\n"); if (mtk_crtc_state->base.event) { mtk_crtc_state->base.event->pipe = drm_crtc_index(crtc); WARN_ON(drm_crtc_vblank_get(crtc) != 0); spin_lock_irqsave(&crtc->dev->event_lock, flags); mtk_crtc->event = mtk_crtc_state->base.event; spin_unlock_irqrestore(&crtc->dev->event_lock, flags); mtk_crtc_state->base.event = NULL; } } static void mtk_crtc_atomic_flush(struct drm_crtc *crtc, struct drm_atomic_state *state) { struct mtk_crtc *mtk_crtc = to_mtk_crtc(crtc); int i; if (crtc->state->color_mgmt_changed) for (i = 0; i < mtk_crtc->ddp_comp_nr; i++) { mtk_ddp_gamma_set(mtk_crtc->ddp_comp[i], crtc->state); mtk_ddp_ctm_set(mtk_crtc->ddp_comp[i], crtc->state); } mtk_crtc_update_config(mtk_crtc, !!mtk_crtc->event); } static const struct drm_crtc_funcs mtk_crtc_funcs = { .set_config = drm_atomic_helper_set_config, .page_flip = drm_atomic_helper_page_flip, .destroy = mtk_crtc_destroy, .reset = mtk_crtc_reset, .atomic_duplicate_state = mtk_crtc_duplicate_state, .atomic_destroy_state = mtk_crtc_destroy_state, .enable_vblank = mtk_crtc_enable_vblank, .disable_vblank = mtk_crtc_disable_vblank, }; static const struct drm_crtc_helper_funcs mtk_crtc_helper_funcs = { .mode_fixup = mtk_crtc_mode_fixup, .mode_set_nofb = mtk_crtc_mode_set_nofb, .mode_valid = mtk_crtc_mode_valid, .atomic_begin = mtk_crtc_atomic_begin, .atomic_flush = mtk_crtc_atomic_flush, .atomic_enable = mtk_crtc_atomic_enable, .atomic_disable = mtk_crtc_atomic_disable, }; static int mtk_crtc_init(struct drm_device *drm, struct mtk_crtc *mtk_crtc, unsigned int pipe) { struct drm_plane *primary = NULL; struct drm_plane *cursor = NULL; int i, ret; for (i = 0; i < mtk_crtc->layer_nr; i++) { if (mtk_crtc->planes[i].type == DRM_PLANE_TYPE_PRIMARY) primary = &mtk_crtc->planes[i]; else if (mtk_crtc->planes[i].type == DRM_PLANE_TYPE_CURSOR) cursor = &mtk_crtc->planes[i]; } ret = drm_crtc_init_with_planes(drm, &mtk_crtc->base, primary, cursor, &mtk_crtc_funcs, NULL); if (ret) goto err_cleanup_crtc; drm_crtc_helper_add(&mtk_crtc->base, &mtk_crtc_helper_funcs); return 0; err_cleanup_crtc: drm_crtc_cleanup(&mtk_crtc->base); return ret; } static int mtk_crtc_num_comp_planes(struct mtk_crtc *mtk_crtc, int comp_idx) { struct mtk_ddp_comp *comp; if (comp_idx > 1) return 0; comp = mtk_crtc->ddp_comp[comp_idx]; if (!comp->funcs) return 0; if (comp_idx == 1 && !comp->funcs->bgclr_in_on) return 0; return mtk_ddp_comp_layer_nr(comp); } static inline enum drm_plane_type mtk_crtc_plane_type(unsigned int plane_idx, unsigned int num_planes) { if (plane_idx == 0) return DRM_PLANE_TYPE_PRIMARY; else if (plane_idx == (num_planes - 1)) return DRM_PLANE_TYPE_CURSOR; else return DRM_PLANE_TYPE_OVERLAY; } static int mtk_crtc_init_comp_planes(struct drm_device *drm_dev, struct mtk_crtc *mtk_crtc, int comp_idx, int pipe) { int num_planes = mtk_crtc_num_comp_planes(mtk_crtc, comp_idx); struct mtk_ddp_comp *comp = mtk_crtc->ddp_comp[comp_idx]; int i, ret; for (i = 0; i < num_planes; i++) { ret = mtk_plane_init(drm_dev, &mtk_crtc->planes[mtk_crtc->layer_nr], BIT(pipe), mtk_crtc_plane_type(mtk_crtc->layer_nr, num_planes), mtk_ddp_comp_supported_rotations(comp), mtk_ddp_comp_get_formats(comp), mtk_ddp_comp_get_num_formats(comp)); if (ret) return ret; mtk_crtc->layer_nr++; } return 0; } struct device *mtk_crtc_dma_dev_get(struct drm_crtc *crtc) { struct mtk_crtc *mtk_crtc = NULL; if (!crtc) return NULL; mtk_crtc = to_mtk_crtc(crtc); if (!mtk_crtc) return NULL; return mtk_crtc->dma_dev; } int mtk_crtc_create(struct drm_device *drm_dev, const unsigned int *path, unsigned int path_len, int priv_data_index, const struct mtk_drm_route *conn_routes, unsigned int num_conn_routes) { struct mtk_drm_private *priv = drm_dev->dev_private; struct device *dev = drm_dev->dev; struct mtk_crtc *mtk_crtc; unsigned int num_comp_planes = 0; int ret; int i; bool has_ctm = false; uint gamma_lut_size = 0; struct drm_crtc *tmp; int crtc_i = 0; if (!path) return 0; priv = priv->all_drm_private[priv_data_index]; drm_for_each_crtc(tmp, drm_dev) crtc_i++; for (i = 0; i < path_len; i++) { enum mtk_ddp_comp_id comp_id = path[i]; struct device_node *node; struct mtk_ddp_comp *comp; node = priv->comp_node[comp_id]; comp = &priv->ddp_comp[comp_id]; /* Not all drm components have a DTS device node, such as ovl_adaptor, * which is the drm bring up sub driver */ if (!node && comp_id != DDP_COMPONENT_DRM_OVL_ADAPTOR) { dev_info(dev, "Not creating crtc %d because component %d is disabled or missing\n", crtc_i, comp_id); return 0; } if (!comp->dev) { dev_err(dev, "Component %pOF not initialized\n", node); return -ENODEV; } } mtk_crtc = devm_kzalloc(dev, sizeof(*mtk_crtc), GFP_KERNEL); if (!mtk_crtc) return -ENOMEM; mtk_crtc->mmsys_dev = priv->mmsys_dev; mtk_crtc->ddp_comp_nr = path_len; mtk_crtc->ddp_comp = devm_kcalloc(dev, mtk_crtc->ddp_comp_nr + (conn_routes ? 1 : 0), sizeof(*mtk_crtc->ddp_comp), GFP_KERNEL); if (!mtk_crtc->ddp_comp) return -ENOMEM; mtk_crtc->mutex = mtk_mutex_get(priv->mutex_dev); if (IS_ERR(mtk_crtc->mutex)) { ret = PTR_ERR(mtk_crtc->mutex); dev_err(dev, "Failed to get mutex: %d\n", ret); return ret; } for (i = 0; i < mtk_crtc->ddp_comp_nr; i++) { unsigned int comp_id = path[i]; struct mtk_ddp_comp *comp; comp = &priv->ddp_comp[comp_id]; mtk_crtc->ddp_comp[i] = comp; if (comp->funcs) { if (comp->funcs->gamma_set && comp->funcs->gamma_get_lut_size) { unsigned int lut_sz = mtk_ddp_gamma_get_lut_size(comp); if (lut_sz) gamma_lut_size = lut_sz; } if (comp->funcs->ctm_set) has_ctm = true; } mtk_ddp_comp_register_vblank_cb(comp, mtk_crtc_ddp_irq, &mtk_crtc->base); } for (i = 0; i < mtk_crtc->ddp_comp_nr; i++) num_comp_planes += mtk_crtc_num_comp_planes(mtk_crtc, i); mtk_crtc->planes = devm_kcalloc(dev, num_comp_planes, sizeof(struct drm_plane), GFP_KERNEL); if (!mtk_crtc->planes) return -ENOMEM; for (i = 0; i < mtk_crtc->ddp_comp_nr; i++) { ret = mtk_crtc_init_comp_planes(drm_dev, mtk_crtc, i, crtc_i); if (ret) return ret; } /* * Default to use the first component as the dma dev. * In the case of ovl_adaptor sub driver, it needs to use the * dma_dev_get function to get representative dma dev. */ mtk_crtc->dma_dev = mtk_ddp_comp_dma_dev_get(&priv->ddp_comp[path[0]]); ret = mtk_crtc_init(drm_dev, mtk_crtc, crtc_i); if (ret < 0) return ret; if (gamma_lut_size) drm_mode_crtc_set_gamma_size(&mtk_crtc->base, gamma_lut_size); drm_crtc_enable_color_mgmt(&mtk_crtc->base, 0, has_ctm, gamma_lut_size); mutex_init(&mtk_crtc->hw_lock); #if IS_REACHABLE(CONFIG_MTK_CMDQ) i = priv->mbox_index++; mtk_crtc->cmdq_client.client.dev = mtk_crtc->mmsys_dev; mtk_crtc->cmdq_client.client.tx_block = false; mtk_crtc->cmdq_client.client.knows_txdone = true; mtk_crtc->cmdq_client.client.rx_callback = ddp_cmdq_cb; mtk_crtc->cmdq_client.chan = mbox_request_channel(&mtk_crtc->cmdq_client.client, i); if (IS_ERR(mtk_crtc->cmdq_client.chan)) { dev_dbg(dev, "mtk_crtc %d failed to create mailbox client, writing register by CPU now\n", drm_crtc_index(&mtk_crtc->base)); mtk_crtc->cmdq_client.chan = NULL; } if (mtk_crtc->cmdq_client.chan) { ret = of_property_read_u32_index(priv->mutex_node, "mediatek,gce-events", i, &mtk_crtc->cmdq_event); if (ret) { dev_dbg(dev, "mtk_crtc %d failed to get mediatek,gce-events property\n", drm_crtc_index(&mtk_crtc->base)); mbox_free_channel(mtk_crtc->cmdq_client.chan); mtk_crtc->cmdq_client.chan = NULL; } else { ret = mtk_drm_cmdq_pkt_create(&mtk_crtc->cmdq_client, &mtk_crtc->cmdq_handle, PAGE_SIZE); if (ret) { dev_dbg(dev, "mtk_crtc %d failed to create cmdq packet\n", drm_crtc_index(&mtk_crtc->base)); mbox_free_channel(mtk_crtc->cmdq_client.chan); mtk_crtc->cmdq_client.chan = NULL; } } /* for sending blocking cmd in crtc disable */ init_waitqueue_head(&mtk_crtc->cb_blocking_queue); } #endif if (conn_routes) { for (i = 0; i < num_conn_routes; i++) { unsigned int comp_id = conn_routes[i].route_ddp; struct device_node *node = priv->comp_node[comp_id]; struct mtk_ddp_comp *comp = &priv->ddp_comp[comp_id]; if (!comp->dev) { dev_dbg(dev, "comp_id:%d, Component %pOF not initialized\n", comp_id, node); /* mark encoder_index to -1, if route comp device is not enabled */ comp->encoder_index = -1; continue; } mtk_ddp_comp_encoder_index_set(&priv->ddp_comp[comp_id]); } mtk_crtc->num_conn_routes = num_conn_routes; mtk_crtc->conn_routes = conn_routes; /* increase ddp_comp_nr at the end of mtk_crtc_create */ mtk_crtc->ddp_comp_nr++; } return 0; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1