Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Qingqing Zhuo | 4480 | 88.59% | 2 | 5.71% |
Alvin lee | 135 | 2.67% | 6 | 17.14% |
Dillon Varone | 121 | 2.39% | 4 | 11.43% |
Alex Deucher | 101 | 2.00% | 3 | 8.57% |
Charlene Liu | 61 | 1.21% | 3 | 8.57% |
Bhawanpreet Lakha | 38 | 0.75% | 2 | 5.71% |
Harry Wentland | 25 | 0.49% | 1 | 2.86% |
Nicholas Kazlauskas | 24 | 0.47% | 2 | 5.71% |
rodrigosiqueira | 20 | 0.40% | 1 | 2.86% |
Srinivasan S | 20 | 0.40% | 1 | 2.86% |
Aurabindo Pillai | 13 | 0.26% | 1 | 2.86% |
Dmytro Laktyushkin | 6 | 0.12% | 1 | 2.86% |
Hersen Wu | 6 | 0.12% | 1 | 2.86% |
Jerry (Fangzhi) Zuo | 1 | 0.02% | 1 | 2.86% |
Eric Yang | 1 | 0.02% | 1 | 2.86% |
Stylon Wang | 1 | 0.02% | 1 | 2.86% |
Taimur Hassan | 1 | 0.02% | 1 | 2.86% |
Jun Lei | 1 | 0.02% | 1 | 2.86% |
Joshua Aberback | 1 | 0.02% | 1 | 2.86% |
Natanel Roizenman | 1 | 0.02% | 1 | 2.86% |
Total | 5057 | 35 |
/* SPDX-License-Identifier: MIT */ /* * Copyright 2023 Advanced Micro Devices, Inc. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR * OTHER DEALINGS IN THE SOFTWARE. * * Authors: AMD * */ #include "display_mode_core.h" #include "dml2_internal_types.h" #include "dml2_utils.h" #include "dml2_policy.h" #include "dml2_translation_helper.h" #include "dml2_mall_phantom.h" #include "dml2_dc_resource_mgmt.h" static void initialize_dml2_ip_params(struct dml2_context *dml2, const struct dc *in_dc, struct ip_params_st *out) { if (dml2->config.use_native_soc_bb_construction) dml2_init_ip_params(dml2, in_dc, out); else dml2_translate_ip_params(in_dc, out); } static void initialize_dml2_soc_bbox(struct dml2_context *dml2, const struct dc *in_dc, struct soc_bounding_box_st *out) { if (dml2->config.use_native_soc_bb_construction) dml2_init_socbb_params(dml2, in_dc, out); else dml2_translate_socbb_params(in_dc, out); } static void initialize_dml2_soc_states(struct dml2_context *dml2, const struct dc *in_dc, const struct soc_bounding_box_st *in_bbox, struct soc_states_st *out) { if (dml2->config.use_native_soc_bb_construction) dml2_init_soc_states(dml2, in_dc, in_bbox, out); else dml2_translate_soc_states(in_dc, out, in_dc->dml.soc.num_states); } static void map_hw_resources(struct dml2_context *dml2, struct dml_display_cfg_st *in_out_display_cfg, struct dml_mode_support_info_st *mode_support_info) { unsigned int num_pipes = 0; int i, j; for (i = 0; i < __DML_NUM_PLANES__; i++) { in_out_display_cfg->hw.ODMMode[i] = mode_support_info->ODMMode[i]; in_out_display_cfg->hw.DPPPerSurface[i] = mode_support_info->DPPPerSurface[i]; in_out_display_cfg->hw.DSCEnabled[i] = mode_support_info->DSCEnabled[i]; in_out_display_cfg->hw.NumberOfDSCSlices[i] = mode_support_info->NumberOfDSCSlices[i]; in_out_display_cfg->hw.DLGRefClkFreqMHz = 24; if (dml2->v20.dml_core_ctx.project != dml_project_dcn35 && dml2->v20.dml_core_ctx.project != dml_project_dcn351) { /*dGPU default as 50Mhz*/ in_out_display_cfg->hw.DLGRefClkFreqMHz = 50; } for (j = 0; j < mode_support_info->DPPPerSurface[i]; j++) { if (i >= __DML2_WRAPPER_MAX_STREAMS_PLANES__) { dml_print("DML::%s: Index out of bounds: i=%d, __DML2_WRAPPER_MAX_STREAMS_PLANES__=%d\n", __func__, i, __DML2_WRAPPER_MAX_STREAMS_PLANES__); break; } dml2->v20.scratch.dml_to_dc_pipe_mapping.dml_pipe_idx_to_stream_id[num_pipes] = dml2->v20.scratch.dml_to_dc_pipe_mapping.disp_cfg_to_stream_id[i]; dml2->v20.scratch.dml_to_dc_pipe_mapping.dml_pipe_idx_to_stream_id_valid[num_pipes] = true; dml2->v20.scratch.dml_to_dc_pipe_mapping.dml_pipe_idx_to_plane_id[num_pipes] = dml2->v20.scratch.dml_to_dc_pipe_mapping.disp_cfg_to_plane_id[i]; dml2->v20.scratch.dml_to_dc_pipe_mapping.dml_pipe_idx_to_plane_id_valid[num_pipes] = true; num_pipes++; } } } static unsigned int pack_and_call_dml_mode_support_ex(struct dml2_context *dml2, const struct dml_display_cfg_st *display_cfg, struct dml_mode_support_info_st *evaluation_info) { struct dml2_wrapper_scratch *s = &dml2->v20.scratch; s->mode_support_params.mode_lib = &dml2->v20.dml_core_ctx; s->mode_support_params.in_display_cfg = display_cfg; s->mode_support_params.out_evaluation_info = evaluation_info; memset(evaluation_info, 0, sizeof(struct dml_mode_support_info_st)); s->mode_support_params.out_lowest_state_idx = 0; return dml_mode_support_ex(&s->mode_support_params); } static bool optimize_configuration(struct dml2_context *dml2, struct dml2_wrapper_optimize_configuration_params *p) { int unused_dpps = p->ip_params->max_num_dpp; int i, j; int odms_needed, refresh_rate_hz, dpps_needed, subvp_height, pstate_width_fw_delay_lines, surface_count; int subvp_timing_to_add, new_timing_index, subvp_surface_to_add, new_surface_index; float frame_time_sec, max_frame_time_sec; int largest_blend_and_timing = 0; bool optimization_done = false; for (i = 0; i < (int) p->cur_display_config->num_timings; i++) { if (p->cur_display_config->plane.BlendingAndTiming[i] > largest_blend_and_timing) largest_blend_and_timing = p->cur_display_config->plane.BlendingAndTiming[i]; } if (p->new_policy != p->cur_policy) *p->new_policy = *p->cur_policy; if (p->new_display_config != p->cur_display_config) *p->new_display_config = *p->cur_display_config; // Optimize P-State Support if (dml2->config.use_native_pstate_optimization) { if (p->cur_mode_support_info->DRAMClockChangeSupport[0] == dml_dram_clock_change_unsupported) { // Find a display with < 120Hz refresh rate with maximal refresh rate that's not already subvp subvp_timing_to_add = -1; subvp_surface_to_add = -1; max_frame_time_sec = 0; surface_count = 0; for (i = 0; i < (int) p->cur_display_config->num_timings; i++) { refresh_rate_hz = (int)div_u64((unsigned long long) p->cur_display_config->timing.PixelClock[i] * 1000 * 1000, (p->cur_display_config->timing.HTotal[i] * p->cur_display_config->timing.VTotal[i])); if (refresh_rate_hz < 120) { // Check its upstream surfaces to see if this one could be converted to subvp. dpps_needed = 0; for (j = 0; j < (int) p->cur_display_config->num_surfaces; j++) { if (p->cur_display_config->plane.BlendingAndTiming[j] == i && p->cur_display_config->plane.UseMALLForPStateChange[j] == dml_use_mall_pstate_change_disable) { dpps_needed += p->cur_mode_support_info->DPPPerSurface[j]; subvp_surface_to_add = j; surface_count++; } } if (surface_count == 1 && dpps_needed > 0 && dpps_needed <= unused_dpps) { frame_time_sec = (float)1 / refresh_rate_hz; if (frame_time_sec > max_frame_time_sec) { max_frame_time_sec = frame_time_sec; subvp_timing_to_add = i; } } } } if (subvp_timing_to_add >= 0) { new_timing_index = p->new_display_config->num_timings++; new_surface_index = p->new_display_config->num_surfaces++; // Add a phantom pipe reflecting the main pipe's timing dml2_util_copy_dml_timing(&p->new_display_config->timing, new_timing_index, subvp_timing_to_add); pstate_width_fw_delay_lines = (int)(((double)(p->config->svp_pstate.subvp_fw_processing_delay_us + p->config->svp_pstate.subvp_pstate_allow_width_us) / 1000000) * (p->new_display_config->timing.PixelClock[subvp_timing_to_add] * 1000 * 1000) / (double)p->new_display_config->timing.HTotal[subvp_timing_to_add]); subvp_height = p->cur_mode_support_info->SubViewportLinesNeededInMALL[subvp_timing_to_add] + pstate_width_fw_delay_lines; p->new_display_config->timing.VActive[new_timing_index] = subvp_height; p->new_display_config->timing.VTotal[new_timing_index] = subvp_height + p->new_display_config->timing.VTotal[subvp_timing_to_add] - p->new_display_config->timing.VActive[subvp_timing_to_add]; p->new_display_config->output.OutputDisabled[new_timing_index] = true; p->new_display_config->plane.UseMALLForPStateChange[subvp_surface_to_add] = dml_use_mall_pstate_change_sub_viewport; dml2_util_copy_dml_plane(&p->new_display_config->plane, new_surface_index, subvp_surface_to_add); dml2_util_copy_dml_surface(&p->new_display_config->surface, new_surface_index, subvp_surface_to_add); p->new_display_config->plane.ViewportHeight[new_surface_index] = subvp_height; p->new_display_config->plane.ViewportHeightChroma[new_surface_index] = subvp_height; p->new_display_config->plane.ViewportStationary[new_surface_index] = false; p->new_display_config->plane.UseMALLForStaticScreen[new_surface_index] = dml_use_mall_static_screen_disable; p->new_display_config->plane.UseMALLForPStateChange[new_surface_index] = dml_use_mall_pstate_change_phantom_pipe; p->new_display_config->plane.NumberOfCursors[new_surface_index] = 0; p->new_policy->ImmediateFlipRequirement[new_surface_index] = dml_immediate_flip_not_required; p->new_display_config->plane.BlendingAndTiming[new_surface_index] = new_timing_index; optimization_done = true; } } } // Optimize Clocks if (!optimization_done) { if (largest_blend_and_timing == 0 && p->cur_policy->ODMUse[0] == dml_odm_use_policy_combine_as_needed && dml2->config.minimize_dispclk_using_odm) { odms_needed = dml2_util_get_maximum_odm_combine_for_output(dml2->config.optimize_odm_4to1, p->cur_display_config->output.OutputEncoder[0], p->cur_mode_support_info->DSCEnabled[0]) - 1; if (odms_needed <= unused_dpps) { unused_dpps -= odms_needed; if (odms_needed == 1) { p->new_policy->ODMUse[0] = dml_odm_use_policy_combine_2to1; optimization_done = true; } else if (odms_needed == 3) { p->new_policy->ODMUse[0] = dml_odm_use_policy_combine_4to1; optimization_done = true; } else optimization_done = false; } } } return optimization_done; } static int calculate_lowest_supported_state_for_temp_read(struct dml2_context *dml2, struct dc_state *display_state) { struct dml2_calculate_lowest_supported_state_for_temp_read_scratch *s = &dml2->v20.scratch.dml2_calculate_lowest_supported_state_for_temp_read_scratch; struct dml2_wrapper_scratch *s_global = &dml2->v20.scratch; unsigned int dml_result = 0; int result = -1, i, j; build_unoptimized_policy_settings(dml2->v20.dml_core_ctx.project, &dml2->v20.dml_core_ctx.policy); /* Zero out before each call before proceeding */ memset(s, 0, sizeof(struct dml2_calculate_lowest_supported_state_for_temp_read_scratch)); memset(&s_global->mode_support_params, 0, sizeof(struct dml_mode_support_ex_params_st)); memset(&s_global->dml_to_dc_pipe_mapping, 0, sizeof(struct dml2_dml_to_dc_pipe_mapping)); for (i = 0; i < dml2->config.dcn_pipe_count; i++) { /* Calling resource_build_scaling_params will populate the pipe params * with the necessary information needed for correct DML calculations * This is also done in DML1 driver code path and hence display_state * cannot be const. */ struct pipe_ctx *pipe = &display_state->res_ctx.pipe_ctx[i]; if (pipe->plane_state) { if (!dml2->config.callbacks.build_scaling_params(pipe)) { ASSERT(false); return false; } } } map_dc_state_into_dml_display_cfg(dml2, display_state, &s->cur_display_config); for (i = 0; i < dml2->v20.dml_core_ctx.states.num_states; i++) { s->uclk_change_latencies[i] = dml2->v20.dml_core_ctx.states.state_array[i].dram_clock_change_latency_us; } for (i = 0; i < 4; i++) { for (j = 0; j < dml2->v20.dml_core_ctx.states.num_states; j++) { dml2->v20.dml_core_ctx.states.state_array[j].dram_clock_change_latency_us = s_global->dummy_pstate_table[i].dummy_pstate_latency_us; } dml_result = pack_and_call_dml_mode_support_ex(dml2, &s->cur_display_config, &s->evaluation_info); if (dml_result && s->evaluation_info.DRAMClockChangeSupport[0] == dml_dram_clock_change_vactive) { map_hw_resources(dml2, &s->cur_display_config, &s->evaluation_info); dml_result = dml_mode_programming(&dml2->v20.dml_core_ctx, s_global->mode_support_params.out_lowest_state_idx, &s->cur_display_config, true); ASSERT(dml_result); dml2_extract_watermark_set(&dml2->v20.g6_temp_read_watermark_set, &dml2->v20.dml_core_ctx); dml2->v20.g6_temp_read_watermark_set.cstate_pstate.fclk_pstate_change_ns = dml2->v20.g6_temp_read_watermark_set.cstate_pstate.pstate_change_ns; result = s_global->mode_support_params.out_lowest_state_idx; while (dml2->v20.dml_core_ctx.states.state_array[result].dram_speed_mts < s_global->dummy_pstate_table[i].dram_speed_mts) result++; break; } } for (i = 0; i < dml2->v20.dml_core_ctx.states.num_states; i++) { dml2->v20.dml_core_ctx.states.state_array[i].dram_clock_change_latency_us = s->uclk_change_latencies[i]; } return result; } static void copy_dummy_pstate_table(struct dummy_pstate_entry *dest, struct dummy_pstate_entry *src, unsigned int num_entries) { for (int i = 0; i < num_entries; i++) { dest[i] = src[i]; } } static bool are_timings_requiring_odm_doing_blending(const struct dml_display_cfg_st *display_cfg, const struct dml_mode_support_info_st *evaluation_info) { unsigned int planes_per_timing[__DML_NUM_PLANES__] = {0}; int i; for (i = 0; i < display_cfg->num_surfaces; i++) planes_per_timing[display_cfg->plane.BlendingAndTiming[i]]++; for (i = 0; i < __DML_NUM_PLANES__; i++) { if (planes_per_timing[i] > 1 && evaluation_info->ODMMode[i] != dml_odm_mode_bypass) return true; } return false; } static bool does_configuration_meet_sw_policies(struct dml2_context *ctx, const struct dml_display_cfg_st *display_cfg, const struct dml_mode_support_info_st *evaluation_info) { bool pass = true; if (!ctx->config.enable_windowed_mpo_odm) { if (are_timings_requiring_odm_doing_blending(display_cfg, evaluation_info)) pass = false; } return pass; } static bool dml_mode_support_wrapper(struct dml2_context *dml2, struct dc_state *display_state) { struct dml2_wrapper_scratch *s = &dml2->v20.scratch; unsigned int result = 0, i; unsigned int optimized_result = true; build_unoptimized_policy_settings(dml2->v20.dml_core_ctx.project, &dml2->v20.dml_core_ctx.policy); /* Zero out before each call before proceeding */ memset(&s->cur_display_config, 0, sizeof(struct dml_display_cfg_st)); memset(&s->mode_support_params, 0, sizeof(struct dml_mode_support_ex_params_st)); memset(&s->dml_to_dc_pipe_mapping, 0, sizeof(struct dml2_dml_to_dc_pipe_mapping)); memset(&s->optimize_configuration_params, 0, sizeof(struct dml2_wrapper_optimize_configuration_params)); for (i = 0; i < dml2->config.dcn_pipe_count; i++) { /* Calling resource_build_scaling_params will populate the pipe params * with the necessary information needed for correct DML calculations * This is also done in DML1 driver code path and hence display_state * cannot be const. */ struct pipe_ctx *pipe = &display_state->res_ctx.pipe_ctx[i]; if (pipe->plane_state) { if (!dml2->config.callbacks.build_scaling_params(pipe)) { ASSERT(false); return false; } } } map_dc_state_into_dml_display_cfg(dml2, display_state, &s->cur_display_config); if (!dml2->config.skip_hw_state_mapping) dml2_apply_det_buffer_allocation_policy(dml2, &s->cur_display_config); result = pack_and_call_dml_mode_support_ex(dml2, &s->cur_display_config, &s->mode_support_info); if (result) result = does_configuration_meet_sw_policies(dml2, &s->cur_display_config, &s->mode_support_info); // Try to optimize if (result) { s->cur_policy = dml2->v20.dml_core_ctx.policy; s->optimize_configuration_params.dml_core_ctx = &dml2->v20.dml_core_ctx; s->optimize_configuration_params.config = &dml2->config; s->optimize_configuration_params.ip_params = &dml2->v20.dml_core_ctx.ip; s->optimize_configuration_params.cur_display_config = &s->cur_display_config; s->optimize_configuration_params.cur_mode_support_info = &s->mode_support_info; s->optimize_configuration_params.cur_policy = &s->cur_policy; s->optimize_configuration_params.new_display_config = &s->new_display_config; s->optimize_configuration_params.new_policy = &s->new_policy; while (optimized_result && optimize_configuration(dml2, &s->optimize_configuration_params)) { dml2->v20.dml_core_ctx.policy = s->new_policy; optimized_result = pack_and_call_dml_mode_support_ex(dml2, &s->new_display_config, &s->mode_support_info); if (optimized_result) optimized_result = does_configuration_meet_sw_policies(dml2, &s->new_display_config, &s->mode_support_info); // If the new optimized state is supposed, then set current = new if (optimized_result) { s->cur_display_config = s->new_display_config; s->cur_policy = s->new_policy; } else { // Else, restore policy to current dml2->v20.dml_core_ctx.policy = s->cur_policy; } } // Optimize ended with a failed config, so we need to restore DML state to last passing if (!optimized_result) { result = pack_and_call_dml_mode_support_ex(dml2, &s->cur_display_config, &s->mode_support_info); } } if (result) map_hw_resources(dml2, &s->cur_display_config, &s->mode_support_info); return result; } static int find_drr_eligible_stream(struct dc_state *display_state) { int i; for (i = 0; i < display_state->stream_count; i++) { if (dc_state_get_stream_subvp_type(display_state, display_state->streams[i]) == SUBVP_NONE && display_state->streams[i]->ignore_msa_timing_param) { // Use ignore_msa_timing_param flag to identify as DRR return i; } } return -1; } static bool optimize_pstate_with_svp_and_drr(struct dml2_context *dml2, struct dc_state *display_state) { struct dml2_wrapper_scratch *s = &dml2->v20.scratch; bool pstate_optimization_done = false; bool pstate_optimization_success = false; bool result = false; int drr_display_index = 0, non_svp_streams = 0; bool force_svp = dml2->config.svp_pstate.force_enable_subvp; bool advanced_pstate_switching = false; display_state->bw_ctx.bw.dcn.clk.fw_based_mclk_switching = false; display_state->bw_ctx.bw.dcn.legacy_svp_drr_stream_index_valid = false; result = dml_mode_support_wrapper(dml2, display_state); if (!result) { pstate_optimization_done = true; } else if (!advanced_pstate_switching || (s->mode_support_info.DRAMClockChangeSupport[0] != dml_dram_clock_change_unsupported && !force_svp)) { pstate_optimization_success = true; pstate_optimization_done = true; } if (display_state->stream_count == 1 && dml2->config.callbacks.can_support_mclk_switch_using_fw_based_vblank_stretch(dml2->config.callbacks.dc, display_state)) { display_state->bw_ctx.bw.dcn.clk.fw_based_mclk_switching = true; result = dml_mode_support_wrapper(dml2, display_state); } else { non_svp_streams = display_state->stream_count; while (!pstate_optimization_done) { result = dml_mode_programming(&dml2->v20.dml_core_ctx, s->mode_support_params.out_lowest_state_idx, &s->cur_display_config, true); // Always try adding SVP first if (result) result = dml2_svp_add_phantom_pipe_to_dc_state(dml2, display_state, &s->mode_support_info); else pstate_optimization_done = true; if (result) { result = dml_mode_support_wrapper(dml2, display_state); } else { pstate_optimization_done = true; } if (result) { non_svp_streams--; if (s->mode_support_info.DRAMClockChangeSupport[0] != dml_dram_clock_change_unsupported) { if (dml2_svp_validate_static_schedulability(dml2, display_state, s->mode_support_info.DRAMClockChangeSupport[0])) { pstate_optimization_success = true; pstate_optimization_done = true; } else { pstate_optimization_success = false; pstate_optimization_done = false; } } else { drr_display_index = find_drr_eligible_stream(display_state); // If there is only 1 remaining non SubVP pipe that is DRR, check static // schedulability for SubVP + DRR. if (non_svp_streams == 1 && drr_display_index >= 0) { if (dml2_svp_drr_schedulable(dml2, display_state, &display_state->streams[drr_display_index]->timing)) { display_state->bw_ctx.bw.dcn.legacy_svp_drr_stream_index_valid = true; display_state->bw_ctx.bw.dcn.legacy_svp_drr_stream_index = drr_display_index; result = dml_mode_support_wrapper(dml2, display_state); } if (result && s->mode_support_info.DRAMClockChangeSupport[0] != dml_dram_clock_change_unsupported) { pstate_optimization_success = true; pstate_optimization_done = true; } else { pstate_optimization_success = false; pstate_optimization_done = false; } } if (pstate_optimization_success) { pstate_optimization_done = true; } else { pstate_optimization_done = false; } } } } } if (!pstate_optimization_success) { dml2_svp_remove_all_phantom_pipes(dml2, display_state); display_state->bw_ctx.bw.dcn.clk.fw_based_mclk_switching = false; display_state->bw_ctx.bw.dcn.legacy_svp_drr_stream_index_valid = false; result = dml_mode_support_wrapper(dml2, display_state); } return result; } static bool call_dml_mode_support_and_programming(struct dc_state *context) { unsigned int result = 0; unsigned int min_state; int min_state_for_g6_temp_read = 0; struct dml2_context *dml2 = context->bw_ctx.dml2; struct dml2_wrapper_scratch *s = &dml2->v20.scratch; min_state_for_g6_temp_read = calculate_lowest_supported_state_for_temp_read(dml2, context); ASSERT(min_state_for_g6_temp_read >= 0); if (!dml2->config.use_native_pstate_optimization) { result = optimize_pstate_with_svp_and_drr(dml2, context); } else { result = dml_mode_support_wrapper(dml2, context); } /* Upon trying to sett certain frequencies in FRL, min_state_for_g6_temp_read is reported as -1. This leads to an invalid value of min_state causing crashes later on. * Use the default logic for min_state only when min_state_for_g6_temp_read is a valid value. In other cases, use the value calculated by the DML directly. */ if (min_state_for_g6_temp_read >= 0) min_state = min_state_for_g6_temp_read > s->mode_support_params.out_lowest_state_idx ? min_state_for_g6_temp_read : s->mode_support_params.out_lowest_state_idx; else min_state = s->mode_support_params.out_lowest_state_idx; if (result) result = dml_mode_programming(&dml2->v20.dml_core_ctx, min_state, &s->cur_display_config, true); return result; } static bool dml2_validate_and_build_resource(const struct dc *in_dc, struct dc_state *context) { struct dml2_context *dml2 = context->bw_ctx.dml2; struct dml2_wrapper_scratch *s = &dml2->v20.scratch; struct dml2_dcn_clocks out_clks; unsigned int result = 0; bool need_recalculation = false; uint32_t cstate_enter_plus_exit_z8_ns; if (!context || context->stream_count == 0) return true; /* Zero out before each call before proceeding */ memset(&dml2->v20.scratch, 0, sizeof(struct dml2_wrapper_scratch)); memset(&dml2->v20.dml_core_ctx.policy, 0, sizeof(struct dml_mode_eval_policy_st)); memset(&dml2->v20.dml_core_ctx.ms, 0, sizeof(struct mode_support_st)); memset(&dml2->v20.dml_core_ctx.mp, 0, sizeof(struct mode_program_st)); /* Initialize DET scratch */ dml2_initialize_det_scratch(dml2); copy_dummy_pstate_table(s->dummy_pstate_table, in_dc->clk_mgr->bw_params->dummy_pstate_table, 4); result = call_dml_mode_support_and_programming(context); /* Call map dc pipes to map the pipes based on the DML output. For correctly determining if recalculation * is required or not, the resource context needs to correctly reflect the number of active pipes. We would * only know the correct number if active pipes after dml2_map_dc_pipes is called. */ if (result && !dml2->config.skip_hw_state_mapping) dml2_map_dc_pipes(dml2, context, &s->cur_display_config, &s->dml_to_dc_pipe_mapping, in_dc->current_state); /* Verify and update DET Buffer configuration if needed. dml2_verify_det_buffer_configuration will check if DET Buffer * size needs to be updated. If yes it will update the DETOverride variable and set need_recalculation flag to true. * Based on that flag, run mode support again. Verification needs to be run after dml_mode_programming because the getters * return correct det buffer values only after dml_mode_programming is called. */ if (result && !dml2->config.skip_hw_state_mapping) { need_recalculation = dml2_verify_det_buffer_configuration(dml2, context, &dml2->det_helper_scratch); if (need_recalculation) { /* Engage the DML again if recalculation is required. */ call_dml_mode_support_and_programming(context); if (!dml2->config.skip_hw_state_mapping) { dml2_map_dc_pipes(dml2, context, &s->cur_display_config, &s->dml_to_dc_pipe_mapping, in_dc->current_state); } need_recalculation = dml2_verify_det_buffer_configuration(dml2, context, &dml2->det_helper_scratch); ASSERT(need_recalculation == false); } } if (result) { unsigned int lowest_state_idx = s->mode_support_params.out_lowest_state_idx; out_clks.dispclk_khz = (unsigned int)dml2->v20.dml_core_ctx.mp.Dispclk_calculated * 1000; out_clks.p_state_supported = s->mode_support_info.DRAMClockChangeSupport[0] != dml_dram_clock_change_unsupported; if (in_dc->config.use_default_clock_table && (lowest_state_idx < dml2->v20.dml_core_ctx.states.num_states - 1)) { lowest_state_idx = dml2->v20.dml_core_ctx.states.num_states - 1; out_clks.dispclk_khz = (unsigned int)dml2->v20.dml_core_ctx.states.state_array[lowest_state_idx].dispclk_mhz * 1000; } out_clks.dcfclk_khz = (unsigned int)dml2->v20.dml_core_ctx.states.state_array[lowest_state_idx].dcfclk_mhz * 1000; out_clks.fclk_khz = (unsigned int)dml2->v20.dml_core_ctx.states.state_array[lowest_state_idx].fabricclk_mhz * 1000; out_clks.uclk_mts = (unsigned int)dml2->v20.dml_core_ctx.states.state_array[lowest_state_idx].dram_speed_mts; out_clks.phyclk_khz = (unsigned int)dml2->v20.dml_core_ctx.states.state_array[lowest_state_idx].phyclk_mhz * 1000; out_clks.socclk_khz = (unsigned int)dml2->v20.dml_core_ctx.states.state_array[lowest_state_idx].socclk_mhz * 1000; out_clks.ref_dtbclk_khz = (unsigned int)dml2->v20.dml_core_ctx.states.state_array[lowest_state_idx].dtbclk_mhz * 1000; context->bw_ctx.bw.dcn.clk.dtbclk_en = is_dtbclk_required(in_dc, context); if (!dml2->config.skip_hw_state_mapping) { /* Call dml2_calculate_rq_and_dlg_params */ dml2_calculate_rq_and_dlg_params(in_dc, context, &context->res_ctx, dml2, in_dc->res_pool->pipe_count); } dml2_copy_clocks_to_dc_state(&out_clks, context); dml2_extract_watermark_set(&context->bw_ctx.bw.dcn.watermarks.a, &dml2->v20.dml_core_ctx); dml2_extract_watermark_set(&context->bw_ctx.bw.dcn.watermarks.b, &dml2->v20.dml_core_ctx); memcpy(&context->bw_ctx.bw.dcn.watermarks.c, &dml2->v20.g6_temp_read_watermark_set, sizeof(context->bw_ctx.bw.dcn.watermarks.c)); dml2_extract_watermark_set(&context->bw_ctx.bw.dcn.watermarks.d, &dml2->v20.dml_core_ctx); dml2_extract_writeback_wm(context, &dml2->v20.dml_core_ctx); //copy for deciding zstate use context->bw_ctx.dml.vba.StutterPeriod = context->bw_ctx.dml2->v20.dml_core_ctx.mp.StutterPeriod; cstate_enter_plus_exit_z8_ns = context->bw_ctx.bw.dcn.watermarks.a.cstate_pstate.cstate_enter_plus_exit_z8_ns; if (context->bw_ctx.dml.vba.StutterPeriod < in_dc->debug.minimum_z8_residency_time && cstate_enter_plus_exit_z8_ns < in_dc->debug.minimum_z8_residency_time * 1000) cstate_enter_plus_exit_z8_ns = in_dc->debug.minimum_z8_residency_time * 1000; context->bw_ctx.bw.dcn.watermarks.a.cstate_pstate.cstate_enter_plus_exit_z8_ns = cstate_enter_plus_exit_z8_ns; } return result; } static bool dml2_validate_only(struct dc_state *context) { struct dml2_context *dml2 = context->bw_ctx.dml2; unsigned int result = 0; if (!context || context->stream_count == 0) return true; /* Zero out before each call before proceeding */ memset(&dml2->v20.scratch, 0, sizeof(struct dml2_wrapper_scratch)); memset(&dml2->v20.dml_core_ctx.policy, 0, sizeof(struct dml_mode_eval_policy_st)); memset(&dml2->v20.dml_core_ctx.ms, 0, sizeof(struct mode_support_st)); memset(&dml2->v20.dml_core_ctx.mp, 0, sizeof(struct mode_program_st)); build_unoptimized_policy_settings(dml2->v20.dml_core_ctx.project, &dml2->v20.dml_core_ctx.policy); map_dc_state_into_dml_display_cfg(dml2, context, &dml2->v20.scratch.cur_display_config); result = pack_and_call_dml_mode_support_ex(dml2, &dml2->v20.scratch.cur_display_config, &dml2->v20.scratch.mode_support_info); if (result) result = does_configuration_meet_sw_policies(dml2, &dml2->v20.scratch.cur_display_config, &dml2->v20.scratch.mode_support_info); return (result == 1) ? true : false; } static void dml2_apply_debug_options(const struct dc *dc, struct dml2_context *dml2) { if (dc->debug.override_odm_optimization) { dml2->config.minimize_dispclk_using_odm = dc->debug.minimize_dispclk_using_odm; } } bool dml2_validate(const struct dc *in_dc, struct dc_state *context, struct dml2_context *dml2, bool fast_validate) { bool out = false; if (!dml2) return false; dml2_apply_debug_options(in_dc, dml2); /* Use dml_validate_only for fast_validate path */ if (fast_validate) out = dml2_validate_only(context); else out = dml2_validate_and_build_resource(in_dc, context); return out; } static inline struct dml2_context *dml2_allocate_memory(void) { return (struct dml2_context *) kzalloc(sizeof(struct dml2_context), GFP_KERNEL); } static void dml2_init(const struct dc *in_dc, const struct dml2_configuration_options *config, struct dml2_context **dml2) { // Store config options (*dml2)->config = *config; switch (in_dc->ctx->dce_version) { case DCN_VERSION_3_5: (*dml2)->v20.dml_core_ctx.project = dml_project_dcn35; break; case DCN_VERSION_3_51: (*dml2)->v20.dml_core_ctx.project = dml_project_dcn351; break; case DCN_VERSION_3_2: (*dml2)->v20.dml_core_ctx.project = dml_project_dcn32; break; case DCN_VERSION_3_21: (*dml2)->v20.dml_core_ctx.project = dml_project_dcn321; break; default: (*dml2)->v20.dml_core_ctx.project = dml_project_default; break; } initialize_dml2_ip_params(*dml2, in_dc, &(*dml2)->v20.dml_core_ctx.ip); initialize_dml2_soc_bbox(*dml2, in_dc, &(*dml2)->v20.dml_core_ctx.soc); initialize_dml2_soc_states(*dml2, in_dc, &(*dml2)->v20.dml_core_ctx.soc, &(*dml2)->v20.dml_core_ctx.states); } bool dml2_create(const struct dc *in_dc, const struct dml2_configuration_options *config, struct dml2_context **dml2) { // Allocate Mode Lib Ctx *dml2 = dml2_allocate_memory(); if (!(*dml2)) return false; dml2_init(in_dc, config, dml2); return true; } void dml2_destroy(struct dml2_context *dml2) { if (!dml2) return; kfree(dml2); } void dml2_extract_dram_and_fclk_change_support(struct dml2_context *dml2, unsigned int *fclk_change_support, unsigned int *dram_clk_change_support) { *fclk_change_support = (unsigned int) dml2->v20.dml_core_ctx.ms.support.FCLKChangeSupport[0]; *dram_clk_change_support = (unsigned int) dml2->v20.dml_core_ctx.ms.support.DRAMClockChangeSupport[0]; } void dml2_copy(struct dml2_context *dst_dml2, struct dml2_context *src_dml2) { /* copy Mode Lib Ctx */ memcpy(dst_dml2, src_dml2, sizeof(struct dml2_context)); } bool dml2_create_copy(struct dml2_context **dst_dml2, struct dml2_context *src_dml2) { /* Allocate Mode Lib Ctx */ *dst_dml2 = dml2_allocate_memory(); if (!(*dst_dml2)) return false; /* copy Mode Lib Ctx */ dml2_copy(*dst_dml2, src_dml2); return true; } void dml2_reinit(const struct dc *in_dc, const struct dml2_configuration_options *config, struct dml2_context **dml2) { dml2_init(in_dc, config, dml2); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1