Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Hersen Wu | 3645 | 21.86% | 9 | 6.38% |
Eryk Brol | 3374 | 20.24% | 16 | 11.35% |
Wayne Lin | 1791 | 10.74% | 13 | 9.22% |
Nicholas Kazlauskas | 1621 | 9.72% | 9 | 6.38% |
David Francis | 1403 | 8.41% | 4 | 2.84% |
Mikita Lipski | 1026 | 6.15% | 5 | 3.55% |
Jerry (Fangzhi) Zuo | 804 | 4.82% | 6 | 4.26% |
Aurabindo Pillai | 387 | 2.32% | 8 | 5.67% |
Hamza Mahfooz | 330 | 1.98% | 2 | 1.42% |
Harry Wentland | 318 | 1.91% | 7 | 4.96% |
Roman Li | 305 | 1.83% | 4 | 2.84% |
Tom Chung | 274 | 1.64% | 1 | 0.71% |
Stylon Wang | 262 | 1.57% | 6 | 4.26% |
Bhawanpreet Lakha | 232 | 1.39% | 4 | 2.84% |
Victor Lu | 206 | 1.24% | 2 | 1.42% |
Alex Deucher | 169 | 1.01% | 2 | 1.42% |
Leo (Hanghong) Ma | 142 | 0.85% | 2 | 1.42% |
Nirmoy Das | 57 | 0.34% | 1 | 0.71% |
Shirish S | 50 | 0.30% | 1 | 0.71% |
Alan Liu | 47 | 0.28% | 2 | 1.42% |
Bhanuprakash Modem | 32 | 0.19% | 1 | 0.71% |
Wenjing Liu | 29 | 0.17% | 7 | 4.96% |
Yongzhi Liu | 24 | 0.14% | 1 | 0.71% |
Leo (Sunpeng) Li | 23 | 0.14% | 1 | 0.71% |
Alex Hung | 22 | 0.13% | 2 | 1.42% |
Anson Jacob | 20 | 0.12% | 3 | 2.13% |
Oleg Vasilev | 9 | 0.05% | 1 | 0.71% |
Patrik Jakobsson | 9 | 0.05% | 1 | 0.71% |
Srinivasan S | 8 | 0.05% | 2 | 1.42% |
Lucas De Marchi | 7 | 0.04% | 1 | 0.71% |
Luben Tuikov | 6 | 0.04% | 1 | 0.71% |
Krunoslav Kovac | 6 | 0.04% | 1 | 0.71% |
rodrigosiqueira | 6 | 0.04% | 1 | 0.71% |
Greg Kroah-Hartman | 6 | 0.04% | 2 | 1.42% |
Qingqing Zhuo | 5 | 0.03% | 1 | 0.71% |
Simon Ser | 4 | 0.02% | 1 | 0.71% |
Po-Ting Chen | 3 | 0.02% | 1 | 0.71% |
Jack Chang | 2 | 0.01% | 1 | 0.71% |
Lee Jones | 2 | 0.01% | 1 | 0.71% |
Anthony Koo | 1 | 0.01% | 1 | 0.71% |
Andrey Grodzovsky | 1 | 0.01% | 1 | 0.71% |
Yue haibing | 1 | 0.01% | 1 | 0.71% |
Bhaskar Chowdhury | 1 | 0.01% | 1 | 0.71% |
Thelford Williams | 1 | 0.01% | 1 | 0.71% |
Sam Ravnborg | 1 | 0.01% | 1 | 0.71% |
Stephen Chandler Paul | 1 | 0.01% | 1 | 0.71% |
Total | 16673 | 141 |
/* * Copyright 2018 Advanced Micro Devices, Inc. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR * OTHER DEALINGS IN THE SOFTWARE. * * Authors: AMD * */ #include <linux/string_helpers.h> #include <linux/uaccess.h> #include "dc.h" #include "amdgpu.h" #include "amdgpu_dm.h" #include "amdgpu_dm_debugfs.h" #include "amdgpu_dm_replay.h" #include "dm_helpers.h" #include "dmub/dmub_srv.h" #include "resource.h" #include "dsc.h" #include "link_hwss.h" #include "dc/dc_dmub_srv.h" #include "link/protocols/link_dp_capability.h" #include "inc/hw/dchubbub.h" #ifdef CONFIG_DRM_AMD_SECURE_DISPLAY #include "amdgpu_dm_psr.h" #endif struct dmub_debugfs_trace_header { uint32_t entry_count; uint32_t reserved[3]; }; struct dmub_debugfs_trace_entry { uint32_t trace_code; uint32_t tick_count; uint32_t param0; uint32_t param1; }; static const char *const mst_progress_status[] = { "probe", "remote_edid", "allocate_new_payload", "clear_allocated_payload", }; /* parse_write_buffer_into_params - Helper function to parse debugfs write buffer into an array * * Function takes in attributes passed to debugfs write entry * and writes into param array. * The user passes max_param_num to identify maximum number of * parameters that could be parsed. * */ static int parse_write_buffer_into_params(char *wr_buf, uint32_t wr_buf_size, long *param, const char __user *buf, int max_param_num, uint8_t *param_nums) { char *wr_buf_ptr = NULL; uint32_t wr_buf_count = 0; int r; char *sub_str = NULL; const char delimiter[3] = {' ', '\n', '\0'}; uint8_t param_index = 0; *param_nums = 0; wr_buf_ptr = wr_buf; /* r is bytes not be copied */ if (copy_from_user(wr_buf_ptr, buf, wr_buf_size)) { DRM_DEBUG_DRIVER("user data could not be read successfully\n"); return -EFAULT; } /* check number of parameters. isspace could not differ space and \n */ while ((*wr_buf_ptr != 0xa) && (wr_buf_count < wr_buf_size)) { /* skip space*/ while (isspace(*wr_buf_ptr) && (wr_buf_count < wr_buf_size)) { wr_buf_ptr++; wr_buf_count++; } if (wr_buf_count == wr_buf_size) break; /* skip non-space*/ while ((!isspace(*wr_buf_ptr)) && (wr_buf_count < wr_buf_size)) { wr_buf_ptr++; wr_buf_count++; } (*param_nums)++; if (wr_buf_count == wr_buf_size) break; } if (*param_nums > max_param_num) *param_nums = max_param_num; wr_buf_ptr = wr_buf; /* reset buf pointer */ wr_buf_count = 0; /* number of char already checked */ while (isspace(*wr_buf_ptr) && (wr_buf_count < wr_buf_size)) { wr_buf_ptr++; wr_buf_count++; } while (param_index < *param_nums) { /* after strsep, wr_buf_ptr will be moved to after space */ sub_str = strsep(&wr_buf_ptr, delimiter); r = kstrtol(sub_str, 16, &(param[param_index])); if (r) DRM_DEBUG_DRIVER("string to int convert error code: %d\n", r); param_index++; } return 0; } /* function description * get/ set DP configuration: lane_count, link_rate, spread_spectrum * * valid lane count value: 1, 2, 4 * valid link rate value: * 06h = 1.62Gbps per lane * 0Ah = 2.7Gbps per lane * 0Ch = 3.24Gbps per lane * 14h = 5.4Gbps per lane * 1Eh = 8.1Gbps per lane * * debugfs is located at /sys/kernel/debug/dri/0/DP-x/link_settings * * --- to get dp configuration * * cat /sys/kernel/debug/dri/0/DP-x/link_settings * * It will list current, verified, reported, preferred dp configuration. * current -- for current video mode * verified --- maximum configuration which pass link training * reported --- DP rx report caps (DPCD register offset 0, 1 2) * preferred --- user force settings * * --- set (or force) dp configuration * * echo <lane_count> <link_rate> > link_settings * * for example, to force to 2 lane, 2.7GHz, * echo 4 0xa > /sys/kernel/debug/dri/0/DP-x/link_settings * * spread_spectrum could not be changed dynamically. * * in case invalid lane count, link rate are force, no hw programming will be * done. please check link settings after force operation to see if HW get * programming. * * cat /sys/kernel/debug/dri/0/DP-x/link_settings * * check current and preferred settings. * */ static ssize_t dp_link_settings_read(struct file *f, char __user *buf, size_t size, loff_t *pos) { struct amdgpu_dm_connector *connector = file_inode(f)->i_private; struct dc_link *link = connector->dc_link; char *rd_buf = NULL; char *rd_buf_ptr = NULL; const uint32_t rd_buf_size = 100; uint32_t result = 0; uint8_t str_len = 0; int r; if (*pos & 3 || size & 3) return -EINVAL; rd_buf = kcalloc(rd_buf_size, sizeof(char), GFP_KERNEL); if (!rd_buf) return 0; rd_buf_ptr = rd_buf; str_len = strlen("Current: %d 0x%x %d "); snprintf(rd_buf_ptr, str_len, "Current: %d 0x%x %d ", link->cur_link_settings.lane_count, link->cur_link_settings.link_rate, link->cur_link_settings.link_spread); rd_buf_ptr += str_len; str_len = strlen("Verified: %d 0x%x %d "); snprintf(rd_buf_ptr, str_len, "Verified: %d 0x%x %d ", link->verified_link_cap.lane_count, link->verified_link_cap.link_rate, link->verified_link_cap.link_spread); rd_buf_ptr += str_len; str_len = strlen("Reported: %d 0x%x %d "); snprintf(rd_buf_ptr, str_len, "Reported: %d 0x%x %d ", link->reported_link_cap.lane_count, link->reported_link_cap.link_rate, link->reported_link_cap.link_spread); rd_buf_ptr += str_len; str_len = strlen("Preferred: %d 0x%x %d "); snprintf(rd_buf_ptr, str_len, "Preferred: %d 0x%x %d\n", link->preferred_link_setting.lane_count, link->preferred_link_setting.link_rate, link->preferred_link_setting.link_spread); while (size) { if (*pos >= rd_buf_size) break; r = put_user(*(rd_buf + result), buf); if (r) { kfree(rd_buf); return r; /* r = -EFAULT */ } buf += 1; size -= 1; *pos += 1; result += 1; } kfree(rd_buf); return result; } static ssize_t dp_link_settings_write(struct file *f, const char __user *buf, size_t size, loff_t *pos) { struct amdgpu_dm_connector *connector = file_inode(f)->i_private; struct dc_link *link = connector->dc_link; struct amdgpu_device *adev = drm_to_adev(connector->base.dev); struct dc *dc = (struct dc *)link->dc; struct dc_link_settings prefer_link_settings; char *wr_buf = NULL; const uint32_t wr_buf_size = 40; /* 0: lane_count; 1: link_rate */ int max_param_num = 2; uint8_t param_nums = 0; long param[2]; bool valid_input = true; if (size == 0) return -EINVAL; wr_buf = kcalloc(wr_buf_size, sizeof(char), GFP_KERNEL); if (!wr_buf) return -ENOSPC; if (parse_write_buffer_into_params(wr_buf, wr_buf_size, (long *)param, buf, max_param_num, ¶m_nums)) { kfree(wr_buf); return -EINVAL; } if (param_nums <= 0) { kfree(wr_buf); DRM_DEBUG_DRIVER("user data not be read\n"); return -EINVAL; } switch (param[0]) { case LANE_COUNT_ONE: case LANE_COUNT_TWO: case LANE_COUNT_FOUR: break; default: valid_input = false; break; } switch (param[1]) { case LINK_RATE_LOW: case LINK_RATE_HIGH: case LINK_RATE_RBR2: case LINK_RATE_HIGH2: case LINK_RATE_HIGH3: case LINK_RATE_UHBR10: case LINK_RATE_UHBR13_5: case LINK_RATE_UHBR20: break; default: valid_input = false; break; } if (!valid_input) { kfree(wr_buf); DRM_DEBUG_DRIVER("Invalid Input value No HW will be programmed\n"); mutex_lock(&adev->dm.dc_lock); dc_link_set_preferred_training_settings(dc, NULL, NULL, link, false); mutex_unlock(&adev->dm.dc_lock); return size; } /* save user force lane_count, link_rate to preferred settings * spread spectrum will not be changed */ prefer_link_settings.link_spread = link->cur_link_settings.link_spread; prefer_link_settings.use_link_rate_set = false; prefer_link_settings.lane_count = param[0]; prefer_link_settings.link_rate = param[1]; mutex_lock(&adev->dm.dc_lock); dc_link_set_preferred_training_settings(dc, &prefer_link_settings, NULL, link, false); mutex_unlock(&adev->dm.dc_lock); kfree(wr_buf); return size; } static bool dp_mst_is_end_device(struct amdgpu_dm_connector *aconnector) { bool is_end_device = false; struct drm_dp_mst_topology_mgr *mgr = NULL; struct drm_dp_mst_port *port = NULL; if (aconnector->mst_root && aconnector->mst_root->mst_mgr.mst_state) { mgr = &aconnector->mst_root->mst_mgr; port = aconnector->mst_output_port; drm_modeset_lock(&mgr->base.lock, NULL); if (port->pdt == DP_PEER_DEVICE_SST_SINK || port->pdt == DP_PEER_DEVICE_DP_LEGACY_CONV) is_end_device = true; drm_modeset_unlock(&mgr->base.lock); } return is_end_device; } /* Change MST link setting * * valid lane count value: 1, 2, 4 * valid link rate value: * 06h = 1.62Gbps per lane * 0Ah = 2.7Gbps per lane * 0Ch = 3.24Gbps per lane * 14h = 5.4Gbps per lane * 1Eh = 8.1Gbps per lane * 3E8h = 10.0Gbps per lane * 546h = 13.5Gbps per lane * 7D0h = 20.0Gbps per lane * * debugfs is located at /sys/kernel/debug/dri/0/DP-x/mst_link_settings * * for example, to force to 2 lane, 10.0GHz, * echo 2 0x3e8 > /sys/kernel/debug/dri/0/DP-x/mst_link_settings * * Valid input will trigger hotplug event to get new link setting applied * Invalid input will trigger training setting reset * * The usage can be referred to link_settings entry * */ static ssize_t dp_mst_link_setting(struct file *f, const char __user *buf, size_t size, loff_t *pos) { struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct dc_link *link = aconnector->dc_link; struct amdgpu_device *adev = drm_to_adev(aconnector->base.dev); struct dc *dc = (struct dc *)link->dc; struct dc_link_settings prefer_link_settings; char *wr_buf = NULL; const uint32_t wr_buf_size = 40; /* 0: lane_count; 1: link_rate */ int max_param_num = 2; uint8_t param_nums = 0; long param[2]; bool valid_input = true; if (!dp_mst_is_end_device(aconnector)) return -EINVAL; if (size == 0) return -EINVAL; wr_buf = kcalloc(wr_buf_size, sizeof(char), GFP_KERNEL); if (!wr_buf) return -ENOSPC; if (parse_write_buffer_into_params(wr_buf, wr_buf_size, (long *)param, buf, max_param_num, ¶m_nums)) { kfree(wr_buf); return -EINVAL; } if (param_nums <= 0) { kfree(wr_buf); DRM_DEBUG_DRIVER("user data not be read\n"); return -EINVAL; } switch (param[0]) { case LANE_COUNT_ONE: case LANE_COUNT_TWO: case LANE_COUNT_FOUR: break; default: valid_input = false; break; } switch (param[1]) { case LINK_RATE_LOW: case LINK_RATE_HIGH: case LINK_RATE_RBR2: case LINK_RATE_HIGH2: case LINK_RATE_HIGH3: case LINK_RATE_UHBR10: case LINK_RATE_UHBR13_5: case LINK_RATE_UHBR20: break; default: valid_input = false; break; } if (!valid_input) { kfree(wr_buf); DRM_DEBUG_DRIVER("Invalid Input value No HW will be programmed\n"); mutex_lock(&adev->dm.dc_lock); dc_link_set_preferred_training_settings(dc, NULL, NULL, link, false); mutex_unlock(&adev->dm.dc_lock); return -EINVAL; } /* save user force lane_count, link_rate to preferred settings * spread spectrum will not be changed */ prefer_link_settings.link_spread = link->cur_link_settings.link_spread; prefer_link_settings.use_link_rate_set = false; prefer_link_settings.lane_count = param[0]; prefer_link_settings.link_rate = param[1]; /* skip immediate retrain, and train to new link setting after hotplug event triggered */ mutex_lock(&adev->dm.dc_lock); dc_link_set_preferred_training_settings(dc, &prefer_link_settings, NULL, link, true); mutex_unlock(&adev->dm.dc_lock); mutex_lock(&aconnector->base.dev->mode_config.mutex); aconnector->base.force = DRM_FORCE_OFF; mutex_unlock(&aconnector->base.dev->mode_config.mutex); drm_kms_helper_hotplug_event(aconnector->base.dev); msleep(100); mutex_lock(&aconnector->base.dev->mode_config.mutex); aconnector->base.force = DRM_FORCE_UNSPECIFIED; mutex_unlock(&aconnector->base.dev->mode_config.mutex); drm_kms_helper_hotplug_event(aconnector->base.dev); kfree(wr_buf); return size; } /* function: get current DP PHY settings: voltage swing, pre-emphasis, * post-cursor2 (defined by VESA DP specification) * * valid values * voltage swing: 0,1,2,3 * pre-emphasis : 0,1,2,3 * post cursor2 : 0,1,2,3 * * * how to use this debugfs * * debugfs is located at /sys/kernel/debug/dri/0/DP-x * * there will be directories, like DP-1, DP-2,DP-3, etc. for DP display * * To figure out which DP-x is the display for DP to be check, * cd DP-x * ls -ll * There should be debugfs file, like link_settings, phy_settings. * cat link_settings * from lane_count, link_rate to figure which DP-x is for display to be worked * on * * To get current DP PHY settings, * cat phy_settings * * To change DP PHY settings, * echo <voltage_swing> <pre-emphasis> <post_cursor2> > phy_settings * for examle, to change voltage swing to 2, pre-emphasis to 3, post_cursor2 to * 0, * echo 2 3 0 > phy_settings * * To check if change be applied, get current phy settings by * cat phy_settings * * In case invalid values are set by user, like * echo 1 4 0 > phy_settings * * HW will NOT be programmed by these settings. * cat phy_settings will show the previous valid settings. */ static ssize_t dp_phy_settings_read(struct file *f, char __user *buf, size_t size, loff_t *pos) { struct amdgpu_dm_connector *connector = file_inode(f)->i_private; struct dc_link *link = connector->dc_link; char *rd_buf = NULL; const uint32_t rd_buf_size = 20; uint32_t result = 0; int r; if (*pos & 3 || size & 3) return -EINVAL; rd_buf = kcalloc(rd_buf_size, sizeof(char), GFP_KERNEL); if (!rd_buf) return -EINVAL; snprintf(rd_buf, rd_buf_size, " %d %d %d\n", link->cur_lane_setting[0].VOLTAGE_SWING, link->cur_lane_setting[0].PRE_EMPHASIS, link->cur_lane_setting[0].POST_CURSOR2); while (size) { if (*pos >= rd_buf_size) break; r = put_user((*(rd_buf + result)), buf); if (r) { kfree(rd_buf); return r; /* r = -EFAULT */ } buf += 1; size -= 1; *pos += 1; result += 1; } kfree(rd_buf); return result; } static int dp_lttpr_status_show(struct seq_file *m, void *unused) { struct drm_connector *connector = m->private; struct amdgpu_dm_connector *aconnector = to_amdgpu_dm_connector(connector); struct dc_lttpr_caps caps = aconnector->dc_link->dpcd_caps.lttpr_caps; if (connector->status != connector_status_connected) return -ENODEV; seq_printf(m, "phy repeater count: %u (raw: 0x%x)\n", dp_parse_lttpr_repeater_count(caps.phy_repeater_cnt), caps.phy_repeater_cnt); seq_puts(m, "phy repeater mode: "); switch (caps.mode) { case DP_PHY_REPEATER_MODE_TRANSPARENT: seq_puts(m, "transparent"); break; case DP_PHY_REPEATER_MODE_NON_TRANSPARENT: seq_puts(m, "non-transparent"); break; case 0x00: seq_puts(m, "non lttpr"); break; default: seq_printf(m, "read error (raw: 0x%x)", caps.mode); break; } seq_puts(m, "\n"); return 0; } static ssize_t dp_phy_settings_write(struct file *f, const char __user *buf, size_t size, loff_t *pos) { struct amdgpu_dm_connector *connector = file_inode(f)->i_private; struct dc_link *link = connector->dc_link; struct dc *dc = (struct dc *)link->dc; char *wr_buf = NULL; uint32_t wr_buf_size = 40; long param[3]; bool use_prefer_link_setting; struct link_training_settings link_lane_settings; int max_param_num = 3; uint8_t param_nums = 0; int r = 0; if (size == 0) return -EINVAL; wr_buf = kcalloc(wr_buf_size, sizeof(char), GFP_KERNEL); if (!wr_buf) return -ENOSPC; if (parse_write_buffer_into_params(wr_buf, wr_buf_size, (long *)param, buf, max_param_num, ¶m_nums)) { kfree(wr_buf); return -EINVAL; } if (param_nums <= 0) { kfree(wr_buf); DRM_DEBUG_DRIVER("user data not be read\n"); return -EINVAL; } if ((param[0] > VOLTAGE_SWING_MAX_LEVEL) || (param[1] > PRE_EMPHASIS_MAX_LEVEL) || (param[2] > POST_CURSOR2_MAX_LEVEL)) { kfree(wr_buf); DRM_DEBUG_DRIVER("Invalid Input No HW will be programmed\n"); return size; } /* get link settings: lane count, link rate */ use_prefer_link_setting = ((link->preferred_link_setting.link_rate != LINK_RATE_UNKNOWN) && (link->test_pattern_enabled)); memset(&link_lane_settings, 0, sizeof(link_lane_settings)); if (use_prefer_link_setting) { link_lane_settings.link_settings.lane_count = link->preferred_link_setting.lane_count; link_lane_settings.link_settings.link_rate = link->preferred_link_setting.link_rate; link_lane_settings.link_settings.link_spread = link->preferred_link_setting.link_spread; } else { link_lane_settings.link_settings.lane_count = link->cur_link_settings.lane_count; link_lane_settings.link_settings.link_rate = link->cur_link_settings.link_rate; link_lane_settings.link_settings.link_spread = link->cur_link_settings.link_spread; } /* apply phy settings from user */ for (r = 0; r < link_lane_settings.link_settings.lane_count; r++) { link_lane_settings.hw_lane_settings[r].VOLTAGE_SWING = (enum dc_voltage_swing) (param[0]); link_lane_settings.hw_lane_settings[r].PRE_EMPHASIS = (enum dc_pre_emphasis) (param[1]); link_lane_settings.hw_lane_settings[r].POST_CURSOR2 = (enum dc_post_cursor2) (param[2]); } /* program ASIC registers and DPCD registers */ dc_link_set_drive_settings(dc, &link_lane_settings, link); kfree(wr_buf); return size; } /* function description * * set PHY layer or Link layer test pattern * PHY test pattern is used for PHY SI check. * Link layer test will not affect PHY SI. * * Reset Test Pattern: * 0 = DP_TEST_PATTERN_VIDEO_MODE * * PHY test pattern supported: * 1 = DP_TEST_PATTERN_D102 * 2 = DP_TEST_PATTERN_SYMBOL_ERROR * 3 = DP_TEST_PATTERN_PRBS7 * 4 = DP_TEST_PATTERN_80BIT_CUSTOM * 5 = DP_TEST_PATTERN_CP2520_1 * 6 = DP_TEST_PATTERN_CP2520_2 = DP_TEST_PATTERN_HBR2_COMPLIANCE_EYE * 7 = DP_TEST_PATTERN_CP2520_3 * * DP PHY Link Training Patterns * 8 = DP_TEST_PATTERN_TRAINING_PATTERN1 * 9 = DP_TEST_PATTERN_TRAINING_PATTERN2 * a = DP_TEST_PATTERN_TRAINING_PATTERN3 * b = DP_TEST_PATTERN_TRAINING_PATTERN4 * * DP Link Layer Test pattern * c = DP_TEST_PATTERN_COLOR_SQUARES * d = DP_TEST_PATTERN_COLOR_SQUARES_CEA * e = DP_TEST_PATTERN_VERTICAL_BARS * f = DP_TEST_PATTERN_HORIZONTAL_BARS * 10= DP_TEST_PATTERN_COLOR_RAMP * * debugfs phy_test_pattern is located at /syskernel/debug/dri/0/DP-x * * --- set test pattern * echo <test pattern #> > test_pattern * * If test pattern # is not supported, NO HW programming will be done. * for DP_TEST_PATTERN_80BIT_CUSTOM, it needs extra 10 bytes of data * for the user pattern. input 10 bytes data are separated by space * * echo 0x4 0x11 0x22 0x33 0x44 0x55 0x66 0x77 0x88 0x99 0xaa > test_pattern * * --- reset test pattern * echo 0 > test_pattern * * --- HPD detection is disabled when set PHY test pattern * * when PHY test pattern (pattern # within [1,7]) is set, HPD pin of HW ASIC * is disable. User could unplug DP display from DP connected and plug scope to * check test pattern PHY SI. * If there is need unplug scope and plug DP display back, do steps below: * echo 0 > phy_test_pattern * unplug scope * plug DP display. * * "echo 0 > phy_test_pattern" will re-enable HPD pin again so that video sw * driver could detect "unplug scope" and "plug DP display" */ static ssize_t dp_phy_test_pattern_debugfs_write(struct file *f, const char __user *buf, size_t size, loff_t *pos) { struct amdgpu_dm_connector *connector = file_inode(f)->i_private; struct dc_link *link = connector->dc_link; char *wr_buf = NULL; uint32_t wr_buf_size = 100; long param[11] = {0x0}; int max_param_num = 11; enum dp_test_pattern test_pattern = DP_TEST_PATTERN_UNSUPPORTED; bool disable_hpd = false; bool valid_test_pattern = false; uint8_t param_nums = 0; /* init with default 80bit custom pattern */ uint8_t custom_pattern[10] = { 0x1f, 0x7c, 0xf0, 0xc1, 0x07, 0x1f, 0x7c, 0xf0, 0xc1, 0x07 }; struct dc_link_settings prefer_link_settings = {LANE_COUNT_UNKNOWN, LINK_RATE_UNKNOWN, LINK_SPREAD_DISABLED}; struct dc_link_settings cur_link_settings = {LANE_COUNT_UNKNOWN, LINK_RATE_UNKNOWN, LINK_SPREAD_DISABLED}; struct link_training_settings link_training_settings; int i; if (size == 0) return -EINVAL; wr_buf = kcalloc(wr_buf_size, sizeof(char), GFP_KERNEL); if (!wr_buf) return -ENOSPC; if (parse_write_buffer_into_params(wr_buf, wr_buf_size, (long *)param, buf, max_param_num, ¶m_nums)) { kfree(wr_buf); return -EINVAL; } if (param_nums <= 0) { kfree(wr_buf); DRM_DEBUG_DRIVER("user data not be read\n"); return -EINVAL; } test_pattern = param[0]; switch (test_pattern) { case DP_TEST_PATTERN_VIDEO_MODE: case DP_TEST_PATTERN_COLOR_SQUARES: case DP_TEST_PATTERN_COLOR_SQUARES_CEA: case DP_TEST_PATTERN_VERTICAL_BARS: case DP_TEST_PATTERN_HORIZONTAL_BARS: case DP_TEST_PATTERN_COLOR_RAMP: valid_test_pattern = true; break; case DP_TEST_PATTERN_D102: case DP_TEST_PATTERN_SYMBOL_ERROR: case DP_TEST_PATTERN_PRBS7: case DP_TEST_PATTERN_80BIT_CUSTOM: case DP_TEST_PATTERN_HBR2_COMPLIANCE_EYE: case DP_TEST_PATTERN_TRAINING_PATTERN4: disable_hpd = true; valid_test_pattern = true; break; default: valid_test_pattern = false; test_pattern = DP_TEST_PATTERN_UNSUPPORTED; break; } if (!valid_test_pattern) { kfree(wr_buf); DRM_DEBUG_DRIVER("Invalid Test Pattern Parameters\n"); return size; } if (test_pattern == DP_TEST_PATTERN_80BIT_CUSTOM) { for (i = 0; i < 10; i++) { if ((uint8_t) param[i + 1] != 0x0) break; } if (i < 10) { /* not use default value */ for (i = 0; i < 10; i++) custom_pattern[i] = (uint8_t) param[i + 1]; } } /* Usage: set DP physical test pattern using debugfs with normal DP * panel. Then plug out DP panel and connect a scope to measure * For normal video mode and test pattern generated from CRCT, * they are visibile to user. So do not disable HPD. * Video Mode is also set to clear the test pattern, so enable HPD * because it might have been disabled after a test pattern was set. * AUX depends on HPD * sequence dependent, do not move! */ if (!disable_hpd) dc_link_enable_hpd(link); prefer_link_settings.lane_count = link->verified_link_cap.lane_count; prefer_link_settings.link_rate = link->verified_link_cap.link_rate; prefer_link_settings.link_spread = link->verified_link_cap.link_spread; cur_link_settings.lane_count = link->cur_link_settings.lane_count; cur_link_settings.link_rate = link->cur_link_settings.link_rate; cur_link_settings.link_spread = link->cur_link_settings.link_spread; link_training_settings.link_settings = cur_link_settings; if (test_pattern != DP_TEST_PATTERN_VIDEO_MODE) { if (prefer_link_settings.lane_count != LANE_COUNT_UNKNOWN && prefer_link_settings.link_rate != LINK_RATE_UNKNOWN && (prefer_link_settings.lane_count != cur_link_settings.lane_count || prefer_link_settings.link_rate != cur_link_settings.link_rate)) link_training_settings.link_settings = prefer_link_settings; } for (i = 0; i < (unsigned int)(link_training_settings.link_settings.lane_count); i++) link_training_settings.hw_lane_settings[i] = link->cur_lane_setting[i]; dc_link_dp_set_test_pattern( link, test_pattern, DP_TEST_PATTERN_COLOR_SPACE_RGB, &link_training_settings, custom_pattern, 10); /* Usage: Set DP physical test pattern using AMDDP with normal DP panel * Then plug out DP panel and connect a scope to measure DP PHY signal. * Need disable interrupt to avoid SW driver disable DP output. This is * done after the test pattern is set. */ if (valid_test_pattern && disable_hpd) dc_link_disable_hpd(link); kfree(wr_buf); return size; } /* * Returns the DMCUB tracebuffer contents. * Example usage: cat /sys/kernel/debug/dri/0/amdgpu_dm_dmub_tracebuffer */ static int dmub_tracebuffer_show(struct seq_file *m, void *data) { struct amdgpu_device *adev = m->private; struct dmub_srv_fb_info *fb_info = adev->dm.dmub_fb_info; struct dmub_debugfs_trace_entry *entries; uint8_t *tbuf_base; uint32_t tbuf_size, max_entries, num_entries, i; if (!fb_info) return 0; tbuf_base = (uint8_t *)fb_info->fb[DMUB_WINDOW_5_TRACEBUFF].cpu_addr; if (!tbuf_base) return 0; tbuf_size = fb_info->fb[DMUB_WINDOW_5_TRACEBUFF].size; max_entries = (tbuf_size - sizeof(struct dmub_debugfs_trace_header)) / sizeof(struct dmub_debugfs_trace_entry); num_entries = ((struct dmub_debugfs_trace_header *)tbuf_base)->entry_count; num_entries = min(num_entries, max_entries); entries = (struct dmub_debugfs_trace_entry *)(tbuf_base + sizeof(struct dmub_debugfs_trace_header)); for (i = 0; i < num_entries; ++i) { struct dmub_debugfs_trace_entry *entry = &entries[i]; seq_printf(m, "trace_code=%u tick_count=%u param0=%u param1=%u\n", entry->trace_code, entry->tick_count, entry->param0, entry->param1); } return 0; } /* * Returns the DMCUB firmware state contents. * Example usage: cat /sys/kernel/debug/dri/0/amdgpu_dm_dmub_fw_state */ static int dmub_fw_state_show(struct seq_file *m, void *data) { struct amdgpu_device *adev = m->private; struct dmub_srv_fb_info *fb_info = adev->dm.dmub_fb_info; uint8_t *state_base; uint32_t state_size; if (!fb_info) return 0; state_base = (uint8_t *)fb_info->fb[DMUB_WINDOW_6_FW_STATE].cpu_addr; if (!state_base) return 0; state_size = fb_info->fb[DMUB_WINDOW_6_FW_STATE].size; return seq_write(m, state_base, state_size); } /* replay_capability_show() - show eDP panel replay capability * * The read function: replay_capability_show * Shows if sink and driver has Replay capability or not. * * cat /sys/kernel/debug/dri/0/eDP-X/replay_capability * * Expected output: * "Sink support: no\n" - if panel doesn't support Replay * "Sink support: yes\n" - if panel supports Replay * "Driver support: no\n" - if driver doesn't support Replay * "Driver support: yes\n" - if driver supports Replay */ static int replay_capability_show(struct seq_file *m, void *data) { struct drm_connector *connector = m->private; struct amdgpu_dm_connector *aconnector = to_amdgpu_dm_connector(connector); struct dc_link *link = aconnector->dc_link; bool sink_support_replay = false; bool driver_support_replay = false; if (!link) return -ENODEV; if (link->type == dc_connection_none) return -ENODEV; if (!(link->connector_signal & SIGNAL_TYPE_EDP)) return -ENODEV; /* If Replay is already set to support, skip the checks */ if (link->replay_settings.config.replay_supported) { sink_support_replay = true; driver_support_replay = true; } else if ((amdgpu_dc_debug_mask & DC_DISABLE_REPLAY)) { sink_support_replay = amdgpu_dm_link_supports_replay(link, aconnector); } else { struct dc *dc = link->ctx->dc; sink_support_replay = amdgpu_dm_link_supports_replay(link, aconnector); if (dc->ctx->dmub_srv && dc->ctx->dmub_srv->dmub) driver_support_replay = (bool)dc->ctx->dmub_srv->dmub->feature_caps.replay_supported; } seq_printf(m, "Sink support: %s\n", str_yes_no(sink_support_replay)); seq_printf(m, "Driver support: %s\n", str_yes_no(driver_support_replay)); seq_printf(m, "Config support: %s\n", str_yes_no(link->replay_settings.config.replay_supported)); return 0; } /* psr_capability_show() - show eDP panel PSR capability * * The read function: sink_psr_capability_show * Shows if sink has PSR capability or not. * If yes - the PSR version is appended * * cat /sys/kernel/debug/dri/0/eDP-X/psr_capability * * Expected output: * "Sink support: no\n" - if panel doesn't support PSR * "Sink support: yes [0x01]\n" - if panel supports PSR1 * "Driver support: no\n" - if driver doesn't support PSR * "Driver support: yes [0x01]\n" - if driver supports PSR1 */ static int psr_capability_show(struct seq_file *m, void *data) { struct drm_connector *connector = m->private; struct amdgpu_dm_connector *aconnector = to_amdgpu_dm_connector(connector); struct dc_link *link = aconnector->dc_link; if (!link) return -ENODEV; if (link->type == dc_connection_none) return -ENODEV; if (!(link->connector_signal & SIGNAL_TYPE_EDP)) return -ENODEV; seq_printf(m, "Sink support: %s", str_yes_no(link->dpcd_caps.psr_info.psr_version != 0)); if (link->dpcd_caps.psr_info.psr_version) seq_printf(m, " [0x%02x]", link->dpcd_caps.psr_info.psr_version); seq_puts(m, "\n"); seq_printf(m, "Driver support: %s", str_yes_no(link->psr_settings.psr_feature_enabled)); if (link->psr_settings.psr_version) seq_printf(m, " [0x%02x]", link->psr_settings.psr_version); seq_puts(m, "\n"); return 0; } /* * Returns the current bpc for the crtc. * Example usage: cat /sys/kernel/debug/dri/0/crtc-0/amdgpu_current_bpc */ static int amdgpu_current_bpc_show(struct seq_file *m, void *data) { struct drm_crtc *crtc = m->private; struct drm_device *dev = crtc->dev; struct dm_crtc_state *dm_crtc_state = NULL; int res = -ENODEV; unsigned int bpc; mutex_lock(&dev->mode_config.mutex); drm_modeset_lock(&crtc->mutex, NULL); if (crtc->state == NULL) goto unlock; dm_crtc_state = to_dm_crtc_state(crtc->state); if (dm_crtc_state->stream == NULL) goto unlock; switch (dm_crtc_state->stream->timing.display_color_depth) { case COLOR_DEPTH_666: bpc = 6; break; case COLOR_DEPTH_888: bpc = 8; break; case COLOR_DEPTH_101010: bpc = 10; break; case COLOR_DEPTH_121212: bpc = 12; break; case COLOR_DEPTH_161616: bpc = 16; break; default: goto unlock; } seq_printf(m, "Current: %u\n", bpc); res = 0; unlock: drm_modeset_unlock(&crtc->mutex); mutex_unlock(&dev->mode_config.mutex); return res; } DEFINE_SHOW_ATTRIBUTE(amdgpu_current_bpc); /* * Returns the current colorspace for the crtc. * Example usage: cat /sys/kernel/debug/dri/0/crtc-0/amdgpu_current_colorspace */ static int amdgpu_current_colorspace_show(struct seq_file *m, void *data) { struct drm_crtc *crtc = m->private; struct drm_device *dev = crtc->dev; struct dm_crtc_state *dm_crtc_state = NULL; int res = -ENODEV; mutex_lock(&dev->mode_config.mutex); drm_modeset_lock(&crtc->mutex, NULL); if (crtc->state == NULL) goto unlock; dm_crtc_state = to_dm_crtc_state(crtc->state); if (dm_crtc_state->stream == NULL) goto unlock; switch (dm_crtc_state->stream->output_color_space) { case COLOR_SPACE_SRGB: seq_puts(m, "sRGB"); break; case COLOR_SPACE_YCBCR601: case COLOR_SPACE_YCBCR601_LIMITED: seq_puts(m, "BT601_YCC"); break; case COLOR_SPACE_YCBCR709: case COLOR_SPACE_YCBCR709_LIMITED: seq_puts(m, "BT709_YCC"); break; case COLOR_SPACE_ADOBERGB: seq_puts(m, "opRGB"); break; case COLOR_SPACE_2020_RGB_FULLRANGE: seq_puts(m, "BT2020_RGB"); break; case COLOR_SPACE_2020_YCBCR: seq_puts(m, "BT2020_YCC"); break; default: goto unlock; } res = 0; unlock: drm_modeset_unlock(&crtc->mutex); mutex_unlock(&dev->mode_config.mutex); return res; } DEFINE_SHOW_ATTRIBUTE(amdgpu_current_colorspace); /* * Example usage: * Disable dsc passthrough, i.e.,: have dsc decoding at converver, not external RX * echo 1 /sys/kernel/debug/dri/0/DP-1/dsc_disable_passthrough * Enable dsc passthrough, i.e.,: have dsc passthrough to external RX * echo 0 /sys/kernel/debug/dri/0/DP-1/dsc_disable_passthrough */ static ssize_t dp_dsc_passthrough_set(struct file *f, const char __user *buf, size_t size, loff_t *pos) { struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; char *wr_buf = NULL; uint32_t wr_buf_size = 42; int max_param_num = 1; long param; uint8_t param_nums = 0; if (size == 0) return -EINVAL; wr_buf = kcalloc(wr_buf_size, sizeof(char), GFP_KERNEL); if (!wr_buf) { DRM_DEBUG_DRIVER("no memory to allocate write buffer\n"); return -ENOSPC; } if (parse_write_buffer_into_params(wr_buf, wr_buf_size, ¶m, buf, max_param_num, ¶m_nums)) { kfree(wr_buf); return -EINVAL; } aconnector->dsc_settings.dsc_force_disable_passthrough = param; kfree(wr_buf); return 0; } /* * Returns the HDCP capability of the Display (1.4 for now). * * NOTE* Not all HDMI displays report their HDCP caps even when they are capable. * Since its rare for a display to not be HDCP 1.4 capable, we set HDMI as always capable. * * Example usage: cat /sys/kernel/debug/dri/0/DP-1/hdcp_sink_capability * or cat /sys/kernel/debug/dri/0/HDMI-A-1/hdcp_sink_capability */ static int hdcp_sink_capability_show(struct seq_file *m, void *data) { struct drm_connector *connector = m->private; struct amdgpu_dm_connector *aconnector = to_amdgpu_dm_connector(connector); bool hdcp_cap, hdcp2_cap; if (connector->status != connector_status_connected) return -ENODEV; seq_printf(m, "%s:%d HDCP version: ", connector->name, connector->base.id); hdcp_cap = dc_link_is_hdcp14(aconnector->dc_link, aconnector->dc_sink->sink_signal); hdcp2_cap = dc_link_is_hdcp22(aconnector->dc_link, aconnector->dc_sink->sink_signal); if (hdcp_cap) seq_printf(m, "%s ", "HDCP1.4"); if (hdcp2_cap) seq_printf(m, "%s ", "HDCP2.2"); if (!hdcp_cap && !hdcp2_cap) seq_printf(m, "%s ", "None"); seq_puts(m, "\n"); return 0; } /* * Returns whether the connected display is internal and not hotpluggable. * Example usage: cat /sys/kernel/debug/dri/0/DP-1/internal_display */ static int internal_display_show(struct seq_file *m, void *data) { struct drm_connector *connector = m->private; struct amdgpu_dm_connector *aconnector = to_amdgpu_dm_connector(connector); struct dc_link *link = aconnector->dc_link; seq_printf(m, "Internal: %u\n", link->is_internal_display); return 0; } /* * Returns the number of segments used if ODM Combine mode is enabled. * Example usage: cat /sys/kernel/debug/dri/0/DP-1/odm_combine_segments */ static int odm_combine_segments_show(struct seq_file *m, void *unused) { struct drm_connector *connector = m->private; struct amdgpu_dm_connector *aconnector = to_amdgpu_dm_connector(connector); struct dc_link *link = aconnector->dc_link; struct pipe_ctx *pipe_ctx = NULL; int i, segments = -EOPNOTSUPP; for (i = 0; i < MAX_PIPES; i++) { pipe_ctx = &link->dc->current_state->res_ctx.pipe_ctx[i]; if (pipe_ctx->stream && pipe_ctx->stream->link == link) break; } if (connector->status != connector_status_connected) return -ENODEV; if (pipe_ctx != NULL && pipe_ctx->stream_res.tg->funcs->get_odm_combine_segments) pipe_ctx->stream_res.tg->funcs->get_odm_combine_segments(pipe_ctx->stream_res.tg, &segments); seq_printf(m, "%d\n", segments); return 0; } /* function description * * generic SDP message access for testing * * debugfs sdp_message is located at /syskernel/debug/dri/0/DP-x * * SDP header * Hb0 : Secondary-Data Packet ID * Hb1 : Secondary-Data Packet type * Hb2 : Secondary-Data-packet-specific header, Byte 0 * Hb3 : Secondary-Data-packet-specific header, Byte 1 * * for using custom sdp message: input 4 bytes SDP header and 32 bytes raw data */ static ssize_t dp_sdp_message_debugfs_write(struct file *f, const char __user *buf, size_t size, loff_t *pos) { int r; uint8_t data[36] = {0}; struct amdgpu_dm_connector *connector = file_inode(f)->i_private; struct dm_crtc_state *acrtc_state; uint32_t write_size = 36; if (connector->base.status != connector_status_connected) return -ENODEV; if (size == 0) return 0; acrtc_state = to_dm_crtc_state(connector->base.state->crtc->state); r = copy_from_user(data, buf, write_size); write_size -= r; dc_stream_send_dp_sdp(acrtc_state->stream, data, write_size); return write_size; } /* function: Read link's DSC & FEC capabilities * * * Access it with the following command (you need to specify * connector like DP-1): * * cat /sys/kernel/debug/dri/0/DP-X/dp_dsc_fec_support * */ static int dp_dsc_fec_support_show(struct seq_file *m, void *data) { struct drm_connector *connector = m->private; struct drm_modeset_acquire_ctx ctx; struct drm_device *dev = connector->dev; struct amdgpu_dm_connector *aconnector = to_amdgpu_dm_connector(connector); int ret = 0; bool try_again = false; bool is_fec_supported = false; bool is_dsc_supported = false; struct dpcd_caps dpcd_caps; drm_modeset_acquire_init(&ctx, DRM_MODESET_ACQUIRE_INTERRUPTIBLE); do { try_again = false; ret = drm_modeset_lock(&dev->mode_config.connection_mutex, &ctx); if (ret) { if (ret == -EDEADLK) { ret = drm_modeset_backoff(&ctx); if (!ret) { try_again = true; continue; } } break; } if (connector->status != connector_status_connected) { ret = -ENODEV; break; } dpcd_caps = aconnector->dc_link->dpcd_caps; if (aconnector->mst_output_port) { /* aconnector sets dsc_aux during get_modes call * if MST connector has it means it can either * enable DSC on the sink device or on MST branch * its connected to. */ if (aconnector->dsc_aux) { is_fec_supported = true; is_dsc_supported = true; } } else { is_fec_supported = dpcd_caps.fec_cap.raw & 0x1; is_dsc_supported = dpcd_caps.dsc_caps.dsc_basic_caps.raw[0] & 0x1; } } while (try_again); drm_modeset_drop_locks(&ctx); drm_modeset_acquire_fini(&ctx); seq_printf(m, "FEC_Sink_Support: %s\n", str_yes_no(is_fec_supported)); seq_printf(m, "DSC_Sink_Support: %s\n", str_yes_no(is_dsc_supported)); return ret; } /* function: Trigger virtual HPD redetection on connector * * This function will perform link rediscovery, link disable * and enable, and dm connector state update. * * Retrigger HPD on an existing connector by echoing 1 into * its respectful "trigger_hotplug" debugfs entry: * * echo 1 > /sys/kernel/debug/dri/0/DP-X/trigger_hotplug * * This function can perform HPD unplug: * * echo 0 > /sys/kernel/debug/dri/0/DP-X/trigger_hotplug * */ static ssize_t trigger_hotplug(struct file *f, const char __user *buf, size_t size, loff_t *pos) { struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct drm_connector *connector = &aconnector->base; struct dc_link *link = NULL; struct drm_device *dev = connector->dev; struct amdgpu_device *adev = drm_to_adev(dev); enum dc_connection_type new_connection_type = dc_connection_none; char *wr_buf = NULL; uint32_t wr_buf_size = 42; int max_param_num = 1; long param[1] = {0}; uint8_t param_nums = 0; bool ret = false; if (!aconnector->dc_link) return -EINVAL; if (size == 0) return -EINVAL; wr_buf = kcalloc(wr_buf_size, sizeof(char), GFP_KERNEL); if (!wr_buf) { DRM_DEBUG_DRIVER("no memory to allocate write buffer\n"); return -ENOSPC; } if (parse_write_buffer_into_params(wr_buf, wr_buf_size, (long *)param, buf, max_param_num, ¶m_nums)) { kfree(wr_buf); return -EINVAL; } kfree(wr_buf); if (param_nums <= 0) { DRM_DEBUG_DRIVER("user data not be read\n"); return -EINVAL; } mutex_lock(&aconnector->hpd_lock); /* Don't support for mst end device*/ if (aconnector->mst_root) { mutex_unlock(&aconnector->hpd_lock); return -EINVAL; } if (param[0] == 1) { if (!dc_link_detect_connection_type(aconnector->dc_link, &new_connection_type) && new_connection_type != dc_connection_none) goto unlock; mutex_lock(&adev->dm.dc_lock); ret = dc_link_detect(aconnector->dc_link, DETECT_REASON_HPD); mutex_unlock(&adev->dm.dc_lock); if (!ret) goto unlock; amdgpu_dm_update_connector_after_detect(aconnector); drm_modeset_lock_all(dev); dm_restore_drm_connector_state(dev, connector); drm_modeset_unlock_all(dev); drm_kms_helper_connector_hotplug_event(connector); } else if (param[0] == 0) { if (!aconnector->dc_link) goto unlock; link = aconnector->dc_link; if (link->local_sink) { dc_sink_release(link->local_sink); link->local_sink = NULL; } link->dpcd_sink_count = 0; link->type = dc_connection_none; link->dongle_max_pix_clk = 0; amdgpu_dm_update_connector_after_detect(aconnector); /* If the aconnector is the root node in mst topology */ if (aconnector->mst_mgr.mst_state == true) dc_link_reset_cur_dp_mst_topology(link); drm_modeset_lock_all(dev); dm_restore_drm_connector_state(dev, connector); drm_modeset_unlock_all(dev); drm_kms_helper_connector_hotplug_event(connector); } unlock: mutex_unlock(&aconnector->hpd_lock); return size; } /* function: read DSC status on the connector * * The read function: dp_dsc_clock_en_read * returns current status of DSC clock on the connector. * The return is a boolean flag: 1 or 0. * * Access it with the following command (you need to specify * connector like DP-1): * * cat /sys/kernel/debug/dri/0/DP-X/dsc_clock_en * * Expected output: * 1 - means that DSC is currently enabled * 0 - means that DSC is disabled */ static ssize_t dp_dsc_clock_en_read(struct file *f, char __user *buf, size_t size, loff_t *pos) { char *rd_buf = NULL; char *rd_buf_ptr = NULL; struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct display_stream_compressor *dsc; struct dcn_dsc_state dsc_state = {0}; const uint32_t rd_buf_size = 10; struct pipe_ctx *pipe_ctx; ssize_t result = 0; int i, r, str_len = 10; rd_buf = kcalloc(rd_buf_size, sizeof(char), GFP_KERNEL); if (!rd_buf) return -ENOMEM; rd_buf_ptr = rd_buf; for (i = 0; i < MAX_PIPES; i++) { pipe_ctx = &aconnector->dc_link->dc->current_state->res_ctx.pipe_ctx[i]; if (pipe_ctx->stream && pipe_ctx->stream->link == aconnector->dc_link && pipe_ctx->stream->sink && pipe_ctx->stream->sink == aconnector->dc_sink) break; } dsc = pipe_ctx->stream_res.dsc; if (dsc) dsc->funcs->dsc_read_state(dsc, &dsc_state); snprintf(rd_buf_ptr, str_len, "%d\n", dsc_state.dsc_clock_en); rd_buf_ptr += str_len; while (size) { if (*pos >= rd_buf_size) break; r = put_user(*(rd_buf + result), buf); if (r) { kfree(rd_buf); return r; /* r = -EFAULT */ } buf += 1; size -= 1; *pos += 1; result += 1; } kfree(rd_buf); return result; } /* function: write force DSC on the connector * * The write function: dp_dsc_clock_en_write * enables to force DSC on the connector. * User can write to either force enable or force disable DSC * on the next modeset or set it to driver default * * Accepted inputs: * 0 - default DSC enablement policy * 1 - force enable DSC on the connector * 2 - force disable DSC on the connector (might cause fail in atomic_check) * * Writing DSC settings is done with the following command: * - To force enable DSC (you need to specify * connector like DP-1): * * echo 0x1 > /sys/kernel/debug/dri/0/DP-X/dsc_clock_en * * - To return to default state set the flag to zero and * let driver deal with DSC automatically * (you need to specify connector like DP-1): * * echo 0x0 > /sys/kernel/debug/dri/0/DP-X/dsc_clock_en * */ static ssize_t dp_dsc_clock_en_write(struct file *f, const char __user *buf, size_t size, loff_t *pos) { struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct drm_connector *connector = &aconnector->base; struct drm_device *dev = connector->dev; struct drm_crtc *crtc = NULL; struct dm_crtc_state *dm_crtc_state = NULL; struct pipe_ctx *pipe_ctx; int i; char *wr_buf = NULL; uint32_t wr_buf_size = 42; int max_param_num = 1; long param[1] = {0}; uint8_t param_nums = 0; if (size == 0) return -EINVAL; wr_buf = kcalloc(wr_buf_size, sizeof(char), GFP_KERNEL); if (!wr_buf) { DRM_DEBUG_DRIVER("no memory to allocate write buffer\n"); return -ENOSPC; } if (parse_write_buffer_into_params(wr_buf, wr_buf_size, (long *)param, buf, max_param_num, ¶m_nums)) { kfree(wr_buf); return -EINVAL; } if (param_nums <= 0) { DRM_DEBUG_DRIVER("user data not be read\n"); kfree(wr_buf); return -EINVAL; } for (i = 0; i < MAX_PIPES; i++) { pipe_ctx = &aconnector->dc_link->dc->current_state->res_ctx.pipe_ctx[i]; if (pipe_ctx->stream && pipe_ctx->stream->link == aconnector->dc_link && pipe_ctx->stream->sink && pipe_ctx->stream->sink == aconnector->dc_sink) break; } if (!pipe_ctx->stream) goto done; // Get CRTC state mutex_lock(&dev->mode_config.mutex); drm_modeset_lock(&dev->mode_config.connection_mutex, NULL); if (connector->state == NULL) goto unlock; crtc = connector->state->crtc; if (crtc == NULL) goto unlock; drm_modeset_lock(&crtc->mutex, NULL); if (crtc->state == NULL) goto unlock; dm_crtc_state = to_dm_crtc_state(crtc->state); if (dm_crtc_state->stream == NULL) goto unlock; if (param[0] == 1) aconnector->dsc_settings.dsc_force_enable = DSC_CLK_FORCE_ENABLE; else if (param[0] == 2) aconnector->dsc_settings.dsc_force_enable = DSC_CLK_FORCE_DISABLE; else aconnector->dsc_settings.dsc_force_enable = DSC_CLK_FORCE_DEFAULT; dm_crtc_state->dsc_force_changed = true; unlock: if (crtc) drm_modeset_unlock(&crtc->mutex); drm_modeset_unlock(&dev->mode_config.connection_mutex); mutex_unlock(&dev->mode_config.mutex); done: kfree(wr_buf); return size; } /* function: read DSC slice width parameter on the connector * * The read function: dp_dsc_slice_width_read * returns dsc slice width used in the current configuration * The return is an integer: 0 or other positive number * * Access the status with the following command: * * cat /sys/kernel/debug/dri/0/DP-X/dsc_slice_width * * 0 - means that DSC is disabled * * Any other number more than zero represents the * slice width currently used by DSC in pixels * */ static ssize_t dp_dsc_slice_width_read(struct file *f, char __user *buf, size_t size, loff_t *pos) { char *rd_buf = NULL; char *rd_buf_ptr = NULL; struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct display_stream_compressor *dsc; struct dcn_dsc_state dsc_state = {0}; const uint32_t rd_buf_size = 100; struct pipe_ctx *pipe_ctx; ssize_t result = 0; int i, r, str_len = 30; rd_buf = kcalloc(rd_buf_size, sizeof(char), GFP_KERNEL); if (!rd_buf) return -ENOMEM; rd_buf_ptr = rd_buf; for (i = 0; i < MAX_PIPES; i++) { pipe_ctx = &aconnector->dc_link->dc->current_state->res_ctx.pipe_ctx[i]; if (pipe_ctx->stream && pipe_ctx->stream->link == aconnector->dc_link && pipe_ctx->stream->sink && pipe_ctx->stream->sink == aconnector->dc_sink) break; } dsc = pipe_ctx->stream_res.dsc; if (dsc) dsc->funcs->dsc_read_state(dsc, &dsc_state); snprintf(rd_buf_ptr, str_len, "%d\n", dsc_state.dsc_slice_width); rd_buf_ptr += str_len; while (size) { if (*pos >= rd_buf_size) break; r = put_user(*(rd_buf + result), buf); if (r) { kfree(rd_buf); return r; /* r = -EFAULT */ } buf += 1; size -= 1; *pos += 1; result += 1; } kfree(rd_buf); return result; } /* function: write DSC slice width parameter * * The write function: dp_dsc_slice_width_write * overwrites automatically generated DSC configuration * of slice width. * * The user has to write the slice width divisible by the * picture width. * * Also the user has to write width in hexidecimal * rather than in decimal. * * Writing DSC settings is done with the following command: * - To force overwrite slice width: (example sets to 1920 pixels) * * echo 0x780 > /sys/kernel/debug/dri/0/DP-X/dsc_slice_width * * - To stop overwriting and let driver find the optimal size, * set the width to zero: * * echo 0x0 > /sys/kernel/debug/dri/0/DP-X/dsc_slice_width * */ static ssize_t dp_dsc_slice_width_write(struct file *f, const char __user *buf, size_t size, loff_t *pos) { struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct pipe_ctx *pipe_ctx; struct drm_connector *connector = &aconnector->base; struct drm_device *dev = connector->dev; struct drm_crtc *crtc = NULL; struct dm_crtc_state *dm_crtc_state = NULL; int i; char *wr_buf = NULL; uint32_t wr_buf_size = 42; int max_param_num = 1; long param[1] = {0}; uint8_t param_nums = 0; if (size == 0) return -EINVAL; wr_buf = kcalloc(wr_buf_size, sizeof(char), GFP_KERNEL); if (!wr_buf) { DRM_DEBUG_DRIVER("no memory to allocate write buffer\n"); return -ENOSPC; } if (parse_write_buffer_into_params(wr_buf, wr_buf_size, (long *)param, buf, max_param_num, ¶m_nums)) { kfree(wr_buf); return -EINVAL; } if (param_nums <= 0) { DRM_DEBUG_DRIVER("user data not be read\n"); kfree(wr_buf); return -EINVAL; } for (i = 0; i < MAX_PIPES; i++) { pipe_ctx = &aconnector->dc_link->dc->current_state->res_ctx.pipe_ctx[i]; if (pipe_ctx->stream && pipe_ctx->stream->link == aconnector->dc_link && pipe_ctx->stream->sink && pipe_ctx->stream->sink == aconnector->dc_sink) break; } if (!pipe_ctx->stream) goto done; // Safely get CRTC state mutex_lock(&dev->mode_config.mutex); drm_modeset_lock(&dev->mode_config.connection_mutex, NULL); if (connector->state == NULL) goto unlock; crtc = connector->state->crtc; if (crtc == NULL) goto unlock; drm_modeset_lock(&crtc->mutex, NULL); if (crtc->state == NULL) goto unlock; dm_crtc_state = to_dm_crtc_state(crtc->state); if (dm_crtc_state->stream == NULL) goto unlock; if (param[0] > 0) aconnector->dsc_settings.dsc_num_slices_h = DIV_ROUND_UP( pipe_ctx->stream->timing.h_addressable, param[0]); else aconnector->dsc_settings.dsc_num_slices_h = 0; dm_crtc_state->dsc_force_changed = true; unlock: if (crtc) drm_modeset_unlock(&crtc->mutex); drm_modeset_unlock(&dev->mode_config.connection_mutex); mutex_unlock(&dev->mode_config.mutex); done: kfree(wr_buf); return size; } /* function: read DSC slice height parameter on the connector * * The read function: dp_dsc_slice_height_read * returns dsc slice height used in the current configuration * The return is an integer: 0 or other positive number * * Access the status with the following command: * * cat /sys/kernel/debug/dri/0/DP-X/dsc_slice_height * * 0 - means that DSC is disabled * * Any other number more than zero represents the * slice height currently used by DSC in pixels * */ static ssize_t dp_dsc_slice_height_read(struct file *f, char __user *buf, size_t size, loff_t *pos) { char *rd_buf = NULL; char *rd_buf_ptr = NULL; struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct display_stream_compressor *dsc; struct dcn_dsc_state dsc_state = {0}; const uint32_t rd_buf_size = 100; struct pipe_ctx *pipe_ctx; ssize_t result = 0; int i, r, str_len = 30; rd_buf = kcalloc(rd_buf_size, sizeof(char), GFP_KERNEL); if (!rd_buf) return -ENOMEM; rd_buf_ptr = rd_buf; for (i = 0; i < MAX_PIPES; i++) { pipe_ctx = &aconnector->dc_link->dc->current_state->res_ctx.pipe_ctx[i]; if (pipe_ctx->stream && pipe_ctx->stream->link == aconnector->dc_link && pipe_ctx->stream->sink && pipe_ctx->stream->sink == aconnector->dc_sink) break; } dsc = pipe_ctx->stream_res.dsc; if (dsc) dsc->funcs->dsc_read_state(dsc, &dsc_state); snprintf(rd_buf_ptr, str_len, "%d\n", dsc_state.dsc_slice_height); rd_buf_ptr += str_len; while (size) { if (*pos >= rd_buf_size) break; r = put_user(*(rd_buf + result), buf); if (r) { kfree(rd_buf); return r; /* r = -EFAULT */ } buf += 1; size -= 1; *pos += 1; result += 1; } kfree(rd_buf); return result; } /* function: write DSC slice height parameter * * The write function: dp_dsc_slice_height_write * overwrites automatically generated DSC configuration * of slice height. * * The user has to write the slice height divisible by the * picture height. * * Also the user has to write height in hexidecimal * rather than in decimal. * * Writing DSC settings is done with the following command: * - To force overwrite slice height (example sets to 128 pixels): * * echo 0x80 > /sys/kernel/debug/dri/0/DP-X/dsc_slice_height * * - To stop overwriting and let driver find the optimal size, * set the height to zero: * * echo 0x0 > /sys/kernel/debug/dri/0/DP-X/dsc_slice_height * */ static ssize_t dp_dsc_slice_height_write(struct file *f, const char __user *buf, size_t size, loff_t *pos) { struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct drm_connector *connector = &aconnector->base; struct drm_device *dev = connector->dev; struct drm_crtc *crtc = NULL; struct dm_crtc_state *dm_crtc_state = NULL; struct pipe_ctx *pipe_ctx; int i; char *wr_buf = NULL; uint32_t wr_buf_size = 42; int max_param_num = 1; uint8_t param_nums = 0; long param[1] = {0}; if (size == 0) return -EINVAL; wr_buf = kcalloc(wr_buf_size, sizeof(char), GFP_KERNEL); if (!wr_buf) { DRM_DEBUG_DRIVER("no memory to allocate write buffer\n"); return -ENOSPC; } if (parse_write_buffer_into_params(wr_buf, wr_buf_size, (long *)param, buf, max_param_num, ¶m_nums)) { kfree(wr_buf); return -EINVAL; } if (param_nums <= 0) { DRM_DEBUG_DRIVER("user data not be read\n"); kfree(wr_buf); return -EINVAL; } for (i = 0; i < MAX_PIPES; i++) { pipe_ctx = &aconnector->dc_link->dc->current_state->res_ctx.pipe_ctx[i]; if (pipe_ctx->stream && pipe_ctx->stream->link == aconnector->dc_link && pipe_ctx->stream->sink && pipe_ctx->stream->sink == aconnector->dc_sink) break; } if (!pipe_ctx->stream) goto done; // Get CRTC state mutex_lock(&dev->mode_config.mutex); drm_modeset_lock(&dev->mode_config.connection_mutex, NULL); if (connector->state == NULL) goto unlock; crtc = connector->state->crtc; if (crtc == NULL) goto unlock; drm_modeset_lock(&crtc->mutex, NULL); if (crtc->state == NULL) goto unlock; dm_crtc_state = to_dm_crtc_state(crtc->state); if (dm_crtc_state->stream == NULL) goto unlock; if (param[0] > 0) aconnector->dsc_settings.dsc_num_slices_v = DIV_ROUND_UP( pipe_ctx->stream->timing.v_addressable, param[0]); else aconnector->dsc_settings.dsc_num_slices_v = 0; dm_crtc_state->dsc_force_changed = true; unlock: if (crtc) drm_modeset_unlock(&crtc->mutex); drm_modeset_unlock(&dev->mode_config.connection_mutex); mutex_unlock(&dev->mode_config.mutex); done: kfree(wr_buf); return size; } /* function: read DSC target rate on the connector in bits per pixel * * The read function: dp_dsc_bits_per_pixel_read * returns target rate of compression in bits per pixel * The return is an integer: 0 or other positive integer * * Access it with the following command: * * cat /sys/kernel/debug/dri/0/DP-X/dsc_bits_per_pixel * * 0 - means that DSC is disabled */ static ssize_t dp_dsc_bits_per_pixel_read(struct file *f, char __user *buf, size_t size, loff_t *pos) { char *rd_buf = NULL; char *rd_buf_ptr = NULL; struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct display_stream_compressor *dsc; struct dcn_dsc_state dsc_state = {0}; const uint32_t rd_buf_size = 100; struct pipe_ctx *pipe_ctx; ssize_t result = 0; int i, r, str_len = 30; rd_buf = kcalloc(rd_buf_size, sizeof(char), GFP_KERNEL); if (!rd_buf) return -ENOMEM; rd_buf_ptr = rd_buf; for (i = 0; i < MAX_PIPES; i++) { pipe_ctx = &aconnector->dc_link->dc->current_state->res_ctx.pipe_ctx[i]; if (pipe_ctx->stream && pipe_ctx->stream->link == aconnector->dc_link && pipe_ctx->stream->sink && pipe_ctx->stream->sink == aconnector->dc_sink) break; } dsc = pipe_ctx->stream_res.dsc; if (dsc) dsc->funcs->dsc_read_state(dsc, &dsc_state); snprintf(rd_buf_ptr, str_len, "%d\n", dsc_state.dsc_bits_per_pixel); rd_buf_ptr += str_len; while (size) { if (*pos >= rd_buf_size) break; r = put_user(*(rd_buf + result), buf); if (r) { kfree(rd_buf); return r; /* r = -EFAULT */ } buf += 1; size -= 1; *pos += 1; result += 1; } kfree(rd_buf); return result; } /* function: write DSC target rate in bits per pixel * * The write function: dp_dsc_bits_per_pixel_write * overwrites automatically generated DSC configuration * of DSC target bit rate. * * Also the user has to write bpp in hexidecimal * rather than in decimal. * * Writing DSC settings is done with the following command: * - To force overwrite rate (example sets to 256 bpp x 1/16): * * echo 0x100 > /sys/kernel/debug/dri/0/DP-X/dsc_bits_per_pixel * * - To stop overwriting and let driver find the optimal rate, * set the rate to zero: * * echo 0x0 > /sys/kernel/debug/dri/0/DP-X/dsc_bits_per_pixel * */ static ssize_t dp_dsc_bits_per_pixel_write(struct file *f, const char __user *buf, size_t size, loff_t *pos) { struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct drm_connector *connector = &aconnector->base; struct drm_device *dev = connector->dev; struct drm_crtc *crtc = NULL; struct dm_crtc_state *dm_crtc_state = NULL; struct pipe_ctx *pipe_ctx; int i; char *wr_buf = NULL; uint32_t wr_buf_size = 42; int max_param_num = 1; uint8_t param_nums = 0; long param[1] = {0}; if (size == 0) return -EINVAL; wr_buf = kcalloc(wr_buf_size, sizeof(char), GFP_KERNEL); if (!wr_buf) { DRM_DEBUG_DRIVER("no memory to allocate write buffer\n"); return -ENOSPC; } if (parse_write_buffer_into_params(wr_buf, wr_buf_size, (long *)param, buf, max_param_num, ¶m_nums)) { kfree(wr_buf); return -EINVAL; } if (param_nums <= 0) { DRM_DEBUG_DRIVER("user data not be read\n"); kfree(wr_buf); return -EINVAL; } for (i = 0; i < MAX_PIPES; i++) { pipe_ctx = &aconnector->dc_link->dc->current_state->res_ctx.pipe_ctx[i]; if (pipe_ctx->stream && pipe_ctx->stream->link == aconnector->dc_link && pipe_ctx->stream->sink && pipe_ctx->stream->sink == aconnector->dc_sink) break; } if (!pipe_ctx->stream) goto done; // Get CRTC state mutex_lock(&dev->mode_config.mutex); drm_modeset_lock(&dev->mode_config.connection_mutex, NULL); if (connector->state == NULL) goto unlock; crtc = connector->state->crtc; if (crtc == NULL) goto unlock; drm_modeset_lock(&crtc->mutex, NULL); if (crtc->state == NULL) goto unlock; dm_crtc_state = to_dm_crtc_state(crtc->state); if (dm_crtc_state->stream == NULL) goto unlock; aconnector->dsc_settings.dsc_bits_per_pixel = param[0]; dm_crtc_state->dsc_force_changed = true; unlock: if (crtc) drm_modeset_unlock(&crtc->mutex); drm_modeset_unlock(&dev->mode_config.connection_mutex); mutex_unlock(&dev->mode_config.mutex); done: kfree(wr_buf); return size; } /* function: read DSC picture width parameter on the connector * * The read function: dp_dsc_pic_width_read * returns dsc picture width used in the current configuration * It is the same as h_addressable of the current * display's timing * The return is an integer: 0 or other positive integer * If 0 then DSC is disabled. * * Access it with the following command: * * cat /sys/kernel/debug/dri/0/DP-X/dsc_pic_width * * 0 - means that DSC is disabled */ static ssize_t dp_dsc_pic_width_read(struct file *f, char __user *buf, size_t size, loff_t *pos) { char *rd_buf = NULL; char *rd_buf_ptr = NULL; struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct display_stream_compressor *dsc; struct dcn_dsc_state dsc_state = {0}; const uint32_t rd_buf_size = 100; struct pipe_ctx *pipe_ctx; ssize_t result = 0; int i, r, str_len = 30; rd_buf = kcalloc(rd_buf_size, sizeof(char), GFP_KERNEL); if (!rd_buf) return -ENOMEM; rd_buf_ptr = rd_buf; for (i = 0; i < MAX_PIPES; i++) { pipe_ctx = &aconnector->dc_link->dc->current_state->res_ctx.pipe_ctx[i]; if (pipe_ctx->stream && pipe_ctx->stream->link == aconnector->dc_link && pipe_ctx->stream->sink && pipe_ctx->stream->sink == aconnector->dc_sink) break; } dsc = pipe_ctx->stream_res.dsc; if (dsc) dsc->funcs->dsc_read_state(dsc, &dsc_state); snprintf(rd_buf_ptr, str_len, "%d\n", dsc_state.dsc_pic_width); rd_buf_ptr += str_len; while (size) { if (*pos >= rd_buf_size) break; r = put_user(*(rd_buf + result), buf); if (r) { kfree(rd_buf); return r; /* r = -EFAULT */ } buf += 1; size -= 1; *pos += 1; result += 1; } kfree(rd_buf); return result; } static ssize_t dp_dsc_pic_height_read(struct file *f, char __user *buf, size_t size, loff_t *pos) { char *rd_buf = NULL; char *rd_buf_ptr = NULL; struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct display_stream_compressor *dsc; struct dcn_dsc_state dsc_state = {0}; const uint32_t rd_buf_size = 100; struct pipe_ctx *pipe_ctx; ssize_t result = 0; int i, r, str_len = 30; rd_buf = kcalloc(rd_buf_size, sizeof(char), GFP_KERNEL); if (!rd_buf) return -ENOMEM; rd_buf_ptr = rd_buf; for (i = 0; i < MAX_PIPES; i++) { pipe_ctx = &aconnector->dc_link->dc->current_state->res_ctx.pipe_ctx[i]; if (pipe_ctx->stream && pipe_ctx->stream->link == aconnector->dc_link && pipe_ctx->stream->sink && pipe_ctx->stream->sink == aconnector->dc_sink) break; } dsc = pipe_ctx->stream_res.dsc; if (dsc) dsc->funcs->dsc_read_state(dsc, &dsc_state); snprintf(rd_buf_ptr, str_len, "%d\n", dsc_state.dsc_pic_height); rd_buf_ptr += str_len; while (size) { if (*pos >= rd_buf_size) break; r = put_user(*(rd_buf + result), buf); if (r) { kfree(rd_buf); return r; /* r = -EFAULT */ } buf += 1; size -= 1; *pos += 1; result += 1; } kfree(rd_buf); return result; } /* function: read DSC chunk size parameter on the connector * * The read function: dp_dsc_chunk_size_read * returns dsc chunk size set in the current configuration * The value is calculated automatically by DSC code * and depends on slice parameters and bpp target rate * The return is an integer: 0 or other positive integer * If 0 then DSC is disabled. * * Access it with the following command: * * cat /sys/kernel/debug/dri/0/DP-X/dsc_chunk_size * * 0 - means that DSC is disabled */ static ssize_t dp_dsc_chunk_size_read(struct file *f, char __user *buf, size_t size, loff_t *pos) { char *rd_buf = NULL; char *rd_buf_ptr = NULL; struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct display_stream_compressor *dsc; struct dcn_dsc_state dsc_state = {0}; const uint32_t rd_buf_size = 100; struct pipe_ctx *pipe_ctx; ssize_t result = 0; int i, r, str_len = 30; rd_buf = kcalloc(rd_buf_size, sizeof(char), GFP_KERNEL); if (!rd_buf) return -ENOMEM; rd_buf_ptr = rd_buf; for (i = 0; i < MAX_PIPES; i++) { pipe_ctx = &aconnector->dc_link->dc->current_state->res_ctx.pipe_ctx[i]; if (pipe_ctx->stream && pipe_ctx->stream->link == aconnector->dc_link && pipe_ctx->stream->sink && pipe_ctx->stream->sink == aconnector->dc_sink) break; } dsc = pipe_ctx->stream_res.dsc; if (dsc) dsc->funcs->dsc_read_state(dsc, &dsc_state); snprintf(rd_buf_ptr, str_len, "%d\n", dsc_state.dsc_chunk_size); rd_buf_ptr += str_len; while (size) { if (*pos >= rd_buf_size) break; r = put_user(*(rd_buf + result), buf); if (r) { kfree(rd_buf); return r; /* r = -EFAULT */ } buf += 1; size -= 1; *pos += 1; result += 1; } kfree(rd_buf); return result; } /* function: read DSC slice bpg offset on the connector * * The read function: dp_dsc_slice_bpg_offset_read * returns dsc bpg slice offset set in the current configuration * The value is calculated automatically by DSC code * and depends on slice parameters and bpp target rate * The return is an integer: 0 or other positive integer * If 0 then DSC is disabled. * * Access it with the following command: * * cat /sys/kernel/debug/dri/0/DP-X/dsc_slice_bpg_offset * * 0 - means that DSC is disabled */ static ssize_t dp_dsc_slice_bpg_offset_read(struct file *f, char __user *buf, size_t size, loff_t *pos) { char *rd_buf = NULL; char *rd_buf_ptr = NULL; struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct display_stream_compressor *dsc; struct dcn_dsc_state dsc_state = {0}; const uint32_t rd_buf_size = 100; struct pipe_ctx *pipe_ctx; ssize_t result = 0; int i, r, str_len = 30; rd_buf = kcalloc(rd_buf_size, sizeof(char), GFP_KERNEL); if (!rd_buf) return -ENOMEM; rd_buf_ptr = rd_buf; for (i = 0; i < MAX_PIPES; i++) { pipe_ctx = &aconnector->dc_link->dc->current_state->res_ctx.pipe_ctx[i]; if (pipe_ctx->stream && pipe_ctx->stream->link == aconnector->dc_link && pipe_ctx->stream->sink && pipe_ctx->stream->sink == aconnector->dc_sink) break; } dsc = pipe_ctx->stream_res.dsc; if (dsc) dsc->funcs->dsc_read_state(dsc, &dsc_state); snprintf(rd_buf_ptr, str_len, "%d\n", dsc_state.dsc_slice_bpg_offset); rd_buf_ptr += str_len; while (size) { if (*pos >= rd_buf_size) break; r = put_user(*(rd_buf + result), buf); if (r) { kfree(rd_buf); return r; /* r = -EFAULT */ } buf += 1; size -= 1; *pos += 1; result += 1; } kfree(rd_buf); return result; } /* * function description: Read max_requested_bpc property from the connector * * Access it with the following command: * * cat /sys/kernel/debug/dri/0/DP-X/max_bpc * */ static ssize_t dp_max_bpc_read(struct file *f, char __user *buf, size_t size, loff_t *pos) { struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct drm_connector *connector = &aconnector->base; struct drm_device *dev = connector->dev; struct dm_connector_state *state; ssize_t result = 0; char *rd_buf = NULL; char *rd_buf_ptr = NULL; const uint32_t rd_buf_size = 10; int r; rd_buf = kcalloc(rd_buf_size, sizeof(char), GFP_KERNEL); if (!rd_buf) return -ENOMEM; mutex_lock(&dev->mode_config.mutex); drm_modeset_lock(&dev->mode_config.connection_mutex, NULL); if (connector->state == NULL) goto unlock; state = to_dm_connector_state(connector->state); rd_buf_ptr = rd_buf; snprintf(rd_buf_ptr, rd_buf_size, "%u\n", state->base.max_requested_bpc); while (size) { if (*pos >= rd_buf_size) break; r = put_user(*(rd_buf + result), buf); if (r) { result = r; /* r = -EFAULT */ goto unlock; } buf += 1; size -= 1; *pos += 1; result += 1; } unlock: drm_modeset_unlock(&dev->mode_config.connection_mutex); mutex_unlock(&dev->mode_config.mutex); kfree(rd_buf); return result; } /* * function description: Set max_requested_bpc property on the connector * * This function will not force the input BPC on connector, it will only * change the max value. This is equivalent to setting max_bpc through * xrandr. * * The BPC value written must be >= 6 and <= 16. Values outside of this * range will result in errors. * * BPC values: * 0x6 - 6 BPC * 0x8 - 8 BPC * 0xa - 10 BPC * 0xc - 12 BPC * 0x10 - 16 BPC * * Write the max_bpc in the following way: * * echo 0x6 > /sys/kernel/debug/dri/0/DP-X/max_bpc * */ static ssize_t dp_max_bpc_write(struct file *f, const char __user *buf, size_t size, loff_t *pos) { struct amdgpu_dm_connector *aconnector = file_inode(f)->i_private; struct drm_connector *connector = &aconnector->base; struct dm_connector_state *state; struct drm_device *dev = connector->dev; char *wr_buf = NULL; uint32_t wr_buf_size = 42; int max_param_num = 1; long param[1] = {0}; uint8_t param_nums = 0; if (size == 0) return -EINVAL; wr_buf = kcalloc(wr_buf_size, sizeof(char), GFP_KERNEL); if (!wr_buf) { DRM_DEBUG_DRIVER("no memory to allocate write buffer\n"); return -ENOSPC; } if (parse_write_buffer_into_params(wr_buf, wr_buf_size, (long *)param, buf, max_param_num, ¶m_nums)) { kfree(wr_buf); return -EINVAL; } if (param_nums <= 0) { DRM_DEBUG_DRIVER("user data not be read\n"); kfree(wr_buf); return -EINVAL; } if (param[0] < 6 || param[0] > 16) { DRM_DEBUG_DRIVER("bad max_bpc value\n"); kfree(wr_buf); return -EINVAL; } mutex_lock(&dev->mode_config.mutex); drm_modeset_lock(&dev->mode_config.connection_mutex, NULL); if (connector->state == NULL) goto unlock; state = to_dm_connector_state(connector->state); state->base.max_requested_bpc = param[0]; unlock: drm_modeset_unlock(&dev->mode_config.connection_mutex); mutex_unlock(&dev->mode_config.mutex); kfree(wr_buf); return size; } /* * IPS status. Read only. * * Example usage: cat /sys/kernel/debug/dri/0/amdgpu_dm_ips_status */ static int ips_status_show(struct seq_file *m, void *unused) { struct amdgpu_device *adev = m->private; struct dc *dc = adev->dm.dc; struct dc_dmub_srv *dc_dmub_srv; seq_printf(m, "IPS config: %d\n", dc->config.disable_ips); seq_printf(m, "Idle optimization: %d\n", dc->idle_optimizations_allowed); if (adev->dm.idle_workqueue) { seq_printf(m, "Idle workqueue - enabled: %d\n", adev->dm.idle_workqueue->enable); seq_printf(m, "Idle workqueue - running: %d\n", adev->dm.idle_workqueue->running); } dc_dmub_srv = dc->ctx->dmub_srv; if (dc_dmub_srv && dc_dmub_srv->dmub) { uint32_t rcg_count, ips1_count, ips2_count; volatile const struct dmub_shared_state_ips_fw *ips_fw = &dc_dmub_srv->dmub->shared_state[DMUB_SHARED_SHARE_FEATURE__IPS_FW].data.ips_fw; rcg_count = ips_fw->rcg_entry_count; ips1_count = ips_fw->ips1_entry_count; ips2_count = ips_fw->ips2_entry_count; seq_printf(m, "entry counts: rcg=%u ips1=%u ips2=%u\n", rcg_count, ips1_count, ips2_count); rcg_count = ips_fw->rcg_exit_count; ips1_count = ips_fw->ips1_exit_count; ips2_count = ips_fw->ips2_exit_count; seq_printf(m, "exit counts: rcg=%u ips1=%u ips2=%u", rcg_count, ips1_count, ips2_count); seq_puts(m, "\n"); } return 0; } /* * Backlight at this moment. Read only. * As written to display, taking ABM and backlight lut into account. * Ranges from 0x0 to 0x10000 (= 100% PWM) * * Example usage: cat /sys/kernel/debug/dri/0/eDP-1/current_backlight */ static int current_backlight_show(struct seq_file *m, void *unused) { struct amdgpu_dm_connector *aconnector = to_amdgpu_dm_connector(m->private); struct dc_link *link = aconnector->dc_link; unsigned int backlight; backlight = dc_link_get_backlight_level(link); seq_printf(m, "0x%x\n", backlight); return 0; } /* * Backlight value that is being approached. Read only. * As written to display, taking ABM and backlight lut into account. * Ranges from 0x0 to 0x10000 (= 100% PWM) * * Example usage: cat /sys/kernel/debug/dri/0/eDP-1/target_backlight */ static int target_backlight_show(struct seq_file *m, void *unused) { struct amdgpu_dm_connector *aconnector = to_amdgpu_dm_connector(m->private); struct dc_link *link = aconnector->dc_link; unsigned int backlight; backlight = dc_link_get_target_backlight_pwm(link); seq_printf(m, "0x%x\n", backlight); return 0; } /* * function description: Determine if the connector is mst connector * * This function helps to determine whether a connector is a mst connector. * - "root" stands for the root connector of the topology * - "branch" stands for branch device of the topology * - "end" stands for leaf node connector of the topology * - "no" stands for the connector is not a device of a mst topology * Access it with the following command: * * cat /sys/kernel/debug/dri/0/DP-X/is_mst_connector * */ static int dp_is_mst_connector_show(struct seq_file *m, void *unused) { struct drm_connector *connector = m->private; struct amdgpu_dm_connector *aconnector = to_amdgpu_dm_connector(connector); struct drm_dp_mst_topology_mgr *mgr = NULL; struct drm_dp_mst_port *port = NULL; char *role = NULL; mutex_lock(&aconnector->hpd_lock); if (aconnector->mst_mgr.mst_state) { role = "root"; } else if (aconnector->mst_root && aconnector->mst_root->mst_mgr.mst_state) { role = "end"; mgr = &aconnector->mst_root->mst_mgr; port = aconnector->mst_output_port; drm_modeset_lock(&mgr->base.lock, NULL); if (port->pdt == DP_PEER_DEVICE_MST_BRANCHING && port->mcs) role = "branch"; drm_modeset_unlock(&mgr->base.lock); } else { role = "no"; } seq_printf(m, "%s\n", role); mutex_unlock(&aconnector->hpd_lock); return 0; } /* * function description: Read out the mst progress status * * This function helps to determine the mst progress status of * a mst connector. * * Access it with the following command: * * cat /sys/kernel/debug/dri/0/DP-X/mst_progress_status * */ static int dp_mst_progress_status_show(struct seq_file *m, void *unused) { struct drm_connector *connector = m->private; struct amdgpu_dm_connector *aconnector = to_amdgpu_dm_connector(connector); struct amdgpu_device *adev = drm_to_adev(connector->dev); int i; mutex_lock(&aconnector->hpd_lock); mutex_lock(&adev->dm.dc_lock); if (aconnector->mst_status == MST_STATUS_DEFAULT) { seq_puts(m, "disabled\n"); } else { for (i = 0; i < sizeof(mst_progress_status)/sizeof(char *); i++) seq_printf(m, "%s:%s\n", mst_progress_status[i], aconnector->mst_status & BIT(i) ? "done" : "not_done"); } mutex_unlock(&adev->dm.dc_lock); mutex_unlock(&aconnector->hpd_lock); return 0; } /* * Reports whether the connected display is a USB4 DPIA tunneled display * Example usage: cat /sys/kernel/debug/dri/0/DP-8/is_dpia_link */ static int is_dpia_link_show(struct seq_file *m, void *data) { struct drm_connector *connector = m->private; struct amdgpu_dm_connector *aconnector = to_amdgpu_dm_connector(connector); struct dc_link *link = aconnector->dc_link; if (connector->status != connector_status_connected) return -ENODEV; seq_printf(m, "%s\n", (link->ep_type == DISPLAY_ENDPOINT_USB4_DPIA) ? "yes" : (link->ep_type == DISPLAY_ENDPOINT_PHY) ? "no" : "unknown"); return 0; } DEFINE_SHOW_ATTRIBUTE(dp_dsc_fec_support); DEFINE_SHOW_ATTRIBUTE(dmub_fw_state); DEFINE_SHOW_ATTRIBUTE(dmub_tracebuffer); DEFINE_SHOW_ATTRIBUTE(dp_lttpr_status); DEFINE_SHOW_ATTRIBUTE(hdcp_sink_capability); DEFINE_SHOW_ATTRIBUTE(internal_display); DEFINE_SHOW_ATTRIBUTE(odm_combine_segments); DEFINE_SHOW_ATTRIBUTE(replay_capability); DEFINE_SHOW_ATTRIBUTE(psr_capability); DEFINE_SHOW_ATTRIBUTE(dp_is_mst_connector); DEFINE_SHOW_ATTRIBUTE(dp_mst_progress_status); DEFINE_SHOW_ATTRIBUTE(is_dpia_link); static const struct file_operations dp_dsc_clock_en_debugfs_fops = { .owner = THIS_MODULE, .read = dp_dsc_clock_en_read, .write = dp_dsc_clock_en_write, .llseek = default_llseek }; static const struct file_operations dp_dsc_slice_width_debugfs_fops = { .owner = THIS_MODULE, .read = dp_dsc_slice_width_read, .write = dp_dsc_slice_width_write, .llseek = default_llseek }; static const struct file_operations dp_dsc_slice_height_debugfs_fops = { .owner = THIS_MODULE, .read = dp_dsc_slice_height_read, .write = dp_dsc_slice_height_write, .llseek = default_llseek }; static const struct file_operations dp_dsc_bits_per_pixel_debugfs_fops = { .owner = THIS_MODULE, .read = dp_dsc_bits_per_pixel_read, .write = dp_dsc_bits_per_pixel_write, .llseek = default_llseek }; static const struct file_operations dp_dsc_pic_width_debugfs_fops = { .owner = THIS_MODULE, .read = dp_dsc_pic_width_read, .llseek = default_llseek }; static const struct file_operations dp_dsc_pic_height_debugfs_fops = { .owner = THIS_MODULE, .read = dp_dsc_pic_height_read, .llseek = default_llseek }; static const struct file_operations dp_dsc_chunk_size_debugfs_fops = { .owner = THIS_MODULE, .read = dp_dsc_chunk_size_read, .llseek = default_llseek }; static const struct file_operations dp_dsc_slice_bpg_offset_debugfs_fops = { .owner = THIS_MODULE, .read = dp_dsc_slice_bpg_offset_read, .llseek = default_llseek }; static const struct file_operations trigger_hotplug_debugfs_fops = { .owner = THIS_MODULE, .write = trigger_hotplug, .llseek = default_llseek }; static const struct file_operations dp_link_settings_debugfs_fops = { .owner = THIS_MODULE, .read = dp_link_settings_read, .write = dp_link_settings_write, .llseek = default_llseek }; static const struct file_operations dp_phy_settings_debugfs_fop = { .owner = THIS_MODULE, .read = dp_phy_settings_read, .write = dp_phy_settings_write, .llseek = default_llseek }; static const struct file_operations dp_phy_test_pattern_fops = { .owner = THIS_MODULE, .write = dp_phy_test_pattern_debugfs_write, .llseek = default_llseek }; static const struct file_operations sdp_message_fops = { .owner = THIS_MODULE, .write = dp_sdp_message_debugfs_write, .llseek = default_llseek }; static const struct file_operations dp_max_bpc_debugfs_fops = { .owner = THIS_MODULE, .read = dp_max_bpc_read, .write = dp_max_bpc_write, .llseek = default_llseek }; static const struct file_operations dp_dsc_disable_passthrough_debugfs_fops = { .owner = THIS_MODULE, .write = dp_dsc_passthrough_set, .llseek = default_llseek }; static const struct file_operations dp_mst_link_settings_debugfs_fops = { .owner = THIS_MODULE, .write = dp_mst_link_setting, .llseek = default_llseek }; static const struct { char *name; const struct file_operations *fops; } dp_debugfs_entries[] = { {"link_settings", &dp_link_settings_debugfs_fops}, {"phy_settings", &dp_phy_settings_debugfs_fop}, {"lttpr_status", &dp_lttpr_status_fops}, {"test_pattern", &dp_phy_test_pattern_fops}, {"hdcp_sink_capability", &hdcp_sink_capability_fops}, {"sdp_message", &sdp_message_fops}, {"dsc_clock_en", &dp_dsc_clock_en_debugfs_fops}, {"dsc_slice_width", &dp_dsc_slice_width_debugfs_fops}, {"dsc_slice_height", &dp_dsc_slice_height_debugfs_fops}, {"dsc_bits_per_pixel", &dp_dsc_bits_per_pixel_debugfs_fops}, {"dsc_pic_width", &dp_dsc_pic_width_debugfs_fops}, {"dsc_pic_height", &dp_dsc_pic_height_debugfs_fops}, {"dsc_chunk_size", &dp_dsc_chunk_size_debugfs_fops}, {"dsc_slice_bpg", &dp_dsc_slice_bpg_offset_debugfs_fops}, {"dp_dsc_fec_support", &dp_dsc_fec_support_fops}, {"max_bpc", &dp_max_bpc_debugfs_fops}, {"dsc_disable_passthrough", &dp_dsc_disable_passthrough_debugfs_fops}, {"is_mst_connector", &dp_is_mst_connector_fops}, {"mst_progress_status", &dp_mst_progress_status_fops}, {"is_dpia_link", &is_dpia_link_fops}, {"mst_link_settings", &dp_mst_link_settings_debugfs_fops} }; static const struct { char *name; const struct file_operations *fops; } hdmi_debugfs_entries[] = { {"hdcp_sink_capability", &hdcp_sink_capability_fops} }; /* * Force YUV420 output if available from the given mode */ static int force_yuv420_output_set(void *data, u64 val) { struct amdgpu_dm_connector *connector = data; connector->force_yuv420_output = (bool)val; return 0; } /* * Check if YUV420 is forced when available from the given mode */ static int force_yuv420_output_get(void *data, u64 *val) { struct amdgpu_dm_connector *connector = data; *val = connector->force_yuv420_output; return 0; } DEFINE_DEBUGFS_ATTRIBUTE(force_yuv420_output_fops, force_yuv420_output_get, force_yuv420_output_set, "%llu\n"); /* * Read Replay state */ static int replay_get_state(void *data, u64 *val) { struct amdgpu_dm_connector *connector = data; struct dc_link *link = connector->dc_link; uint64_t state = REPLAY_STATE_INVALID; dc_link_get_replay_state(link, &state); *val = state; return 0; } /* * Read PSR state */ static int psr_get(void *data, u64 *val) { struct amdgpu_dm_connector *connector = data; struct dc_link *link = connector->dc_link; enum dc_psr_state state = PSR_STATE0; dc_link_get_psr_state(link, &state); *val = state; return 0; } /* * Read PSR state residency */ static int psr_read_residency(void *data, u64 *val) { struct amdgpu_dm_connector *connector = data; struct dc_link *link = connector->dc_link; u32 residency = 0; link->dc->link_srv->edp_get_psr_residency(link, &residency, PSR_RESIDENCY_MODE_PHY); *val = (u64)residency; return 0; } /* read allow_edp_hotplug_detection */ static int allow_edp_hotplug_detection_get(void *data, u64 *val) { struct amdgpu_dm_connector *aconnector = data; struct drm_connector *connector = &aconnector->base; struct drm_device *dev = connector->dev; struct amdgpu_device *adev = drm_to_adev(dev); *val = adev->dm.dc->config.allow_edp_hotplug_detection; return 0; } /* set allow_edp_hotplug_detection */ static int allow_edp_hotplug_detection_set(void *data, u64 val) { struct amdgpu_dm_connector *aconnector = data; struct drm_connector *connector = &aconnector->base; struct drm_device *dev = connector->dev; struct amdgpu_device *adev = drm_to_adev(dev); adev->dm.dc->config.allow_edp_hotplug_detection = (uint32_t) val; return 0; } /* check if kernel disallow eDP enter psr state * cat /sys/kernel/debug/dri/0/eDP-X/disallow_edp_enter_psr * 0: allow edp enter psr; 1: disallow */ static int disallow_edp_enter_psr_get(void *data, u64 *val) { struct amdgpu_dm_connector *aconnector = data; *val = (u64) aconnector->disallow_edp_enter_psr; return 0; } /* set kernel disallow eDP enter psr state * echo 0x0 /sys/kernel/debug/dri/0/eDP-X/disallow_edp_enter_psr * 0: allow edp enter psr; 1: disallow * * usage: test app read crc from PSR eDP rx. * * during kernel boot up, kernel write dpcd 0x170 = 5. * this notify eDP rx psr enable and let rx check crc. * rx fw will start checking crc for rx internal logic. * crc read count within dpcd 0x246 is not updated and * value is 0. when eDP tx driver wants to read rx crc * from dpcd 0x246, 0x270, read count 0 lead tx driver * timeout. * * to avoid this, we add this debugfs to let test app to disbable * rx crc checking for rx internal logic. then test app can read * non-zero crc read count. * * expected app sequence is as below: * 1. disable eDP PHY and notify eDP rx with dpcd 0x600 = 2. * 2. echo 0x1 /sys/kernel/debug/dri/0/eDP-X/disallow_edp_enter_psr * 3. enable eDP PHY and notify eDP rx with dpcd 0x600 = 1 but * without dpcd 0x170 = 5. * 4. read crc from rx dpcd 0x270, 0x246, etc. * 5. echo 0x0 /sys/kernel/debug/dri/0/eDP-X/disallow_edp_enter_psr. * this will let eDP back to normal with psr setup dpcd 0x170 = 5. */ static int disallow_edp_enter_psr_set(void *data, u64 val) { struct amdgpu_dm_connector *aconnector = data; aconnector->disallow_edp_enter_psr = val ? true : false; return 0; } static int dmub_trace_mask_set(void *data, u64 val) { struct amdgpu_device *adev = data; struct dmub_srv *srv = adev->dm.dc->ctx->dmub_srv->dmub; enum dmub_gpint_command cmd; u64 mask = 0xffff; u8 shift = 0; u32 res; int i; if (!srv->fw_version) return -EINVAL; for (i = 0; i < 4; i++) { res = (val & mask) >> shift; switch (i) { case 0: cmd = DMUB_GPINT__SET_TRACE_BUFFER_MASK_WORD0; break; case 1: cmd = DMUB_GPINT__SET_TRACE_BUFFER_MASK_WORD1; break; case 2: cmd = DMUB_GPINT__SET_TRACE_BUFFER_MASK_WORD2; break; case 3: cmd = DMUB_GPINT__SET_TRACE_BUFFER_MASK_WORD3; break; } if (!dc_wake_and_execute_gpint(adev->dm.dc->ctx, cmd, res, NULL, DM_DMUB_WAIT_TYPE_WAIT)) return -EIO; usleep_range(100, 1000); mask <<= 16; shift += 16; } return 0; } static int dmub_trace_mask_show(void *data, u64 *val) { enum dmub_gpint_command cmd = DMUB_GPINT__GET_TRACE_BUFFER_MASK_WORD0; struct amdgpu_device *adev = data; struct dmub_srv *srv = adev->dm.dc->ctx->dmub_srv->dmub; u8 shift = 0; u64 raw = 0; u64 res = 0; int i = 0; if (!srv->fw_version) return -EINVAL; while (i < 4) { uint32_t response; if (!dc_wake_and_execute_gpint(adev->dm.dc->ctx, cmd, 0, &response, DM_DMUB_WAIT_TYPE_WAIT_WITH_REPLY)) return -EIO; raw = response; usleep_range(100, 1000); cmd++; res |= (raw << shift); shift += 16; i++; } *val = res; return 0; } DEFINE_DEBUGFS_ATTRIBUTE(dmub_trace_mask_fops, dmub_trace_mask_show, dmub_trace_mask_set, "0x%llx\n"); /* * Set dmcub trace event IRQ enable or disable. * Usage to enable dmcub trace event IRQ: echo 1 > /sys/kernel/debug/dri/0/amdgpu_dm_dmcub_trace_event_en * Usage to disable dmcub trace event IRQ: echo 0 > /sys/kernel/debug/dri/0/amdgpu_dm_dmcub_trace_event_en */ static int dmcub_trace_event_state_set(void *data, u64 val) { struct amdgpu_device *adev = data; if (val == 1 || val == 0) { dc_dmub_trace_event_control(adev->dm.dc, val); adev->dm.dmcub_trace_event_en = (bool)val; } else return 0; return 0; } /* * The interface doesn't need get function, so it will return the * value of zero * Usage: cat /sys/kernel/debug/dri/0/amdgpu_dm_dmcub_trace_event_en */ static int dmcub_trace_event_state_get(void *data, u64 *val) { struct amdgpu_device *adev = data; *val = adev->dm.dmcub_trace_event_en; return 0; } DEFINE_DEBUGFS_ATTRIBUTE(dmcub_trace_event_state_fops, dmcub_trace_event_state_get, dmcub_trace_event_state_set, "%llu\n"); DEFINE_DEBUGFS_ATTRIBUTE(replay_state_fops, replay_get_state, NULL, "%llu\n"); DEFINE_DEBUGFS_ATTRIBUTE(psr_fops, psr_get, NULL, "%llu\n"); DEFINE_DEBUGFS_ATTRIBUTE(psr_residency_fops, psr_read_residency, NULL, "%llu\n"); DEFINE_DEBUGFS_ATTRIBUTE(allow_edp_hotplug_detection_fops, allow_edp_hotplug_detection_get, allow_edp_hotplug_detection_set, "%llu\n"); DEFINE_DEBUGFS_ATTRIBUTE(disallow_edp_enter_psr_fops, disallow_edp_enter_psr_get, disallow_edp_enter_psr_set, "%llu\n"); DEFINE_SHOW_ATTRIBUTE(current_backlight); DEFINE_SHOW_ATTRIBUTE(target_backlight); DEFINE_SHOW_ATTRIBUTE(ips_status); static const struct { char *name; const struct file_operations *fops; } connector_debugfs_entries[] = { {"force_yuv420_output", &force_yuv420_output_fops}, {"trigger_hotplug", &trigger_hotplug_debugfs_fops}, {"internal_display", &internal_display_fops}, {"odm_combine_segments", &odm_combine_segments_fops} }; /* * Returns supported customized link rates by this eDP panel. * Example usage: cat /sys/kernel/debug/dri/0/eDP-x/ilr_setting */ static int edp_ilr_show(struct seq_file *m, void *unused) { struct amdgpu_dm_connector *aconnector = to_amdgpu_dm_connector(m->private); struct dc_link *link = aconnector->dc_link; uint8_t supported_link_rates[16]; uint32_t link_rate_in_khz; uint32_t entry = 0; uint8_t dpcd_rev; memset(supported_link_rates, 0, sizeof(supported_link_rates)); dm_helpers_dp_read_dpcd(link->ctx, link, DP_SUPPORTED_LINK_RATES, supported_link_rates, sizeof(supported_link_rates)); dpcd_rev = link->dpcd_caps.dpcd_rev.raw; if (dpcd_rev >= DP_DPCD_REV_13 && (supported_link_rates[entry+1] != 0 || supported_link_rates[entry] != 0)) { for (entry = 0; entry < 16; entry += 2) { link_rate_in_khz = (supported_link_rates[entry+1] * 0x100 + supported_link_rates[entry]) * 200; seq_printf(m, "[%d] %d kHz\n", entry/2, link_rate_in_khz); } } else { seq_puts(m, "ILR is not supported by this eDP panel.\n"); } return 0; } /* * Set supported customized link rate to eDP panel. * * echo <lane_count> <link_rate option> > ilr_setting * * for example, supported ILR : [0] 1620000 kHz [1] 2160000 kHz [2] 2430000 kHz ... * echo 4 1 > /sys/kernel/debug/dri/0/eDP-x/ilr_setting * to set 4 lanes and 2.16 GHz */ static ssize_t edp_ilr_write(struct file *f, const char __user *buf, size_t size, loff_t *pos) { struct amdgpu_dm_connector *connector = file_inode(f)->i_private; struct dc_link *link = connector->dc_link; struct amdgpu_device *adev = drm_to_adev(connector->base.dev); struct dc *dc = (struct dc *)link->dc; struct dc_link_settings prefer_link_settings; char *wr_buf = NULL; const uint32_t wr_buf_size = 40; /* 0: lane_count; 1: link_rate */ int max_param_num = 2; uint8_t param_nums = 0; long param[2]; bool valid_input = true; if (size == 0) return -EINVAL; wr_buf = kcalloc(wr_buf_size, sizeof(char), GFP_KERNEL); if (!wr_buf) return -ENOMEM; if (parse_write_buffer_into_params(wr_buf, wr_buf_size, (long *)param, buf, max_param_num, ¶m_nums)) { kfree(wr_buf); return -EINVAL; } if (param_nums <= 0) { kfree(wr_buf); return -EINVAL; } switch (param[0]) { case LANE_COUNT_ONE: case LANE_COUNT_TWO: case LANE_COUNT_FOUR: break; default: valid_input = false; break; } if (param[1] >= link->dpcd_caps.edp_supported_link_rates_count) valid_input = false; if (!valid_input) { kfree(wr_buf); DRM_DEBUG_DRIVER("Invalid Input value. No HW will be programmed\n"); prefer_link_settings.use_link_rate_set = false; mutex_lock(&adev->dm.dc_lock); dc_link_set_preferred_training_settings(dc, NULL, NULL, link, false); mutex_unlock(&adev->dm.dc_lock); return size; } /* save user force lane_count, link_rate to preferred settings * spread spectrum will not be changed */ prefer_link_settings.link_spread = link->cur_link_settings.link_spread; prefer_link_settings.lane_count = param[0]; prefer_link_settings.use_link_rate_set = true; prefer_link_settings.link_rate_set = param[1]; prefer_link_settings.link_rate = link->dpcd_caps.edp_supported_link_rates[param[1]]; mutex_lock(&adev->dm.dc_lock); dc_link_set_preferred_training_settings(dc, &prefer_link_settings, NULL, link, false); mutex_unlock(&adev->dm.dc_lock); kfree(wr_buf); return size; } static int edp_ilr_open(struct inode *inode, struct file *file) { return single_open(file, edp_ilr_show, inode->i_private); } static const struct file_operations edp_ilr_debugfs_fops = { .owner = THIS_MODULE, .open = edp_ilr_open, .read = seq_read, .llseek = seq_lseek, .release = single_release, .write = edp_ilr_write }; void connector_debugfs_init(struct amdgpu_dm_connector *connector) { int i; struct dentry *dir = connector->base.debugfs_entry; if (connector->base.connector_type == DRM_MODE_CONNECTOR_DisplayPort || connector->base.connector_type == DRM_MODE_CONNECTOR_eDP) { for (i = 0; i < ARRAY_SIZE(dp_debugfs_entries); i++) { debugfs_create_file(dp_debugfs_entries[i].name, 0644, dir, connector, dp_debugfs_entries[i].fops); } } if (connector->base.connector_type == DRM_MODE_CONNECTOR_eDP) { debugfs_create_file("replay_capability", 0444, dir, connector, &replay_capability_fops); debugfs_create_file("replay_state", 0444, dir, connector, &replay_state_fops); debugfs_create_file_unsafe("psr_capability", 0444, dir, connector, &psr_capability_fops); debugfs_create_file_unsafe("psr_state", 0444, dir, connector, &psr_fops); debugfs_create_file_unsafe("psr_residency", 0444, dir, connector, &psr_residency_fops); debugfs_create_file("amdgpu_current_backlight_pwm", 0444, dir, connector, ¤t_backlight_fops); debugfs_create_file("amdgpu_target_backlight_pwm", 0444, dir, connector, &target_backlight_fops); debugfs_create_file("ilr_setting", 0644, dir, connector, &edp_ilr_debugfs_fops); debugfs_create_file("allow_edp_hotplug_detection", 0644, dir, connector, &allow_edp_hotplug_detection_fops); debugfs_create_file("disallow_edp_enter_psr", 0644, dir, connector, &disallow_edp_enter_psr_fops); } for (i = 0; i < ARRAY_SIZE(connector_debugfs_entries); i++) { debugfs_create_file(connector_debugfs_entries[i].name, 0644, dir, connector, connector_debugfs_entries[i].fops); } if (connector->base.connector_type == DRM_MODE_CONNECTOR_HDMIA) { for (i = 0; i < ARRAY_SIZE(hdmi_debugfs_entries); i++) { debugfs_create_file(hdmi_debugfs_entries[i].name, 0644, dir, connector, hdmi_debugfs_entries[i].fops); } } } #ifdef CONFIG_DRM_AMD_SECURE_DISPLAY /* * Set crc window coordinate x start */ static int crc_win_x_start_set(void *data, u64 val) { struct drm_crtc *crtc = data; struct drm_device *drm_dev = crtc->dev; struct amdgpu_crtc *acrtc = to_amdgpu_crtc(crtc); spin_lock_irq(&drm_dev->event_lock); acrtc->dm_irq_params.window_param.x_start = (uint16_t) val; acrtc->dm_irq_params.window_param.update_win = false; spin_unlock_irq(&drm_dev->event_lock); return 0; } /* * Get crc window coordinate x start */ static int crc_win_x_start_get(void *data, u64 *val) { struct drm_crtc *crtc = data; struct drm_device *drm_dev = crtc->dev; struct amdgpu_crtc *acrtc = to_amdgpu_crtc(crtc); spin_lock_irq(&drm_dev->event_lock); *val = acrtc->dm_irq_params.window_param.x_start; spin_unlock_irq(&drm_dev->event_lock); return 0; } DEFINE_DEBUGFS_ATTRIBUTE(crc_win_x_start_fops, crc_win_x_start_get, crc_win_x_start_set, "%llu\n"); /* * Set crc window coordinate y start */ static int crc_win_y_start_set(void *data, u64 val) { struct drm_crtc *crtc = data; struct drm_device *drm_dev = crtc->dev; struct amdgpu_crtc *acrtc = to_amdgpu_crtc(crtc); spin_lock_irq(&drm_dev->event_lock); acrtc->dm_irq_params.window_param.y_start = (uint16_t) val; acrtc->dm_irq_params.window_param.update_win = false; spin_unlock_irq(&drm_dev->event_lock); return 0; } /* * Get crc window coordinate y start */ static int crc_win_y_start_get(void *data, u64 *val) { struct drm_crtc *crtc = data; struct drm_device *drm_dev = crtc->dev; struct amdgpu_crtc *acrtc = to_amdgpu_crtc(crtc); spin_lock_irq(&drm_dev->event_lock); *val = acrtc->dm_irq_params.window_param.y_start; spin_unlock_irq(&drm_dev->event_lock); return 0; } DEFINE_DEBUGFS_ATTRIBUTE(crc_win_y_start_fops, crc_win_y_start_get, crc_win_y_start_set, "%llu\n"); /* * Set crc window coordinate x end */ static int crc_win_x_end_set(void *data, u64 val) { struct drm_crtc *crtc = data; struct drm_device *drm_dev = crtc->dev; struct amdgpu_crtc *acrtc = to_amdgpu_crtc(crtc); spin_lock_irq(&drm_dev->event_lock); acrtc->dm_irq_params.window_param.x_end = (uint16_t) val; acrtc->dm_irq_params.window_param.update_win = false; spin_unlock_irq(&drm_dev->event_lock); return 0; } /* * Get crc window coordinate x end */ static int crc_win_x_end_get(void *data, u64 *val) { struct drm_crtc *crtc = data; struct drm_device *drm_dev = crtc->dev; struct amdgpu_crtc *acrtc = to_amdgpu_crtc(crtc); spin_lock_irq(&drm_dev->event_lock); *val = acrtc->dm_irq_params.window_param.x_end; spin_unlock_irq(&drm_dev->event_lock); return 0; } DEFINE_DEBUGFS_ATTRIBUTE(crc_win_x_end_fops, crc_win_x_end_get, crc_win_x_end_set, "%llu\n"); /* * Set crc window coordinate y end */ static int crc_win_y_end_set(void *data, u64 val) { struct drm_crtc *crtc = data; struct drm_device *drm_dev = crtc->dev; struct amdgpu_crtc *acrtc = to_amdgpu_crtc(crtc); spin_lock_irq(&drm_dev->event_lock); acrtc->dm_irq_params.window_param.y_end = (uint16_t) val; acrtc->dm_irq_params.window_param.update_win = false; spin_unlock_irq(&drm_dev->event_lock); return 0; } /* * Get crc window coordinate y end */ static int crc_win_y_end_get(void *data, u64 *val) { struct drm_crtc *crtc = data; struct drm_device *drm_dev = crtc->dev; struct amdgpu_crtc *acrtc = to_amdgpu_crtc(crtc); spin_lock_irq(&drm_dev->event_lock); *val = acrtc->dm_irq_params.window_param.y_end; spin_unlock_irq(&drm_dev->event_lock); return 0; } DEFINE_DEBUGFS_ATTRIBUTE(crc_win_y_end_fops, crc_win_y_end_get, crc_win_y_end_set, "%llu\n"); /* * Trigger to commit crc window */ static int crc_win_update_set(void *data, u64 val) { struct drm_crtc *crtc = data; struct amdgpu_crtc *acrtc; struct amdgpu_device *adev = drm_to_adev(crtc->dev); if (val) { acrtc = to_amdgpu_crtc(crtc); mutex_lock(&adev->dm.dc_lock); /* PSR may write to OTG CRC window control register, * so close it before starting secure_display. */ amdgpu_dm_psr_disable(acrtc->dm_irq_params.stream); spin_lock_irq(&adev_to_drm(adev)->event_lock); acrtc->dm_irq_params.window_param.activated = true; acrtc->dm_irq_params.window_param.update_win = true; acrtc->dm_irq_params.window_param.skip_frame_cnt = 0; spin_unlock_irq(&adev_to_drm(adev)->event_lock); mutex_unlock(&adev->dm.dc_lock); } return 0; } /* * Get crc window update flag */ static int crc_win_update_get(void *data, u64 *val) { *val = 0; return 0; } DEFINE_DEBUGFS_ATTRIBUTE(crc_win_update_fops, crc_win_update_get, crc_win_update_set, "%llu\n"); #endif void crtc_debugfs_init(struct drm_crtc *crtc) { #ifdef CONFIG_DRM_AMD_SECURE_DISPLAY struct dentry *dir = debugfs_lookup("crc", crtc->debugfs_entry); if (!dir) return; debugfs_create_file_unsafe("crc_win_x_start", 0644, dir, crtc, &crc_win_x_start_fops); debugfs_create_file_unsafe("crc_win_y_start", 0644, dir, crtc, &crc_win_y_start_fops); debugfs_create_file_unsafe("crc_win_x_end", 0644, dir, crtc, &crc_win_x_end_fops); debugfs_create_file_unsafe("crc_win_y_end", 0644, dir, crtc, &crc_win_y_end_fops); debugfs_create_file_unsafe("crc_win_update", 0644, dir, crtc, &crc_win_update_fops); dput(dir); #endif debugfs_create_file("amdgpu_current_bpc", 0644, crtc->debugfs_entry, crtc, &amdgpu_current_bpc_fops); debugfs_create_file("amdgpu_current_colorspace", 0644, crtc->debugfs_entry, crtc, &amdgpu_current_colorspace_fops); } /* * Writes DTN log state to the user supplied buffer. * Example usage: cat /sys/kernel/debug/dri/0/amdgpu_dm_dtn_log */ static ssize_t dtn_log_read( struct file *f, char __user *buf, size_t size, loff_t *pos) { struct amdgpu_device *adev = file_inode(f)->i_private; struct dc *dc = adev->dm.dc; struct dc_log_buffer_ctx log_ctx = { 0 }; ssize_t result = 0; if (!buf || !size) return -EINVAL; if (!dc->hwss.log_hw_state) return 0; dc->hwss.log_hw_state(dc, &log_ctx); if (*pos < log_ctx.pos) { size_t to_copy = log_ctx.pos - *pos; to_copy = min(to_copy, size); if (!copy_to_user(buf, log_ctx.buf + *pos, to_copy)) { *pos += to_copy; result = to_copy; } } kfree(log_ctx.buf); return result; } /* * Writes DTN log state to dmesg when triggered via a write. * Example usage: echo 1 > /sys/kernel/debug/dri/0/amdgpu_dm_dtn_log */ static ssize_t dtn_log_write( struct file *f, const char __user *buf, size_t size, loff_t *pos) { struct amdgpu_device *adev = file_inode(f)->i_private; struct dc *dc = adev->dm.dc; /* Write triggers log output via dmesg. */ if (size == 0) return 0; if (dc->hwss.log_hw_state) dc->hwss.log_hw_state(dc, NULL); return size; } static int mst_topo_show(struct seq_file *m, void *unused) { struct amdgpu_device *adev = (struct amdgpu_device *)m->private; struct drm_device *dev = adev_to_drm(adev); struct drm_connector *connector; struct drm_connector_list_iter conn_iter; struct amdgpu_dm_connector *aconnector; drm_connector_list_iter_begin(dev, &conn_iter); drm_for_each_connector_iter(connector, &conn_iter) { if (connector->connector_type != DRM_MODE_CONNECTOR_DisplayPort) continue; aconnector = to_amdgpu_dm_connector(connector); /* Ensure we're only dumping the topology of a root mst node */ if (!aconnector->mst_mgr.mst_state) continue; seq_printf(m, "\nMST topology for connector %d\n", aconnector->connector_id); drm_dp_mst_dump_topology(m, &aconnector->mst_mgr); } drm_connector_list_iter_end(&conn_iter); return 0; } /* * Sets trigger hpd for MST topologies. * All connected connectors will be rediscovered and re started as needed if val of 1 is sent. * All topologies will be disconnected if val of 0 is set . * Usage to enable topologies: echo 1 > /sys/kernel/debug/dri/0/amdgpu_dm_trigger_hpd_mst * Usage to disable topologies: echo 0 > /sys/kernel/debug/dri/0/amdgpu_dm_trigger_hpd_mst */ static int trigger_hpd_mst_set(void *data, u64 val) { struct amdgpu_device *adev = data; struct drm_device *dev = adev_to_drm(adev); struct drm_connector_list_iter iter; struct amdgpu_dm_connector *aconnector; struct drm_connector *connector; struct dc_link *link = NULL; int ret; if (val == 1) { drm_connector_list_iter_begin(dev, &iter); drm_for_each_connector_iter(connector, &iter) { aconnector = to_amdgpu_dm_connector(connector); if (aconnector->dc_link->type == dc_connection_mst_branch && aconnector->mst_mgr.aux) { mutex_lock(&adev->dm.dc_lock); dc_link_detect(aconnector->dc_link, DETECT_REASON_HPD); mutex_unlock(&adev->dm.dc_lock); ret = drm_dp_mst_topology_mgr_set_mst(&aconnector->mst_mgr, true); if (ret < 0) DRM_ERROR("DM_MST: Failed to set the device into MST mode!"); } } } else if (val == 0) { drm_connector_list_iter_begin(dev, &iter); drm_for_each_connector_iter(connector, &iter) { aconnector = to_amdgpu_dm_connector(connector); if (!aconnector->dc_link) continue; if (!aconnector->mst_root) continue; link = aconnector->dc_link; dc_link_dp_receiver_power_ctrl(link, false); drm_dp_mst_topology_mgr_set_mst(&aconnector->mst_root->mst_mgr, false); link->mst_stream_alloc_table.stream_count = 0; memset(link->mst_stream_alloc_table.stream_allocations, 0, sizeof(link->mst_stream_alloc_table.stream_allocations)); } } else { return 0; } drm_kms_helper_hotplug_event(dev); return 0; } /* * The interface doesn't need get function, so it will return the * value of zero * Usage: cat /sys/kernel/debug/dri/0/amdgpu_dm_trigger_hpd_mst */ static int trigger_hpd_mst_get(void *data, u64 *val) { *val = 0; return 0; } DEFINE_DEBUGFS_ATTRIBUTE(trigger_hpd_mst_ops, trigger_hpd_mst_get, trigger_hpd_mst_set, "%llu\n"); /* * Sets the force_timing_sync debug option from the given string. * All connected displays will be force synchronized immediately. * Usage: echo 1 > /sys/kernel/debug/dri/0/amdgpu_dm_force_timing_sync */ static int force_timing_sync_set(void *data, u64 val) { struct amdgpu_device *adev = data; adev->dm.force_timing_sync = (bool)val; amdgpu_dm_trigger_timing_sync(adev_to_drm(adev)); return 0; } /* * Gets the force_timing_sync debug option value into the given buffer. * Usage: cat /sys/kernel/debug/dri/0/amdgpu_dm_force_timing_sync */ static int force_timing_sync_get(void *data, u64 *val) { struct amdgpu_device *adev = data; *val = adev->dm.force_timing_sync; return 0; } DEFINE_DEBUGFS_ATTRIBUTE(force_timing_sync_ops, force_timing_sync_get, force_timing_sync_set, "%llu\n"); /* * Disables all HPD and HPD RX interrupt handling in the * driver when set to 1. Default is 0. */ static int disable_hpd_set(void *data, u64 val) { struct amdgpu_device *adev = data; adev->dm.disable_hpd_irq = (bool)val; return 0; } /* * Returns 1 if HPD and HPRX interrupt handling is disabled, * 0 otherwise. */ static int disable_hpd_get(void *data, u64 *val) { struct amdgpu_device *adev = data; *val = adev->dm.disable_hpd_irq; return 0; } DEFINE_DEBUGFS_ATTRIBUTE(disable_hpd_ops, disable_hpd_get, disable_hpd_set, "%llu\n"); /* * Prints hardware capabilities. These are used for IGT testing. */ static int capabilities_show(struct seq_file *m, void *unused) { struct amdgpu_device *adev = (struct amdgpu_device *)m->private; struct dc *dc = adev->dm.dc; bool mall_supported = dc->caps.mall_size_total; bool subvp_supported = dc->caps.subvp_fw_processing_delay_us; unsigned int mall_in_use = false; unsigned int subvp_in_use = false; struct hubbub *hubbub = dc->res_pool->hubbub; if (hubbub->funcs->get_mall_en) hubbub->funcs->get_mall_en(hubbub, &mall_in_use); if (dc->cap_funcs.get_subvp_en) subvp_in_use = dc->cap_funcs.get_subvp_en(dc, dc->current_state); seq_printf(m, "mall supported: %s, enabled: %s\n", mall_supported ? "yes" : "no", mall_in_use ? "yes" : "no"); seq_printf(m, "sub-viewport supported: %s, enabled: %s\n", subvp_supported ? "yes" : "no", subvp_in_use ? "yes" : "no"); return 0; } DEFINE_SHOW_ATTRIBUTE(capabilities); /* * Temporary w/a to force sst sequence in M42D DP2 mst receiver * Example usage: echo 1 > /sys/kernel/debug/dri/0/amdgpu_dm_dp_set_mst_en_for_sst */ static int dp_force_sst_set(void *data, u64 val) { struct amdgpu_device *adev = data; adev->dm.dc->debug.set_mst_en_for_sst = val; return 0; } static int dp_force_sst_get(void *data, u64 *val) { struct amdgpu_device *adev = data; *val = adev->dm.dc->debug.set_mst_en_for_sst; return 0; } DEFINE_DEBUGFS_ATTRIBUTE(dp_set_mst_en_for_sst_ops, dp_force_sst_get, dp_force_sst_set, "%llu\n"); /* * Force DP2 sequence without VESA certified cable. * Example usage: echo 1 > /sys/kernel/debug/dri/0/amdgpu_dm_dp_ignore_cable_id */ static int dp_ignore_cable_id_set(void *data, u64 val) { struct amdgpu_device *adev = data; adev->dm.dc->debug.ignore_cable_id = val; return 0; } static int dp_ignore_cable_id_get(void *data, u64 *val) { struct amdgpu_device *adev = data; *val = adev->dm.dc->debug.ignore_cable_id; return 0; } DEFINE_DEBUGFS_ATTRIBUTE(dp_ignore_cable_id_ops, dp_ignore_cable_id_get, dp_ignore_cable_id_set, "%llu\n"); /* * Sets the DC visual confirm debug option from the given string. * Example usage: echo 1 > /sys/kernel/debug/dri/0/amdgpu_visual_confirm */ static int visual_confirm_set(void *data, u64 val) { struct amdgpu_device *adev = data; adev->dm.dc->debug.visual_confirm = (enum visual_confirm)val; return 0; } /* * Reads the DC visual confirm debug option value into the given buffer. * Example usage: cat /sys/kernel/debug/dri/0/amdgpu_dm_visual_confirm */ static int visual_confirm_get(void *data, u64 *val) { struct amdgpu_device *adev = data; *val = adev->dm.dc->debug.visual_confirm; return 0; } DEFINE_SHOW_ATTRIBUTE(mst_topo); DEFINE_DEBUGFS_ATTRIBUTE(visual_confirm_fops, visual_confirm_get, visual_confirm_set, "%llu\n"); /* * Sets the DC skip_detection_link_training debug option from the given string. * Example usage: echo 1 > /sys/kernel/debug/dri/0/amdgpu_skip_detection_link_training */ static int skip_detection_link_training_set(void *data, u64 val) { struct amdgpu_device *adev = data; if (val == 0) adev->dm.dc->debug.skip_detection_link_training = false; else adev->dm.dc->debug.skip_detection_link_training = true; return 0; } /* * Reads the DC skip_detection_link_training debug option value into the given buffer. * Example usage: cat /sys/kernel/debug/dri/0/amdgpu_dm_skip_detection_link_training */ static int skip_detection_link_training_get(void *data, u64 *val) { struct amdgpu_device *adev = data; *val = adev->dm.dc->debug.skip_detection_link_training; return 0; } DEFINE_DEBUGFS_ATTRIBUTE(skip_detection_link_training_fops, skip_detection_link_training_get, skip_detection_link_training_set, "%llu\n"); /* * Dumps the DCC_EN bit for each pipe. * Example usage: cat /sys/kernel/debug/dri/0/amdgpu_dm_dcc_en */ static ssize_t dcc_en_bits_read( struct file *f, char __user *buf, size_t size, loff_t *pos) { struct amdgpu_device *adev = file_inode(f)->i_private; struct dc *dc = adev->dm.dc; char *rd_buf = NULL; const uint32_t rd_buf_size = 32; uint32_t result = 0; int offset = 0; int num_pipes = dc->res_pool->pipe_count; int *dcc_en_bits; int i, r; dcc_en_bits = kcalloc(num_pipes, sizeof(int), GFP_KERNEL); if (!dcc_en_bits) return -ENOMEM; if (!dc->hwss.get_dcc_en_bits) { kfree(dcc_en_bits); return 0; } dc->hwss.get_dcc_en_bits(dc, dcc_en_bits); rd_buf = kcalloc(rd_buf_size, sizeof(char), GFP_KERNEL); if (!rd_buf) { kfree(dcc_en_bits); return -ENOMEM; } for (i = 0; i < num_pipes; i++) offset += snprintf(rd_buf + offset, rd_buf_size - offset, "%d ", dcc_en_bits[i]); rd_buf[strlen(rd_buf)] = '\n'; kfree(dcc_en_bits); while (size) { if (*pos >= rd_buf_size) break; r = put_user(*(rd_buf + result), buf); if (r) { kfree(rd_buf); return r; /* r = -EFAULT */ } buf += 1; size -= 1; *pos += 1; result += 1; } kfree(rd_buf); return result; } void dtn_debugfs_init(struct amdgpu_device *adev) { static const struct file_operations dtn_log_fops = { .owner = THIS_MODULE, .read = dtn_log_read, .write = dtn_log_write, .llseek = default_llseek }; static const struct file_operations dcc_en_bits_fops = { .owner = THIS_MODULE, .read = dcc_en_bits_read, .llseek = default_llseek }; struct drm_minor *minor = adev_to_drm(adev)->primary; struct dentry *root = minor->debugfs_root; debugfs_create_file("amdgpu_mst_topology", 0444, root, adev, &mst_topo_fops); debugfs_create_file("amdgpu_dm_capabilities", 0444, root, adev, &capabilities_fops); debugfs_create_file("amdgpu_dm_dtn_log", 0644, root, adev, &dtn_log_fops); debugfs_create_file("amdgpu_dm_dp_set_mst_en_for_sst", 0644, root, adev, &dp_set_mst_en_for_sst_ops); debugfs_create_file("amdgpu_dm_dp_ignore_cable_id", 0644, root, adev, &dp_ignore_cable_id_ops); debugfs_create_file_unsafe("amdgpu_dm_visual_confirm", 0644, root, adev, &visual_confirm_fops); debugfs_create_file_unsafe("amdgpu_dm_skip_detection_link_training", 0644, root, adev, &skip_detection_link_training_fops); debugfs_create_file_unsafe("amdgpu_dm_dmub_tracebuffer", 0644, root, adev, &dmub_tracebuffer_fops); debugfs_create_file_unsafe("amdgpu_dm_dmub_fw_state", 0644, root, adev, &dmub_fw_state_fops); debugfs_create_file_unsafe("amdgpu_dm_force_timing_sync", 0644, root, adev, &force_timing_sync_ops); debugfs_create_file_unsafe("amdgpu_dm_dmub_trace_mask", 0644, root, adev, &dmub_trace_mask_fops); debugfs_create_file_unsafe("amdgpu_dm_dmcub_trace_event_en", 0644, root, adev, &dmcub_trace_event_state_fops); debugfs_create_file_unsafe("amdgpu_dm_trigger_hpd_mst", 0644, root, adev, &trigger_hpd_mst_ops); debugfs_create_file_unsafe("amdgpu_dm_dcc_en", 0644, root, adev, &dcc_en_bits_fops); debugfs_create_file_unsafe("amdgpu_dm_disable_hpd", 0644, root, adev, &disable_hpd_ops); if (adev->dm.dc->caps.ips_support) debugfs_create_file_unsafe("amdgpu_dm_ips_status", 0644, root, adev, &ips_status_fops); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1