Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Mika Westerberg | 939 | 60.35% | 24 | 60.00% |
Gil Fine | 340 | 21.85% | 7 | 17.50% |
Rajmohan Mani | 223 | 14.33% | 4 | 10.00% |
Andreas Noever | 53 | 3.41% | 4 | 10.00% |
Greg Kroah-Hartman | 1 | 0.06% | 1 | 2.50% |
Total | 1556 | 40 |
// SPDX-License-Identifier: GPL-2.0 /* * CLx support * * Copyright (C) 2020 - 2023, Intel Corporation * Authors: Gil Fine <gil.fine@intel.com> * Mika Westerberg <mika.westerberg@linux.intel.com> */ #include <linux/module.h> #include "tb.h" static bool clx_enabled = true; module_param_named(clx, clx_enabled, bool, 0444); MODULE_PARM_DESC(clx, "allow low power states on the high-speed lanes (default: true)"); static const char *clx_name(unsigned int clx) { switch (clx) { case TB_CL0S | TB_CL1 | TB_CL2: return "CL0s/CL1/CL2"; case TB_CL1 | TB_CL2: return "CL1/CL2"; case TB_CL0S | TB_CL2: return "CL0s/CL2"; case TB_CL0S | TB_CL1: return "CL0s/CL1"; case TB_CL0S: return "CL0s"; case 0: return "disabled"; default: return "unknown"; } } static int tb_port_pm_secondary_set(struct tb_port *port, bool secondary) { u32 phy; int ret; ret = tb_port_read(port, &phy, TB_CFG_PORT, port->cap_phy + LANE_ADP_CS_1, 1); if (ret) return ret; if (secondary) phy |= LANE_ADP_CS_1_PMS; else phy &= ~LANE_ADP_CS_1_PMS; return tb_port_write(port, &phy, TB_CFG_PORT, port->cap_phy + LANE_ADP_CS_1, 1); } static int tb_port_pm_secondary_enable(struct tb_port *port) { return tb_port_pm_secondary_set(port, true); } static int tb_port_pm_secondary_disable(struct tb_port *port) { return tb_port_pm_secondary_set(port, false); } /* Called for USB4 or Titan Ridge routers only */ static bool tb_port_clx_supported(struct tb_port *port, unsigned int clx) { u32 val, mask = 0; bool ret; /* Don't enable CLx in case of two single-lane links */ if (!port->bonded && port->dual_link_port) return false; /* Don't enable CLx in case of inter-domain link */ if (port->xdomain) return false; if (tb_switch_is_usb4(port->sw)) { if (!usb4_port_clx_supported(port)) return false; } else if (!tb_lc_is_clx_supported(port)) { return false; } if (clx & TB_CL0S) mask |= LANE_ADP_CS_0_CL0S_SUPPORT; if (clx & TB_CL1) mask |= LANE_ADP_CS_0_CL1_SUPPORT; if (clx & TB_CL2) mask |= LANE_ADP_CS_0_CL2_SUPPORT; ret = tb_port_read(port, &val, TB_CFG_PORT, port->cap_phy + LANE_ADP_CS_0, 1); if (ret) return false; return !!(val & mask); } static int tb_port_clx_set(struct tb_port *port, unsigned int clx, bool enable) { u32 phy, mask = 0; int ret; if (clx & TB_CL0S) mask |= LANE_ADP_CS_1_CL0S_ENABLE; if (clx & TB_CL1) mask |= LANE_ADP_CS_1_CL1_ENABLE; if (clx & TB_CL2) mask |= LANE_ADP_CS_1_CL2_ENABLE; if (!mask) return -EOPNOTSUPP; ret = tb_port_read(port, &phy, TB_CFG_PORT, port->cap_phy + LANE_ADP_CS_1, 1); if (ret) return ret; if (enable) phy |= mask; else phy &= ~mask; return tb_port_write(port, &phy, TB_CFG_PORT, port->cap_phy + LANE_ADP_CS_1, 1); } static int tb_port_clx_disable(struct tb_port *port, unsigned int clx) { return tb_port_clx_set(port, clx, false); } static int tb_port_clx_enable(struct tb_port *port, unsigned int clx) { return tb_port_clx_set(port, clx, true); } static int tb_port_clx(struct tb_port *port) { u32 val; int ret; if (!tb_port_clx_supported(port, TB_CL0S | TB_CL1 | TB_CL2)) return 0; ret = tb_port_read(port, &val, TB_CFG_PORT, port->cap_phy + LANE_ADP_CS_1, 1); if (ret) return ret; if (val & LANE_ADP_CS_1_CL0S_ENABLE) ret |= TB_CL0S; if (val & LANE_ADP_CS_1_CL1_ENABLE) ret |= TB_CL1; if (val & LANE_ADP_CS_1_CL2_ENABLE) ret |= TB_CL2; return ret; } /** * tb_port_clx_is_enabled() - Is given CL state enabled * @port: USB4 port to check * @clx: Mask of CL states to check * * Returns true if any of the given CL states is enabled for @port. */ bool tb_port_clx_is_enabled(struct tb_port *port, unsigned int clx) { return !!(tb_port_clx(port) & clx); } /** * tb_switch_clx_is_supported() - Is CLx supported on this type of router * @sw: The router to check CLx support for */ static bool tb_switch_clx_is_supported(const struct tb_switch *sw) { if (!clx_enabled) return false; if (sw->quirks & QUIRK_NO_CLX) return false; /* * CLx is not enabled and validated on Intel USB4 platforms * before Alder Lake. */ if (tb_switch_is_tiger_lake(sw)) return false; return tb_switch_is_usb4(sw) || tb_switch_is_titan_ridge(sw); } /** * tb_switch_clx_init() - Initialize router CL states * @sw: Router * * Can be called for any router. Initializes the current CL state by * reading it from the hardware. * * Returns %0 in case of success and negative errno in case of failure. */ int tb_switch_clx_init(struct tb_switch *sw) { struct tb_port *up, *down; unsigned int clx, tmp; if (tb_switch_is_icm(sw)) return 0; if (!tb_route(sw)) return 0; if (!tb_switch_clx_is_supported(sw)) return 0; up = tb_upstream_port(sw); down = tb_switch_downstream_port(sw); clx = tb_port_clx(up); tmp = tb_port_clx(down); if (clx != tmp) tb_sw_warn(sw, "CLx: inconsistent configuration %#x != %#x\n", clx, tmp); tb_sw_dbg(sw, "CLx: current mode: %s\n", clx_name(clx)); sw->clx = clx; return 0; } static int tb_switch_pm_secondary_resolve(struct tb_switch *sw) { struct tb_port *up, *down; int ret; if (!tb_route(sw)) return 0; up = tb_upstream_port(sw); down = tb_switch_downstream_port(sw); ret = tb_port_pm_secondary_enable(up); if (ret) return ret; return tb_port_pm_secondary_disable(down); } static int tb_switch_mask_clx_objections(struct tb_switch *sw) { int up_port = sw->config.upstream_port_number; u32 offset, val[2], mask_obj, unmask_obj; int ret, i; /* Only Titan Ridge of pre-USB4 devices support CLx states */ if (!tb_switch_is_titan_ridge(sw)) return 0; if (!tb_route(sw)) return 0; /* * In Titan Ridge there are only 2 dual-lane Thunderbolt ports: * Port A consists of lane adapters 1,2 and * Port B consists of lane adapters 3,4 * If upstream port is A, (lanes are 1,2), we mask objections from * port B (lanes 3,4) and unmask objections from Port A and vice-versa. */ if (up_port == 1) { mask_obj = TB_LOW_PWR_C0_PORT_B_MASK; unmask_obj = TB_LOW_PWR_C1_PORT_A_MASK; offset = TB_LOW_PWR_C1_CL1; } else { mask_obj = TB_LOW_PWR_C1_PORT_A_MASK; unmask_obj = TB_LOW_PWR_C0_PORT_B_MASK; offset = TB_LOW_PWR_C3_CL1; } ret = tb_sw_read(sw, &val, TB_CFG_SWITCH, sw->cap_lp + offset, ARRAY_SIZE(val)); if (ret) return ret; for (i = 0; i < ARRAY_SIZE(val); i++) { val[i] |= mask_obj; val[i] &= ~unmask_obj; } return tb_sw_write(sw, &val, TB_CFG_SWITCH, sw->cap_lp + offset, ARRAY_SIZE(val)); } static bool validate_mask(unsigned int clx) { /* Previous states need to be enabled */ if (clx & TB_CL1) return (clx & TB_CL0S) == TB_CL0S; return true; } /** * tb_switch_clx_enable() - Enable CLx on upstream port of specified router * @sw: Router to enable CLx for * @clx: The CLx state to enable * * CLx is enabled only if both sides of the link support CLx, and if both sides * of the link are not configured as two single lane links and only if the link * is not inter-domain link. The complete set of conditions is described in CM * Guide 1.0 section 8.1. * * Returns %0 on success or an error code on failure. */ int tb_switch_clx_enable(struct tb_switch *sw, unsigned int clx) { bool up_clx_support, down_clx_support; struct tb_switch *parent_sw; struct tb_port *up, *down; int ret; if (!clx || sw->clx == clx) return 0; if (!validate_mask(clx)) return -EINVAL; parent_sw = tb_switch_parent(sw); if (!parent_sw) return 0; if (!tb_switch_clx_is_supported(parent_sw) || !tb_switch_clx_is_supported(sw)) return 0; /* Only support CL2 for v2 routers */ if ((clx & TB_CL2) && (usb4_switch_version(parent_sw) < 2 || usb4_switch_version(sw) < 2)) return -EOPNOTSUPP; ret = tb_switch_pm_secondary_resolve(sw); if (ret) return ret; up = tb_upstream_port(sw); down = tb_switch_downstream_port(sw); up_clx_support = tb_port_clx_supported(up, clx); down_clx_support = tb_port_clx_supported(down, clx); tb_port_dbg(up, "CLx: %s %ssupported\n", clx_name(clx), up_clx_support ? "" : "not "); tb_port_dbg(down, "CLx: %s %ssupported\n", clx_name(clx), down_clx_support ? "" : "not "); if (!up_clx_support || !down_clx_support) return -EOPNOTSUPP; ret = tb_port_clx_enable(up, clx); if (ret) return ret; ret = tb_port_clx_enable(down, clx); if (ret) { tb_port_clx_disable(up, clx); return ret; } ret = tb_switch_mask_clx_objections(sw); if (ret) { tb_port_clx_disable(up, clx); tb_port_clx_disable(down, clx); return ret; } sw->clx |= clx; tb_sw_dbg(sw, "CLx: %s enabled\n", clx_name(clx)); return 0; } /** * tb_switch_clx_disable() - Disable CLx on upstream port of specified router * @sw: Router to disable CLx for * * Disables all CL states of the given router. Can be called on any * router and if the states were not enabled already does nothing. * * Returns the CL states that were disabled or negative errno in case of * failure. */ int tb_switch_clx_disable(struct tb_switch *sw) { unsigned int clx = sw->clx; struct tb_port *up, *down; int ret; if (!tb_switch_clx_is_supported(sw)) return 0; if (!clx) return 0; if (sw->is_unplugged) return clx; up = tb_upstream_port(sw); down = tb_switch_downstream_port(sw); ret = tb_port_clx_disable(up, clx); if (ret) return ret; ret = tb_port_clx_disable(down, clx); if (ret) return ret; sw->clx = 0; tb_sw_dbg(sw, "CLx: %s disabled\n", clx_name(clx)); return clx; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1