Contributors: 20
Author Tokens Token Proportion Commits Commit Proportion
Mark Brown 80 18.26% 1 2.63%
Juri Lelli 77 17.58% 4 10.53%
Dietmar Eggemann 70 15.98% 3 7.89%
Viresh Kumar 53 12.10% 4 10.53%
Jonathan Cameron 30 6.85% 1 2.63%
Ionela Voinescu 27 6.16% 3 7.89%
Yicong Yang 26 5.94% 2 5.26%
Atish Patra 15 3.42% 2 5.26%
Vincent Guittot 10 2.28% 3 7.89%
Yunhui Cui 10 2.28% 1 2.63%
Sudeep Holla 10 2.28% 2 5.26%
Morten Rasmussen 6 1.37% 1 2.63%
Valentin Schneider 6 1.37% 3 7.89%
Thara Gopinath 5 1.14% 1 2.63%
Lukasz Luba 4 0.91% 1 2.63%
Jeremy Linton 3 0.68% 2 5.26%
David Dai 2 0.46% 1 2.63%
Ben Dooks 2 0.46% 1 2.63%
Greg Kroah-Hartman 1 0.23% 1 2.63%
Bartosz Golaszewski 1 0.23% 1 2.63%
Total 438 38


/* SPDX-License-Identifier: GPL-2.0 */
/*
 * include/linux/arch_topology.h - arch specific cpu topology information
 */
#ifndef _LINUX_ARCH_TOPOLOGY_H_
#define _LINUX_ARCH_TOPOLOGY_H_

#include <linux/types.h>
#include <linux/percpu.h>

void topology_normalize_cpu_scale(void);
int topology_update_cpu_topology(void);

struct device_node;
bool topology_parse_cpu_capacity(struct device_node *cpu_node, int cpu);


DECLARE_PER_CPU(unsigned long, capacity_freq_ref);

static inline unsigned long topology_get_freq_ref(int cpu)
{
	return per_cpu(capacity_freq_ref, cpu);
}

DECLARE_PER_CPU(unsigned long, arch_freq_scale);

static inline unsigned long topology_get_freq_scale(int cpu)
{
	return per_cpu(arch_freq_scale, cpu);
}

void topology_set_freq_scale(const struct cpumask *cpus, unsigned long cur_freq,
			     unsigned long max_freq);
bool topology_scale_freq_invariant(void);

enum scale_freq_source {
	SCALE_FREQ_SOURCE_CPUFREQ = 0,
	SCALE_FREQ_SOURCE_ARCH,
	SCALE_FREQ_SOURCE_CPPC,
	SCALE_FREQ_SOURCE_VIRT,
};

struct scale_freq_data {
	enum scale_freq_source source;
	void (*set_freq_scale)(void);
};

void topology_scale_freq_tick(void);
void topology_set_scale_freq_source(struct scale_freq_data *data, const struct cpumask *cpus);
void topology_clear_scale_freq_source(enum scale_freq_source source, const struct cpumask *cpus);

DECLARE_PER_CPU(unsigned long, hw_pressure);

static inline unsigned long topology_get_hw_pressure(int cpu)
{
	return per_cpu(hw_pressure, cpu);
}

void topology_update_hw_pressure(const struct cpumask *cpus,
				      unsigned long capped_freq);

struct cpu_topology {
	int thread_id;
	int core_id;
	int cluster_id;
	int package_id;
	cpumask_t thread_sibling;
	cpumask_t core_sibling;
	cpumask_t cluster_sibling;
	cpumask_t llc_sibling;
};

#ifdef CONFIG_GENERIC_ARCH_TOPOLOGY
extern struct cpu_topology cpu_topology[NR_CPUS];

#define topology_physical_package_id(cpu)	(cpu_topology[cpu].package_id)
#define topology_cluster_id(cpu)	(cpu_topology[cpu].cluster_id)
#define topology_core_id(cpu)		(cpu_topology[cpu].core_id)
#define topology_core_cpumask(cpu)	(&cpu_topology[cpu].core_sibling)
#define topology_sibling_cpumask(cpu)	(&cpu_topology[cpu].thread_sibling)
#define topology_cluster_cpumask(cpu)	(&cpu_topology[cpu].cluster_sibling)
#define topology_llc_cpumask(cpu)	(&cpu_topology[cpu].llc_sibling)

#ifndef arch_cpu_is_threaded
#define arch_cpu_is_threaded()	(0)
#endif

void init_cpu_topology(void);
void store_cpu_topology(unsigned int cpuid);
const struct cpumask *cpu_coregroup_mask(int cpu);
const struct cpumask *cpu_clustergroup_mask(int cpu);
void update_siblings_masks(unsigned int cpu);
void remove_cpu_topology(unsigned int cpuid);
void reset_cpu_topology(void);
int parse_acpi_topology(void);
void freq_inv_set_max_ratio(int cpu, u64 max_rate);

/*
 * Architectures like ARM64 don't have reliable architectural way to get SMT
 * information and depend on the firmware (ACPI/OF) report. Non-SMT core won't
 * initialize thread_id so we can use this to detect the SMT implementation.
 */
static inline bool topology_core_has_smt(int cpu)
{
	return cpu_topology[cpu].thread_id != -1;
}

#else

static inline bool topology_core_has_smt(int cpu) { return false; }

#endif /* CONFIG_GENERIC_ARCH_TOPOLOGY */

#endif /* _LINUX_ARCH_TOPOLOGY_H_ */