Release 4.7 drivers/base/power/opp/cpu.c
  
  
/*
 * Generic OPP helper interface for CPU device
 *
 * Copyright (C) 2009-2014 Texas Instruments Incorporated.
 *      Nishanth Menon
 *      Romit Dasgupta
 *      Kevin Hilman
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License version 2 as
 * published by the Free Software Foundation.
 */
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
#include <linux/cpu.h>
#include <linux/cpufreq.h>
#include <linux/err.h>
#include <linux/errno.h>
#include <linux/export.h>
#include <linux/slab.h>
#include "opp.h"
#ifdef CONFIG_CPU_FREQ
/**
 * dev_pm_opp_init_cpufreq_table() - create a cpufreq table for a device
 * @dev:        device for which we do this operation
 * @table:      Cpufreq table returned back to caller
 *
 * Generate a cpufreq table for a provided device- this assumes that the
 * opp table is already initialized and ready for usage.
 *
 * This function allocates required memory for the cpufreq table. It is
 * expected that the caller does the required maintenance such as freeing
 * the table as required.
 *
 * Returns -EINVAL for bad pointers, -ENODEV if the device is not found, -ENOMEM
 * if no memory available for the operation (table is not populated), returns 0
 * if successful and table is populated.
 *
 * WARNING: It is  important for the callers to ensure refreshing their copy of
 * the table if any of the mentioned functions have been invoked in the interim.
 *
 * Locking: The internal opp_table and opp structures are RCU protected.
 * Since we just use the regular accessor functions to access the internal data
 * structures, we use RCU read lock inside this function. As a result, users of
 * this function DONOT need to use explicit locks for invoking.
 */
int dev_pm_opp_init_cpufreq_table(struct device *dev,
				  struct cpufreq_frequency_table **table)
{
	struct dev_pm_opp *opp;
	struct cpufreq_frequency_table *freq_table = NULL;
	int i, max_opps, ret = 0;
	unsigned long rate;
	rcu_read_lock();
	max_opps = dev_pm_opp_get_opp_count(dev);
	if (max_opps <= 0) {
		ret = max_opps ? max_opps : -ENODATA;
		goto out;
	}
	freq_table = kcalloc((max_opps + 1), sizeof(*freq_table), GFP_ATOMIC);
	if (!freq_table) {
		ret = -ENOMEM;
		goto out;
	}
	for (i = 0, rate = 0; i < max_opps; i++, rate++) {
		/* find next rate */
		opp = dev_pm_opp_find_freq_ceil(dev, &rate);
		if (IS_ERR(opp)) {
			ret = PTR_ERR(opp);
			goto out;
		}
		freq_table[i].driver_data = i;
		freq_table[i].frequency = rate / 1000;
		/* Is Boost/turbo opp ? */
		if (dev_pm_opp_is_turbo(opp))
			freq_table[i].flags = CPUFREQ_BOOST_FREQ;
	}
	freq_table[i].driver_data = i;
	freq_table[i].frequency = CPUFREQ_TABLE_END;
	*table = &freq_table[0];
out:
	rcu_read_unlock();
	if (ret)
		kfree(freq_table);
	return ret;
}
Contributors
 | Person | Tokens | Prop | Commits | CommitProp | 
| nishanth menon | nishanth menon | 213 | 89.50% | 1 | 25.00% | 
| bartlomiej zolnierkiewicz | bartlomiej zolnierkiewicz | 17 | 7.14% | 1 | 25.00% | 
| anand moon | anand moon | 6 | 2.52% | 1 | 25.00% | 
| stephen boyd | stephen boyd | 2 | 0.84% | 1 | 25.00% | 
 | Total | 238 | 100.00% | 4 | 100.00% | 
EXPORT_SYMBOL_GPL(dev_pm_opp_init_cpufreq_table);
/**
 * dev_pm_opp_free_cpufreq_table() - free the cpufreq table
 * @dev:        device for which we do this operation
 * @table:      table to free
 *
 * Free up the table allocated by dev_pm_opp_init_cpufreq_table
 */
void dev_pm_opp_free_cpufreq_table(struct device *dev,
				   struct cpufreq_frequency_table **table)
{
	if (!table)
		return;
	kfree(*table);
	*table = NULL;
}
Contributors
 | Person | Tokens | Prop | Commits | CommitProp | 
| nishanth menon | nishanth menon | 33 | 100.00% | 1 | 100.00% | 
 | Total | 33 | 100.00% | 1 | 100.00% | 
EXPORT_SYMBOL_GPL(dev_pm_opp_free_cpufreq_table);
#endif	/* CONFIG_CPU_FREQ */
void _dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask, bool of)
{
	struct device *cpu_dev;
	int cpu;
	WARN_ON(cpumask_empty(cpumask));
	for_each_cpu(cpu, cpumask) {
		cpu_dev = get_cpu_device(cpu);
		if (!cpu_dev) {
			pr_err("%s: failed to get cpu%d device\n", __func__,
			       cpu);
			continue;
		}
		if (of)
			dev_pm_opp_of_remove_table(cpu_dev);
		else
			dev_pm_opp_remove_table(cpu_dev);
	}
}
Contributors
 | Person | Tokens | Prop | Commits | CommitProp | 
| viresh kumar | viresh kumar | 57 | 74.03% | 2 | 50.00% | 
| sudeep holla | sudeep holla | 16 | 20.78% | 1 | 25.00% | 
| arnd bergmann | arnd bergmann | 4 | 5.19% | 1 | 25.00% | 
 | Total | 77 | 100.00% | 4 | 100.00% | 
/**
 * dev_pm_opp_cpumask_remove_table() - Removes OPP table for @cpumask
 * @cpumask:    cpumask for which OPP table needs to be removed
 *
 * This removes the OPP tables for CPUs present in the @cpumask.
 * This should be used to remove all the OPPs entries associated with
 * the cpus in @cpumask.
 *
 * Locking: The internal opp_table and opp structures are RCU protected.
 * Hence this function internally uses RCU updater strategy with mutex locks
 * to keep the integrity of the internal data structures. Callers should ensure
 * that this function is *NOT* called under RCU protection or in contexts where
 * mutex cannot be locked.
 */
void dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask)
{
	_dev_pm_opp_cpumask_remove_table(cpumask, false);
}
Contributors
 | Person | Tokens | Prop | Commits | CommitProp | 
| sudeep holla | sudeep holla | 17 | 94.44% | 1 | 50.00% | 
| viresh kumar | viresh kumar | 1 | 5.56% | 1 | 50.00% | 
 | Total | 18 | 100.00% | 2 | 100.00% | 
EXPORT_SYMBOL_GPL(dev_pm_opp_cpumask_remove_table);
/**
 * dev_pm_opp_set_sharing_cpus() - Mark OPP table as shared by few CPUs
 * @cpu_dev:    CPU device for which we do this operation
 * @cpumask:    cpumask of the CPUs which share the OPP table with @cpu_dev
 *
 * This marks OPP table of the @cpu_dev as shared by the CPUs present in
 * @cpumask.
 *
 * Returns -ENODEV if OPP table isn't already present.
 *
 * Locking: The internal opp_table and opp structures are RCU protected.
 * Hence this function internally uses RCU updater strategy with mutex locks
 * to keep the integrity of the internal data structures. Callers should ensure
 * that this function is *NOT* called under RCU protection or in contexts where
 * mutex cannot be locked.
 */
int dev_pm_opp_set_sharing_cpus(struct device *cpu_dev,
				const struct cpumask *cpumask)
{
	struct opp_device *opp_dev;
	struct opp_table *opp_table;
	struct device *dev;
	int cpu, ret = 0;
	mutex_lock(&opp_table_lock);
	opp_table = _find_opp_table(cpu_dev);
	if (IS_ERR(opp_table)) {
		ret = PTR_ERR(opp_table);
		goto unlock;
	}
	for_each_cpu(cpu, cpumask) {
		if (cpu == cpu_dev->id)
			continue;
		dev = get_cpu_device(cpu);
		if (!dev) {
			dev_err(cpu_dev, "%s: failed to get cpu%d device\n",
				__func__, cpu);
			continue;
		}
		opp_dev = _add_opp_dev(dev, opp_table);
		if (!opp_dev) {
			dev_err(dev, "%s: failed to add opp-dev for cpu%d device\n",
				__func__, cpu);
			continue;
		}
		/* Mark opp-table as multiple CPUs are sharing it now */
		opp_table->shared_opp = OPP_TABLE_ACCESS_SHARED;
	}
unlock:
	mutex_unlock(&opp_table_lock);
	return ret;
}
Contributors
 | Person | Tokens | Prop | Commits | CommitProp | 
| viresh kumar | viresh kumar | 156 | 98.11% | 4 | 80.00% | 
| arnd bergmann | arnd bergmann | 3 | 1.89% | 1 | 20.00% | 
 | Total | 159 | 100.00% | 5 | 100.00% | 
EXPORT_SYMBOL_GPL(dev_pm_opp_set_sharing_cpus);
/**
 * dev_pm_opp_get_sharing_cpus() - Get cpumask of CPUs sharing OPPs with @cpu_dev
 * @cpu_dev:    CPU device for which we do this operation
 * @cpumask:    cpumask to update with information of sharing CPUs
 *
 * This updates the @cpumask with CPUs that are sharing OPPs with @cpu_dev.
 *
 * Returns -ENODEV if OPP table isn't already present and -EINVAL if the OPP
 * table's status is access-unknown.
 *
 * Locking: The internal opp_table and opp structures are RCU protected.
 * Hence this function internally uses RCU updater strategy with mutex locks
 * to keep the integrity of the internal data structures. Callers should ensure
 * that this function is *NOT* called under RCU protection or in contexts where
 * mutex cannot be locked.
 */
int dev_pm_opp_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask)
{
	struct opp_device *opp_dev;
	struct opp_table *opp_table;
	int ret = 0;
	mutex_lock(&opp_table_lock);
	opp_table = _find_opp_table(cpu_dev);
	if (IS_ERR(opp_table)) {
		ret = PTR_ERR(opp_table);
		goto unlock;
	}
	if (opp_table->shared_opp == OPP_TABLE_ACCESS_UNKNOWN) {
		ret = -EINVAL;
		goto unlock;
	}
	cpumask_clear(cpumask);
	if (opp_table->shared_opp == OPP_TABLE_ACCESS_SHARED) {
		list_for_each_entry(opp_dev, &opp_table->dev_list, node)
			cpumask_set_cpu(opp_dev->dev->id, cpumask);
	} else {
		cpumask_set_cpu(cpu_dev->id, cpumask);
	}
unlock:
	mutex_unlock(&opp_table_lock);
	return ret;
}
Contributors
 | Person | Tokens | Prop | Commits | CommitProp | 
| viresh kumar | viresh kumar | 134 | 97.81% | 2 | 66.67% | 
| arnd bergmann | arnd bergmann | 3 | 2.19% | 1 | 33.33% | 
 | Total | 137 | 100.00% | 3 | 100.00% | 
EXPORT_SYMBOL_GPL(dev_pm_opp_get_sharing_cpus);
Overall Contributors
 | Person | Tokens | Prop | Commits | CommitProp | 
| viresh kumar | viresh kumar | 379 | 52.13% | 10 | 62.50% | 
| nishanth menon | nishanth menon | 274 | 37.69% | 1 | 6.25% | 
| sudeep holla | sudeep holla | 39 | 5.36% | 1 | 6.25% | 
| bartlomiej zolnierkiewicz | bartlomiej zolnierkiewicz | 17 | 2.34% | 1 | 6.25% | 
| arnd bergmann | arnd bergmann | 10 | 1.38% | 1 | 6.25% | 
| anand moon | anand moon | 6 | 0.83% | 1 | 6.25% | 
| stephen boyd | stephen boyd | 2 | 0.28% | 1 | 6.25% | 
 | Total | 727 | 100.00% | 16 | 100.00% | 
  
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.