* Author:
* Colin Cross <ccross@google.com>
*
- * Copyright (C) 2010-2013 NVIDIA CORPORATION. All rights reserved.
+ * Copyright (C) 2010-2014 NVIDIA CORPORATION. All rights reserved.
*
* This software is licensed under the terms of the GNU General Public
* License version 2, as published by the Free Software Foundation, and
#include <linux/reboot.h>
#include <linux/clk/tegra.h>
#include <linux/tegra-soc.h>
+#include <linux/tegra-fuse.h>
#include "board.h"
#include "clock.h"
#include "dvfs.h"
+#include "tegra_cl_dvfs.h"
#define DVFS_RAIL_STATS_BIN 12500
struct dvfs_rail *tegra_cpu_rail;
struct dvfs_rail *tegra_core_rail;
struct dvfs_rail *tegra_gpu_rail;
+int tegra_override_dfll_range;
static LIST_HEAD(dvfs_rail_list);
static DEFINE_MUTEX(dvfs_lock);
rel = &rels[i];
list_add_tail(&rel->from_node, &rel->to->relationships_from);
list_add_tail(&rel->to_node, &rel->from->relationships_to);
+
+ /* Overriding dependent rail below nominal may not be safe */
+ rel->to->min_override_millivolts = rel->to->nominal_millivolts;
}
mutex_unlock(&dvfs_lock);
/* Limit override range to maximum floor */
if (rail->therm_mv_floors)
rail->min_override_millivolts = rail->therm_mv_floors[0];
+
+ /* Only GPU thermal dvfs is supported */
+ if (rail->vts_cdev && (rail != tegra_gpu_rail)) {
+ rail->vts_cdev = NULL;
+ WARN(1, "%s: thermal dvfs is not supported\n", rail->reg_id);
+ }
+
+ /* Thermal clock switch is only supported for CPU */
+ if (rail->clk_switch_cdev && (rail != tegra_cpu_rail)) {
+ rail->clk_switch_cdev = NULL;
+ WARN(1, "%s: thermal clock switch is not supported\n",
+ rail->reg_id);
+ }
+
+ if (!rail->simon_vmin_offsets != !rail->simon_vmin_offs_num) {
+ rail->simon_vmin_offs_num = 0;
+ rail->simon_vmin_offsets = NULL;
+ WARN(1, "%s: not matching simon offsets/num\n", rail->reg_id);
+ }
}
int tegra_dvfs_init_rails(struct dvfs_rail *rails[], int n)
list_add_tail(&rails[i]->node, &dvfs_rail_list);
- dvfs_validate_cdevs(rails[i]);
-
if (!strcmp("vdd_cpu", rails[i]->reg_id))
tegra_cpu_rail = rails[i];
else if (!strcmp("vdd_gpu", rails[i]->reg_id))
tegra_gpu_rail = rails[i];
else if (!strcmp("vdd_core", rails[i]->reg_id))
tegra_core_rail = rails[i];
+
+ dvfs_validate_cdevs(rails[i]);
}
mutex_unlock(&dvfs_lock);
rail->reg_max_millivolts * 1000);
rail->updating = false;
+ pr_debug("%s: request_mV [%d, %d] from %s regulator (%d)\n", __func__,
+ millivolts, rail->reg_max_millivolts, rail->reg_id, ret);
+
return ret;
}
offset = -step;
}
- /*
- * DFLL adjusts rail voltage automatically, but not exactly to the
- * expected level - update stats, anyway.
- */
- if (rail->dfll_mode) {
- rail->millivolts = rail->new_millivolts = millivolts;
- dvfs_rail_stats_update(rail, millivolts, ktime_get());
- return 0;
- }
-
- if (rail->disabled)
+ /* Voltage change is always happening in DFLL mode */
+ if (rail->disabled && !rail->dfll_mode)
return 0;
rail->resolving_to = true;
jmp_to_zero = rail->jmp_to_zero &&
((millivolts == 0) || (rail->millivolts == 0));
- steps = jmp_to_zero ? 1 :
- DIV_ROUND_UP(abs(millivolts - rail->millivolts), step);
+
+ if (jmp_to_zero || rail->dfll_mode)
+ steps = 1;
+ else
+ steps = DIV_ROUND_UP(abs(millivolts - rail->millivolts), step);
for (i = 0; i < steps; i++) {
- if (!jmp_to_zero &&
- (abs(millivolts - rail->millivolts) > step))
+ if (steps > (i + 1))
rail->new_millivolts = rail->millivolts + offset;
else
rail->new_millivolts = millivolts;
goto out;
}
- ret = dvfs_rail_set_voltage_reg(rail, rail->new_millivolts);
- if (ret) {
- pr_err("Failed to set dvfs regulator %s\n", rail->reg_id);
- goto out;
+ /*
+ * DFLL adjusts voltage automatically - don't touch regulator,
+ * but update stats, anyway.
+ */
+ if (!rail->dfll_mode) {
+ ret = dvfs_rail_set_voltage_reg(rail,
+ rail->new_millivolts);
+ if (ret) {
+ pr_err("Failed to set dvfs regulator %s\n",
+ rail->reg_id);
+ goto out;
+ }
}
rail->millivolts = rail->new_millivolts;
static inline int dvfs_rail_apply_limits(struct dvfs_rail *rail, int millivolts)
{
int min_mv = rail->min_millivolts;
-
- if (rail->therm_mv_floors) {
- int i = rail->therm_floor_idx;
- if (i < rail->therm_mv_floors_num)
- min_mv = rail->therm_mv_floors[i];
- }
+ min_mv = max(min_mv, tegra_dvfs_rail_get_thermal_floor(rail));
if (rail->override_millivolts) {
millivolts = rail->override_millivolts;
/* Apply offset and min/max limits if any clock is requesting voltage */
if (millivolts)
millivolts = dvfs_rail_apply_limits(rail, millivolts);
+ /* Keep current voltage if regulator is to be disabled via explicitly */
+ else if (rail->in_band_pm)
+ return 0;
+ /* Keep current voltage if regulator must not be disabled at run time */
+ else if (!rail->jmp_to_zero) {
+ WARN(1, "%s cannot be turned off by dvfs\n", rail->reg_id);
+ return 0;
+ }
+ /* else: fall thru if regulator is turned off by side band signaling */
/* retry update if limited by from-relationship to account for
circular dependencies */
rail->reg_id);
return PTR_ERR(reg);
}
+ pr_info("tegra_dvfs: %s rail is fixed in pll mode\n",
+ rail->reg_id);
}
rail->reg = reg;
}
rail->reg_id, rail->millivolts, rail->boot_millivolts);
rail->boot_millivolts = rail->millivolts;
}
+
+ pr_info("tegra_dvfs: %s connected to regulator\n", rail->reg_id);
return 0;
}
static inline unsigned long *dvfs_get_freqs(struct dvfs *d)
{
- return d->alt_freqs ? : &d->freqs[0];
+ return d->alt_freqs && d->use_alt_freqs ? d->alt_freqs : &d->freqs[0];
}
static inline const int *dvfs_get_millivolts(struct dvfs *d, unsigned long rate)
if (tegra_dvfs_is_dfll_scale(d, rate))
return d->dfll_millivolts;
- return d->millivolts;
+ return tegra_dvfs_get_millivolts_pll(d);
}
static int
if (rate == 0) {
d->cur_millivolts = 0;
- } else {
+ /*
+ * For single clock GPU rail keep DVFS rate unchanged when clock
+ * is disabled. Rail is turned off explicitly, in any case, but
+ * with non-zero rate voltage level at regulator is updated when
+ * temperature is changes while rail is off.
+ */
+ if (d->dvfs_rail == tegra_gpu_rail)
+ rate = d->cur_rate;
+ }
+
+ if (rate != 0) {
while (i < d->num_freqs && rate > freqs[i])
i++;
- if ((d->max_millivolts) &&
- (millivolts[i] > d->max_millivolts)) {
- pr_warn("tegra_dvfs: voltage %d too high for dvfs on"
- " %s\n", millivolts[i], d->clk_name);
+ mv = millivolts[i];
+
+ if ((d->max_millivolts) && (mv > d->max_millivolts)) {
+ pr_warn("tegra_dvfs: voltage %d too high for dvfs on %s\n",
+ mv, d->clk_name);
return -EINVAL;
}
- mv = millivolts[i];
detach_mv = tegra_dvfs_rail_get_boot_level(d->dvfs_rail);
if (!d->dvfs_rail->reg && (mv > detach_mv)) {
pr_warn("%s: %s: voltage %d above boot limit %d\n",
__func__, d->clk_name, mv, detach_mv);
return -EINVAL;
}
- d->cur_millivolts = millivolts[i];
+
+ detach_mv = d->dvfs_rail->override_millivolts;
+ if (detach_mv && (mv > detach_mv)) {
+ pr_warn("%s: %s: voltage %d above override level %d\n",
+ __func__, d->clk_name, mv, detach_mv);
+ return -EINVAL;
+ }
+ d->cur_millivolts = mv;
}
d->cur_rate = rate;
return ret;
}
+/*
+ * Some clocks may have alternative frequency ladder that provides lower minimum
+ * voltage at the same rate (or complimentary: higher maximum rate at the same
+ * voltage). Interfaces below allows dvfs clients to install such ladder, and
+ * switch between primary and alternative frequencies in flight.
+ */
+static int alt_freqs_validate(struct dvfs *d, unsigned long *alt_freqs)
+{
+ int i;
+
+ if (alt_freqs) {
+ for (i = 0; i < d->num_freqs; i++) {
+ if (d->freqs[i] > alt_freqs[i]) {
+ pr_err("%s: Invalid alt freqs for %s\n",
+ __func__, d->clk_name);
+ return -EINVAL;
+ }
+ }
+ }
+ return 0;
+}
+
+int tegra_dvfs_alt_freqs_install(struct dvfs *d, unsigned long *alt_freqs)
+{
+ int ret = 0;
+
+ mutex_lock(&dvfs_lock);
+
+ ret = alt_freqs_validate(d, alt_freqs);
+ if (!ret)
+ d->alt_freqs = alt_freqs;
+
+ mutex_unlock(&dvfs_lock);
+ return ret;
+}
+
+int tegra_dvfs_use_alt_freqs_on_clk(struct clk *c, bool use_alt_freq)
+{
+ int ret = -ENOENT;
+ struct dvfs *d = c->dvfs;
+
+ mutex_lock(&dvfs_lock);
+
+ if (d && d->alt_freqs) {
+ ret = 0;
+ if (d->use_alt_freqs != use_alt_freq) {
+ d->use_alt_freqs = use_alt_freq;
+ ret = __tegra_dvfs_set_rate(d, d->cur_rate);
+ }
+ }
+
+ mutex_unlock(&dvfs_lock);
+ return ret;
+}
+
int tegra_dvfs_alt_freqs_set(struct dvfs *d, unsigned long *alt_freqs)
{
int ret = 0;
mutex_lock(&dvfs_lock);
if (d->alt_freqs != alt_freqs) {
- d->alt_freqs = alt_freqs;
- ret = __tegra_dvfs_set_rate(d, d->cur_rate);
+ ret = alt_freqs_validate(d, alt_freqs);
+ if (!ret) {
+ d->use_alt_freqs = !!alt_freqs;
+ d->alt_freqs = alt_freqs;
+ ret = __tegra_dvfs_set_rate(d, d->cur_rate);
+ }
}
mutex_unlock(&dvfs_lock);
return ret;
}
-static int predict_millivolts(struct clk *c, const int *millivolts,
- unsigned long rate)
+/*
+ * Some clocks may need run-time voltage ladder replacement. Allow it only if
+ * peak voltages across all possible ladders are specified, and new voltages
+ * do not violate peaks.
+ */
+static int new_voltages_validate(struct dvfs *d, const int *new_millivolts,
+ int freqs_num, int ranges_num)
+{
+ const int *millivolts;
+ int freq_idx, therm_idx;
+
+ for (therm_idx = 0; therm_idx < ranges_num; therm_idx++) {
+ millivolts = new_millivolts + therm_idx * MAX_DVFS_FREQS;
+ for (freq_idx = 0; freq_idx < freqs_num; freq_idx++) {
+ if (millivolts[freq_idx] >
+ d->peak_millivolts[freq_idx]) {
+ pr_err("%s: Invalid new voltages for %s\n",
+ __func__, d->clk_name);
+ return -EINVAL;
+ }
+ }
+ }
+ return 0;
+}
+
+int tegra_dvfs_replace_voltage_table(struct dvfs *d, const int *new_millivolts)
+{
+ int ret = 0;
+ int ranges_num = 1;
+
+ mutex_lock(&dvfs_lock);
+
+ if (!d->peak_millivolts) {
+ ret = -EINVAL;
+ goto out;
+ }
+
+ if (d->therm_dvfs && d->dvfs_rail->vts_cdev)
+ ranges_num += d->dvfs_rail->vts_cdev->trip_temperatures_num;
+
+ if (new_voltages_validate(d, new_millivolts,
+ d->num_freqs, ranges_num)) {
+ ret = -EINVAL;
+ goto out;
+ }
+
+ d->millivolts = new_millivolts;
+ if (__tegra_dvfs_set_rate(d, d->cur_rate))
+ ret = -EAGAIN;
+out:
+ mutex_unlock(&dvfs_lock);
+ return ret;
+}
+
+/*
+ * Using non alt frequencies always results in peak voltage
+ * (enforced by alt_freqs_validate())
+ */
+static int predict_non_alt_millivolts(struct clk *c, const int *millivolts,
+ unsigned long rate)
{
int i;
if (!millivolts)
return -ENODEV;
- /*
- * Predicted voltage can not be used across the switch to alternative
- * frequency limits. For now, just fail the call for clock that has
- * alternative limits initialized.
- */
- if (c->dvfs->alt_freqs)
- return -ENOSYS;
for (i = 0; i < c->dvfs->num_freqs; i++) {
if (rate <= c->dvfs->freqs[i])
}
if (i == c->dvfs->num_freqs)
- return -EINVAL;
+ i--;
return millivolts[i];
}
-int tegra_dvfs_predict_millivolts(struct clk *c, unsigned long rate)
+static int predict_millivolts(struct clk *c, const int *millivolts,
+ unsigned long rate)
{
- const int *millivolts;
-
- if (!rate || !c->dvfs)
- return 0;
+ /*
+ * Predicted voltage can not be used across the switch to alternative
+ * frequency limits. For now, just fail the call for clock that has
+ * alternative limits initialized.
+ */
+ if (c->dvfs->alt_freqs)
+ return -ENOSYS;
- millivolts = dvfs_get_millivolts(c->dvfs, rate);
- return predict_millivolts(c, millivolts, rate);
+ return predict_non_alt_millivolts(c, millivolts, rate);
}
-int tegra_dvfs_predict_millivolts_pll(struct clk *c, unsigned long rate)
+int tegra_dvfs_predict_millivolts(struct clk *c, unsigned long rate)
{
const int *millivolts;
if (!rate || !c->dvfs)
return 0;
- millivolts = c->dvfs->millivolts;
+ millivolts = tegra_dvfs_is_dfll_range(c->dvfs, rate) ?
+ c->dvfs->dfll_millivolts :
+ tegra_dvfs_get_millivolts_pll(c->dvfs);
return predict_millivolts(c, millivolts, rate);
}
+EXPORT_SYMBOL(tegra_dvfs_predict_millivolts);
-int tegra_dvfs_predict_millivolts_dfll(struct clk *c, unsigned long rate)
+int tegra_dvfs_predict_peak_millivolts(struct clk *c, unsigned long rate)
{
+ int mv;
const int *millivolts;
if (!rate || !c->dvfs)
return 0;
- millivolts = c->dvfs->dfll_millivolts;
- return predict_millivolts(c, millivolts, rate);
+ millivolts = tegra_dvfs_is_dfll_range(c->dvfs, rate) ?
+ c->dvfs->dfll_millivolts : c->dvfs->peak_millivolts ? :
+ tegra_dvfs_get_millivolts_pll(c->dvfs);
+
+ mv = predict_non_alt_millivolts(c, millivolts, rate);
+ if (mv < 0)
+ return mv;
+
+ if (c->dvfs->dvfs_rail->therm_mv_floors)
+ mv = max(mv, c->dvfs->dvfs_rail->therm_mv_floors[0]);
+ if (c->dvfs->dvfs_rail->therm_mv_dfll_floors)
+ mv = max(mv, c->dvfs->dvfs_rail->therm_mv_dfll_floors[0]);
+ return mv;
+}
+
+const int *tegra_dvfs_get_millivolts_pll(struct dvfs *d)
+{
+ if (d->therm_dvfs) {
+ int therm_idx = d->dvfs_rail->therm_scale_idx;
+ return d->millivolts + therm_idx * MAX_DVFS_FREQS;
+ }
+ return d->millivolts;
}
int tegra_dvfs_set_rate(struct clk *c, unsigned long rate)
{
int ret;
+ bool suspended;
if (!c->dvfs)
return -EINVAL;
- mutex_lock(&dvfs_lock);
+ suspended = timekeeping_suspended && c->dvfs->dvfs_rail->suspended;
+ if (suspended) {
+ if (mutex_is_locked(&dvfs_lock))
+ WARN(1, "%s: Entered suspend with DVFS mutex locked\n",
+ __func__);
+ } else {
+ mutex_lock(&dvfs_lock);
+ }
+
ret = __tegra_dvfs_set_rate(c->dvfs, rate);
- mutex_unlock(&dvfs_lock);
+
+ if (!suspended)
+ mutex_unlock(&dvfs_lock);
return ret;
}
}
EXPORT_SYMBOL(tegra_dvfs_get_freqs);
+static inline int dvfs_rail_get_override_floor(struct dvfs_rail *rail)
+{
+ return rail->override_unresolved ? rail->nominal_millivolts :
+ rail->min_override_millivolts;
+}
+
#ifdef CONFIG_TEGRA_VDD_CORE_OVERRIDE
static DEFINE_MUTEX(rail_override_lock);
if (rail->fixed_millivolts)
return -ENOSYS;
- floor = rail->min_override_millivolts;
+ mutex_lock(&rail_override_lock);
+
+ floor = dvfs_rail_get_override_floor(rail);
ceiling = rail->nominal_millivolts;
if (override_mv && ((override_mv < floor) || (override_mv > ceiling))) {
pr_err("%s: override level %d outside the range [%d...%d]\n",
__func__, override_mv, floor, ceiling);
+ mutex_unlock(&rail_override_lock);
return -EINVAL;
}
- mutex_lock(&rail_override_lock);
-
if (override_mv == rail->override_millivolts) {
ret = 0;
goto out;
}
if (override_mv) {
- ret = tegra_dvfs_core_cap_level_apply(override_mv);
+ ret = tegra_dvfs_override_core_cap_apply(override_mv);
if (ret) {
pr_err("%s: failed to set cap for override level %d\n",
__func__, override_mv);
mutex_unlock(&dvfs_lock);
if (!override_mv || ret)
- tegra_dvfs_core_cap_level_apply(0);
+ tegra_dvfs_override_core_cap_apply(0);
+out:
+ mutex_unlock(&rail_override_lock);
+ return ret;
+}
+
+int tegra_dvfs_resolve_override(struct clk *c, unsigned long max_rate)
+{
+ int mv;
+ struct dvfs *d = c->dvfs;
+ struct dvfs_rail *rail;
+
+ if (!d)
+ return 0;
+ rail = d->dvfs_rail;
+
+ mutex_lock(&rail_override_lock);
+ mutex_lock(&dvfs_lock);
+
+ if (d->defer_override && rail->override_unresolved) {
+ d->defer_override = false;
+
+ mv = tegra_dvfs_predict_peak_millivolts(c, max_rate);
+ if (rail->min_override_millivolts < mv)
+ rail->min_override_millivolts = mv;
+
+ rail->override_unresolved--;
+ if (!rail->override_unresolved && rail->resolve_override)
+ rail->resolve_override(rail->min_override_millivolts);
+ }
+ mutex_unlock(&dvfs_lock);
+ mutex_unlock(&rail_override_lock);
+ return 0;
+}
+
+int tegra_dvfs_rail_get_override_floor(struct dvfs_rail *rail)
+{
+ if (rail) {
+ int mv;
+ mutex_lock(&rail_override_lock);
+ mv = dvfs_rail_get_override_floor(rail);
+ mutex_unlock(&rail_override_lock);
+ return mv;
+ }
+ return -ENOENT;
+}
+
+static int dvfs_set_fmax_at_vmin(struct clk *c, unsigned long f_max, int v_min)
+{
+ int i, ret = 0;
+ struct dvfs *d = c->dvfs;
+ unsigned long f_min = 1000; /* 1kHz min rate in DVFS tables */
+
+ mutex_lock(&rail_override_lock);
+ mutex_lock(&dvfs_lock);
+
+ if (v_min > d->dvfs_rail->override_millivolts) {
+ pr_err("%s: new %s vmin %dmV is above override voltage %dmV\n",
+ __func__, c->name, v_min,
+ d->dvfs_rail->override_millivolts);
+ ret = -EPERM;
+ goto out;
+ }
+
+ if (v_min >= d->max_millivolts) {
+ pr_err("%s: new %s vmin %dmV is at/above max voltage %dmV\n",
+ __func__, c->name, v_min, d->max_millivolts);
+ ret = -EINVAL;
+ goto out;
+ }
+
+ /*
+ * dvfs table update:
+ * - for voltages below new v_min the respective frequencies are shifted
+ * below new f_max to the levels already present in the table; if the
+ * 1st table entry has frequency above new fmax, all entries below v_min
+ * are filled in with 1kHz (min rate used in DVFS tables).
+ * - for voltages above new v_min, the respective frequencies are
+ * increased to at least new f_max
+ * - if new v_min is already in the table set the respective frequency
+ * to new f_max
+ */
+ for (i = 0; i < d->num_freqs; i++) {
+ int mv = d->millivolts[i];
+ unsigned long f = d->freqs[i];
+
+ if (mv < v_min) {
+ if (d->freqs[i] >= f_max)
+ d->freqs[i] = i ? d->freqs[i-1] : f_min;
+ } else if (mv > v_min) {
+ d->freqs[i] = max(f, f_max);
+ } else {
+ d->freqs[i] = f_max;
+ }
+ ret = __tegra_dvfs_set_rate(d, d->cur_rate);
+ }
out:
+ mutex_unlock(&dvfs_lock);
mutex_unlock(&rail_override_lock);
+
return ret;
}
#else
pr_err("%s: vdd core override is not supported\n", __func__);
return -ENOSYS;
}
+
+static int dvfs_set_fmax_at_vmin(struct clk *c, unsigned long f_max, int v_min)
+{
+ pr_err("%s: vdd core override is not supported\n", __func__);
+ return -ENOSYS;
+}
#endif
int tegra_dvfs_override_core_voltage(struct clk *c, int override_mv)
}
EXPORT_SYMBOL(tegra_dvfs_override_core_voltage);
+int tegra_dvfs_set_fmax_at_vmin(struct clk *c, unsigned long f_max, int v_min)
+{
+ if (!c->dvfs || !c->dvfs->can_override) {
+ pr_err("%s: %s cannot set fmax_at_vmin)\n", __func__, c->name);
+ return -EPERM;
+ }
+ return dvfs_set_fmax_at_vmin(c, f_max, v_min);
+}
+EXPORT_SYMBOL(tegra_dvfs_set_fmax_at_vmin);
+
/* May only be called during clock init, does not take any locks on clock c. */
int __init tegra_enable_dvfs_on_clk(struct clk *c, struct dvfs *d)
{
*/
if (i && c->ops && !c->ops->shared_bus_update &&
!(c->flags & PERIPH_ON_CBUS) && !d->can_override) {
- int mv = tegra_dvfs_predict_millivolts(c, d->freqs[i-1]);
- if (d->dvfs_rail->min_override_millivolts < mv)
- d->dvfs_rail->min_override_millivolts = mv;
+ int mv = tegra_dvfs_predict_peak_millivolts(c, d->freqs[i-1]);
+ struct dvfs_rail *rail = d->dvfs_rail;
+ if (d->defer_override)
+ rail->override_unresolved++;
+ else if (rail->min_override_millivolts < mv)
+ rail->min_override_millivolts =
+ min(mv, rail->nominal_millivolts);
}
mutex_lock(&dvfs_lock);
return all_suspended;
}
+static bool is_solved_at_suspend(struct dvfs_rail *to,
+ struct dvfs_relationship *rel)
+{
+ if (rel->solved_at_suspend)
+ return true;
+
+ if (rel->solved_at_nominal) {
+ int mv = tegra_dvfs_rail_get_suspend_level(to);
+ if (mv == to->nominal_millivolts)
+ return true;
+ }
+ return false;
+}
+
static bool tegra_dvfs_from_rails_suspended_or_solved(struct dvfs_rail *to)
{
struct dvfs_relationship *rel;
list_for_each_entry(rel, &to->relationships_from, from_node)
if (!rel->from->suspended && !rel->from->disabled &&
- !rel->solved_at_nominal)
+ !is_solved_at_suspend(to, rel))
all_suspended = false;
return all_suspended;
list_for_each_entry(rail, &dvfs_rail_list, node) {
if (!rail->suspended && !rail->disabled &&
tegra_dvfs_from_rails_suspended_or_solved(rail)) {
- /* Safe, as pll mode rate is capped to fixed level */
- if (!rail->dfll_mode && rail->fixed_millivolts) {
+ if (rail->dfll_mode) {
+ /* s/w doesn't change voltage in dfll mode */
+ mv = rail->millivolts;
+ } else if (rail->fixed_millivolts) {
+ /* Safe: pll mode rate capped to fixed level */
mv = rail->fixed_millivolts;
} else {
mv = tegra_dvfs_rail_get_suspend_level(rail);
return ret;
}
-bool tegra_dvfs_rail_updating(struct clk *clk)
+int tegra_dvfs_rail_set_mode(struct dvfs_rail *rail, unsigned int mode)
{
- return (!clk ? false :
- (!clk->dvfs ? false :
- (!clk->dvfs->dvfs_rail ? false :
- (clk->dvfs->dvfs_rail->updating ||
- clk->dvfs->dvfs_rail->dfll_mode_updating))));
+ int ret = -ENOENT;
+ unsigned int cur_mode;
+
+ if (!rail || !rail->reg)
+ return ret;
+
+ if (regulator_can_set_mode(rail->reg)) {
+ pr_debug("%s: updating %s mode to %u\n", __func__,
+ rail->reg_id, mode);
+ ret = regulator_set_mode(rail->reg, mode);
+ if (ret)
+ pr_err("%s: failed to set dvfs regulator %s mode %u\n",
+ __func__, rail->reg_id, mode);
+ return ret;
+ }
+
+ /*
+ * Set mode is not supported - check request against current mode
+ * (if the latter is unknown, assume NORMAL).
+ */
+ cur_mode = regulator_get_mode(rail->reg);
+ if (IS_ERR_VALUE(cur_mode))
+ cur_mode = REGULATOR_MODE_NORMAL;
+
+ if (WARN_ONCE(cur_mode != mode,
+ "%s: dvfs regulator %s cannot change mode from %u\n",
+ __func__, rail->reg_id, cur_mode))
+ return -EINVAL;
+
+ return 0;
}
-#ifdef CONFIG_OF
-int __init of_tegra_dvfs_init(const struct of_device_id *matches)
+int tegra_dvfs_rail_register_notifier(struct dvfs_rail *rail,
+ struct notifier_block *nb)
{
- int ret;
- struct device_node *np;
+ if (!rail || !rail->reg)
+ return -ENOENT;
+
+ return regulator_register_notifier(rail->reg, nb);
+}
+
+int tegra_dvfs_rail_unregister_notifier(struct dvfs_rail *rail,
+ struct notifier_block *nb)
+{
+ if (!rail || !rail->reg)
+ return -ENOENT;
+
+ return regulator_unregister_notifier(rail->reg, nb);
+}
+
+bool tegra_dvfs_rail_updating(struct clk *clk)
+{
+ return (!clk ? false :
+ (!clk->dvfs ? false :
+ (!clk->dvfs->dvfs_rail ? false :
+ (clk->dvfs->dvfs_rail->updating ||
+ clk->dvfs->dvfs_rail->dfll_mode_updating))));
+}
+
+#ifdef CONFIG_OF
+int __init of_tegra_dvfs_init(const struct of_device_id *matches)
+{
+ int ret;
+ struct device_node *np;
for_each_matching_node(np, matches) {
const struct of_device_id *match = of_match_node(matches, np);
if (!d->dvfs_rail->dfll_mode) {
d->dvfs_rail->dfll_mode = true;
__tegra_dvfs_set_rate(d, rate);
+
+ /*
+ * Report error, but continue: DFLL is functional, anyway, and
+ * no error with proper regulator driver update
+ */
+ if (regulator_set_vsel_volatile(d->dvfs_rail->reg, true))
+ WARN_ONCE(1, "%s: failed to set vsel volatile\n",
+ __func__);
}
mutex_unlock(&dvfs_lock);
return 0;
mutex_lock(&dvfs_lock);
if (d->dvfs_rail->dfll_mode) {
d->dvfs_rail->dfll_mode = false;
+ regulator_set_vsel_volatile(d->dvfs_rail->reg, false);
+
/* avoid false detection of matching target (voltage in dfll
mode is fluctuating, and recorded level is just estimate) */
d->dvfs_rail->millivolts--;
return NULL;
}
+struct tegra_cooling_device *tegra_dvfs_get_core_vmax_cdev(void)
+{
+ if (tegra_core_rail)
+ return tegra_core_rail->vmax_cdev;
+ return NULL;
+}
+
struct tegra_cooling_device *tegra_dvfs_get_core_vmin_cdev(void)
{
if (tegra_core_rail)
return NULL;
}
+struct tegra_cooling_device *tegra_dvfs_get_gpu_vts_cdev(void)
+{
+ if (tegra_gpu_rail)
+ return tegra_gpu_rail->vts_cdev;
+ return NULL;
+}
+
+struct tegra_cooling_device *tegra_dvfs_get_cpu_clk_switch_cdev(void)
+{
+ if (tegra_cpu_rail)
+ return tegra_cpu_rail->clk_switch_cdev;
+ return NULL;
+}
+
+static void make_safe_thermal_dvfs(struct dvfs_rail *rail)
+{
+ struct dvfs *d;
+
+ mutex_lock(&dvfs_lock);
+ list_for_each_entry(d, &rail->dvfs, reg_node) {
+ if (d->therm_dvfs) {
+ BUG_ON(!d->peak_millivolts);
+ d->millivolts = d->peak_millivolts;
+ d->therm_dvfs = false;
+ }
+ }
+ mutex_unlock(&dvfs_lock);
+}
+
#ifdef CONFIG_THERMAL
/* Cooling device limits minimum rail voltage at cold temperature in pll mode */
static int tegra_dvfs_rail_get_vmin_cdev_max_state(
return 0;
}
-static struct thermal_cooling_device_ops tegra_dvfs_rail_cooling_ops = {
+static struct thermal_cooling_device_ops tegra_dvfs_vmin_cooling_ops = {
.get_max_state = tegra_dvfs_rail_get_vmin_cdev_max_state,
.get_cur_state = tegra_dvfs_rail_get_vmin_cdev_cur_state,
.set_cur_state = tegra_dvfs_rail_set_vmin_cdev_state,
/* just report error - initialized for cold temperature, anyway */
if (IS_ERR_OR_NULL(thermal_cooling_device_register(
rail->vmin_cdev->cdev_type, (void *)rail,
- &tegra_dvfs_rail_cooling_ops)))
+ &tegra_dvfs_vmin_cooling_ops)))
pr_err("tegra cooling device %s failed to register\n",
rail->vmin_cdev->cdev_type);
}
+/*
+ * Cooling device limits frequencies of the clocks in pll mode based on rail
+ * vmax thermal profile. Supported for core rail only, and applied only to
+ * shared buses selected by platform specific code.
+ */
+static int tegra_dvfs_rail_get_vmax_cdev_max_state(
+ struct thermal_cooling_device *cdev, unsigned long *max_state)
+{
+ struct dvfs_rail *rail = (struct dvfs_rail *)cdev->devdata;
+ *max_state = rail->vmax_cdev->trip_temperatures_num;
+ return 0;
+}
+
+static int tegra_dvfs_rail_get_vmax_cdev_cur_state(
+ struct thermal_cooling_device *cdev, unsigned long *cur_state)
+{
+ struct dvfs_rail *rail = (struct dvfs_rail *)cdev->devdata;
+ *cur_state = rail->therm_cap_idx;
+ return 0;
+}
+
+static int tegra_dvfs_rail_set_vmax_cdev_state(
+ struct thermal_cooling_device *cdev, unsigned long cur_state)
+{
+ struct dvfs_rail *rail = (struct dvfs_rail *)cdev->devdata;
+ int cur_cap = cur_state ? rail->therm_mv_caps[cur_state - 1] : 0;
+
+ return tegra_dvfs_therm_vmax_core_cap_apply(&rail->therm_cap_idx,
+ cur_state, cur_cap);
+}
+
+static struct thermal_cooling_device_ops tegra_dvfs_vmax_cooling_ops = {
+ .get_max_state = tegra_dvfs_rail_get_vmax_cdev_max_state,
+ .get_cur_state = tegra_dvfs_rail_get_vmax_cdev_cur_state,
+ .set_cur_state = tegra_dvfs_rail_set_vmax_cdev_state,
+};
+
+void tegra_dvfs_rail_register_vmax_cdev(struct dvfs_rail *rail)
+{
+ struct thermal_cooling_device *dev;
+
+ if (!rail || !rail->vmax_cdev || (rail != tegra_core_rail))
+ return;
+
+ dev = thermal_cooling_device_register(rail->vmax_cdev->cdev_type,
+ (void *)rail, &tegra_dvfs_vmax_cooling_ops);
+
+ if (IS_ERR_OR_NULL(dev) || list_empty(&dev->thermal_instances)) {
+ /* report error & set the most agressive caps */
+ int cur_state = rail->vmax_cdev->trip_temperatures_num;
+ int cur_cap = rail->therm_mv_caps[cur_state - 1];
+ tegra_dvfs_therm_vmax_core_cap_apply(&rail->therm_cap_idx,
+ cur_state, cur_cap);
+ pr_err("tegra cooling device %s failed to register\n",
+ rail->vmax_cdev->cdev_type);
+ }
+}
+
+/* Cooling device to switch the cpu clock source between PLLX and DFLL */
+static int tegra_dvfs_rail_get_clk_switch_cdev_max_state(
+ struct thermal_cooling_device *cdev, unsigned long *max_state)
+{
+ struct dvfs_rail *rail = (struct dvfs_rail *)cdev->devdata;
+ *max_state = rail->clk_switch_cdev->trip_temperatures_num;
+ return 0;
+}
+
+static int tegra_dvfs_rail_get_clk_switch_cdev_cur_state(
+ struct thermal_cooling_device *cdev, unsigned long *cur_state)
+{
+ struct dvfs_rail *rail = (struct dvfs_rail *)cdev->devdata;
+ *cur_state = rail->therm_scale_idx;
+ return 0;
+}
+
+static int tegra_dvfs_rail_set_clk_switch_cdev_state(
+ struct thermal_cooling_device *cdev, unsigned long cur_state)
+{
+ int ret = 0;
+ enum dfll_range use_dfll;
+ struct dvfs_rail *rail = (struct dvfs_rail *)cdev->devdata;
+
+ if (tegra_override_dfll_range == TEGRA_USE_DFLL_CDEV_CNTRL) {
+ if (rail->therm_scale_idx != cur_state) {
+ rail->therm_scale_idx = cur_state;
+ if (rail->therm_scale_idx == 0)
+ use_dfll = DFLL_RANGE_NONE;
+ else
+ use_dfll = DFLL_RANGE_ALL_RATES;
+
+ ret = tegra_clk_dfll_range_control(use_dfll);
+ }
+ } else {
+ pr_warn("\n%s: Not Allowed:", __func__);
+ pr_warn("DFLL is not under thermal cooling device control\n");
+ return -EACCES;
+ }
+ return ret;
+}
+
+static struct thermal_cooling_device_ops tegra_dvfs_clk_cooling_ops = {
+ .get_max_state = tegra_dvfs_rail_get_clk_switch_cdev_max_state,
+ .get_cur_state = tegra_dvfs_rail_get_clk_switch_cdev_cur_state,
+ .set_cur_state = tegra_dvfs_rail_set_clk_switch_cdev_state,
+};
+
+static void tegra_dvfs_rail_register_clk_switch_cdev(struct dvfs_rail *rail)
+{
+ struct thermal_cooling_device *dev;
+
+ if (!rail->clk_switch_cdev)
+ return;
+
+ dev = thermal_cooling_device_register(rail->clk_switch_cdev->cdev_type,
+ (void *)rail, &tegra_dvfs_clk_cooling_ops);
+ /* report error & set max limits across thermal ranges as safe dvfs */
+ if (IS_ERR_OR_NULL(dev) || list_empty(&dev->thermal_instances)) {
+ pr_err("tegra cooling device %s failed to register\n",
+ rail->clk_switch_cdev->cdev_type);
+ make_safe_thermal_dvfs(rail);
+ }
+}
+
+
+/* Cooling device to scale voltage with temperature in pll mode */
+static int tegra_dvfs_rail_get_vts_cdev_max_state(
+ struct thermal_cooling_device *cdev, unsigned long *max_state)
+{
+ struct dvfs_rail *rail = (struct dvfs_rail *)cdev->devdata;
+ *max_state = rail->vts_cdev->trip_temperatures_num;
+ return 0;
+}
+
+static int tegra_dvfs_rail_get_vts_cdev_cur_state(
+ struct thermal_cooling_device *cdev, unsigned long *cur_state)
+{
+ struct dvfs_rail *rail = (struct dvfs_rail *)cdev->devdata;
+ *cur_state = rail->therm_scale_idx;
+ return 0;
+}
+
+static int tegra_dvfs_rail_set_vts_cdev_state(
+ struct thermal_cooling_device *cdev, unsigned long cur_state)
+{
+ struct dvfs_rail *rail = (struct dvfs_rail *)cdev->devdata;
+ struct dvfs *d;
+
+ mutex_lock(&dvfs_lock);
+ if (rail->therm_scale_idx != cur_state) {
+ rail->therm_scale_idx = cur_state;
+ list_for_each_entry(d, &rail->dvfs, reg_node) {
+ if (d->therm_dvfs)
+ __tegra_dvfs_set_rate(d, d->cur_rate);
+ }
+ }
+ mutex_unlock(&dvfs_lock);
+ return 0;
+}
+
+static struct thermal_cooling_device_ops tegra_dvfs_vts_cooling_ops = {
+ .get_max_state = tegra_dvfs_rail_get_vts_cdev_max_state,
+ .get_cur_state = tegra_dvfs_rail_get_vts_cdev_cur_state,
+ .set_cur_state = tegra_dvfs_rail_set_vts_cdev_state,
+};
+
+static void tegra_dvfs_rail_register_vts_cdev(struct dvfs_rail *rail)
+{
+ struct thermal_cooling_device *dev;
+
+ if (!rail->vts_cdev)
+ return;
+
+ dev = thermal_cooling_device_register(rail->vts_cdev->cdev_type,
+ (void *)rail, &tegra_dvfs_vts_cooling_ops);
+ /* report error & set max limits across thermal ranges as safe dvfs */
+ if (IS_ERR_OR_NULL(dev) || list_empty(&dev->thermal_instances)) {
+ pr_err("tegra cooling device %s failed to register\n",
+ rail->vts_cdev->cdev_type);
+ make_safe_thermal_dvfs(rail);
+ }
+}
+
#else
#define tegra_dvfs_rail_register_vmin_cdev(rail)
+void tegra_dvfs_rail_register_vmax_cdev(struct dvfs_rail *rail)
+{ }
+static inline void tegra_dvfs_rail_register_vts_cdev(struct dvfs_rail *rail)
+{
+ make_safe_thermal_dvfs(rail);
+}
+#endif
+
+#ifdef CONFIG_TEGRA_USE_SIMON
+/*
+ * Validate rail SiMon Vmin offsets. Valid offsets should be negative,
+ * descending, starting from zero.
+ */
+void __init tegra_dvfs_rail_init_simon_vmin_offsets(
+ int *offsets, int offs_num, struct dvfs_rail *rail)
+{
+ int i;
+
+ if (!offsets || !offs_num || offsets[0]) {
+ WARN(1, "%s: invalid initial SiMon offset\n", rail->reg_id);
+ return;
+ }
+
+ for (i = 0; i < offs_num - 1; i++) {
+ if (offsets[i] < offsets[i+1]) {
+ WARN(1, "%s: SiMon offsets are not ordered\n",
+ rail->reg_id);
+ return;
+ }
+ }
+ rail->simon_vmin_offsets = offsets;
+ rail->simon_vmin_offs_num = offs_num;
+}
#endif
/*
* - voltage limits are descending with temperature increasing
* - the lowest limit is above rail minimum voltage in pll and
* in dfll mode (if applicable)
- * - the highest limit is below rail nominal voltage
+ * - the highest limit is below rail nominal voltage (required only
+ * for Vmin profile)
*/
static int __init get_thermal_profile_size(
int *trips_table, int *limits_table,
return -EINVAL;
}
- if (limits_table[0] > rail->nominal_millivolts) {
- pr_warn("%s: thermal profile above Vmax\n", rail->reg_id);
- return -EINVAL;
- }
return i + 1;
}
}
}
+int __init tegra_dvfs_rail_init_clk_switch_thermal_profile(
+ int *clk_switch_trips, struct dvfs_rail *rail)
+{
+ int i;
+
+ if (!rail->clk_switch_cdev) {
+ WARN(1, "%s: missing thermal dvfs cooling device\n",
+ rail->reg_id);
+ return -ENOENT;
+ }
+
+ for (i = 0; i < MAX_THERMAL_LIMITS - 1; i++) {
+ if (clk_switch_trips[i] >= clk_switch_trips[i+1])
+ break;
+ }
+
+ /*Only one trip point is allowed for this cdev*/
+ if (i != 0) {
+ WARN(1, "%s: Only one trip point allowed\n", __func__);
+ return -EINVAL;
+ }
+
+ rail->clk_switch_cdev->trip_temperatures_num = i + 1;
+ rail->clk_switch_cdev->trip_temperatures = clk_switch_trips;
+ return 0;
+}
+
void __init tegra_dvfs_rail_init_vmin_thermal_profile(
int *therm_trips_table, int *therm_floors_table,
struct dvfs_rail *rail, struct dvfs_dfll_data *d)
{
int i = get_thermal_profile_size(therm_trips_table,
therm_floors_table, rail, d);
- if (i <= 0) {
+
+ if (i <= 0 || therm_floors_table[0] > rail->nominal_millivolts) {
rail->vmin_cdev = NULL;
WARN(1, "%s: invalid Vmin thermal profile\n", rail->reg_id);
return;
}
}
+/*
+ * Validate thermal dvfs settings:
+ * - trip-points are montonically increasing
+ * - voltages in any temperature range are montonically increasing with
+ * frequency (can go up/down across ranges at iso frequency)
+ * - voltage for any frequency/thermal range combination must be within
+ * rail minimum/maximum limits
+ */
+int __init tegra_dvfs_rail_init_thermal_dvfs_trips(
+ int *therm_trips_table, struct dvfs_rail *rail)
+{
+ int i;
+
+ if (!rail->vts_cdev) {
+ WARN(1, "%s: missing thermal dvfs cooling device\n",
+ rail->reg_id);
+ return -ENOENT;
+ }
+
+ for (i = 0; i < MAX_THERMAL_LIMITS - 1; i++) {
+ if (therm_trips_table[i] >= therm_trips_table[i+1])
+ break;
+ }
+
+ rail->vts_cdev->trip_temperatures_num = i + 1;
+ rail->vts_cdev->trip_temperatures = therm_trips_table;
+ return 0;
+}
+
+int __init tegra_dvfs_init_thermal_dvfs_voltages(int *therm_voltages,
+ int *peak_voltages, int freqs_num, int ranges_num, struct dvfs *d)
+{
+ int *millivolts;
+ int freq_idx, therm_idx;
+
+ for (therm_idx = 0; therm_idx < ranges_num; therm_idx++) {
+ millivolts = therm_voltages + therm_idx * MAX_DVFS_FREQS;
+ for (freq_idx = 0; freq_idx < freqs_num; freq_idx++) {
+ int mv = millivolts[freq_idx];
+ if ((mv > d->dvfs_rail->max_millivolts) ||
+ (mv < d->dvfs_rail->min_millivolts) ||
+ (freq_idx && (mv < millivolts[freq_idx - 1]))) {
+ WARN(1, "%s: invalid thermal dvfs entry %d(%d, %d)\n",
+ d->clk_name, mv, freq_idx, therm_idx);
+ return -EINVAL;
+ }
+ if (mv > peak_voltages[freq_idx])
+ peak_voltages[freq_idx] = mv;
+ }
+ }
+
+ d->millivolts = therm_voltages;
+ d->peak_millivolts = peak_voltages;
+ d->therm_dvfs = true;
+ return 0;
+}
+
/* Directly set cold temperature limit in dfll mode */
-int tegra_dvfs_rail_dfll_mode_set_cold(struct dvfs_rail *rail)
+int tegra_dvfs_rail_dfll_mode_set_cold(struct dvfs_rail *rail,
+ struct clk *dfll_clk)
{
int ret = 0;
return ret;
/*
- * Since cooling thresholds are the same in pll and dfll modes, pll mode
- * thermal index can be used to decide if cold limit should be set in
- * dfll mode.
+ * Compare last set Vmin with requirement based on current temperature,
+ * and set cold limit at regulator only Vmin is below requirement.
*/
mutex_lock(&dvfs_lock);
- if (rail->dfll_mode &&
- (rail->therm_floor_idx < rail->therm_mv_floors_num)) {
- int mv = rail->therm_mv_floors[rail->therm_floor_idx];
+ if (rail->dfll_mode) {
+ int mv, cmp;
+ cmp = tegra_cl_dvfs_vmin_cmp_needed(
+ tegra_dfll_get_cl_dvfs_data(dfll_clk), &mv);
+ if (cmp < 0)
ret = dvfs_rail_set_voltage_reg(rail, mv);
}
mutex_unlock(&dvfs_lock);
return ret;
}
+/* Get current thermal floor */
+int tegra_dvfs_rail_get_thermal_floor(struct dvfs_rail *rail)
+{
+ if (rail && rail->therm_mv_floors &&
+ (rail->therm_floor_idx < rail->therm_mv_floors_num)) {
+ int i = rail->therm_floor_idx;
+ if (rail->dfll_mode) {
+ BUG_ON(!rail->therm_mv_dfll_floors);
+ return rail->therm_mv_dfll_floors[i];
+ }
+ return rail->therm_mv_floors[i];
+ }
+ return 0;
+}
+
/*
* Iterate through all the dvfs regulators, finding the regulator exported
* by the regulator api for each one. Must be called in late init, after
* all the regulator api's regulators are initialized.
*/
-int __init tegra_dvfs_late_init(void)
+
+#ifdef CONFIG_TEGRA_DVFS_RAIL_CONNECT_ALL
+/*
+ * Enable voltage scaling only if all the rails connect successfully
+ */
+int __init tegra_dvfs_rail_connect_regulators(void)
{
bool connected = true;
struct dvfs_rail *rail;
if (dvfs_rail_connect_to_regulator(rail))
connected = false;
- list_for_each_entry(rail, &dvfs_rail_list, node)
- if (connected)
+ list_for_each_entry(rail, &dvfs_rail_list, node) {
+ if (connected) {
dvfs_rail_update(rail);
- else
- __tegra_dvfs_rail_disable(rail);
-
+ if (!rail->disabled)
+ continue;
+ /* Don't rely on boot level - force disabled voltage */
+ rail->disabled = false;
+ }
+ __tegra_dvfs_rail_disable(rail);
+ }
mutex_unlock(&dvfs_lock);
if (!connected && tegra_platform_is_silicon()) {
return -ENODEV;
}
+ return 0;
+}
+#else
+int __init tegra_dvfs_rail_connect_regulators(void)
+{
+ struct dvfs_rail *rail;
+
+ mutex_lock(&dvfs_lock);
+
+ list_for_each_entry(rail, &dvfs_rail_list, node) {
+ if (!dvfs_rail_connect_to_regulator(rail)) {
+ dvfs_rail_update(rail);
+ if (!rail->disabled)
+ continue;
+ /* Don't rely on boot level - force disabled voltage */
+ rail->disabled = false;
+ }
+ __tegra_dvfs_rail_disable(rail);
+ }
+
+ mutex_unlock(&dvfs_lock);
+
+ return 0;
+}
+#endif
+
+int __init tegra_dvfs_rail_register_notifiers(void)
+{
+ struct dvfs_rail *rail;
+
register_pm_notifier(&tegra_dvfs_suspend_nb);
register_pm_notifier(&tegra_dvfs_resume_nb);
register_reboot_notifier(&tegra_dvfs_reboot_nb);
- list_for_each_entry(rail, &dvfs_rail_list, node)
- tegra_dvfs_rail_register_vmin_cdev(rail);
+ list_for_each_entry(rail, &dvfs_rail_list, node) {
+ tegra_dvfs_rail_register_vmin_cdev(rail);
+ tegra_dvfs_rail_register_vts_cdev(rail);
+ tegra_dvfs_rail_register_clk_switch_cdev(rail);
+
+ }
return 0;
}
return strcmp(da->clk_name, db->clk_name);
}
+/* To emulate and show rail relations with 0 mV on dependent rail-to */
+static struct dvfs_rail show_to;
+static struct dvfs_relationship show_rel;
+
static int dvfs_tree_show(struct seq_file *s, void *data)
{
struct dvfs *d;
list_for_each_entry(rail, &dvfs_rail_list, node) {
int thermal_mv_floor = 0;
- seq_printf(s, "%s %d mV%s:\n", rail->reg_id, rail->millivolts,
+ seq_printf(s, "%s %d mV%s:\n", rail->reg_id,
+ rail->stats.off ? 0 : rail->millivolts,
rail->dfll_mode ? " dfll mode" :
rail->disabled ? " disabled" : "");
list_for_each_entry(rel, &rail->relationships_from, from_node) {
- seq_printf(s, " %-10s %-7d mV %-4d mV\n",
+ show_rel = *rel;
+ show_rel.to = &show_to;
+ show_to = *rel->to;
+ show_to.millivolts = show_to.new_millivolts = 0;
+ seq_printf(s, " %-10s %-7d mV %-4d mV .. %-4d mV\n",
rel->from->reg_id, rel->from->millivolts,
+ dvfs_solve_relationship(&show_rel),
dvfs_solve_relationship(rel));
}
+ seq_printf(s, " nominal %-7d mV\n",
+ rail->nominal_millivolts);
seq_printf(s, " offset %-7d mV\n", rail->dbg_mv_offs);
- if (rail->therm_mv_floors) {
- int i = rail->therm_floor_idx;
- if (i < rail->therm_mv_floors_num)
- thermal_mv_floor = rail->therm_mv_floors[i];
- }
+ thermal_mv_floor = tegra_dvfs_rail_get_thermal_floor(rail);
seq_printf(s, " thermal %-7d mV\n", thermal_mv_floor);
if (rail == tegra_core_rail) {
- seq_printf(s, " override %-7d mV [%-4d...%-4d]\n",
+ seq_printf(s, " override %-7d mV [%-4d...%-4d]",
rail->override_millivolts,
- rail->min_override_millivolts,
+ dvfs_rail_get_override_floor(rail),
rail->nominal_millivolts);
+ if (rail->override_unresolved)
+ seq_printf(s, " unresolved %d",
+ rail->override_unresolved);
+ seq_putc(s, '\n');
}
list_sort(NULL, &rail->dvfs, dvfs_tree_sort_cmp);
.release = single_release,
};
-static int gpu_dvfs_show(struct seq_file *s, void *data)
-{
- int idx;
- int *millivolts;
- unsigned long *freqs;
-
- if (read_gpu_dvfs_table(&millivolts, &freqs)) {
- seq_printf(s, "Only supported for T124 or higher\n");
- return 0;
- }
-
- seq_printf(s, "millivolts \t \t frequency\n");
- seq_printf(s, "=====================================\n");
-
- for (idx = 0; millivolts[idx]; idx++)
- seq_printf(s, "%d mV \t \t %lu Hz\n", millivolts[idx],
- freqs[idx]);
-
- return 0;
-}
-
-static int gpu_dvfs_open(struct inode *inode, struct file *file)
-{
- return single_open(file, gpu_dvfs_show, NULL);
-}
-
-static const struct file_operations gpu_dvfs_fops = {
- .open = gpu_dvfs_open,
- .read = seq_read,
- .llseek = seq_lseek,
- .release = single_release,
-};
-
static int rail_offs_set(struct dvfs_rail *rail, int offs)
{
if (rail) {
DEFINE_SIMPLE_ATTRIBUTE(core_override_fops,
core_override_get, core_override_set, "%llu\n");
+static int rail_mv_get(void *data, u64 *val)
+{
+ struct dvfs_rail *rail = data;
+ if (rail) {
+ *val = rail->stats.off ? 0 : rail->millivolts;
+ return 0;
+ }
+ *val = 0;
+ return -ENOENT;
+}
+DEFINE_SIMPLE_ATTRIBUTE(rail_mv_fops, rail_mv_get, NULL, "%llu\n");
+
+static int gpu_dvfs_t_show(struct seq_file *s, void *data)
+{
+ int i, j;
+ int num_ranges = 1;
+ int *trips = NULL;
+ struct dvfs *d;
+ struct dvfs_rail *rail = tegra_gpu_rail;
+ int max_mv[MAX_DVFS_FREQS] = {};
+
+ if (!tegra_gpu_rail) {
+ seq_printf(s, "Only supported for T124 or higher\n");
+ return -ENOSYS;
+ }
+
+ mutex_lock(&dvfs_lock);
+
+ d = list_first_entry(&rail->dvfs, struct dvfs, reg_node);
+ if (rail->vts_cdev && d->therm_dvfs) {
+ num_ranges = rail->vts_cdev->trip_temperatures_num + 1;
+ trips = rail->vts_cdev->trip_temperatures;
+ }
+
+ seq_printf(s, "%-11s", "T(C)\\F(kHz)");
+ for (i = 0; i < d->num_freqs; i++) {
+ unsigned int f = d->freqs[i]/1000;
+ seq_printf(s, " %7u", f);
+ }
+ seq_printf(s, "\n");
+
+ for (j = 0; j < num_ranges; j++) {
+ seq_printf(s, "%s", j == rail->therm_scale_idx ? ">" : " ");
+
+ if (!trips || (num_ranges == 1))
+ seq_printf(s, "%4s..%-4s", "", "");
+ else if (j == 0)
+ seq_printf(s, "%4s..%-4d", "", trips[j]);
+ else if (j == num_ranges - 1)
+ seq_printf(s, "%4d..%-4s", trips[j], "");
+ else
+ seq_printf(s, "%4d..%-4d", trips[j-1], trips[j]);
+
+ for (i = 0; i < d->num_freqs; i++) {
+ int mv = *(d->millivolts + j * MAX_DVFS_FREQS + i);
+ seq_printf(s, " %7d", mv);
+ max_mv[i] = max(max_mv[i], mv);
+ }
+ seq_printf(s, " mV\n");
+ }
+
+ seq_printf(s, "%3s%-8s\n", "", "------");
+ seq_printf(s, "%3s%-8s", "", "max(T)");
+ for (i = 0; i < d->num_freqs; i++)
+ seq_printf(s, " %7d", max_mv[i]);
+ seq_printf(s, " mV\n");
+
+ mutex_unlock(&dvfs_lock);
+
+ return 0;
+}
+
+static int gpu_dvfs_t_open(struct inode *inode, struct file *file)
+{
+ return single_open(file, gpu_dvfs_t_show, NULL);
+}
+
+static const struct file_operations gpu_dvfs_t_fops = {
+ .open = gpu_dvfs_t_open,
+ .read = seq_read,
+ .llseek = seq_lseek,
+ .release = single_release,
+};
+
static int dvfs_table_show(struct seq_file *s, void *data)
{
int i;
struct dvfs *d;
struct dvfs_rail *rail;
+ const int *v_pll, *last_v_pll = NULL;
+ const int *v_dfll, *last_v_dfll = NULL;
- seq_printf(s, "DVFS tables: units mV/MHz\n\n");
+ seq_printf(s, "DVFS tables: units mV/MHz\n");
mutex_lock(&dvfs_lock);
list_for_each_entry(rail, &dvfs_rail_list, node) {
- bool mv_done = false;
+ if (rail->version) {
+ seq_printf(s, "%-9s table version: ", rail->reg_id);
+ seq_printf(s, "%-16s\n", rail->version);
+ }
+ }
+
+ list_for_each_entry(rail, &dvfs_rail_list, node) {
list_for_each_entry(d, &rail->dvfs, reg_node) {
- if (!mv_done) {
- mv_done = true;
- seq_printf(s, "%-16s", rail->reg_id);
- for (i = 0; i < d->num_freqs; i++) {
- int mv = d->millivolts[i];
- seq_printf(s, "%7d", mv);
+ bool mv_done = false;
+ v_pll = tegra_dvfs_get_millivolts_pll(d);
+ v_dfll = d->dfll_millivolts;
+
+ if (v_pll && (last_v_pll != v_pll)) {
+ if (!mv_done) {
+ seq_printf(s, "\n");
+ mv_done = true;
}
+ last_v_pll = v_pll;
+ seq_printf(s, "%-16s", rail->reg_id);
+ for (i = 0; i < d->num_freqs; i++)
+ seq_printf(s, "%7d", v_pll[i]);
seq_printf(s, "\n");
- if (d->dfll_millivolts) {
- seq_printf(s, "%-8s (dfll) ",
- rail->reg_id);
- for (i = 0; i < d->num_freqs; i++) {
- int mv = d->dfll_millivolts[i];
- seq_printf(s, "%7d", mv);
- }
+ }
+
+ if (v_dfll && (last_v_dfll != v_dfll)) {
+ if (!mv_done) {
seq_printf(s, "\n");
+ mv_done = true;
}
+ last_v_dfll = v_dfll;
+ seq_printf(s, "%-8s (dfll) ", rail->reg_id);
+ for (i = 0; i < d->num_freqs; i++)
+ seq_printf(s, "%7d", v_dfll[i]);
+ seq_printf(s, "\n");
}
seq_printf(s, "%-16s", d->clk_name);
for (i = 0; i < d->num_freqs; i++) {
- unsigned int f = d->freqs[i]/100000;
+ unsigned long *freqs = dvfs_get_freqs(d);
+ unsigned int f = freqs[i]/100000;
seq_printf(s, " %4u.%u", f/10, f%10);
}
seq_printf(s, "\n");
}
- seq_printf(s, "\n");
}
mutex_unlock(&dvfs_lock);
if (!d)
return -ENOMEM;
- d = debugfs_create_file("gpu_dvfs", S_IRUGO | S_IWUSR,
- clk_debugfs_root, NULL, &gpu_dvfs_fops);
+ d = debugfs_create_file("vdd_cpu_mv", S_IRUGO, clk_debugfs_root,
+ tegra_cpu_rail, &rail_mv_fops);
+ if (!d)
+ return -ENOMEM;
+
+ d = debugfs_create_file("vdd_gpu_mv", S_IRUGO, clk_debugfs_root,
+ tegra_gpu_rail, &rail_mv_fops);
+ if (!d)
+ return -ENOMEM;
+
+ d = debugfs_create_file("vdd_core_mv", S_IRUGO, clk_debugfs_root,
+ tegra_core_rail, &rail_mv_fops);
+ if (!d)
+ return -ENOMEM;
+
+ d = debugfs_create_file("gpu_dvfs_t", S_IRUGO | S_IWUSR,
+ clk_debugfs_root, NULL, &gpu_dvfs_t_fops);
if (!d)
return -ENOMEM;