1297 lines
31 KiB
C
Executable File
1297 lines
31 KiB
C
Executable File
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Copyright (c) 2016-2021, The Linux Foundation. All rights reserved.
|
|
* Copyright (c) 2022-2024, Qualcomm Innovation Center, Inc. All rights reserved.
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/io.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/err.h>
|
|
#include <linux/of.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/regulator/debug-regulator.h>
|
|
#include <linux/regulator/driver.h>
|
|
#include <linux/regulator/machine.h>
|
|
#include <linux/regulator/of_regulator.h>
|
|
#include <linux/regulator/proxy-consumer.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/clk.h>
|
|
#include <linux/regmap.h>
|
|
#include <linux/reset.h>
|
|
#include <linux/mfd/syscon.h>
|
|
#include <linux/mailbox_client.h>
|
|
#include <linux/mailbox_controller.h>
|
|
#include <linux/mailbox/qmp.h>
|
|
#include <linux/interconnect.h>
|
|
#include <linux/pm.h>
|
|
#include <linux/suspend.h>
|
|
|
|
#include "../../regulator/internal.h"
|
|
#include "gdsc-debug.h"
|
|
|
|
#define CREATE_TRACE_POINTS
|
|
#include "trace-gdsc.h"
|
|
|
|
/* GDSCR */
|
|
#define PWR_ON_MASK BIT(31)
|
|
#define CLK_DIS_WAIT_MASK (0xF << 12)
|
|
#define CLK_DIS_WAIT_SHIFT (12)
|
|
#define RETAIN_FF_ENABLE_MASK BIT(11)
|
|
#define SW_OVERRIDE_MASK BIT(2)
|
|
#define HW_CONTROL_MASK BIT(1)
|
|
#define SW_COLLAPSE_MASK BIT(0)
|
|
|
|
/* CFG_GDSCR */
|
|
#define POWER_DOWN_COMPLETE_MASK BIT(15)
|
|
#define POWER_UP_COMPLETE_MASK BIT(16)
|
|
|
|
/* Domain Address */
|
|
#define GMEM_CLAMP_IO_MASK BIT(0)
|
|
#define GMEM_RESET_MASK BIT(4)
|
|
|
|
/* SW Reset */
|
|
#define BCR_BLK_ARES_BIT BIT(0)
|
|
|
|
/* Register Offset */
|
|
#define REG_OFFSET 0x0
|
|
#define CFG_GDSCR_OFFSET (REG_OFFSET + 0x4)
|
|
|
|
/* Timeout Delay */
|
|
#define TIMEOUT_US 1500
|
|
|
|
#define MBOX_TOUT_MS 500
|
|
|
|
struct collapse_vote {
|
|
struct regmap *regmap;
|
|
u32 vote_bit;
|
|
};
|
|
|
|
struct clk_ctrl {
|
|
struct regmap *regmap;
|
|
unsigned int offset;
|
|
unsigned int bit;
|
|
bool en_inverted;
|
|
};
|
|
|
|
struct gdsc {
|
|
struct regulator_dev *rdev;
|
|
struct regulator_desc rdesc;
|
|
void __iomem *gdscr;
|
|
struct regmap *regmap;
|
|
struct regmap *domain_addr;
|
|
struct regmap *hw_ctrl;
|
|
struct regmap **sw_resets;
|
|
struct clk_ctrl *clk_ctrl;
|
|
struct collapse_vote collapse_vote;
|
|
struct clk **clocks;
|
|
struct mbox_client mbox_client;
|
|
struct mbox_chan *mbox;
|
|
struct reset_control **reset_clocks;
|
|
struct icc_path **paths;
|
|
bool toggle_logic;
|
|
bool retain_ff_enable;
|
|
bool resets_asserted;
|
|
bool root_en;
|
|
bool force_root_en;
|
|
bool no_status_check_on_disable;
|
|
bool is_gdsc_enabled;
|
|
bool is_gdsc_hw_ctrl_mode;
|
|
bool is_root_clk_voted;
|
|
bool reset_aon;
|
|
bool pm_ops;
|
|
int clock_count;
|
|
int reset_count;
|
|
int root_clk_idx;
|
|
int sw_reset_count;
|
|
int path_count;
|
|
u32 clk_dis_wait_val;
|
|
int clk_ctrl_count;
|
|
u32 gds_timeout;
|
|
bool skip_disable_before_enable;
|
|
bool skip_disable;
|
|
bool bypass_skip_disable;
|
|
bool cfg_gdscr;
|
|
};
|
|
|
|
enum gdscr_status {
|
|
DISABLED,
|
|
ENABLED,
|
|
};
|
|
|
|
static inline u32 gdsc_mb(struct gdsc *gds)
|
|
{
|
|
u32 reg;
|
|
|
|
regmap_read(gds->regmap, REG_OFFSET, ®);
|
|
return reg;
|
|
}
|
|
|
|
static int poll_gdsc_status(struct gdsc *sc, enum gdscr_status status)
|
|
{
|
|
struct regmap *regmap;
|
|
int count = sc->gds_timeout;
|
|
u32 val, reg_offset;
|
|
|
|
if (sc->hw_ctrl && !sc->cfg_gdscr)
|
|
regmap = sc->hw_ctrl;
|
|
else
|
|
regmap = sc->regmap;
|
|
|
|
if (sc->cfg_gdscr)
|
|
reg_offset = CFG_GDSCR_OFFSET;
|
|
else
|
|
reg_offset = REG_OFFSET;
|
|
|
|
for (; count > 0; count--) {
|
|
regmap_read(regmap, reg_offset, &val);
|
|
|
|
switch (status) {
|
|
case ENABLED:
|
|
if (sc->cfg_gdscr)
|
|
val &= POWER_UP_COMPLETE_MASK;
|
|
else
|
|
val &= PWR_ON_MASK;
|
|
break;
|
|
case DISABLED:
|
|
if (sc->cfg_gdscr) {
|
|
val &= POWER_DOWN_COMPLETE_MASK;
|
|
} else {
|
|
val &= PWR_ON_MASK;
|
|
val = !val;
|
|
}
|
|
break;
|
|
}
|
|
|
|
if (val) {
|
|
trace_gdsc_time(sc->rdesc.name, status,
|
|
sc->gds_timeout - count, 0);
|
|
return 0;
|
|
}
|
|
/*
|
|
* There is no guarantee about the delay needed for the enable
|
|
* bit in the GDSCR to be set or reset after the GDSC state
|
|
* changes. Hence, keep on checking for a reasonable number
|
|
* of times until the bit is set with the least possible delay
|
|
* between successive tries.
|
|
*/
|
|
udelay(1);
|
|
}
|
|
|
|
trace_gdsc_time(sc->rdesc.name, status,
|
|
sc->gds_timeout - count, 1);
|
|
return -ETIMEDOUT;
|
|
}
|
|
|
|
static int check_gdsc_status(struct gdsc *sc, struct device *pdev,
|
|
enum gdscr_status state)
|
|
{
|
|
u32 regval, hw_ctrl_regval;
|
|
int ret;
|
|
static const char * const gdsc_states[] = {
|
|
"disable",
|
|
"enable",
|
|
};
|
|
|
|
if (!sc || !sc->regmap || !pdev)
|
|
return -EINVAL;
|
|
|
|
ret = poll_gdsc_status(sc, state);
|
|
if (ret) {
|
|
regmap_read(sc->regmap, REG_OFFSET, ®val);
|
|
if (sc->hw_ctrl) {
|
|
regmap_read(sc->hw_ctrl, REG_OFFSET,
|
|
&hw_ctrl_regval);
|
|
dev_warn(pdev, "%s %s state (after %d us timeout): 0x%x, GDS_HW_CTRL: 0x%x. Re-polling.\n",
|
|
sc->rdesc.name, gdsc_states[state], sc->gds_timeout,
|
|
regval, hw_ctrl_regval);
|
|
|
|
ret = poll_gdsc_status(sc, state);
|
|
if (ret) {
|
|
regmap_read(sc->regmap, REG_OFFSET,
|
|
®val);
|
|
regmap_read(sc->hw_ctrl, REG_OFFSET,
|
|
&hw_ctrl_regval);
|
|
dev_err(pdev, "%s %s final state (after additional %d us timeout): 0x%x, GDS_HW_CTRL: 0x%x\n",
|
|
sc->rdesc.name, gdsc_states[state], sc->gds_timeout,
|
|
regval, hw_ctrl_regval);
|
|
return ret;
|
|
}
|
|
} else {
|
|
dev_err(pdev, "%s %s timed out: 0x%x\n",
|
|
sc->rdesc.name, gdsc_states[state], regval);
|
|
|
|
udelay(sc->gds_timeout);
|
|
regmap_read(sc->regmap, REG_OFFSET, ®val);
|
|
dev_err(pdev, "%s %s final state: 0x%x (%d us timeout)\n",
|
|
sc->rdesc.name, gdsc_states[state], regval, sc->gds_timeout);
|
|
return ret;
|
|
}
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static int gdsc_init_is_enabled(struct gdsc *sc)
|
|
{
|
|
struct regmap *regmap;
|
|
uint32_t regval, mask;
|
|
int ret;
|
|
|
|
if (!sc->toggle_logic) {
|
|
sc->is_gdsc_enabled = !sc->resets_asserted;
|
|
return 0;
|
|
}
|
|
|
|
if (sc->collapse_vote.regmap) {
|
|
regmap = sc->collapse_vote.regmap;
|
|
mask = BIT(sc->collapse_vote.vote_bit);
|
|
} else {
|
|
regmap = sc->regmap;
|
|
mask = SW_COLLAPSE_MASK;
|
|
}
|
|
|
|
ret = regmap_read(regmap, REG_OFFSET, ®val);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
sc->is_gdsc_enabled = !(regval & mask);
|
|
|
|
if (sc->is_gdsc_enabled && sc->retain_ff_enable)
|
|
regmap_update_bits(sc->regmap, REG_OFFSET,
|
|
RETAIN_FF_ENABLE_MASK, RETAIN_FF_ENABLE_MASK);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int gdsc_is_enabled(struct regulator_dev *rdev)
|
|
{
|
|
struct gdsc *sc = rdev_get_drvdata(rdev);
|
|
|
|
/*
|
|
* Return the logical GDSC enable state given that it will only be
|
|
* physically disabled by AOP during system sleep.
|
|
*/
|
|
if (sc->skip_disable)
|
|
return sc->is_gdsc_enabled;
|
|
|
|
if (!sc->toggle_logic)
|
|
return !sc->resets_asserted;
|
|
|
|
if (sc->skip_disable_before_enable)
|
|
return false;
|
|
|
|
return sc->is_gdsc_enabled;
|
|
}
|
|
|
|
#define MAX_LEN 96
|
|
|
|
static int gdsc_qmp_enable(struct gdsc *sc)
|
|
{
|
|
char buf[MAX_LEN] = "{class: clock, res: gpu_noc_wa}";
|
|
struct qmp_pkt pkt;
|
|
uint32_t regval;
|
|
int ret;
|
|
|
|
regmap_read(sc->regmap, REG_OFFSET, ®val);
|
|
if (!(regval & SW_COLLAPSE_MASK)) {
|
|
/*
|
|
* Do not enable via a QMP request if the GDSC is already
|
|
* enabled by software.
|
|
*/
|
|
return 0;
|
|
}
|
|
|
|
pkt.size = MAX_LEN;
|
|
pkt.data = buf;
|
|
|
|
ret = mbox_send_message(sc->mbox, &pkt);
|
|
if (ret < 0)
|
|
dev_err(&sc->rdev->dev, "qmp message send failed, ret=%d\n",
|
|
ret);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void gdsc_clk_ctrl(struct gdsc *sc, bool en)
|
|
{
|
|
uint32_t clk_ctrl_mask, clk_ctrl_val;
|
|
int i;
|
|
|
|
for (i = 0; i < sc->clk_ctrl_count; i++) {
|
|
clk_ctrl_mask = BIT(sc->clk_ctrl[i].bit);
|
|
|
|
if (sc->clk_ctrl[i].en_inverted ^ en)
|
|
clk_ctrl_val = clk_ctrl_mask;
|
|
else
|
|
clk_ctrl_val = 0;
|
|
|
|
regmap_update_bits(sc->clk_ctrl[i].regmap,
|
|
sc->clk_ctrl[i].offset, clk_ctrl_mask, clk_ctrl_val);
|
|
}
|
|
}
|
|
|
|
static int gdsc_enable(struct regulator_dev *rdev)
|
|
{
|
|
struct gdsc *sc = rdev_get_drvdata(rdev);
|
|
u32 regval;
|
|
int i, ret = 0;
|
|
|
|
if (sc->skip_disable_before_enable)
|
|
return 0;
|
|
|
|
if (sc->root_en || sc->force_root_en) {
|
|
clk_prepare_enable(sc->clocks[sc->root_clk_idx]);
|
|
sc->is_root_clk_voted = true;
|
|
}
|
|
|
|
regmap_read(sc->regmap, REG_OFFSET, ®val);
|
|
if (regval & HW_CONTROL_MASK) {
|
|
dev_warn(&rdev->dev, "Invalid enable while %s is under HW control\n",
|
|
sc->rdesc.name);
|
|
return -EBUSY;
|
|
}
|
|
|
|
if (sc->clk_ctrl_count)
|
|
gdsc_clk_ctrl(sc, true);
|
|
|
|
if (sc->toggle_logic) {
|
|
for (i = 0; i < sc->path_count; i++) {
|
|
ret = icc_set_bw(sc->paths[i], 1, 1);
|
|
if (ret) {
|
|
dev_err(&rdev->dev, "Failed to vote BW for %d, ret=%d\n", i, ret);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
if (sc->sw_reset_count) {
|
|
for (i = 0; i < sc->sw_reset_count; i++)
|
|
regmap_set_bits(sc->sw_resets[i], REG_OFFSET,
|
|
BCR_BLK_ARES_BIT);
|
|
|
|
/*
|
|
* BLK_ARES should be kept asserted for 1us before
|
|
* being de-asserted.
|
|
*/
|
|
gdsc_mb(sc);
|
|
udelay(1);
|
|
|
|
for (i = 0; i < sc->sw_reset_count; i++)
|
|
regmap_clear_bits(sc->sw_resets[i], REG_OFFSET,
|
|
BCR_BLK_ARES_BIT);
|
|
|
|
/* Make sure de-assert goes through before continuing */
|
|
gdsc_mb(sc);
|
|
}
|
|
|
|
if (sc->domain_addr) {
|
|
if (sc->reset_aon) {
|
|
regmap_read(sc->domain_addr, REG_OFFSET,
|
|
®val);
|
|
regval |= GMEM_RESET_MASK;
|
|
regmap_write(sc->domain_addr, REG_OFFSET,
|
|
regval);
|
|
/*
|
|
* Keep reset asserted for at-least 1us before
|
|
* continuing.
|
|
*/
|
|
gdsc_mb(sc);
|
|
udelay(1);
|
|
|
|
regval &= ~GMEM_RESET_MASK;
|
|
regmap_write(sc->domain_addr, REG_OFFSET,
|
|
regval);
|
|
/*
|
|
* Make sure GMEM_RESET is de-asserted before
|
|
* continuing.
|
|
*/
|
|
gdsc_mb(sc);
|
|
}
|
|
|
|
regmap_read(sc->domain_addr, REG_OFFSET, ®val);
|
|
regval &= ~GMEM_CLAMP_IO_MASK;
|
|
regmap_write(sc->domain_addr, REG_OFFSET, regval);
|
|
|
|
/*
|
|
* Make sure CLAMP_IO is de-asserted before continuing.
|
|
*/
|
|
gdsc_mb(sc);
|
|
}
|
|
|
|
/* Enable gdsc */
|
|
if (sc->mbox) {
|
|
ret = gdsc_qmp_enable(sc);
|
|
if (ret < 0)
|
|
return ret;
|
|
} else if (sc->collapse_vote.regmap) {
|
|
regmap_update_bits(sc->collapse_vote.regmap, REG_OFFSET,
|
|
BIT(sc->collapse_vote.vote_bit),
|
|
~BIT(sc->collapse_vote.vote_bit));
|
|
} else {
|
|
regmap_read(sc->regmap, REG_OFFSET, ®val);
|
|
regval &= ~SW_COLLAPSE_MASK;
|
|
regmap_write(sc->regmap, REG_OFFSET, regval);
|
|
}
|
|
|
|
/* Wait for 8 XO cycles before polling the status bit. */
|
|
gdsc_mb(sc);
|
|
udelay(1);
|
|
|
|
ret = check_gdsc_status(sc, &rdev->dev, ENABLED);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (sc->retain_ff_enable && !(regval & RETAIN_FF_ENABLE_MASK)) {
|
|
regval |= RETAIN_FF_ENABLE_MASK;
|
|
regmap_write(sc->regmap, REG_OFFSET, regval);
|
|
}
|
|
} else {
|
|
for (i = 0; i < sc->reset_count; i++)
|
|
reset_control_deassert(sc->reset_clocks[i]);
|
|
sc->resets_asserted = false;
|
|
}
|
|
|
|
/*
|
|
* If clocks to this power domain were already on, they will take an
|
|
* additional 4 clock cycles to re-enable after the rail is enabled.
|
|
* Delay to account for this. A delay is also needed to ensure clocks
|
|
* are not enabled within 400ns of enabling power to the memories.
|
|
*/
|
|
udelay(1);
|
|
|
|
/* Delay to account for staggered memory powerup. */
|
|
udelay(1);
|
|
|
|
if (sc->force_root_en) {
|
|
clk_disable_unprepare(sc->clocks[sc->root_clk_idx]);
|
|
sc->is_root_clk_voted = false;
|
|
}
|
|
|
|
sc->is_gdsc_enabled = true;
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int gdsc_disable(struct regulator_dev *rdev)
|
|
{
|
|
struct gdsc *sc = rdev_get_drvdata(rdev);
|
|
struct regulator_dev *parent_rdev;
|
|
uint32_t regval;
|
|
int i, ret = 0;
|
|
bool lock = false;
|
|
|
|
if (rdev->supply) {
|
|
parent_rdev = rdev->supply->rdev;
|
|
|
|
/*
|
|
* At this point, it can be assumed that parent supply's mutex is always
|
|
* locked by regulator framework before calling this callback but there
|
|
* are code paths where it isn't locked (e.g regulator_late_cleanup()).
|
|
*
|
|
* If parent supply is not locked, lock the parent supply mutex before
|
|
* checking it's enable count, so it won't get disabled while in the
|
|
* middle of GDSC operations
|
|
*/
|
|
if (ww_mutex_trylock(&parent_rdev->mutex, NULL))
|
|
lock = true;
|
|
|
|
if (!parent_rdev->use_count) {
|
|
dev_err(&rdev->dev, "%s cannot disable GDSC while parent is disabled\n",
|
|
sc->rdesc.name);
|
|
ret = -EIO;
|
|
goto done;
|
|
}
|
|
}
|
|
|
|
regmap_read(sc->regmap, REG_OFFSET, ®val);
|
|
if (regval & HW_CONTROL_MASK) {
|
|
dev_warn(&rdev->dev, "Invalid Disable while %s is under HW control\n",
|
|
sc->rdesc.name);
|
|
return -EBUSY;
|
|
}
|
|
|
|
if (sc->clk_ctrl_count)
|
|
gdsc_clk_ctrl(sc, false);
|
|
|
|
if (sc->force_root_en) {
|
|
clk_prepare_enable(sc->clocks[sc->root_clk_idx]);
|
|
sc->is_root_clk_voted = true;
|
|
}
|
|
|
|
/* Delay to account for staggered memory powerdown. */
|
|
udelay(1);
|
|
|
|
if (sc->skip_disable && !sc->bypass_skip_disable) {
|
|
/*
|
|
* Don't change the GDSCR register state on disable. AOP will
|
|
* handle this during system sleep.
|
|
*/
|
|
} else if (sc->toggle_logic) {
|
|
/* Disable gdsc */
|
|
if (sc->collapse_vote.regmap) {
|
|
regmap_update_bits(sc->collapse_vote.regmap, REG_OFFSET,
|
|
BIT(sc->collapse_vote.vote_bit),
|
|
BIT(sc->collapse_vote.vote_bit));
|
|
} else {
|
|
regmap_read(sc->regmap, REG_OFFSET, ®val);
|
|
regval |= SW_COLLAPSE_MASK;
|
|
regmap_write(sc->regmap, REG_OFFSET, regval);
|
|
}
|
|
|
|
/* Wait for 8 XO cycles before polling the status bit. */
|
|
gdsc_mb(sc);
|
|
udelay(1);
|
|
|
|
if (sc->no_status_check_on_disable) {
|
|
/*
|
|
* Add a short delay here to ensure that gdsc_enable
|
|
* right after it was disabled does not put it in a
|
|
* weird state.
|
|
*/
|
|
udelay(100);
|
|
} else {
|
|
ret = check_gdsc_status(sc, &rdev->dev, DISABLED);
|
|
if (ret)
|
|
goto done;
|
|
}
|
|
|
|
if (sc->domain_addr) {
|
|
regmap_read(sc->domain_addr, REG_OFFSET, ®val);
|
|
regval |= GMEM_CLAMP_IO_MASK;
|
|
regmap_write(sc->domain_addr, REG_OFFSET, regval);
|
|
}
|
|
|
|
for (i = 0; i < sc->path_count; i++) {
|
|
ret = icc_set_bw(sc->paths[i], 0, 0);
|
|
if (ret) {
|
|
dev_err(&rdev->dev, "Failed to unvote BW for %d: %d\n", i, ret);
|
|
goto done;
|
|
}
|
|
}
|
|
} else {
|
|
for (i = sc->reset_count - 1; i >= 0; i--)
|
|
reset_control_assert(sc->reset_clocks[i]);
|
|
sc->resets_asserted = true;
|
|
}
|
|
|
|
/*
|
|
* Check if gdsc_enable was called for this GDSC. If not, the root
|
|
* clock will not have been enabled prior to this.
|
|
*/
|
|
if ((sc->is_root_clk_voted && sc->root_en) || sc->force_root_en) {
|
|
clk_disable_unprepare(sc->clocks[sc->root_clk_idx]);
|
|
sc->is_root_clk_voted = false;
|
|
}
|
|
|
|
sc->is_gdsc_enabled = false;
|
|
|
|
done:
|
|
if (rdev->supply && lock)
|
|
ww_mutex_unlock(&parent_rdev->mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int gdsc_init_hw_ctrl_mode(struct gdsc *sc)
|
|
{
|
|
uint32_t regval;
|
|
int ret;
|
|
|
|
ret = regmap_read(sc->regmap, REG_OFFSET, ®val);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
sc->is_gdsc_hw_ctrl_mode = regval & HW_CONTROL_MASK;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static unsigned int gdsc_get_mode(struct regulator_dev *rdev)
|
|
{
|
|
struct gdsc *sc = rdev_get_drvdata(rdev);
|
|
|
|
if (sc->skip_disable) {
|
|
if (sc->bypass_skip_disable)
|
|
return REGULATOR_MODE_IDLE;
|
|
return REGULATOR_MODE_NORMAL;
|
|
}
|
|
|
|
return sc->is_gdsc_hw_ctrl_mode ? REGULATOR_MODE_FAST
|
|
: REGULATOR_MODE_NORMAL;
|
|
}
|
|
|
|
static int gdsc_set_mode(struct regulator_dev *rdev, unsigned int mode)
|
|
{
|
|
struct gdsc *sc = rdev_get_drvdata(rdev);
|
|
struct regulator_dev *parent_rdev;
|
|
uint32_t regval;
|
|
int ret = 0;
|
|
|
|
if (sc->skip_disable) {
|
|
switch (mode) {
|
|
case REGULATOR_MODE_IDLE:
|
|
sc->bypass_skip_disable = true;
|
|
break;
|
|
case REGULATOR_MODE_NORMAL:
|
|
sc->bypass_skip_disable = false;
|
|
break;
|
|
default:
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
if (rdev->supply) {
|
|
parent_rdev = rdev->supply->rdev;
|
|
/*
|
|
* Ensure that the GDSC parent supply is enabled before
|
|
* continuing. This is needed to avoid an unclocked access
|
|
* of the GDSC control register for GDSCs whose register access
|
|
* is gated by the parent supply enable state in hardware.
|
|
*/
|
|
ww_mutex_lock(&parent_rdev->mutex, NULL);
|
|
|
|
if (!parent_rdev->use_count) {
|
|
dev_err(&rdev->dev,
|
|
"%s cannot change GDSC HW/SW control mode while parent is disabled\n",
|
|
sc->rdesc.name);
|
|
ret = -EIO;
|
|
goto done;
|
|
}
|
|
}
|
|
|
|
ret = regmap_read(sc->regmap, REG_OFFSET, ®val);
|
|
if (ret < 0)
|
|
goto done;
|
|
|
|
switch (mode) {
|
|
case REGULATOR_MODE_FAST:
|
|
/* Turn on HW trigger mode */
|
|
regval |= HW_CONTROL_MASK;
|
|
ret = regmap_write(sc->regmap, REG_OFFSET, regval);
|
|
if (ret < 0)
|
|
goto done;
|
|
/*
|
|
* There may be a race with internal HW trigger signal,
|
|
* that will result in GDSC going through a power down and
|
|
* up cycle. In case HW trigger signal is controlled by
|
|
* firmware that also poll same status bits as we do, FW
|
|
* might read an 'on' status before the GDSC can finish
|
|
* power cycle. We wait 1us before returning to ensure
|
|
* FW can't immediately poll the status bit.
|
|
*/
|
|
gdsc_mb(sc);
|
|
udelay(1);
|
|
sc->is_gdsc_hw_ctrl_mode = true;
|
|
break;
|
|
case REGULATOR_MODE_NORMAL:
|
|
/* Turn off HW trigger mode */
|
|
regval &= ~HW_CONTROL_MASK;
|
|
ret = regmap_write(sc->regmap, REG_OFFSET, regval);
|
|
if (ret < 0)
|
|
goto done;
|
|
/*
|
|
* There may be a race with internal HW trigger signal,
|
|
* that will result in GDSC going through a power down and
|
|
* up cycle. Account for this case by waiting 1us before
|
|
* proceeding.
|
|
*/
|
|
gdsc_mb(sc);
|
|
udelay(1);
|
|
/*
|
|
* While switching from HW to SW mode, HW may be busy
|
|
* updating internal required signals. Polling for PWR_ON
|
|
* ensures that the GDSC switches to SW mode before software
|
|
* starts to use SW mode.
|
|
*/
|
|
if (sc->is_gdsc_enabled) {
|
|
ret = check_gdsc_status(sc, &rdev->dev, ENABLED);
|
|
if (ret)
|
|
goto done;
|
|
}
|
|
sc->is_gdsc_hw_ctrl_mode = false;
|
|
break;
|
|
default:
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
|
|
done:
|
|
if (rdev->supply)
|
|
ww_mutex_unlock(&parent_rdev->mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static const struct regulator_ops gdsc_ops = {
|
|
.is_enabled = gdsc_is_enabled,
|
|
.enable = gdsc_enable,
|
|
.disable = gdsc_disable,
|
|
.set_mode = gdsc_set_mode,
|
|
.get_mode = gdsc_get_mode,
|
|
};
|
|
|
|
static struct regmap_config gdsc_regmap_config = {
|
|
.reg_bits = 32,
|
|
.reg_stride = 4,
|
|
.val_bits = 32,
|
|
.max_register = 0x8,
|
|
.fast_io = true,
|
|
};
|
|
|
|
void gdsc_debug_print_regs(struct regulator *regulator)
|
|
{
|
|
struct regulator_dev *rdev = regulator->rdev;
|
|
struct gdsc *sc = rdev_get_drvdata(rdev);
|
|
struct regulator *reg;
|
|
const char *supply_name;
|
|
uint32_t regvals[3] = {0};
|
|
int ret;
|
|
|
|
if (!sc) {
|
|
pr_err("Failed to get GDSC Handle\n");
|
|
return;
|
|
}
|
|
|
|
ww_mutex_lock(&rdev->mutex, NULL);
|
|
|
|
if (rdev->open_count)
|
|
pr_info("%-32s EN\n", "Device-Supply");
|
|
|
|
list_for_each_entry(reg, &rdev->consumer_list, list) {
|
|
if (reg->supply_name)
|
|
supply_name = reg->supply_name;
|
|
else
|
|
supply_name = "(null)-(null)";
|
|
|
|
pr_info("%-32s %c\n", supply_name,
|
|
(reg->enable_count ? 'Y' : 'N'));
|
|
}
|
|
|
|
ww_mutex_unlock(&rdev->mutex);
|
|
|
|
ret = regmap_bulk_read(sc->regmap, REG_OFFSET, regvals,
|
|
gdsc_regmap_config.max_register ? 3 : 1);
|
|
if (ret) {
|
|
pr_err("Failed to read %s registers\n", sc->rdesc.name);
|
|
return;
|
|
}
|
|
|
|
pr_info("Dumping %s Registers:\n", sc->rdesc.name);
|
|
pr_info("GDSCR: 0x%.8x CFG: 0x%.8x CFG2: 0x%.8x\n",
|
|
regvals[0], regvals[1], regvals[2]);
|
|
}
|
|
EXPORT_SYMBOL_GPL(gdsc_debug_print_regs);
|
|
|
|
static int gdsc_parse_dt_clk_ctrl_data(struct gdsc *sc, struct device *dev)
|
|
{
|
|
struct of_phandle_args args;
|
|
int ret, i, clk_ctrl_len;
|
|
|
|
if (of_find_property(dev->of_node, "qcom,clk-ctrl", NULL)) {
|
|
|
|
clk_ctrl_len = of_count_phandle_with_args(dev->of_node, "qcom,clk-ctrl", NULL);
|
|
|
|
if (clk_ctrl_len % 4) {
|
|
dev_err(dev, "Invalid length of clk-ctrl arguments\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
sc->clk_ctrl_count = clk_ctrl_len / 4;
|
|
|
|
sc->clk_ctrl = devm_kmalloc_array(dev, sc->clk_ctrl_count,
|
|
sizeof(*sc->clk_ctrl), GFP_KERNEL);
|
|
if (!sc->clk_ctrl)
|
|
return -ENOMEM;
|
|
|
|
for (i = 0; i < sc->clk_ctrl_count; i++) {
|
|
ret = of_parse_phandle_with_fixed_args(dev->of_node,
|
|
"qcom,clk-ctrl", 3, i, &args);
|
|
if (ret) {
|
|
dev_err(dev, "Failed to get clk-ctrl arguments for index:%d\n", i);
|
|
return ret;
|
|
}
|
|
|
|
sc->clk_ctrl[i].regmap = syscon_node_to_regmap(args.np);
|
|
of_node_put(args.np);
|
|
if (IS_ERR(sc->clk_ctrl[i].regmap)) {
|
|
dev_err(dev, "Failed to get clk-ctrl regmap for index:%d\n", i);
|
|
return PTR_ERR(sc->clk_ctrl[i].regmap);
|
|
}
|
|
|
|
sc->clk_ctrl[i].offset = args.args[0];
|
|
sc->clk_ctrl[i].bit = args.args[1];
|
|
sc->clk_ctrl[i].en_inverted = !!args.args[2];
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int gdsc_parse_dt_data(struct gdsc *sc, struct device *dev,
|
|
struct regulator_init_data **init_data)
|
|
{
|
|
struct device_node *np;
|
|
int ret, i;
|
|
|
|
*init_data = of_get_regulator_init_data(dev, dev->of_node, &sc->rdesc);
|
|
if (*init_data == NULL)
|
|
return -ENOMEM;
|
|
|
|
if (of_get_property(dev->of_node, "parent-supply", NULL))
|
|
(*init_data)->supply_regulator = "parent";
|
|
|
|
ret = of_property_read_string(dev->of_node, "regulator-name",
|
|
&sc->rdesc.name);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (of_find_property(dev->of_node, "domain-addr", NULL)) {
|
|
sc->domain_addr = syscon_regmap_lookup_by_phandle(dev->of_node,
|
|
"domain-addr");
|
|
if (IS_ERR(sc->domain_addr))
|
|
return PTR_ERR(sc->domain_addr);
|
|
}
|
|
|
|
if (of_find_property(dev->of_node, "sw-reset", NULL)) {
|
|
sc->sw_reset_count = of_count_phandle_with_args(dev->of_node,
|
|
"sw-reset", NULL);
|
|
sc->sw_resets = devm_kmalloc_array(dev, sc->sw_reset_count,
|
|
sizeof(*sc->sw_resets), GFP_KERNEL);
|
|
if (!sc->sw_resets)
|
|
return -ENOMEM;
|
|
|
|
for (i = 0; i < sc->sw_reset_count; i++) {
|
|
np = of_parse_phandle(dev->of_node, "sw-reset", i);
|
|
if (!np)
|
|
return -ENODEV;
|
|
|
|
sc->sw_resets[i] = syscon_node_to_regmap(np);
|
|
of_node_put(np);
|
|
if (IS_ERR(sc->sw_resets[i]))
|
|
return PTR_ERR(sc->sw_resets[i]);
|
|
}
|
|
}
|
|
|
|
if (of_find_property(dev->of_node, "hw-ctrl-addr", NULL)) {
|
|
sc->hw_ctrl = syscon_regmap_lookup_by_phandle(dev->of_node,
|
|
"hw-ctrl-addr");
|
|
if (IS_ERR(sc->hw_ctrl))
|
|
return PTR_ERR(sc->hw_ctrl);
|
|
}
|
|
|
|
|
|
ret = gdsc_parse_dt_clk_ctrl_data(sc, dev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
sc->gds_timeout = TIMEOUT_US;
|
|
of_property_read_u32(dev->of_node, "qcom,gds-timeout",
|
|
&sc->gds_timeout);
|
|
|
|
sc->clock_count = of_property_count_strings(dev->of_node,
|
|
"clock-names");
|
|
if (sc->clock_count == -EINVAL) {
|
|
sc->clock_count = 0;
|
|
} else if (sc->clock_count < 0) {
|
|
dev_err(dev, "Failed to get clock names, ret=%d\n",
|
|
sc->clock_count);
|
|
return sc->clock_count;
|
|
}
|
|
|
|
sc->path_count = of_property_count_strings(dev->of_node,
|
|
"interconnect-names");
|
|
if (sc->path_count == -EINVAL) {
|
|
sc->path_count = 0;
|
|
} else if (sc->path_count < 0) {
|
|
dev_err(dev, "Failed to get interconnect names, ret=%d\n",
|
|
sc->path_count);
|
|
return sc->path_count;
|
|
}
|
|
|
|
sc->root_en = of_property_read_bool(dev->of_node,
|
|
"qcom,enable-root-clk");
|
|
sc->force_root_en = of_property_read_bool(dev->of_node,
|
|
"qcom,force-enable-root-clk");
|
|
sc->reset_aon = of_property_read_bool(dev->of_node,
|
|
"qcom,reset-aon-logic");
|
|
sc->no_status_check_on_disable = of_property_read_bool(dev->of_node,
|
|
"qcom,no-status-check-on-disable");
|
|
sc->retain_ff_enable = of_property_read_bool(dev->of_node,
|
|
"qcom,retain-regs");
|
|
sc->skip_disable_before_enable = of_property_read_bool(dev->of_node,
|
|
"qcom,skip-disable-before-sw-enable");
|
|
|
|
sc->cfg_gdscr = of_property_read_bool(dev->of_node,
|
|
"qcom,support-cfg-gdscr");
|
|
|
|
if (of_find_property(dev->of_node, "qcom,collapse-vote", NULL)) {
|
|
ret = of_property_count_u32_elems(dev->of_node,
|
|
"qcom,collapse-vote");
|
|
if (ret != 2) {
|
|
dev_err(dev, "qcom,collapse-vote needs two values\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
sc->collapse_vote.regmap =
|
|
syscon_regmap_lookup_by_phandle(dev->of_node,
|
|
"qcom,collapse-vote");
|
|
if (IS_ERR(sc->collapse_vote.regmap))
|
|
return PTR_ERR(sc->collapse_vote.regmap);
|
|
ret = of_property_read_u32_index(dev->of_node,
|
|
"qcom,collapse-vote", 1,
|
|
&sc->collapse_vote.vote_bit);
|
|
if (ret || sc->collapse_vote.vote_bit > 31) {
|
|
dev_err(dev, "qcom,collapse-vote vote_bit error\n");
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
sc->skip_disable = of_property_read_bool(dev->of_node,
|
|
"qcom,skip-disable");
|
|
if (sc->skip_disable) {
|
|
/*
|
|
* If the disable skipping feature is allowed, then use mode
|
|
* control to enable and disable the feature at runtime instead
|
|
* of using it to enable and disable hardware triggering.
|
|
*/
|
|
(*init_data)->constraints.valid_ops_mask |=
|
|
REGULATOR_CHANGE_MODE;
|
|
(*init_data)->constraints.valid_modes_mask =
|
|
REGULATOR_MODE_NORMAL | REGULATOR_MODE_IDLE;
|
|
}
|
|
|
|
sc->toggle_logic = !of_property_read_bool(dev->of_node,
|
|
"qcom,skip-logic-collapse");
|
|
if (!sc->toggle_logic) {
|
|
sc->reset_count = of_property_count_strings(dev->of_node,
|
|
"reset-names");
|
|
if (sc->reset_count == -EINVAL) {
|
|
sc->reset_count = 0;
|
|
} else if (sc->reset_count < 0) {
|
|
dev_err(dev, "Failed to get reset clock names\n");
|
|
return sc->reset_count;
|
|
}
|
|
}
|
|
|
|
if (of_find_property(dev->of_node, "qcom,support-hw-trigger", NULL)) {
|
|
(*init_data)->constraints.valid_ops_mask |=
|
|
REGULATOR_CHANGE_MODE;
|
|
(*init_data)->constraints.valid_modes_mask |=
|
|
REGULATOR_MODE_NORMAL | REGULATOR_MODE_FAST;
|
|
sc->pm_ops = true;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int gdsc_get_resources(struct gdsc *sc, struct platform_device *pdev)
|
|
{
|
|
struct device *dev = &pdev->dev;
|
|
struct resource *res;
|
|
int ret, i;
|
|
|
|
res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
|
|
if (res == NULL) {
|
|
dev_err(dev, "Failed to get address resource\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
sc->gdscr = devm_ioremap(dev, res->start, resource_size(res));
|
|
if (sc->gdscr == NULL)
|
|
return -ENOMEM;
|
|
|
|
if (of_property_read_bool(dev->of_node, "qcom,no-config-gdscr"))
|
|
gdsc_regmap_config.max_register = 0;
|
|
|
|
sc->regmap = devm_regmap_init_mmio(dev, sc->gdscr, &gdsc_regmap_config);
|
|
if (!sc->regmap) {
|
|
dev_err(dev, "Couldn't get regmap\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
sc->clocks = devm_kcalloc(dev, sc->clock_count, sizeof(*sc->clocks),
|
|
GFP_KERNEL);
|
|
if (sc->clock_count && !sc->clocks)
|
|
return -ENOMEM;
|
|
|
|
sc->root_clk_idx = -1;
|
|
for (i = 0; i < sc->clock_count; i++) {
|
|
const char *clock_name;
|
|
|
|
of_property_read_string_index(dev->of_node, "clock-names", i,
|
|
&clock_name);
|
|
|
|
sc->clocks[i] = devm_clk_get(dev, clock_name);
|
|
if (IS_ERR(sc->clocks[i])) {
|
|
ret = PTR_ERR(sc->clocks[i]);
|
|
if (ret != -EPROBE_DEFER)
|
|
dev_err(dev, "Failed to get %s, ret=%d\n",
|
|
clock_name, ret);
|
|
return ret;
|
|
}
|
|
|
|
if (!strcmp(clock_name, "core_root_clk"))
|
|
sc->root_clk_idx = i;
|
|
}
|
|
|
|
sc->paths = devm_kcalloc(dev, sc->path_count, sizeof(*sc->paths), GFP_KERNEL);
|
|
if (sc->path_count && !sc->paths)
|
|
return -ENOMEM;
|
|
|
|
for (i = 0; i < sc->path_count; i++) {
|
|
const char *name;
|
|
|
|
of_property_read_string_index(dev->of_node, "interconnect-names", i,
|
|
&name);
|
|
|
|
sc->paths[i] = of_icc_get(dev, name);
|
|
if (IS_ERR(sc->paths[i])) {
|
|
ret = PTR_ERR(sc->paths[i]);
|
|
if (ret != -EPROBE_DEFER)
|
|
dev_err(dev, "Failed to get path %s, ret=%d\n",
|
|
name, ret);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
if ((sc->root_en || sc->force_root_en) && (sc->root_clk_idx == -1)) {
|
|
dev_err(dev, "Failed to get root clock name\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!sc->toggle_logic) {
|
|
sc->reset_clocks = devm_kcalloc(&pdev->dev, sc->reset_count,
|
|
sizeof(*sc->reset_clocks),
|
|
GFP_KERNEL);
|
|
if (sc->reset_count && !sc->reset_clocks)
|
|
return -ENOMEM;
|
|
|
|
for (i = 0; i < sc->reset_count; i++) {
|
|
const char *reset_name;
|
|
|
|
of_property_read_string_index(pdev->dev.of_node,
|
|
"reset-names", i, &reset_name);
|
|
sc->reset_clocks[i] = devm_reset_control_get(&pdev->dev,
|
|
reset_name);
|
|
if (IS_ERR(sc->reset_clocks[i])) {
|
|
ret = PTR_ERR(sc->reset_clocks[i]);
|
|
if (ret != -EPROBE_DEFER)
|
|
dev_err(&pdev->dev, "Failed to get %s, ret=%d\n",
|
|
reset_name, ret);
|
|
return ret;
|
|
}
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int restore_hw_trig_clk_dis(struct device *dev)
|
|
{
|
|
struct gdsc *sc = dev_get_drvdata(dev);
|
|
uint32_t regval;
|
|
int ret;
|
|
|
|
if (sc->rdev->supply) {
|
|
ret = regulator_enable(sc->rdev->supply);
|
|
if (ret) {
|
|
dev_err(&sc->rdev->dev, "reg enable failed\n");
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
regmap_read(sc->regmap, REG_OFFSET, ®val);
|
|
if (sc->is_gdsc_hw_ctrl_mode)
|
|
regval |= HW_CONTROL_MASK;
|
|
|
|
if (sc->clk_dis_wait_val) {
|
|
regval &= ~(CLK_DIS_WAIT_MASK);
|
|
regval |= sc->clk_dis_wait_val;
|
|
}
|
|
|
|
ret = regmap_write(sc->regmap, REG_OFFSET, regval);
|
|
|
|
if (sc->rdev->supply)
|
|
regulator_disable(sc->rdev->supply);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int gdsc_pm_resume_early(struct device *dev)
|
|
{
|
|
#ifdef CONFIG_DEEPSLEEP
|
|
if (pm_suspend_via_firmware())
|
|
return restore_hw_trig_clk_dis(dev);
|
|
#endif
|
|
return 0;
|
|
}
|
|
|
|
static int gdsc_pm_restore_early(struct device *dev)
|
|
{
|
|
return restore_hw_trig_clk_dis(dev);
|
|
}
|
|
|
|
static const struct dev_pm_ops gdsc_pm_ops = {
|
|
.resume_early = gdsc_pm_resume_early,
|
|
.restore_early = gdsc_pm_restore_early,
|
|
};
|
|
|
|
static int gdsc_probe(struct platform_device *pdev)
|
|
{
|
|
static atomic_t gdsc_count = ATOMIC_INIT(-1);
|
|
struct regulator_config reg_config = {};
|
|
struct regulator_init_data *init_data = NULL;
|
|
struct device *dev = &pdev->dev;
|
|
struct gdsc *sc;
|
|
uint32_t regval, clk_dis_wait_val = 0;
|
|
int ret;
|
|
|
|
sc = devm_kzalloc(dev, sizeof(*sc), GFP_KERNEL);
|
|
if (sc == NULL)
|
|
return -ENOMEM;
|
|
|
|
ret = gdsc_parse_dt_data(sc, dev, &init_data);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = gdsc_get_resources(sc, pdev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/*
|
|
* Disable HW trigger: collapse/restore occur based on registers writes.
|
|
* Disable SW override: Use hardware state-machine for sequencing.
|
|
*/
|
|
regmap_read(sc->regmap, REG_OFFSET, ®val);
|
|
regval &= ~(HW_CONTROL_MASK | SW_OVERRIDE_MASK);
|
|
|
|
if (of_find_property(pdev->dev.of_node, "mboxes", NULL)) {
|
|
sc->mbox_client.dev = &pdev->dev;
|
|
sc->mbox_client.tx_block = true;
|
|
sc->mbox_client.tx_tout = MBOX_TOUT_MS;
|
|
sc->mbox_client.knows_txdone = false;
|
|
|
|
sc->mbox = mbox_request_channel(&sc->mbox_client, 0);
|
|
if (IS_ERR(sc->mbox)) {
|
|
ret = PTR_ERR(sc->mbox);
|
|
dev_err(&pdev->dev, "mailbox channel request failed, ret=%d\n",
|
|
ret);
|
|
if (ret == -EAGAIN)
|
|
ret = -EPROBE_DEFER;
|
|
sc->mbox = NULL;
|
|
goto err;
|
|
}
|
|
}
|
|
|
|
if (!of_property_read_u32(pdev->dev.of_node, "qcom,clk-dis-wait-val",
|
|
&clk_dis_wait_val)) {
|
|
clk_dis_wait_val = clk_dis_wait_val << CLK_DIS_WAIT_SHIFT;
|
|
sc->clk_dis_wait_val = clk_dis_wait_val;
|
|
|
|
/* Configure wait time between states. */
|
|
regval &= ~(CLK_DIS_WAIT_MASK);
|
|
regval |= clk_dis_wait_val;
|
|
sc->pm_ops = true;
|
|
}
|
|
|
|
regmap_write(sc->regmap, REG_OFFSET, regval);
|
|
|
|
if (!sc->toggle_logic) {
|
|
regval &= ~SW_COLLAPSE_MASK;
|
|
regmap_write(sc->regmap, REG_OFFSET, regval);
|
|
|
|
ret = check_gdsc_status(sc, dev, ENABLED);
|
|
if (ret)
|
|
goto err;
|
|
}
|
|
|
|
ret = gdsc_init_is_enabled(sc);
|
|
if (ret) {
|
|
dev_err(dev, "%s failed to get initial enable state, ret=%d\n",
|
|
sc->rdesc.name, ret);
|
|
goto err;
|
|
}
|
|
|
|
ret = gdsc_init_hw_ctrl_mode(sc);
|
|
if (ret) {
|
|
dev_err(dev, "%s failed to get initial hw_ctrl state, ret=%d\n",
|
|
sc->rdesc.name, ret);
|
|
goto err;
|
|
}
|
|
|
|
if (sc->pm_ops)
|
|
pdev->dev.driver->pm = &gdsc_pm_ops;
|
|
|
|
sc->rdesc.id = atomic_inc_return(&gdsc_count);
|
|
sc->rdesc.ops = &gdsc_ops;
|
|
sc->rdesc.type = REGULATOR_VOLTAGE;
|
|
sc->rdesc.owner = THIS_MODULE;
|
|
|
|
reg_config.dev = dev;
|
|
reg_config.init_data = init_data;
|
|
reg_config.driver_data = sc;
|
|
reg_config.of_node = dev->of_node;
|
|
reg_config.regmap = sc->regmap;
|
|
|
|
sc->rdev = devm_regulator_register(dev, &sc->rdesc, ®_config);
|
|
if (IS_ERR(sc->rdev)) {
|
|
ret = PTR_ERR(sc->rdev);
|
|
dev_err(dev, "regulator_register(\"%s\") failed, ret=%d\n",
|
|
sc->rdesc.name, ret);
|
|
goto err;
|
|
}
|
|
|
|
ret = devm_regulator_proxy_consumer_register(dev, dev->of_node);
|
|
if (ret) {
|
|
dev_err(dev, "failed to register proxy consumer, ret=%d\n",
|
|
ret);
|
|
goto err;
|
|
}
|
|
|
|
ret = devm_regulator_debug_register(dev, sc->rdev);
|
|
if (ret)
|
|
dev_err(dev, "failed to register debug regulator, ret=%d\n",
|
|
ret);
|
|
|
|
platform_set_drvdata(pdev, sc);
|
|
|
|
return 0;
|
|
err:
|
|
if (sc->mbox)
|
|
mbox_free_channel(sc->mbox);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static const struct of_device_id gdsc_match_table[] = {
|
|
{ .compatible = "qcom,gdsc" },
|
|
{}
|
|
};
|
|
|
|
static struct platform_driver gdsc_driver = {
|
|
.probe = gdsc_probe,
|
|
.driver = {
|
|
.name = "gdsc",
|
|
.of_match_table = gdsc_match_table,
|
|
.sync_state = regulator_proxy_consumer_sync_state,
|
|
},
|
|
};
|
|
|
|
static int __init gdsc_init(void)
|
|
{
|
|
return platform_driver_register(&gdsc_driver);
|
|
}
|
|
subsys_initcall(gdsc_init);
|
|
|
|
static void __exit gdsc_exit(void)
|
|
{
|
|
platform_driver_unregister(&gdsc_driver);
|
|
}
|
|
module_exit(gdsc_exit);
|
|
|
|
MODULE_DESCRIPTION("GDSC regulator control library");
|
|
MODULE_LICENSE("GPL");
|