509 lines
12 KiB
C
509 lines
12 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (c) 2018 MediaTek Inc.
|
|
* Author: Owen Chen <owen.chen@mediatek.com>
|
|
*/
|
|
|
|
#include <linux/of.h>
|
|
#include <linux/of_address.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/mfd/syscon.h>
|
|
#include <linux/module.h>
|
|
#include <linux/time64.h>
|
|
#include <linux/timekeeping.h>
|
|
#include <linux/sched/clock.h>
|
|
|
|
#include "clk-mtk.h"
|
|
#include "clk-mux.h"
|
|
|
|
static bool is_registered;
|
|
|
|
static inline struct mtk_clk_mux *to_mtk_clk_mux(struct clk_hw *hw)
|
|
{
|
|
return container_of(hw, struct mtk_clk_mux, hw);
|
|
}
|
|
|
|
static int mtk_clk_mux_enable(struct clk_hw *hw)
|
|
{
|
|
struct mtk_clk_mux *mux = to_mtk_clk_mux(hw);
|
|
u32 mask = BIT(mux->data->gate_shift);
|
|
|
|
return regmap_update_bits(mux->regmap, mux->data->mux_ofs,
|
|
mask, ~mask);
|
|
}
|
|
|
|
static void mtk_clk_mux_disable(struct clk_hw *hw)
|
|
{
|
|
struct mtk_clk_mux *mux = to_mtk_clk_mux(hw);
|
|
u32 mask = BIT(mux->data->gate_shift);
|
|
|
|
regmap_update_bits(mux->regmap, mux->data->mux_ofs, mask, mask);
|
|
}
|
|
|
|
static int mtk_clk_mux_enable_setclr(struct clk_hw *hw)
|
|
{
|
|
struct mtk_clk_mux *mux = to_mtk_clk_mux(hw);
|
|
unsigned long flags = 0;
|
|
|
|
if (mux->lock)
|
|
spin_lock_irqsave(mux->lock, flags);
|
|
else
|
|
__acquire(mux->lock);
|
|
|
|
regmap_write(mux->regmap, mux->data->clr_ofs,
|
|
BIT(mux->data->gate_shift));
|
|
|
|
/*
|
|
* If mux setting restore after vcore resume, it will
|
|
* not be effective yet. Set the update bit to ensure the mux gets
|
|
* updated.
|
|
*/
|
|
regmap_write(mux->regmap, mux->data->upd_ofs,
|
|
BIT(mux->data->upd_shift));
|
|
|
|
if (mux->lock)
|
|
spin_unlock_irqrestore(mux->lock, flags);
|
|
else
|
|
__release(mux->lock);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void mtk_clk_mux_disable_setclr(struct clk_hw *hw)
|
|
{
|
|
struct mtk_clk_mux *mux = to_mtk_clk_mux(hw);
|
|
|
|
regmap_write(mux->regmap, mux->data->set_ofs,
|
|
BIT(mux->data->gate_shift));
|
|
}
|
|
|
|
static int mtk_clk_mux_is_enabled(struct clk_hw *hw)
|
|
{
|
|
struct mtk_clk_mux *mux = to_mtk_clk_mux(hw);
|
|
u32 val = 0;
|
|
|
|
if (!is_registered)
|
|
return 0;
|
|
|
|
regmap_read(mux->regmap, mux->data->mux_ofs, &val);
|
|
|
|
return (val & BIT(mux->data->gate_shift)) == 0;
|
|
}
|
|
|
|
static int mtk_clk_hwv_mux_is_enabled(struct clk_hw *hw)
|
|
{
|
|
struct mtk_clk_mux *mux = to_mtk_clk_mux(hw);
|
|
u32 val = 0;
|
|
|
|
if (!is_registered)
|
|
return 0;
|
|
|
|
regmap_read(mux->hwv_regmap, mux->data->hwv_set_ofs, &val);
|
|
|
|
return (val & BIT(mux->data->gate_shift)) != 0;
|
|
}
|
|
|
|
static int mtk_clk_hwv_mux_is_done(struct clk_hw *hw)
|
|
{
|
|
struct mtk_clk_mux *mux = to_mtk_clk_mux(hw);
|
|
u32 val = 0;
|
|
|
|
regmap_read(mux->hwv_regmap, mux->data->hwv_sta_ofs, &val);
|
|
|
|
return (val & BIT(mux->data->gate_shift)) != 0;
|
|
}
|
|
|
|
static int mtk_clk_hwv_mux_enable(struct clk_hw *hw)
|
|
{
|
|
struct mtk_clk_mux *mux = to_mtk_clk_mux(hw);
|
|
u32 val, val2;
|
|
bool is_done = false;
|
|
int i = 0;
|
|
|
|
regmap_write(mux->hwv_regmap, mux->data->hwv_set_ofs,
|
|
BIT(mux->data->gate_shift));
|
|
|
|
while (!mtk_clk_hwv_mux_is_enabled(hw)) {
|
|
if (i < MTK_WAIT_HWV_PREPARE_CNT)
|
|
udelay(MTK_WAIT_HWV_PREPARE_US);
|
|
else
|
|
goto hwv_prepare_fail;
|
|
i++;
|
|
}
|
|
|
|
i = 0;
|
|
|
|
while (1) {
|
|
if (!is_done)
|
|
regmap_read(mux->hwv_regmap, mux->data->hwv_sta_ofs, &val);
|
|
|
|
if (((val & BIT(mux->data->gate_shift)) != 0))
|
|
is_done = true;
|
|
|
|
if (is_done) {
|
|
regmap_read(mux->regmap, mux->data->mux_ofs, &val2);
|
|
if ((val2 & BIT(mux->data->gate_shift)) == 0)
|
|
break;
|
|
}
|
|
|
|
if (i < MTK_WAIT_HWV_DONE_CNT)
|
|
udelay(MTK_WAIT_HWV_DONE_US);
|
|
else
|
|
goto hwv_done_fail;
|
|
|
|
i++;
|
|
}
|
|
|
|
return 0;
|
|
|
|
hwv_done_fail:
|
|
regmap_read(mux->regmap, mux->data->mux_ofs, &val);
|
|
regmap_read(mux->hwv_regmap, mux->data->hwv_sta_ofs, &val2);
|
|
pr_err("%s mux enable timeout(%x %x)\n", clk_hw_get_name(hw), val, val2);
|
|
hwv_prepare_fail:
|
|
regmap_read(mux->regmap, mux->data->hwv_sta_ofs, &val);
|
|
pr_err("%s mux prepare timeout(%x)\n", clk_hw_get_name(hw), val);
|
|
|
|
mtk_clk_notify(mux->regmap, mux->hwv_regmap, NULL,
|
|
mux->data->mux_ofs, (mux->data->hwv_set_ofs / MTK_HWV_ID_OFS),
|
|
mux->data->gate_shift, CLK_EVT_HWV_CG_TIMEOUT);
|
|
|
|
return -EBUSY;
|
|
}
|
|
|
|
static void mtk_clk_hwv_mux_disable(struct clk_hw *hw)
|
|
{
|
|
struct mtk_clk_mux *mux = to_mtk_clk_mux(hw);
|
|
u32 val;
|
|
int i = 0;
|
|
|
|
regmap_write(mux->hwv_regmap, mux->data->hwv_clr_ofs,
|
|
BIT(mux->data->gate_shift));
|
|
|
|
while (mtk_clk_hwv_mux_is_enabled(hw)) {
|
|
if (i < MTK_WAIT_HWV_PREPARE_CNT)
|
|
udelay(MTK_WAIT_HWV_PREPARE_US);
|
|
else
|
|
goto hwv_prepare_fail;
|
|
i++;
|
|
}
|
|
|
|
i = 0;
|
|
|
|
while (!mtk_clk_hwv_mux_is_done(hw)) {
|
|
if (i < MTK_WAIT_HWV_DONE_CNT)
|
|
udelay(MTK_WAIT_HWV_DONE_US);
|
|
else
|
|
goto hwv_done_fail;
|
|
i++;
|
|
}
|
|
|
|
return;
|
|
|
|
hwv_done_fail:
|
|
regmap_read(mux->regmap, mux->data->mux_ofs, &val);
|
|
pr_err("%s mux disable timeout(%d ns)(%x)\n", clk_hw_get_name(hw),
|
|
i * MTK_WAIT_HWV_DONE_US, val);
|
|
hwv_prepare_fail:
|
|
pr_err("%s mux unprepare timeout(%d ns)\n", clk_hw_get_name(hw),
|
|
i * MTK_WAIT_HWV_PREPARE_US);
|
|
|
|
mtk_clk_notify(mux->regmap, mux->hwv_regmap, clk_hw_get_name(hw),
|
|
mux->data->mux_ofs, (mux->data->hwv_set_ofs / MTK_HWV_ID_OFS),
|
|
mux->data->gate_shift, CLK_EVT_HWV_CG_TIMEOUT);
|
|
return;
|
|
}
|
|
|
|
static int mtk_clk_ipi_mux_enable(struct clk_hw *hw)
|
|
{
|
|
struct mtk_clk_mux *mux = to_mtk_clk_mux(hw);
|
|
struct ipi_callbacks *cb;
|
|
u32 mask;
|
|
u32 val = 0;
|
|
u32 orig = 0;
|
|
int ret = 0;
|
|
|
|
cb = mtk_clk_get_ipi_cb();
|
|
|
|
if (cb && cb->clk_enable) {
|
|
ret = cb->clk_enable(mux->data->ipi_shift);
|
|
if (((mux->flags & CLK_IPI_EN_DEFAULT_OPP) == CLK_IPI_EN_DEFAULT_OPP)
|
|
&& ret == -ENODEV) {
|
|
regmap_read(mux->regmap, mux->data->mux_ofs, &orig);
|
|
mask = GENMASK(mux->data->mux_width - 1, 0);
|
|
val = (orig >> mux->data->mux_shift) & ~mask;
|
|
pr_notice("mux ipi send fail for vcp suspend(%d %d)\n",
|
|
val, mux->data->default_opp);
|
|
if (val != mux->data->default_opp) {
|
|
regmap_write(mux->regmap, mux->data->clr_ofs,
|
|
mask << mux->data->mux_shift);
|
|
regmap_write(mux->regmap, mux->data->set_ofs,
|
|
mux->data->default_opp << mux->data->mux_shift);
|
|
regmap_write(mux->regmap, mux->data->upd_ofs,
|
|
BIT(mux->data->upd_shift));
|
|
}
|
|
} else if (ret) {
|
|
regmap_read(mux->regmap, mux->data->mux_ofs, &val);
|
|
pr_err("Failed to send enable ipi to VCP %s: 0x%x(%d)\n",
|
|
clk_hw_get_name(hw), val, ret);
|
|
goto err;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
err:
|
|
mtk_clk_notify(mux->regmap, mux->hwv_regmap, clk_hw_get_name(hw),
|
|
mux->data->mux_ofs, (mux->data->hwv_set_ofs / MTK_HWV_ID_OFS),
|
|
mux->data->gate_shift, CLK_EVT_IPI_CG_TIMEOUT);
|
|
return ret;
|
|
}
|
|
|
|
static void mtk_clk_ipi_mux_disable(struct clk_hw *hw)
|
|
{
|
|
struct mtk_clk_mux *mux = to_mtk_clk_mux(hw);
|
|
struct ipi_callbacks *cb;
|
|
u32 val = 0;
|
|
int ret = 0;
|
|
|
|
cb = mtk_clk_get_ipi_cb();
|
|
|
|
if (cb && cb->clk_disable) {
|
|
ret = cb->clk_disable(mux->data->ipi_shift);
|
|
if (ret) {
|
|
regmap_read(mux->regmap, mux->data->mux_ofs, &val);
|
|
pr_err("Failed to send disable ipi to VCP %s: 0x%x(%d)\n",
|
|
clk_hw_get_name(hw), val, ret);
|
|
goto err;
|
|
}
|
|
}
|
|
|
|
return;
|
|
|
|
err:
|
|
mtk_clk_notify(mux->regmap, mux->hwv_regmap, clk_hw_get_name(hw),
|
|
mux->data->mux_ofs, (mux->data->hwv_set_ofs / MTK_HWV_ID_OFS),
|
|
mux->data->gate_shift, CLK_EVT_IPI_CG_TIMEOUT);
|
|
}
|
|
|
|
static u8 mtk_clk_mux_get_parent(struct clk_hw *hw)
|
|
{
|
|
struct mtk_clk_mux *mux = to_mtk_clk_mux(hw);
|
|
u32 mask = GENMASK(mux->data->mux_width - 1, 0);
|
|
u32 val = 0;
|
|
|
|
regmap_read(mux->regmap, mux->data->mux_ofs, &val);
|
|
val = (val >> mux->data->mux_shift) & mask;
|
|
|
|
return val;
|
|
}
|
|
|
|
static int __mtk_clk_mux_set_parent_lock(struct clk_hw *hw, u8 index, bool setclr, bool upd)
|
|
{
|
|
struct mtk_clk_mux *mux = to_mtk_clk_mux(hw);
|
|
struct clk_hw *qs_hw;
|
|
u32 mask = GENMASK(mux->data->mux_width - 1, 0);
|
|
u32 val = 0, orig = 0;
|
|
unsigned long flags = 0;
|
|
bool qs_pll_need_off = false;
|
|
|
|
if (mux->lock)
|
|
spin_lock_irqsave(mux->lock, flags);
|
|
else
|
|
__acquire(mux->lock);
|
|
|
|
if (setclr) {
|
|
regmap_read(mux->regmap, mux->data->mux_ofs, &orig);
|
|
|
|
val = (orig & ~(mask << mux->data->mux_shift))
|
|
| (index << mux->data->mux_shift);
|
|
|
|
if (val != orig) {
|
|
if ((mux->flags & CLK_ENABLE_QUICK_SWITCH) == CLK_ENABLE_QUICK_SWITCH) {
|
|
val = mux->data->qs_shift;
|
|
qs_hw = clk_hw_get_parent(clk_hw_get_parent_by_index(hw, val));
|
|
if (!mtk_hwv_pll_is_on(qs_hw)) {
|
|
/* mainpll2 enable for quick switch */
|
|
mtk_hwv_pll_on(qs_hw);
|
|
qs_pll_need_off = true;
|
|
}
|
|
}
|
|
|
|
regmap_write(mux->regmap, mux->data->clr_ofs,
|
|
mask << mux->data->mux_shift);
|
|
regmap_write(mux->regmap, mux->data->set_ofs,
|
|
index << mux->data->mux_shift);
|
|
}
|
|
|
|
if (upd)
|
|
regmap_write(mux->regmap, mux->data->upd_ofs,
|
|
BIT(mux->data->upd_shift));
|
|
|
|
if (qs_pll_need_off && (mux->flags & CLK_ENABLE_QUICK_SWITCH)
|
|
== CLK_ENABLE_QUICK_SWITCH) {
|
|
/* mainpll2 disable for quick switch */
|
|
mtk_hwv_pll_off(qs_hw);
|
|
}
|
|
} else
|
|
regmap_update_bits(mux->regmap, mux->data->mux_ofs, mask,
|
|
index << mux->data->mux_shift);
|
|
|
|
if (mux->lock)
|
|
spin_unlock_irqrestore(mux->lock, flags);
|
|
else
|
|
__release(mux->lock);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mtk_clk_mux_set_parent_lock(struct clk_hw *hw, u8 index)
|
|
{
|
|
return __mtk_clk_mux_set_parent_lock(hw, index, false, false);
|
|
}
|
|
|
|
static int mtk_clk_mux_set_parent_setclr_lock(struct clk_hw *hw, u8 index)
|
|
{
|
|
return __mtk_clk_mux_set_parent_lock(hw, index, true, false);
|
|
}
|
|
|
|
static int mtk_clk_mux_set_parent_setclr_upd_lock(struct clk_hw *hw, u8 index)
|
|
{
|
|
return __mtk_clk_mux_set_parent_lock(hw, index, true, true);
|
|
}
|
|
|
|
const struct clk_ops mtk_mux_ops = {
|
|
.get_parent = mtk_clk_mux_get_parent,
|
|
.set_parent = mtk_clk_mux_set_parent_lock,
|
|
};
|
|
EXPORT_SYMBOL_GPL(mtk_mux_ops);
|
|
|
|
|
|
const struct clk_ops mtk_mux_clr_set_ops = {
|
|
.get_parent = mtk_clk_mux_get_parent,
|
|
.set_parent = mtk_clk_mux_set_parent_setclr_lock,
|
|
};
|
|
EXPORT_SYMBOL_GPL(mtk_mux_clr_set_ops);
|
|
|
|
const struct clk_ops mtk_mux_clr_set_upd_ops = {
|
|
.get_parent = mtk_clk_mux_get_parent,
|
|
.set_parent = mtk_clk_mux_set_parent_setclr_upd_lock,
|
|
};
|
|
EXPORT_SYMBOL_GPL(mtk_mux_clr_set_upd_ops);
|
|
|
|
const struct clk_ops mtk_mux_gate_ops = {
|
|
.enable = mtk_clk_mux_enable,
|
|
.disable = mtk_clk_mux_disable,
|
|
.is_enabled = mtk_clk_mux_is_enabled,
|
|
.get_parent = mtk_clk_mux_get_parent,
|
|
.set_parent = mtk_clk_mux_set_parent_lock,
|
|
};
|
|
EXPORT_SYMBOL_GPL(mtk_mux_gate_ops);
|
|
|
|
const struct clk_ops mtk_mux_gate_clr_set_upd_ops = {
|
|
.enable = mtk_clk_mux_enable_setclr,
|
|
.disable = mtk_clk_mux_disable_setclr,
|
|
.is_enabled = mtk_clk_mux_is_enabled,
|
|
.get_parent = mtk_clk_mux_get_parent,
|
|
.set_parent = mtk_clk_mux_set_parent_setclr_upd_lock,
|
|
};
|
|
EXPORT_SYMBOL_GPL(mtk_mux_gate_clr_set_upd_ops);
|
|
|
|
const struct clk_ops mtk_hwv_mux_ops = {
|
|
.enable = mtk_clk_hwv_mux_enable,
|
|
.disable = mtk_clk_hwv_mux_disable,
|
|
.is_enabled = mtk_clk_mux_is_enabled,
|
|
.get_parent = mtk_clk_mux_get_parent,
|
|
.set_parent = mtk_clk_mux_set_parent_setclr_upd_lock,
|
|
};
|
|
EXPORT_SYMBOL_GPL(mtk_hwv_mux_ops);
|
|
|
|
const struct clk_ops mtk_ipi_mux_ops = {
|
|
.prepare = mtk_clk_ipi_mux_enable,
|
|
.unprepare = mtk_clk_ipi_mux_disable,
|
|
.enable = mtk_clk_hwv_mux_enable,
|
|
.disable = mtk_clk_hwv_mux_disable,
|
|
.is_enabled = mtk_clk_mux_is_enabled,
|
|
.get_parent = mtk_clk_mux_get_parent,
|
|
.set_parent = mtk_clk_mux_set_parent_setclr_upd_lock,
|
|
};
|
|
EXPORT_SYMBOL_GPL(mtk_ipi_mux_ops);
|
|
|
|
static struct clk *mtk_clk_register_mux(const struct mtk_mux *mux,
|
|
struct regmap *regmap,
|
|
struct regmap *hw_voter_regmap,
|
|
spinlock_t *lock)
|
|
{
|
|
struct mtk_clk_mux *clk_mux;
|
|
struct clk_init_data init = {};
|
|
struct clk *clk;
|
|
|
|
clk_mux = kzalloc(sizeof(*clk_mux), GFP_KERNEL);
|
|
if (!clk_mux)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
init.name = mux->name;
|
|
init.flags = mux->flags | CLK_SET_RATE_PARENT;
|
|
init.parent_names = mux->parent_names;
|
|
init.num_parents = mux->num_parents;
|
|
init.ops = mux->ops;
|
|
|
|
clk_mux->regmap = regmap;
|
|
clk_mux->hwv_regmap = hw_voter_regmap;
|
|
clk_mux->data = mux;
|
|
clk_mux->lock = lock;
|
|
clk_mux->flags = mux->flags;
|
|
clk_mux->hw.init = &init;
|
|
|
|
clk = clk_register(NULL, &clk_mux->hw);
|
|
if (IS_ERR(clk)) {
|
|
kfree(clk_mux);
|
|
return clk;
|
|
}
|
|
|
|
return clk;
|
|
}
|
|
|
|
int mtk_clk_register_muxes(const struct mtk_mux *muxes,
|
|
int num, struct device_node *node,
|
|
spinlock_t *lock,
|
|
struct clk_onecell_data *clk_data)
|
|
{
|
|
struct regmap *regmap, *hw_voter_regmap;
|
|
struct clk *clk;
|
|
int i;
|
|
|
|
is_registered = false;
|
|
|
|
regmap = syscon_node_to_regmap(node);
|
|
if (IS_ERR(regmap)) {
|
|
pr_err("Cannot find regmap for %pOF: %ld\n", node,
|
|
PTR_ERR(regmap));
|
|
return PTR_ERR(regmap);
|
|
}
|
|
|
|
hw_voter_regmap = syscon_regmap_lookup_by_phandle(node, "hw-voter-regmap");
|
|
if (IS_ERR(hw_voter_regmap))
|
|
hw_voter_regmap = NULL;
|
|
|
|
for (i = 0; i < num; i++) {
|
|
const struct mtk_mux *mux = &muxes[i];
|
|
|
|
if (IS_ERR_OR_NULL(clk_data->clks[mux->id])) {
|
|
clk = mtk_clk_register_mux(mux, regmap, hw_voter_regmap, lock);
|
|
|
|
if (IS_ERR(clk)) {
|
|
pr_err("Failed to register clk %s: %ld\n",
|
|
mux->name, PTR_ERR(clk));
|
|
continue;
|
|
}
|
|
|
|
clk_data->clks[mux->id] = clk;
|
|
}
|
|
}
|
|
|
|
is_registered = true;
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(mtk_clk_register_muxes);
|
|
|
|
MODULE_LICENSE("GPL");
|