clk-mtk.c 15 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627
  1. // SPDX-License-Identifier: GPL-2.0
  2. /*
  3. * MediaTek common clock driver
  4. *
  5. * Copyright (C) 2018 MediaTek Inc.
  6. * Author: Ryder Lee <ryder.lee@mediatek.com>
  7. */
  8. #include <common.h>
  9. #include <clk-uclass.h>
  10. #include <div64.h>
  11. #include <dm.h>
  12. #include <asm/io.h>
  13. #include <linux/bitops.h>
  14. #include <linux/delay.h>
  15. #include "clk-mtk.h"
  16. #define REG_CON0 0
  17. #define REG_CON1 4
  18. #define CON0_BASE_EN BIT(0)
  19. #define CON0_PWR_ON BIT(0)
  20. #define CON0_ISO_EN BIT(1)
  21. #define CON1_PCW_CHG BIT(31)
  22. #define POSTDIV_MASK 0x7
  23. #define INTEGER_BITS 7
  24. /* scpsys clock off control */
  25. #define CLK_SCP_CFG0 0x200
  26. #define CLK_SCP_CFG1 0x204
  27. #define SCP_ARMCK_OFF_EN GENMASK(9, 0)
  28. #define SCP_AXICK_DCM_DIS_EN BIT(0)
  29. #define SCP_AXICK_26M_SEL_EN BIT(4)
  30. /* shared functions */
  31. /*
  32. * In case the rate change propagation to parent clocks is undesirable,
  33. * this function is recursively called to find the parent to calculate
  34. * the accurate frequency.
  35. */
  36. static ulong mtk_clk_find_parent_rate(struct clk *clk, int id,
  37. struct udevice *pdev)
  38. {
  39. struct clk parent = { .id = id, };
  40. if (pdev)
  41. parent.dev = pdev;
  42. else
  43. parent.dev = clk->dev;
  44. return clk_get_rate(&parent);
  45. }
  46. static int mtk_clk_mux_set_parent(void __iomem *base, u32 parent,
  47. const struct mtk_composite *mux)
  48. {
  49. u32 val, index = 0;
  50. while (mux->parent[index] != parent)
  51. if (++index == mux->num_parents)
  52. return -EINVAL;
  53. if (mux->flags & CLK_MUX_SETCLR_UPD) {
  54. val = (mux->mux_mask << mux->mux_shift);
  55. writel(val, base + mux->mux_clr_reg);
  56. val = (index << mux->mux_shift);
  57. writel(val, base + mux->mux_set_reg);
  58. if (mux->upd_shift >= 0)
  59. writel(BIT(mux->upd_shift), base + mux->upd_reg);
  60. } else {
  61. /* switch mux to a select parent */
  62. val = readl(base + mux->mux_reg);
  63. val &= ~(mux->mux_mask << mux->mux_shift);
  64. val |= index << mux->mux_shift;
  65. writel(val, base + mux->mux_reg);
  66. }
  67. return 0;
  68. }
  69. /* apmixedsys functions */
  70. static unsigned long __mtk_pll_recalc_rate(const struct mtk_pll_data *pll,
  71. u32 fin, u32 pcw, int postdiv)
  72. {
  73. int pcwbits = pll->pcwbits;
  74. int pcwfbits;
  75. int ibits;
  76. u64 vco;
  77. u8 c = 0;
  78. /* The fractional part of the PLL divider. */
  79. ibits = pll->pcwibits ? pll->pcwibits : INTEGER_BITS;
  80. pcwfbits = pcwbits > ibits ? pcwbits - ibits : 0;
  81. vco = (u64)fin * pcw;
  82. if (pcwfbits && (vco & GENMASK(pcwfbits - 1, 0)))
  83. c = 1;
  84. vco >>= pcwfbits;
  85. if (c)
  86. vco++;
  87. return ((unsigned long)vco + postdiv - 1) / postdiv;
  88. }
  89. /**
  90. * MediaTek PLLs are configured through their pcw value. The pcw value
  91. * describes a divider in the PLL feedback loop which consists of 7 bits
  92. * for the integer part and the remaining bits (if present) for the
  93. * fractional part. Also they have a 3 bit power-of-two post divider.
  94. */
  95. static void mtk_pll_set_rate_regs(struct clk *clk, u32 pcw, int postdiv)
  96. {
  97. struct mtk_clk_priv *priv = dev_get_priv(clk->dev);
  98. const struct mtk_pll_data *pll = &priv->tree->plls[clk->id];
  99. u32 val, chg;
  100. /* set postdiv */
  101. val = readl(priv->base + pll->pd_reg);
  102. val &= ~(POSTDIV_MASK << pll->pd_shift);
  103. val |= (ffs(postdiv) - 1) << pll->pd_shift;
  104. /* postdiv and pcw need to set at the same time if on same register */
  105. if (pll->pd_reg != pll->pcw_reg) {
  106. writel(val, priv->base + pll->pd_reg);
  107. val = readl(priv->base + pll->pcw_reg);
  108. }
  109. /* set pcw */
  110. val &= ~GENMASK(pll->pcw_shift + pll->pcwbits - 1, pll->pcw_shift);
  111. val |= pcw << pll->pcw_shift;
  112. if (pll->pcw_chg_reg) {
  113. chg = readl(priv->base + pll->pcw_chg_reg);
  114. chg |= CON1_PCW_CHG;
  115. writel(val, priv->base + pll->pcw_reg);
  116. writel(chg, priv->base + pll->pcw_chg_reg);
  117. } else {
  118. val |= CON1_PCW_CHG;
  119. writel(val, priv->base + pll->pcw_reg);
  120. }
  121. udelay(20);
  122. }
  123. /**
  124. * mtk_pll_calc_values - calculate good values for a given input frequency.
  125. * @clk: The clk
  126. * @pcw: The pcw value (output)
  127. * @postdiv: The post divider (output)
  128. * @freq: The desired target frequency
  129. */
  130. static void mtk_pll_calc_values(struct clk *clk, u32 *pcw, u32 *postdiv,
  131. u32 freq)
  132. {
  133. struct mtk_clk_priv *priv = dev_get_priv(clk->dev);
  134. const struct mtk_pll_data *pll = &priv->tree->plls[clk->id];
  135. unsigned long fmin = pll->fmin ? pll->fmin : 1000 * MHZ;
  136. u64 _pcw;
  137. int ibits;
  138. u32 val;
  139. if (freq > pll->fmax)
  140. freq = pll->fmax;
  141. for (val = 0; val < 5; val++) {
  142. *postdiv = 1 << val;
  143. if ((u64)freq * *postdiv >= fmin)
  144. break;
  145. }
  146. /* _pcw = freq * postdiv / xtal_rate * 2^pcwfbits */
  147. ibits = pll->pcwibits ? pll->pcwibits : INTEGER_BITS;
  148. _pcw = ((u64)freq << val) << (pll->pcwbits - ibits);
  149. do_div(_pcw, priv->tree->xtal2_rate);
  150. *pcw = (u32)_pcw;
  151. }
  152. static ulong mtk_apmixedsys_set_rate(struct clk *clk, ulong rate)
  153. {
  154. u32 pcw = 0;
  155. u32 postdiv;
  156. mtk_pll_calc_values(clk, &pcw, &postdiv, rate);
  157. mtk_pll_set_rate_regs(clk, pcw, postdiv);
  158. return 0;
  159. }
  160. static ulong mtk_apmixedsys_get_rate(struct clk *clk)
  161. {
  162. struct mtk_clk_priv *priv = dev_get_priv(clk->dev);
  163. const struct mtk_pll_data *pll = &priv->tree->plls[clk->id];
  164. u32 postdiv;
  165. u32 pcw;
  166. postdiv = (readl(priv->base + pll->pd_reg) >> pll->pd_shift) &
  167. POSTDIV_MASK;
  168. postdiv = 1 << postdiv;
  169. pcw = readl(priv->base + pll->pcw_reg) >> pll->pcw_shift;
  170. pcw &= GENMASK(pll->pcwbits - 1, 0);
  171. return __mtk_pll_recalc_rate(pll, priv->tree->xtal2_rate,
  172. pcw, postdiv);
  173. }
  174. static int mtk_apmixedsys_enable(struct clk *clk)
  175. {
  176. struct mtk_clk_priv *priv = dev_get_priv(clk->dev);
  177. const struct mtk_pll_data *pll = &priv->tree->plls[clk->id];
  178. u32 r;
  179. r = readl(priv->base + pll->pwr_reg) | CON0_PWR_ON;
  180. writel(r, priv->base + pll->pwr_reg);
  181. udelay(1);
  182. r = readl(priv->base + pll->pwr_reg) & ~CON0_ISO_EN;
  183. writel(r, priv->base + pll->pwr_reg);
  184. udelay(1);
  185. r = readl(priv->base + pll->reg + REG_CON0);
  186. r |= pll->en_mask;
  187. writel(r, priv->base + pll->reg + REG_CON0);
  188. udelay(20);
  189. if (pll->flags & HAVE_RST_BAR) {
  190. r = readl(priv->base + pll->reg + REG_CON0);
  191. r |= pll->rst_bar_mask;
  192. writel(r, priv->base + pll->reg + REG_CON0);
  193. }
  194. return 0;
  195. }
  196. static int mtk_apmixedsys_disable(struct clk *clk)
  197. {
  198. struct mtk_clk_priv *priv = dev_get_priv(clk->dev);
  199. const struct mtk_pll_data *pll = &priv->tree->plls[clk->id];
  200. u32 r;
  201. if (pll->flags & HAVE_RST_BAR) {
  202. r = readl(priv->base + pll->reg + REG_CON0);
  203. r &= ~pll->rst_bar_mask;
  204. writel(r, priv->base + pll->reg + REG_CON0);
  205. }
  206. r = readl(priv->base + pll->reg + REG_CON0);
  207. r &= ~CON0_BASE_EN;
  208. writel(r, priv->base + pll->reg + REG_CON0);
  209. r = readl(priv->base + pll->pwr_reg) | CON0_ISO_EN;
  210. writel(r, priv->base + pll->pwr_reg);
  211. r = readl(priv->base + pll->pwr_reg) & ~CON0_PWR_ON;
  212. writel(r, priv->base + pll->pwr_reg);
  213. return 0;
  214. }
  215. /* topckgen functions */
  216. static ulong mtk_factor_recalc_rate(const struct mtk_fixed_factor *fdiv,
  217. ulong parent_rate)
  218. {
  219. u64 rate = parent_rate * fdiv->mult;
  220. do_div(rate, fdiv->div);
  221. return rate;
  222. }
  223. static ulong mtk_topckgen_get_factor_rate(struct clk *clk, u32 off)
  224. {
  225. struct mtk_clk_priv *priv = dev_get_priv(clk->dev);
  226. const struct mtk_fixed_factor *fdiv = &priv->tree->fdivs[off];
  227. ulong rate;
  228. switch (fdiv->flags & CLK_PARENT_MASK) {
  229. case CLK_PARENT_APMIXED:
  230. rate = mtk_clk_find_parent_rate(clk, fdiv->parent,
  231. priv->parent);
  232. break;
  233. case CLK_PARENT_TOPCKGEN:
  234. rate = mtk_clk_find_parent_rate(clk, fdiv->parent, NULL);
  235. break;
  236. case CLK_PARENT_XTAL:
  237. default:
  238. rate = priv->tree->xtal_rate;
  239. }
  240. return mtk_factor_recalc_rate(fdiv, rate);
  241. }
  242. static ulong mtk_infrasys_get_factor_rate(struct clk *clk, u32 off)
  243. {
  244. struct mtk_clk_priv *priv = dev_get_priv(clk->dev);
  245. const struct mtk_fixed_factor *fdiv = &priv->tree->fdivs[off];
  246. ulong rate;
  247. switch (fdiv->flags & CLK_PARENT_MASK) {
  248. case CLK_PARENT_TOPCKGEN:
  249. rate = mtk_clk_find_parent_rate(clk, fdiv->parent,
  250. priv->parent);
  251. break;
  252. case CLK_PARENT_XTAL:
  253. rate = priv->tree->xtal_rate;
  254. break;
  255. default:
  256. rate = mtk_clk_find_parent_rate(clk, fdiv->parent, NULL);
  257. }
  258. return mtk_factor_recalc_rate(fdiv, rate);
  259. }
  260. static ulong mtk_topckgen_get_mux_rate(struct clk *clk, u32 off)
  261. {
  262. struct mtk_clk_priv *priv = dev_get_priv(clk->dev);
  263. const struct mtk_composite *mux = &priv->tree->muxes[off];
  264. u32 index;
  265. index = readl(priv->base + mux->mux_reg);
  266. index &= mux->mux_mask << mux->mux_shift;
  267. index = index >> mux->mux_shift;
  268. if (mux->parent[index] > 0 ||
  269. (mux->parent[index] == CLK_XTAL &&
  270. priv->tree->flags & CLK_BYPASS_XTAL)) {
  271. switch (mux->flags & CLK_PARENT_MASK) {
  272. case CLK_PARENT_APMIXED:
  273. return mtk_clk_find_parent_rate(clk, mux->parent[index],
  274. priv->parent);
  275. break;
  276. default:
  277. return mtk_clk_find_parent_rate(clk, mux->parent[index],
  278. NULL);
  279. break;
  280. }
  281. }
  282. return priv->tree->xtal_rate;
  283. }
  284. static ulong mtk_infrasys_get_mux_rate(struct clk *clk, u32 off)
  285. {
  286. struct mtk_clk_priv *priv = dev_get_priv(clk->dev);
  287. const struct mtk_composite *mux = &priv->tree->muxes[off];
  288. u32 index;
  289. index = readl(priv->base + mux->mux_reg);
  290. index &= mux->mux_mask << mux->mux_shift;
  291. index = index >> mux->mux_shift;
  292. if (mux->parent[index] > 0 ||
  293. (mux->parent[index] == CLK_XTAL &&
  294. priv->tree->flags & CLK_BYPASS_XTAL)) {
  295. switch (mux->flags & CLK_PARENT_MASK) {
  296. case CLK_PARENT_TOPCKGEN:
  297. return mtk_clk_find_parent_rate(clk, mux->parent[index],
  298. priv->parent);
  299. break;
  300. default:
  301. return mtk_clk_find_parent_rate(clk, mux->parent[index],
  302. NULL);
  303. break;
  304. }
  305. }
  306. return 0;
  307. }
  308. static ulong mtk_topckgen_get_rate(struct clk *clk)
  309. {
  310. struct mtk_clk_priv *priv = dev_get_priv(clk->dev);
  311. if (clk->id < priv->tree->fdivs_offs)
  312. return priv->tree->fclks[clk->id].rate;
  313. else if (clk->id < priv->tree->muxes_offs)
  314. return mtk_topckgen_get_factor_rate(clk, clk->id -
  315. priv->tree->fdivs_offs);
  316. else
  317. return mtk_topckgen_get_mux_rate(clk, clk->id -
  318. priv->tree->muxes_offs);
  319. }
  320. static ulong mtk_infrasys_get_rate(struct clk *clk)
  321. {
  322. struct mtk_clk_priv *priv = dev_get_priv(clk->dev);
  323. ulong rate;
  324. if (clk->id < priv->tree->fdivs_offs) {
  325. rate = priv->tree->fclks[clk->id].rate;
  326. } else if (clk->id < priv->tree->muxes_offs) {
  327. rate = mtk_infrasys_get_factor_rate(clk, clk->id -
  328. priv->tree->fdivs_offs);
  329. } else {
  330. rate = mtk_infrasys_get_mux_rate(clk, clk->id -
  331. priv->tree->muxes_offs);
  332. }
  333. return rate;
  334. }
  335. static int mtk_clk_mux_enable(struct clk *clk)
  336. {
  337. struct mtk_clk_priv *priv = dev_get_priv(clk->dev);
  338. const struct mtk_composite *mux;
  339. u32 val;
  340. if (clk->id < priv->tree->muxes_offs)
  341. return 0;
  342. mux = &priv->tree->muxes[clk->id - priv->tree->muxes_offs];
  343. if (mux->gate_shift < 0)
  344. return 0;
  345. /* enable clock gate */
  346. if (mux->flags & CLK_MUX_SETCLR_UPD) {
  347. val = BIT(mux->gate_shift);
  348. writel(val, priv->base + mux->mux_clr_reg);
  349. } else {
  350. val = readl(priv->base + mux->gate_reg);
  351. val &= ~BIT(mux->gate_shift);
  352. writel(val, priv->base + mux->gate_reg);
  353. }
  354. if (mux->flags & CLK_DOMAIN_SCPSYS) {
  355. /* enable scpsys clock off control */
  356. writel(SCP_ARMCK_OFF_EN, priv->base + CLK_SCP_CFG0);
  357. writel(SCP_AXICK_DCM_DIS_EN | SCP_AXICK_26M_SEL_EN,
  358. priv->base + CLK_SCP_CFG1);
  359. }
  360. return 0;
  361. }
  362. static int mtk_clk_mux_disable(struct clk *clk)
  363. {
  364. struct mtk_clk_priv *priv = dev_get_priv(clk->dev);
  365. const struct mtk_composite *mux;
  366. u32 val;
  367. if (clk->id < priv->tree->muxes_offs)
  368. return 0;
  369. mux = &priv->tree->muxes[clk->id - priv->tree->muxes_offs];
  370. if (mux->gate_shift < 0)
  371. return 0;
  372. /* disable clock gate */
  373. if (mux->flags & CLK_MUX_SETCLR_UPD) {
  374. val = BIT(mux->gate_shift);
  375. writel(val, priv->base + mux->mux_set_reg);
  376. } else {
  377. val = readl(priv->base + mux->gate_reg);
  378. val |= BIT(mux->gate_shift);
  379. writel(val, priv->base + mux->gate_reg);
  380. }
  381. return 0;
  382. }
  383. static int mtk_common_clk_set_parent(struct clk *clk, struct clk *parent)
  384. {
  385. struct mtk_clk_priv *priv = dev_get_priv(clk->dev);
  386. if (clk->id < priv->tree->muxes_offs)
  387. return 0;
  388. return mtk_clk_mux_set_parent(priv->base, parent->id,
  389. &priv->tree->muxes[clk->id - priv->tree->muxes_offs]);
  390. }
  391. /* CG functions */
  392. static int mtk_clk_gate_enable(struct clk *clk)
  393. {
  394. struct mtk_cg_priv *priv = dev_get_priv(clk->dev);
  395. const struct mtk_gate *gate = &priv->gates[clk->id];
  396. u32 bit = BIT(gate->shift);
  397. switch (gate->flags & CLK_GATE_MASK) {
  398. case CLK_GATE_SETCLR:
  399. writel(bit, priv->base + gate->regs->clr_ofs);
  400. break;
  401. case CLK_GATE_SETCLR_INV:
  402. writel(bit, priv->base + gate->regs->set_ofs);
  403. break;
  404. case CLK_GATE_NO_SETCLR:
  405. clrsetbits_le32(priv->base + gate->regs->sta_ofs, bit, 0);
  406. break;
  407. case CLK_GATE_NO_SETCLR_INV:
  408. clrsetbits_le32(priv->base + gate->regs->sta_ofs, bit, bit);
  409. break;
  410. default:
  411. return -EINVAL;
  412. }
  413. return 0;
  414. }
  415. static int mtk_clk_gate_disable(struct clk *clk)
  416. {
  417. struct mtk_cg_priv *priv = dev_get_priv(clk->dev);
  418. const struct mtk_gate *gate = &priv->gates[clk->id];
  419. u32 bit = BIT(gate->shift);
  420. switch (gate->flags & CLK_GATE_MASK) {
  421. case CLK_GATE_SETCLR:
  422. writel(bit, priv->base + gate->regs->set_ofs);
  423. break;
  424. case CLK_GATE_SETCLR_INV:
  425. writel(bit, priv->base + gate->regs->clr_ofs);
  426. break;
  427. case CLK_GATE_NO_SETCLR:
  428. clrsetbits_le32(priv->base + gate->regs->sta_ofs, bit, bit);
  429. break;
  430. case CLK_GATE_NO_SETCLR_INV:
  431. clrsetbits_le32(priv->base + gate->regs->sta_ofs, bit, 0);
  432. break;
  433. default:
  434. return -EINVAL;
  435. }
  436. return 0;
  437. }
  438. static ulong mtk_clk_gate_get_rate(struct clk *clk)
  439. {
  440. struct mtk_cg_priv *priv = dev_get_priv(clk->dev);
  441. const struct mtk_gate *gate = &priv->gates[clk->id];
  442. return mtk_clk_find_parent_rate(clk, gate->parent, priv->parent);
  443. }
  444. const struct clk_ops mtk_clk_apmixedsys_ops = {
  445. .enable = mtk_apmixedsys_enable,
  446. .disable = mtk_apmixedsys_disable,
  447. .set_rate = mtk_apmixedsys_set_rate,
  448. .get_rate = mtk_apmixedsys_get_rate,
  449. };
  450. const struct clk_ops mtk_clk_topckgen_ops = {
  451. .enable = mtk_clk_mux_enable,
  452. .disable = mtk_clk_mux_disable,
  453. .get_rate = mtk_topckgen_get_rate,
  454. .set_parent = mtk_common_clk_set_parent,
  455. };
  456. const struct clk_ops mtk_clk_infrasys_ops = {
  457. .enable = mtk_clk_mux_enable,
  458. .disable = mtk_clk_mux_disable,
  459. .get_rate = mtk_infrasys_get_rate,
  460. .set_parent = mtk_common_clk_set_parent,
  461. };
  462. const struct clk_ops mtk_clk_gate_ops = {
  463. .enable = mtk_clk_gate_enable,
  464. .disable = mtk_clk_gate_disable,
  465. .get_rate = mtk_clk_gate_get_rate,
  466. };
  467. int mtk_common_clk_init(struct udevice *dev,
  468. const struct mtk_clk_tree *tree)
  469. {
  470. struct mtk_clk_priv *priv = dev_get_priv(dev);
  471. struct udevice *parent;
  472. int ret;
  473. priv->base = dev_read_addr_ptr(dev);
  474. if (!priv->base)
  475. return -ENOENT;
  476. ret = uclass_get_device_by_phandle(UCLASS_CLK, dev, "clock-parent", &parent);
  477. if (ret || !parent) {
  478. ret = uclass_get_device_by_driver(UCLASS_CLK,
  479. DM_DRIVER_GET(mtk_clk_apmixedsys), &parent);
  480. if (ret || !parent)
  481. return -ENOENT;
  482. }
  483. priv->parent = parent;
  484. priv->tree = tree;
  485. return 0;
  486. }
  487. int mtk_common_clk_gate_init(struct udevice *dev,
  488. const struct mtk_clk_tree *tree,
  489. const struct mtk_gate *gates)
  490. {
  491. struct mtk_cg_priv *priv = dev_get_priv(dev);
  492. struct udevice *parent;
  493. int ret;
  494. priv->base = dev_read_addr_ptr(dev);
  495. if (!priv->base)
  496. return -ENOENT;
  497. ret = uclass_get_device_by_phandle(UCLASS_CLK, dev, "clock-parent", &parent);
  498. if (ret || !parent) {
  499. ret = uclass_get_device_by_driver(UCLASS_CLK,
  500. DM_DRIVER_GET(mtk_clk_topckgen), &parent);
  501. if (ret || !parent)
  502. return -ENOENT;
  503. }
  504. priv->parent = parent;
  505. priv->tree = tree;
  506. priv->gates = gates;
  507. return 0;
  508. }