Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
1// SPDX-License-Identifier: GPL-2.0+
2/*
3 * Copyright 2018 NXP
4 * Dong Aisheng <aisheng.dong@nxp.com>
5 */
6
7#include <dt-bindings/firmware/imx/rsrc.h>
8#include <linux/arm-smccc.h>
9#include <linux/clk-provider.h>
10#include <linux/err.h>
11#include <linux/of_platform.h>
12#include <linux/platform_device.h>
13#include <linux/pm_domain.h>
14#include <linux/pm_runtime.h>
15#include <linux/slab.h>
16
17#include "clk-scu.h"
18
19#define IMX_SIP_CPUFREQ 0xC2000001
20#define IMX_SIP_SET_CPUFREQ 0x00
21
22static struct imx_sc_ipc *ccm_ipc_handle;
23static struct device_node *pd_np;
24static struct platform_driver imx_clk_scu_driver;
25
26struct imx_scu_clk_node {
27 const char *name;
28 u32 rsrc;
29 u8 clk_type;
30 const char * const *parents;
31 int num_parents;
32
33 struct clk_hw *hw;
34 struct list_head node;
35};
36
37struct list_head imx_scu_clks[IMX_SC_R_LAST];
38
39/*
40 * struct clk_scu - Description of one SCU clock
41 * @hw: the common clk_hw
42 * @rsrc_id: resource ID of this SCU clock
43 * @clk_type: type of this clock resource
44 */
45struct clk_scu {
46 struct clk_hw hw;
47 u16 rsrc_id;
48 u8 clk_type;
49
50 /* for state save&restore */
51 bool is_enabled;
52 u32 rate;
53};
54
55/*
56 * struct imx_sc_msg_req_set_clock_rate - clock set rate protocol
57 * @hdr: SCU protocol header
58 * @rate: rate to set
59 * @resource: clock resource to set rate
60 * @clk: clk type of this resource
61 *
62 * This structure describes the SCU protocol of clock rate set
63 */
64struct imx_sc_msg_req_set_clock_rate {
65 struct imx_sc_rpc_msg hdr;
66 __le32 rate;
67 __le16 resource;
68 u8 clk;
69} __packed __aligned(4);
70
71struct req_get_clock_rate {
72 __le16 resource;
73 u8 clk;
74} __packed __aligned(4);
75
76struct resp_get_clock_rate {
77 __le32 rate;
78};
79
80/*
81 * struct imx_sc_msg_get_clock_rate - clock get rate protocol
82 * @hdr: SCU protocol header
83 * @req: get rate request protocol
84 * @resp: get rate response protocol
85 *
86 * This structure describes the SCU protocol of clock rate get
87 */
88struct imx_sc_msg_get_clock_rate {
89 struct imx_sc_rpc_msg hdr;
90 union {
91 struct req_get_clock_rate req;
92 struct resp_get_clock_rate resp;
93 } data;
94};
95
96/*
97 * struct imx_sc_msg_get_clock_parent - clock get parent protocol
98 * @hdr: SCU protocol header
99 * @req: get parent request protocol
100 * @resp: get parent response protocol
101 *
102 * This structure describes the SCU protocol of clock get parent
103 */
104struct imx_sc_msg_get_clock_parent {
105 struct imx_sc_rpc_msg hdr;
106 union {
107 struct req_get_clock_parent {
108 __le16 resource;
109 u8 clk;
110 } __packed __aligned(4) req;
111 struct resp_get_clock_parent {
112 u8 parent;
113 } resp;
114 } data;
115};
116
117/*
118 * struct imx_sc_msg_set_clock_parent - clock set parent protocol
119 * @hdr: SCU protocol header
120 * @req: set parent request protocol
121 *
122 * This structure describes the SCU protocol of clock set parent
123 */
124struct imx_sc_msg_set_clock_parent {
125 struct imx_sc_rpc_msg hdr;
126 __le16 resource;
127 u8 clk;
128 u8 parent;
129} __packed;
130
131/*
132 * struct imx_sc_msg_req_clock_enable - clock gate protocol
133 * @hdr: SCU protocol header
134 * @resource: clock resource to gate
135 * @clk: clk type of this resource
136 * @enable: whether gate off the clock
137 * @autog: HW auto gate enable
138 *
139 * This structure describes the SCU protocol of clock gate
140 */
141struct imx_sc_msg_req_clock_enable {
142 struct imx_sc_rpc_msg hdr;
143 __le16 resource;
144 u8 clk;
145 u8 enable;
146 u8 autog;
147} __packed __aligned(4);
148
149static inline struct clk_scu *to_clk_scu(struct clk_hw *hw)
150{
151 return container_of(hw, struct clk_scu, hw);
152}
153
154int imx_clk_scu_init(struct device_node *np)
155{
156 u32 clk_cells;
157 int ret, i;
158
159 ret = imx_scu_get_handle(&ccm_ipc_handle);
160 if (ret)
161 return ret;
162
163 of_property_read_u32(np, "#clock-cells", &clk_cells);
164
165 if (clk_cells == 2) {
166 for (i = 0; i < IMX_SC_R_LAST; i++)
167 INIT_LIST_HEAD(&imx_scu_clks[i]);
168
169 /* pd_np will be used to attach power domains later */
170 pd_np = of_find_compatible_node(NULL, NULL, "fsl,scu-pd");
171 if (!pd_np)
172 return -EINVAL;
173 }
174
175 return platform_driver_register(&imx_clk_scu_driver);
176}
177
178/*
179 * clk_scu_recalc_rate - Get clock rate for a SCU clock
180 * @hw: clock to get rate for
181 * @parent_rate: parent rate provided by common clock framework, not used
182 *
183 * Gets the current clock rate of a SCU clock. Returns the current
184 * clock rate, or zero in failure.
185 */
186static unsigned long clk_scu_recalc_rate(struct clk_hw *hw,
187 unsigned long parent_rate)
188{
189 struct clk_scu *clk = to_clk_scu(hw);
190 struct imx_sc_msg_get_clock_rate msg;
191 struct imx_sc_rpc_msg *hdr = &msg.hdr;
192 int ret;
193
194 hdr->ver = IMX_SC_RPC_VERSION;
195 hdr->svc = IMX_SC_RPC_SVC_PM;
196 hdr->func = IMX_SC_PM_FUNC_GET_CLOCK_RATE;
197 hdr->size = 2;
198
199 msg.data.req.resource = cpu_to_le16(clk->rsrc_id);
200 msg.data.req.clk = clk->clk_type;
201
202 ret = imx_scu_call_rpc(ccm_ipc_handle, &msg, true);
203 if (ret) {
204 pr_err("%s: failed to get clock rate %d\n",
205 clk_hw_get_name(hw), ret);
206 return 0;
207 }
208
209 return le32_to_cpu(msg.data.resp.rate);
210}
211
212/*
213 * clk_scu_round_rate - Round clock rate for a SCU clock
214 * @hw: clock to round rate for
215 * @rate: rate to round
216 * @parent_rate: parent rate provided by common clock framework, not used
217 *
218 * Returns the current clock rate, or zero in failure.
219 */
220static long clk_scu_round_rate(struct clk_hw *hw, unsigned long rate,
221 unsigned long *parent_rate)
222{
223 /*
224 * Assume we support all the requested rate and let the SCU firmware
225 * to handle the left work
226 */
227 return rate;
228}
229
230static int clk_scu_atf_set_cpu_rate(struct clk_hw *hw, unsigned long rate,
231 unsigned long parent_rate)
232{
233 struct clk_scu *clk = to_clk_scu(hw);
234 struct arm_smccc_res res;
235 unsigned long cluster_id;
236
237 if (clk->rsrc_id == IMX_SC_R_A35)
238 cluster_id = 0;
239 else
240 return -EINVAL;
241
242 /* CPU frequency scaling can ONLY be done by ARM-Trusted-Firmware */
243 arm_smccc_smc(IMX_SIP_CPUFREQ, IMX_SIP_SET_CPUFREQ,
244 cluster_id, rate, 0, 0, 0, 0, &res);
245
246 return 0;
247}
248
249/*
250 * clk_scu_set_rate - Set rate for a SCU clock
251 * @hw: clock to change rate for
252 * @rate: target rate for the clock
253 * @parent_rate: rate of the clock parent, not used for SCU clocks
254 *
255 * Sets a clock frequency for a SCU clock. Returns the SCU
256 * protocol status.
257 */
258static int clk_scu_set_rate(struct clk_hw *hw, unsigned long rate,
259 unsigned long parent_rate)
260{
261 struct clk_scu *clk = to_clk_scu(hw);
262 struct imx_sc_msg_req_set_clock_rate msg;
263 struct imx_sc_rpc_msg *hdr = &msg.hdr;
264
265 hdr->ver = IMX_SC_RPC_VERSION;
266 hdr->svc = IMX_SC_RPC_SVC_PM;
267 hdr->func = IMX_SC_PM_FUNC_SET_CLOCK_RATE;
268 hdr->size = 3;
269
270 msg.rate = cpu_to_le32(rate);
271 msg.resource = cpu_to_le16(clk->rsrc_id);
272 msg.clk = clk->clk_type;
273
274 return imx_scu_call_rpc(ccm_ipc_handle, &msg, true);
275}
276
277static u8 clk_scu_get_parent(struct clk_hw *hw)
278{
279 struct clk_scu *clk = to_clk_scu(hw);
280 struct imx_sc_msg_get_clock_parent msg;
281 struct imx_sc_rpc_msg *hdr = &msg.hdr;
282 int ret;
283
284 hdr->ver = IMX_SC_RPC_VERSION;
285 hdr->svc = IMX_SC_RPC_SVC_PM;
286 hdr->func = IMX_SC_PM_FUNC_GET_CLOCK_PARENT;
287 hdr->size = 2;
288
289 msg.data.req.resource = cpu_to_le16(clk->rsrc_id);
290 msg.data.req.clk = clk->clk_type;
291
292 ret = imx_scu_call_rpc(ccm_ipc_handle, &msg, true);
293 if (ret) {
294 pr_err("%s: failed to get clock parent %d\n",
295 clk_hw_get_name(hw), ret);
296 return 0;
297 }
298
299 return msg.data.resp.parent;
300}
301
302static int clk_scu_set_parent(struct clk_hw *hw, u8 index)
303{
304 struct clk_scu *clk = to_clk_scu(hw);
305 struct imx_sc_msg_set_clock_parent msg;
306 struct imx_sc_rpc_msg *hdr = &msg.hdr;
307
308 hdr->ver = IMX_SC_RPC_VERSION;
309 hdr->svc = IMX_SC_RPC_SVC_PM;
310 hdr->func = IMX_SC_PM_FUNC_SET_CLOCK_PARENT;
311 hdr->size = 2;
312
313 msg.resource = cpu_to_le16(clk->rsrc_id);
314 msg.clk = clk->clk_type;
315 msg.parent = index;
316
317 return imx_scu_call_rpc(ccm_ipc_handle, &msg, true);
318}
319
320static int sc_pm_clock_enable(struct imx_sc_ipc *ipc, u16 resource,
321 u8 clk, bool enable, bool autog)
322{
323 struct imx_sc_msg_req_clock_enable msg;
324 struct imx_sc_rpc_msg *hdr = &msg.hdr;
325
326 hdr->ver = IMX_SC_RPC_VERSION;
327 hdr->svc = IMX_SC_RPC_SVC_PM;
328 hdr->func = IMX_SC_PM_FUNC_CLOCK_ENABLE;
329 hdr->size = 3;
330
331 msg.resource = cpu_to_le16(resource);
332 msg.clk = clk;
333 msg.enable = enable;
334 msg.autog = autog;
335
336 return imx_scu_call_rpc(ccm_ipc_handle, &msg, true);
337}
338
339/*
340 * clk_scu_prepare - Enable a SCU clock
341 * @hw: clock to enable
342 *
343 * Enable the clock at the DSC slice level
344 */
345static int clk_scu_prepare(struct clk_hw *hw)
346{
347 struct clk_scu *clk = to_clk_scu(hw);
348
349 return sc_pm_clock_enable(ccm_ipc_handle, clk->rsrc_id,
350 clk->clk_type, true, false);
351}
352
353/*
354 * clk_scu_unprepare - Disable a SCU clock
355 * @hw: clock to enable
356 *
357 * Disable the clock at the DSC slice level
358 */
359static void clk_scu_unprepare(struct clk_hw *hw)
360{
361 struct clk_scu *clk = to_clk_scu(hw);
362 int ret;
363
364 ret = sc_pm_clock_enable(ccm_ipc_handle, clk->rsrc_id,
365 clk->clk_type, false, false);
366 if (ret)
367 pr_warn("%s: clk unprepare failed %d\n", clk_hw_get_name(hw),
368 ret);
369}
370
371static const struct clk_ops clk_scu_ops = {
372 .recalc_rate = clk_scu_recalc_rate,
373 .round_rate = clk_scu_round_rate,
374 .set_rate = clk_scu_set_rate,
375 .get_parent = clk_scu_get_parent,
376 .set_parent = clk_scu_set_parent,
377 .prepare = clk_scu_prepare,
378 .unprepare = clk_scu_unprepare,
379};
380
381static const struct clk_ops clk_scu_cpu_ops = {
382 .recalc_rate = clk_scu_recalc_rate,
383 .round_rate = clk_scu_round_rate,
384 .set_rate = clk_scu_atf_set_cpu_rate,
385 .prepare = clk_scu_prepare,
386 .unprepare = clk_scu_unprepare,
387};
388
389struct clk_hw *__imx_clk_scu(struct device *dev, const char *name,
390 const char * const *parents, int num_parents,
391 u32 rsrc_id, u8 clk_type)
392{
393 struct clk_init_data init;
394 struct clk_scu *clk;
395 struct clk_hw *hw;
396 int ret;
397
398 clk = kzalloc(sizeof(*clk), GFP_KERNEL);
399 if (!clk)
400 return ERR_PTR(-ENOMEM);
401
402 clk->rsrc_id = rsrc_id;
403 clk->clk_type = clk_type;
404
405 init.name = name;
406 init.ops = &clk_scu_ops;
407 if (rsrc_id == IMX_SC_R_A35)
408 init.ops = &clk_scu_cpu_ops;
409 else
410 init.ops = &clk_scu_ops;
411 init.parent_names = parents;
412 init.num_parents = num_parents;
413
414 /*
415 * Note on MX8, the clocks are tightly coupled with power domain
416 * that once the power domain is off, the clock status may be
417 * lost. So we make it NOCACHE to let user to retrieve the real
418 * clock status from HW instead of using the possible invalid
419 * cached rate.
420 */
421 init.flags = CLK_GET_RATE_NOCACHE;
422 clk->hw.init = &init;
423
424 hw = &clk->hw;
425 ret = clk_hw_register(dev, hw);
426 if (ret) {
427 kfree(clk);
428 hw = ERR_PTR(ret);
429 return hw;
430 }
431
432 if (dev)
433 dev_set_drvdata(dev, clk);
434
435 return hw;
436}
437
438struct clk_hw *imx_scu_of_clk_src_get(struct of_phandle_args *clkspec,
439 void *data)
440{
441 unsigned int rsrc = clkspec->args[0];
442 unsigned int idx = clkspec->args[1];
443 struct list_head *scu_clks = data;
444 struct imx_scu_clk_node *clk;
445
446 list_for_each_entry(clk, &scu_clks[rsrc], node) {
447 if (clk->clk_type == idx)
448 return clk->hw;
449 }
450
451 return ERR_PTR(-ENODEV);
452}
453
454static int imx_clk_scu_probe(struct platform_device *pdev)
455{
456 struct device *dev = &pdev->dev;
457 struct imx_scu_clk_node *clk = dev_get_platdata(dev);
458 struct clk_hw *hw;
459 int ret;
460
461 pm_runtime_set_suspended(dev);
462 pm_runtime_set_autosuspend_delay(dev, 50);
463 pm_runtime_use_autosuspend(&pdev->dev);
464 pm_runtime_enable(dev);
465
466 ret = pm_runtime_get_sync(dev);
467 if (ret) {
468 pm_runtime_disable(dev);
469 return ret;
470 }
471
472 hw = __imx_clk_scu(dev, clk->name, clk->parents, clk->num_parents,
473 clk->rsrc, clk->clk_type);
474 if (IS_ERR(hw)) {
475 pm_runtime_disable(dev);
476 return PTR_ERR(hw);
477 }
478
479 clk->hw = hw;
480 list_add_tail(&clk->node, &imx_scu_clks[clk->rsrc]);
481
482 pm_runtime_mark_last_busy(&pdev->dev);
483 pm_runtime_put_autosuspend(&pdev->dev);
484
485 dev_dbg(dev, "register SCU clock rsrc:%d type:%d\n", clk->rsrc,
486 clk->clk_type);
487
488 return 0;
489}
490
491static int __maybe_unused imx_clk_scu_suspend(struct device *dev)
492{
493 struct clk_scu *clk = dev_get_drvdata(dev);
494
495 clk->rate = clk_hw_get_rate(&clk->hw);
496 clk->is_enabled = clk_hw_is_enabled(&clk->hw);
497
498 if (clk->rate)
499 dev_dbg(dev, "save rate %d\n", clk->rate);
500
501 if (clk->is_enabled)
502 dev_dbg(dev, "save enabled state\n");
503
504 return 0;
505}
506
507static int __maybe_unused imx_clk_scu_resume(struct device *dev)
508{
509 struct clk_scu *clk = dev_get_drvdata(dev);
510 int ret = 0;
511
512 if (clk->rate) {
513 ret = clk_scu_set_rate(&clk->hw, clk->rate, 0);
514 dev_dbg(dev, "restore rate %d %s\n", clk->rate,
515 !ret ? "success" : "failed");
516 }
517
518 if (clk->is_enabled) {
519 ret = clk_scu_prepare(&clk->hw);
520 dev_dbg(dev, "restore enabled state %s\n",
521 !ret ? "success" : "failed");
522 }
523
524 return ret;
525}
526
527static const struct dev_pm_ops imx_clk_scu_pm_ops = {
528 SET_NOIRQ_SYSTEM_SLEEP_PM_OPS(imx_clk_scu_suspend,
529 imx_clk_scu_resume)
530};
531
532static struct platform_driver imx_clk_scu_driver = {
533 .driver = {
534 .name = "imx-scu-clk",
535 .suppress_bind_attrs = true,
536 .pm = &imx_clk_scu_pm_ops,
537 },
538 .probe = imx_clk_scu_probe,
539};
540
541static int imx_clk_scu_attach_pd(struct device *dev, u32 rsrc_id)
542{
543 struct of_phandle_args genpdspec = {
544 .np = pd_np,
545 .args_count = 1,
546 .args[0] = rsrc_id,
547 };
548
549 if (rsrc_id == IMX_SC_R_A35 || rsrc_id == IMX_SC_R_A53 ||
550 rsrc_id == IMX_SC_R_A72)
551 return 0;
552
553 return of_genpd_add_device(&genpdspec, dev);
554}
555
556struct clk_hw *imx_clk_scu_alloc_dev(const char *name,
557 const char * const *parents,
558 int num_parents, u32 rsrc_id, u8 clk_type)
559{
560 struct imx_scu_clk_node clk = {
561 .name = name,
562 .rsrc = rsrc_id,
563 .clk_type = clk_type,
564 .parents = parents,
565 .num_parents = num_parents,
566 };
567 struct platform_device *pdev;
568 int ret;
569
570 pdev = platform_device_alloc(name, PLATFORM_DEVID_NONE);
571 if (!pdev) {
572 pr_err("%s: failed to allocate scu clk dev rsrc %d type %d\n",
573 name, rsrc_id, clk_type);
574 return ERR_PTR(-ENOMEM);
575 }
576
577 ret = platform_device_add_data(pdev, &clk, sizeof(clk));
578 if (ret) {
579 platform_device_put(pdev);
580 return ERR_PTR(ret);
581 }
582
583 pdev->driver_override = "imx-scu-clk";
584
585 ret = imx_clk_scu_attach_pd(&pdev->dev, rsrc_id);
586 if (ret)
587 pr_warn("%s: failed to attached the power domain %d\n",
588 name, ret);
589
590 platform_device_add(pdev);
591
592 /* For API backwards compatiblilty, simply return NULL for success */
593 return NULL;
594}
595
596void imx_clk_scu_unregister(void)
597{
598 struct imx_scu_clk_node *clk;
599 int i;
600
601 for (i = 0; i < IMX_SC_R_LAST; i++) {
602 list_for_each_entry(clk, &imx_scu_clks[i], node) {
603 clk_hw_unregister(clk->hw);
604 kfree(clk);
605 }
606 }
607}