Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at v6.17-rc5 513 lines 14 kB view raw
1// SPDX-License-Identifier: GPL-2.0 2/* 3 * System Control and Power Interface (SCMI) Protocol based clock driver 4 * 5 * Copyright (C) 2018-2024 ARM Ltd. 6 */ 7 8#include <linux/bits.h> 9#include <linux/clk-provider.h> 10#include <linux/device.h> 11#include <linux/err.h> 12#include <linux/of.h> 13#include <linux/module.h> 14#include <linux/scmi_protocol.h> 15#include <asm/div64.h> 16 17#define NOT_ATOMIC false 18#define ATOMIC true 19 20enum scmi_clk_feats { 21 SCMI_CLK_ATOMIC_SUPPORTED, 22 SCMI_CLK_STATE_CTRL_SUPPORTED, 23 SCMI_CLK_RATE_CTRL_SUPPORTED, 24 SCMI_CLK_PARENT_CTRL_SUPPORTED, 25 SCMI_CLK_DUTY_CYCLE_SUPPORTED, 26 SCMI_CLK_FEATS_COUNT 27}; 28 29#define SCMI_MAX_CLK_OPS BIT(SCMI_CLK_FEATS_COUNT) 30 31static const struct scmi_clk_proto_ops *scmi_proto_clk_ops; 32 33struct scmi_clk { 34 u32 id; 35 struct device *dev; 36 struct clk_hw hw; 37 const struct scmi_clock_info *info; 38 const struct scmi_protocol_handle *ph; 39 struct clk_parent_data *parent_data; 40}; 41 42#define to_scmi_clk(clk) container_of(clk, struct scmi_clk, hw) 43 44static unsigned long scmi_clk_recalc_rate(struct clk_hw *hw, 45 unsigned long parent_rate) 46{ 47 int ret; 48 u64 rate; 49 struct scmi_clk *clk = to_scmi_clk(hw); 50 51 ret = scmi_proto_clk_ops->rate_get(clk->ph, clk->id, &rate); 52 if (ret) 53 return 0; 54 return rate; 55} 56 57static long scmi_clk_round_rate(struct clk_hw *hw, unsigned long rate, 58 unsigned long *parent_rate) 59{ 60 u64 fmin, fmax, ftmp; 61 struct scmi_clk *clk = to_scmi_clk(hw); 62 63 /* 64 * We can't figure out what rate it will be, so just return the 65 * rate back to the caller. scmi_clk_recalc_rate() will be called 66 * after the rate is set and we'll know what rate the clock is 67 * running at then. 68 */ 69 if (clk->info->rate_discrete) 70 return rate; 71 72 fmin = clk->info->range.min_rate; 73 fmax = clk->info->range.max_rate; 74 if (rate <= fmin) 75 return fmin; 76 else if (rate >= fmax) 77 return fmax; 78 79 ftmp = rate - fmin; 80 ftmp += clk->info->range.step_size - 1; /* to round up */ 81 do_div(ftmp, clk->info->range.step_size); 82 83 return ftmp * clk->info->range.step_size + fmin; 84} 85 86static int scmi_clk_set_rate(struct clk_hw *hw, unsigned long rate, 87 unsigned long parent_rate) 88{ 89 struct scmi_clk *clk = to_scmi_clk(hw); 90 91 return scmi_proto_clk_ops->rate_set(clk->ph, clk->id, rate); 92} 93 94static int scmi_clk_set_parent(struct clk_hw *hw, u8 parent_index) 95{ 96 struct scmi_clk *clk = to_scmi_clk(hw); 97 98 return scmi_proto_clk_ops->parent_set(clk->ph, clk->id, parent_index); 99} 100 101static u8 scmi_clk_get_parent(struct clk_hw *hw) 102{ 103 struct scmi_clk *clk = to_scmi_clk(hw); 104 u32 parent_id, p_idx; 105 int ret; 106 107 ret = scmi_proto_clk_ops->parent_get(clk->ph, clk->id, &parent_id); 108 if (ret) 109 return 0; 110 111 for (p_idx = 0; p_idx < clk->info->num_parents; p_idx++) { 112 if (clk->parent_data[p_idx].index == parent_id) 113 break; 114 } 115 116 if (p_idx == clk->info->num_parents) 117 return 0; 118 119 return p_idx; 120} 121 122static int scmi_clk_determine_rate(struct clk_hw *hw, struct clk_rate_request *req) 123{ 124 /* 125 * Suppose all the requested rates are supported, and let firmware 126 * to handle the left work. 127 */ 128 return 0; 129} 130 131static int scmi_clk_enable(struct clk_hw *hw) 132{ 133 struct scmi_clk *clk = to_scmi_clk(hw); 134 135 return scmi_proto_clk_ops->enable(clk->ph, clk->id, NOT_ATOMIC); 136} 137 138static void scmi_clk_disable(struct clk_hw *hw) 139{ 140 struct scmi_clk *clk = to_scmi_clk(hw); 141 142 scmi_proto_clk_ops->disable(clk->ph, clk->id, NOT_ATOMIC); 143} 144 145static int scmi_clk_atomic_enable(struct clk_hw *hw) 146{ 147 struct scmi_clk *clk = to_scmi_clk(hw); 148 149 return scmi_proto_clk_ops->enable(clk->ph, clk->id, ATOMIC); 150} 151 152static void scmi_clk_atomic_disable(struct clk_hw *hw) 153{ 154 struct scmi_clk *clk = to_scmi_clk(hw); 155 156 scmi_proto_clk_ops->disable(clk->ph, clk->id, ATOMIC); 157} 158 159static int __scmi_clk_is_enabled(struct clk_hw *hw, bool atomic) 160{ 161 int ret; 162 bool enabled = false; 163 struct scmi_clk *clk = to_scmi_clk(hw); 164 165 ret = scmi_proto_clk_ops->state_get(clk->ph, clk->id, &enabled, atomic); 166 if (ret) 167 dev_warn(clk->dev, 168 "Failed to get state for clock ID %d\n", clk->id); 169 170 return !!enabled; 171} 172 173static int scmi_clk_atomic_is_enabled(struct clk_hw *hw) 174{ 175 return __scmi_clk_is_enabled(hw, ATOMIC); 176} 177 178static int scmi_clk_is_enabled(struct clk_hw *hw) 179{ 180 return __scmi_clk_is_enabled(hw, NOT_ATOMIC); 181} 182 183static int scmi_clk_get_duty_cycle(struct clk_hw *hw, struct clk_duty *duty) 184{ 185 int ret; 186 u32 val; 187 struct scmi_clk *clk = to_scmi_clk(hw); 188 189 ret = scmi_proto_clk_ops->config_oem_get(clk->ph, clk->id, 190 SCMI_CLOCK_CFG_DUTY_CYCLE, 191 &val, NULL, false); 192 if (!ret) { 193 duty->num = val; 194 duty->den = 100; 195 } else { 196 dev_warn(clk->dev, 197 "Failed to get duty cycle for clock ID %d\n", clk->id); 198 } 199 200 return ret; 201} 202 203static int scmi_clk_set_duty_cycle(struct clk_hw *hw, struct clk_duty *duty) 204{ 205 int ret; 206 u32 val; 207 struct scmi_clk *clk = to_scmi_clk(hw); 208 209 /* SCMI OEM Duty Cycle is expressed as a percentage */ 210 val = (duty->num * 100) / duty->den; 211 ret = scmi_proto_clk_ops->config_oem_set(clk->ph, clk->id, 212 SCMI_CLOCK_CFG_DUTY_CYCLE, 213 val, false); 214 if (ret) 215 dev_warn(clk->dev, 216 "Failed to set duty cycle(%u/%u) for clock ID %d\n", 217 duty->num, duty->den, clk->id); 218 219 return ret; 220} 221 222static int scmi_clk_ops_init(struct device *dev, struct scmi_clk *sclk, 223 const struct clk_ops *scmi_ops) 224{ 225 int ret; 226 unsigned long min_rate, max_rate; 227 228 struct clk_init_data init = { 229 .flags = CLK_GET_RATE_NOCACHE, 230 .num_parents = sclk->info->num_parents, 231 .ops = scmi_ops, 232 .name = sclk->info->name, 233 .parent_data = sclk->parent_data, 234 }; 235 236 sclk->hw.init = &init; 237 ret = devm_clk_hw_register(dev, &sclk->hw); 238 if (ret) 239 return ret; 240 241 if (sclk->info->rate_discrete) { 242 int num_rates = sclk->info->list.num_rates; 243 244 if (num_rates <= 0) 245 return -EINVAL; 246 247 min_rate = sclk->info->list.rates[0]; 248 max_rate = sclk->info->list.rates[num_rates - 1]; 249 } else { 250 min_rate = sclk->info->range.min_rate; 251 max_rate = sclk->info->range.max_rate; 252 } 253 254 clk_hw_set_rate_range(&sclk->hw, min_rate, max_rate); 255 return ret; 256} 257 258/** 259 * scmi_clk_ops_alloc() - Alloc and configure clock operations 260 * @dev: A device reference for devres 261 * @feats_key: A bitmap representing the desired clk_ops capabilities 262 * 263 * Allocate and configure a proper set of clock operations depending on the 264 * specifically required SCMI clock features. 265 * 266 * Return: A pointer to the allocated and configured clk_ops on success, 267 * or NULL on allocation failure. 268 */ 269static const struct clk_ops * 270scmi_clk_ops_alloc(struct device *dev, unsigned long feats_key) 271{ 272 struct clk_ops *ops; 273 274 ops = devm_kzalloc(dev, sizeof(*ops), GFP_KERNEL); 275 if (!ops) 276 return NULL; 277 /* 278 * We can provide enable/disable/is_enabled atomic callbacks only if the 279 * underlying SCMI transport for an SCMI instance is configured to 280 * handle SCMI commands in an atomic manner. 281 * 282 * When no SCMI atomic transport support is available we instead provide 283 * only the prepare/unprepare API, as allowed by the clock framework 284 * when atomic calls are not available. 285 */ 286 if (feats_key & BIT(SCMI_CLK_STATE_CTRL_SUPPORTED)) { 287 if (feats_key & BIT(SCMI_CLK_ATOMIC_SUPPORTED)) { 288 ops->enable = scmi_clk_atomic_enable; 289 ops->disable = scmi_clk_atomic_disable; 290 } else { 291 ops->prepare = scmi_clk_enable; 292 ops->unprepare = scmi_clk_disable; 293 } 294 } 295 296 if (feats_key & BIT(SCMI_CLK_ATOMIC_SUPPORTED)) 297 ops->is_enabled = scmi_clk_atomic_is_enabled; 298 else 299 ops->is_prepared = scmi_clk_is_enabled; 300 301 /* Rate ops */ 302 ops->recalc_rate = scmi_clk_recalc_rate; 303 ops->round_rate = scmi_clk_round_rate; 304 ops->determine_rate = scmi_clk_determine_rate; 305 if (feats_key & BIT(SCMI_CLK_RATE_CTRL_SUPPORTED)) 306 ops->set_rate = scmi_clk_set_rate; 307 308 /* Parent ops */ 309 ops->get_parent = scmi_clk_get_parent; 310 if (feats_key & BIT(SCMI_CLK_PARENT_CTRL_SUPPORTED)) 311 ops->set_parent = scmi_clk_set_parent; 312 313 /* Duty cycle */ 314 if (feats_key & BIT(SCMI_CLK_DUTY_CYCLE_SUPPORTED)) { 315 ops->get_duty_cycle = scmi_clk_get_duty_cycle; 316 ops->set_duty_cycle = scmi_clk_set_duty_cycle; 317 } 318 319 return ops; 320} 321 322/** 323 * scmi_clk_ops_select() - Select a proper set of clock operations 324 * @sclk: A reference to an SCMI clock descriptor 325 * @atomic_capable: A flag to indicate if atomic mode is supported by the 326 * transport 327 * @atomic_threshold_us: Platform atomic threshold value in microseconds: 328 * clk_ops are atomic when clock enable latency is less 329 * than this threshold 330 * @clk_ops_db: A reference to the array used as a database to store all the 331 * created clock operations combinations. 332 * @db_size: Maximum number of entries held by @clk_ops_db 333 * 334 * After having built a bitmap descriptor to represent the set of features 335 * needed by this SCMI clock, at first use it to lookup into the set of 336 * previously allocated clk_ops to check if a suitable combination of clock 337 * operations was already created; when no match is found allocate a brand new 338 * set of clk_ops satisfying the required combination of features and save it 339 * for future references. 340 * 341 * In this way only one set of clk_ops is ever created for each different 342 * combination that is effectively needed by a driver instance. 343 * 344 * Return: A pointer to the allocated and configured clk_ops on success, or 345 * NULL otherwise. 346 */ 347static const struct clk_ops * 348scmi_clk_ops_select(struct scmi_clk *sclk, bool atomic_capable, 349 unsigned int atomic_threshold_us, 350 const struct clk_ops **clk_ops_db, size_t db_size) 351{ 352 const struct scmi_clock_info *ci = sclk->info; 353 unsigned int feats_key = 0; 354 const struct clk_ops *ops; 355 356 /* 357 * Note that when transport is atomic but SCMI protocol did not 358 * specify (or support) an enable_latency associated with a 359 * clock, we default to use atomic operations mode. 360 */ 361 if (atomic_capable && ci->enable_latency <= atomic_threshold_us) 362 feats_key |= BIT(SCMI_CLK_ATOMIC_SUPPORTED); 363 364 if (!ci->state_ctrl_forbidden) 365 feats_key |= BIT(SCMI_CLK_STATE_CTRL_SUPPORTED); 366 367 if (!ci->rate_ctrl_forbidden) 368 feats_key |= BIT(SCMI_CLK_RATE_CTRL_SUPPORTED); 369 370 if (!ci->parent_ctrl_forbidden) 371 feats_key |= BIT(SCMI_CLK_PARENT_CTRL_SUPPORTED); 372 373 if (ci->extended_config) 374 feats_key |= BIT(SCMI_CLK_DUTY_CYCLE_SUPPORTED); 375 376 if (WARN_ON(feats_key >= db_size)) 377 return NULL; 378 379 /* Lookup previously allocated ops */ 380 ops = clk_ops_db[feats_key]; 381 if (ops) 382 return ops; 383 384 /* Did not find a pre-allocated clock_ops */ 385 ops = scmi_clk_ops_alloc(sclk->dev, feats_key); 386 if (!ops) 387 return NULL; 388 389 /* Store new ops combinations */ 390 clk_ops_db[feats_key] = ops; 391 392 return ops; 393} 394 395static int scmi_clocks_probe(struct scmi_device *sdev) 396{ 397 int idx, count, err; 398 unsigned int atomic_threshold_us; 399 bool transport_is_atomic; 400 struct clk_hw **hws; 401 struct clk_hw_onecell_data *clk_data; 402 struct device *dev = &sdev->dev; 403 struct device_node *np = dev->of_node; 404 const struct scmi_handle *handle = sdev->handle; 405 struct scmi_protocol_handle *ph; 406 const struct clk_ops *scmi_clk_ops_db[SCMI_MAX_CLK_OPS] = {}; 407 struct scmi_clk *sclks; 408 409 if (!handle) 410 return -ENODEV; 411 412 scmi_proto_clk_ops = 413 handle->devm_protocol_get(sdev, SCMI_PROTOCOL_CLOCK, &ph); 414 if (IS_ERR(scmi_proto_clk_ops)) 415 return PTR_ERR(scmi_proto_clk_ops); 416 417 count = scmi_proto_clk_ops->count_get(ph); 418 if (count < 0) { 419 dev_err(dev, "%pOFn: invalid clock output count\n", np); 420 return -EINVAL; 421 } 422 423 clk_data = devm_kzalloc(dev, struct_size(clk_data, hws, count), 424 GFP_KERNEL); 425 if (!clk_data) 426 return -ENOMEM; 427 428 clk_data->num = count; 429 hws = clk_data->hws; 430 431 transport_is_atomic = handle->is_transport_atomic(handle, 432 &atomic_threshold_us); 433 434 sclks = devm_kcalloc(dev, count, sizeof(*sclks), GFP_KERNEL); 435 if (!sclks) 436 return -ENOMEM; 437 438 for (idx = 0; idx < count; idx++) 439 hws[idx] = &sclks[idx].hw; 440 441 for (idx = 0; idx < count; idx++) { 442 struct scmi_clk *sclk = &sclks[idx]; 443 const struct clk_ops *scmi_ops; 444 445 sclk->info = scmi_proto_clk_ops->info_get(ph, idx); 446 if (!sclk->info) { 447 dev_dbg(dev, "invalid clock info for idx %d\n", idx); 448 hws[idx] = NULL; 449 continue; 450 } 451 452 sclk->id = idx; 453 sclk->ph = ph; 454 sclk->dev = dev; 455 456 /* 457 * Note that the scmi_clk_ops_db is on the stack, not global, 458 * because it cannot be shared between multiple probe-sequences 459 * to avoid sharing the devm_ allocated clk_ops between multiple 460 * SCMI clk driver instances. 461 */ 462 scmi_ops = scmi_clk_ops_select(sclk, transport_is_atomic, 463 atomic_threshold_us, 464 scmi_clk_ops_db, 465 ARRAY_SIZE(scmi_clk_ops_db)); 466 if (!scmi_ops) 467 return -ENOMEM; 468 469 /* Initialize clock parent data. */ 470 if (sclk->info->num_parents > 0) { 471 sclk->parent_data = devm_kcalloc(dev, sclk->info->num_parents, 472 sizeof(*sclk->parent_data), GFP_KERNEL); 473 if (!sclk->parent_data) 474 return -ENOMEM; 475 476 for (int i = 0; i < sclk->info->num_parents; i++) { 477 sclk->parent_data[i].index = sclk->info->parents[i]; 478 sclk->parent_data[i].hw = hws[sclk->info->parents[i]]; 479 } 480 } 481 482 err = scmi_clk_ops_init(dev, sclk, scmi_ops); 483 if (err) { 484 dev_err(dev, "failed to register clock %d\n", idx); 485 devm_kfree(dev, sclk->parent_data); 486 hws[idx] = NULL; 487 } else { 488 dev_dbg(dev, "Registered clock:%s%s\n", 489 sclk->info->name, 490 scmi_ops->enable ? " (atomic ops)" : ""); 491 } 492 } 493 494 return devm_of_clk_add_hw_provider(dev, of_clk_hw_onecell_get, 495 clk_data); 496} 497 498static const struct scmi_device_id scmi_id_table[] = { 499 { SCMI_PROTOCOL_CLOCK, "clocks" }, 500 { }, 501}; 502MODULE_DEVICE_TABLE(scmi, scmi_id_table); 503 504static struct scmi_driver scmi_clocks_driver = { 505 .name = "scmi-clocks", 506 .probe = scmi_clocks_probe, 507 .id_table = scmi_id_table, 508}; 509module_scmi_driver(scmi_clocks_driver); 510 511MODULE_AUTHOR("Sudeep Holla <sudeep.holla@arm.com>"); 512MODULE_DESCRIPTION("ARM SCMI clock driver"); 513MODULE_LICENSE("GPL v2");