Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

perf/amlogic: Add support for Amlogic meson G12 SoC DDR PMU driver

Add support for Amlogic Meson G12 Series SOC - DDR bandwidth PMU driver
framework and interfaces. The PMU can not only monitor the total DDR
bandwidth, but also individual IP module bandwidth.

Signed-off-by: Jiucheng Xu <jiucheng.xu@amlogic.com>
Tested-by: Chris Healy <healych@amazon.com>
Link: https://lore.kernel.org/r/20221121021602.3306998-1-jiucheng.xu@amlogic.com
Signed-off-by: Will Deacon <will@kernel.org>

authored by

Jiucheng Xu and committed by
Will Deacon
2016e211 dcc7f001

+1048
+8
MAINTAINERS
··· 1093 1093 F: Documentation/hid/amd-sfh* 1094 1094 F: drivers/hid/amd-sfh-hid/ 1095 1095 1096 + AMLOGIC DDR PMU DRIVER 1097 + M: Jiucheng Xu <jiucheng.xu@amlogic.com> 1098 + L: linux-amlogic@lists.infradead.org 1099 + S: Supported 1100 + W: http://www.amlogic.com 1101 + F: drivers/perf/amlogic/ 1102 + F: include/soc/amlogic/ 1103 + 1096 1104 AMPHION VPU CODEC V4L2 DRIVER 1097 1105 M: Ming Qian <ming.qian@nxp.com> 1098 1106 M: Shijie Qin <shijie.qin@nxp.com>
+2
drivers/perf/Kconfig
··· 201 201 202 202 source "drivers/perf/arm_cspmu/Kconfig" 203 203 204 + source "drivers/perf/amlogic/Kconfig" 205 + 204 206 endmenu
+1
drivers/perf/Makefile
··· 22 22 obj-$(CONFIG_APPLE_M1_CPU_PMU) += apple_m1_cpu_pmu.o 23 23 obj-$(CONFIG_ALIBABA_UNCORE_DRW_PMU) += alibaba_uncore_drw_pmu.o 24 24 obj-$(CONFIG_ARM_CORESIGHT_PMU_ARCH_SYSTEM_PMU) += arm_cspmu/ 25 + obj-$(CONFIG_MESON_DDR_PMU) += amlogic/
+10
drivers/perf/amlogic/Kconfig
··· 1 + # SPDX-License-Identifier: GPL-2.0-only 2 + config MESON_DDR_PMU 3 + tristate "Amlogic DDR Bandwidth Performance Monitor" 4 + depends on ARCH_MESON || COMPILE_TEST 5 + help 6 + Provides support for the DDR performance monitor 7 + in Amlogic SoCs, which can give information about 8 + memory throughput and other related events. It 9 + supports multiple channels to monitor the memory 10 + bandwidth simultaneously.
+5
drivers/perf/amlogic/Makefile
··· 1 + # SPDX-License-Identifier: GPL-2.0-only 2 + 3 + obj-$(CONFIG_MESON_DDR_PMU) += meson_ddr_pmu_g12.o 4 + 5 + meson_ddr_pmu_g12-y := meson_ddr_pmu_core.o meson_g12_ddr_pmu.o
+562
drivers/perf/amlogic/meson_ddr_pmu_core.c
··· 1 + // SPDX-License-Identifier: GPL-2.0 2 + /* 3 + * Copyright (c) 2022 Amlogic, Inc. All rights reserved. 4 + */ 5 + 6 + #include <linux/bitfield.h> 7 + #include <linux/init.h> 8 + #include <linux/irqreturn.h> 9 + #include <linux/kernel.h> 10 + #include <linux/module.h> 11 + #include <linux/of.h> 12 + #include <linux/of_device.h> 13 + #include <linux/of_irq.h> 14 + #include <linux/perf_event.h> 15 + #include <linux/platform_device.h> 16 + #include <linux/printk.h> 17 + #include <linux/sysfs.h> 18 + #include <linux/types.h> 19 + #include <linux/version.h> 20 + 21 + #include <soc/amlogic/meson_ddr_pmu.h> 22 + 23 + struct ddr_pmu { 24 + struct pmu pmu; 25 + struct dmc_info info; 26 + struct dmc_counter counters; /* save counters from hw */ 27 + bool pmu_enabled; 28 + struct device *dev; 29 + char *name; 30 + struct hlist_node node; 31 + enum cpuhp_state cpuhp_state; 32 + int cpu; /* for cpu hotplug */ 33 + }; 34 + 35 + #define DDR_PERF_DEV_NAME "meson_ddr_bw" 36 + #define MAX_AXI_PORTS_OF_CHANNEL 4 /* A DMC channel can monitor max 4 axi ports */ 37 + 38 + #define to_ddr_pmu(p) container_of(p, struct ddr_pmu, pmu) 39 + #define dmc_info_to_pmu(p) container_of(p, struct ddr_pmu, info) 40 + 41 + static void dmc_pmu_enable(struct ddr_pmu *pmu) 42 + { 43 + if (!pmu->pmu_enabled) 44 + pmu->info.hw_info->enable(&pmu->info); 45 + 46 + pmu->pmu_enabled = true; 47 + } 48 + 49 + static void dmc_pmu_disable(struct ddr_pmu *pmu) 50 + { 51 + if (pmu->pmu_enabled) 52 + pmu->info.hw_info->disable(&pmu->info); 53 + 54 + pmu->pmu_enabled = false; 55 + } 56 + 57 + static void meson_ddr_set_axi_filter(struct perf_event *event, u8 axi_id) 58 + { 59 + struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); 60 + int chann; 61 + 62 + if (event->attr.config > ALL_CHAN_COUNTER_ID && 63 + event->attr.config < COUNTER_MAX_ID) { 64 + chann = event->attr.config - CHAN1_COUNTER_ID; 65 + 66 + pmu->info.hw_info->set_axi_filter(&pmu->info, axi_id, chann); 67 + } 68 + } 69 + 70 + static void ddr_cnt_addition(struct dmc_counter *sum, 71 + struct dmc_counter *add1, 72 + struct dmc_counter *add2, 73 + int chann_nr) 74 + { 75 + int i; 76 + u64 cnt1, cnt2; 77 + 78 + sum->all_cnt = add1->all_cnt + add2->all_cnt; 79 + sum->all_req = add1->all_req + add2->all_req; 80 + for (i = 0; i < chann_nr; i++) { 81 + cnt1 = add1->channel_cnt[i]; 82 + cnt2 = add2->channel_cnt[i]; 83 + 84 + sum->channel_cnt[i] = cnt1 + cnt2; 85 + } 86 + } 87 + 88 + static void meson_ddr_perf_event_update(struct perf_event *event) 89 + { 90 + struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); 91 + u64 new_raw_count = 0; 92 + struct dmc_counter dc = {0}, sum_dc = {0}; 93 + int idx; 94 + int chann_nr = pmu->info.hw_info->chann_nr; 95 + 96 + /* get the remain counters in register. */ 97 + pmu->info.hw_info->get_counters(&pmu->info, &dc); 98 + 99 + ddr_cnt_addition(&sum_dc, &pmu->counters, &dc, chann_nr); 100 + 101 + switch (event->attr.config) { 102 + case ALL_CHAN_COUNTER_ID: 103 + new_raw_count = sum_dc.all_cnt; 104 + break; 105 + case CHAN1_COUNTER_ID: 106 + case CHAN2_COUNTER_ID: 107 + case CHAN3_COUNTER_ID: 108 + case CHAN4_COUNTER_ID: 109 + case CHAN5_COUNTER_ID: 110 + case CHAN6_COUNTER_ID: 111 + case CHAN7_COUNTER_ID: 112 + case CHAN8_COUNTER_ID: 113 + idx = event->attr.config - CHAN1_COUNTER_ID; 114 + new_raw_count = sum_dc.channel_cnt[idx]; 115 + break; 116 + } 117 + 118 + local64_set(&event->count, new_raw_count); 119 + } 120 + 121 + static int meson_ddr_perf_event_init(struct perf_event *event) 122 + { 123 + struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); 124 + u64 config1 = event->attr.config1; 125 + u64 config2 = event->attr.config2; 126 + 127 + if (event->attr.type != event->pmu->type) 128 + return -ENOENT; 129 + 130 + if (is_sampling_event(event) || event->attach_state & PERF_ATTACH_TASK) 131 + return -EOPNOTSUPP; 132 + 133 + if (event->cpu < 0) 134 + return -EOPNOTSUPP; 135 + 136 + /* check if the number of parameters is too much */ 137 + if (event->attr.config != ALL_CHAN_COUNTER_ID && 138 + hweight64(config1) + hweight64(config2) > MAX_AXI_PORTS_OF_CHANNEL) 139 + return -EOPNOTSUPP; 140 + 141 + event->cpu = pmu->cpu; 142 + 143 + return 0; 144 + } 145 + 146 + static void meson_ddr_perf_event_start(struct perf_event *event, int flags) 147 + { 148 + struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); 149 + 150 + memset(&pmu->counters, 0, sizeof(pmu->counters)); 151 + dmc_pmu_enable(pmu); 152 + } 153 + 154 + static int meson_ddr_perf_event_add(struct perf_event *event, int flags) 155 + { 156 + u64 config1 = event->attr.config1; 157 + u64 config2 = event->attr.config2; 158 + int i; 159 + 160 + for_each_set_bit(i, (const unsigned long *)&config1, sizeof(config1)) 161 + meson_ddr_set_axi_filter(event, i); 162 + 163 + for_each_set_bit(i, (const unsigned long *)&config2, sizeof(config2)) 164 + meson_ddr_set_axi_filter(event, i + 64); 165 + 166 + if (flags & PERF_EF_START) 167 + meson_ddr_perf_event_start(event, flags); 168 + 169 + return 0; 170 + } 171 + 172 + static void meson_ddr_perf_event_stop(struct perf_event *event, int flags) 173 + { 174 + struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); 175 + 176 + if (flags & PERF_EF_UPDATE) 177 + meson_ddr_perf_event_update(event); 178 + 179 + dmc_pmu_disable(pmu); 180 + } 181 + 182 + static void meson_ddr_perf_event_del(struct perf_event *event, int flags) 183 + { 184 + meson_ddr_perf_event_stop(event, PERF_EF_UPDATE); 185 + } 186 + 187 + static ssize_t meson_ddr_perf_cpumask_show(struct device *dev, 188 + struct device_attribute *attr, 189 + char *buf) 190 + { 191 + struct ddr_pmu *pmu = dev_get_drvdata(dev); 192 + 193 + return cpumap_print_to_pagebuf(true, buf, cpumask_of(pmu->cpu)); 194 + } 195 + 196 + static struct device_attribute meson_ddr_perf_cpumask_attr = 197 + __ATTR(cpumask, 0444, meson_ddr_perf_cpumask_show, NULL); 198 + 199 + static struct attribute *meson_ddr_perf_cpumask_attrs[] = { 200 + &meson_ddr_perf_cpumask_attr.attr, 201 + NULL, 202 + }; 203 + 204 + static const struct attribute_group ddr_perf_cpumask_attr_group = { 205 + .attrs = meson_ddr_perf_cpumask_attrs, 206 + }; 207 + 208 + static ssize_t 209 + pmu_event_show(struct device *dev, struct device_attribute *attr, 210 + char *page) 211 + { 212 + struct perf_pmu_events_attr *pmu_attr; 213 + 214 + pmu_attr = container_of(attr, struct perf_pmu_events_attr, attr); 215 + return sysfs_emit(page, "event=0x%02llx\n", pmu_attr->id); 216 + } 217 + 218 + static ssize_t 219 + event_show_unit(struct device *dev, struct device_attribute *attr, 220 + char *page) 221 + { 222 + return sysfs_emit(page, "MB\n"); 223 + } 224 + 225 + static ssize_t 226 + event_show_scale(struct device *dev, struct device_attribute *attr, 227 + char *page) 228 + { 229 + /* one count = 16byte = 1.52587890625e-05 MB */ 230 + return sysfs_emit(page, "1.52587890625e-05\n"); 231 + } 232 + 233 + #define AML_DDR_PMU_EVENT_ATTR(_name, _id) \ 234 + { \ 235 + .attr = __ATTR(_name, 0444, pmu_event_show, NULL), \ 236 + .id = _id, \ 237 + } 238 + 239 + #define AML_DDR_PMU_EVENT_UNIT_ATTR(_name) \ 240 + __ATTR(_name.unit, 0444, event_show_unit, NULL) 241 + 242 + #define AML_DDR_PMU_EVENT_SCALE_ATTR(_name) \ 243 + __ATTR(_name.scale, 0444, event_show_scale, NULL) 244 + 245 + static struct device_attribute event_unit_attrs[] = { 246 + AML_DDR_PMU_EVENT_UNIT_ATTR(total_rw_bytes), 247 + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_1_rw_bytes), 248 + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_2_rw_bytes), 249 + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_3_rw_bytes), 250 + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_4_rw_bytes), 251 + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_5_rw_bytes), 252 + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_6_rw_bytes), 253 + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_7_rw_bytes), 254 + AML_DDR_PMU_EVENT_UNIT_ATTR(chan_8_rw_bytes), 255 + }; 256 + 257 + static struct device_attribute event_scale_attrs[] = { 258 + AML_DDR_PMU_EVENT_SCALE_ATTR(total_rw_bytes), 259 + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_1_rw_bytes), 260 + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_2_rw_bytes), 261 + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_3_rw_bytes), 262 + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_4_rw_bytes), 263 + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_5_rw_bytes), 264 + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_6_rw_bytes), 265 + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_7_rw_bytes), 266 + AML_DDR_PMU_EVENT_SCALE_ATTR(chan_8_rw_bytes), 267 + }; 268 + 269 + static struct perf_pmu_events_attr event_attrs[] = { 270 + AML_DDR_PMU_EVENT_ATTR(total_rw_bytes, ALL_CHAN_COUNTER_ID), 271 + AML_DDR_PMU_EVENT_ATTR(chan_1_rw_bytes, CHAN1_COUNTER_ID), 272 + AML_DDR_PMU_EVENT_ATTR(chan_2_rw_bytes, CHAN2_COUNTER_ID), 273 + AML_DDR_PMU_EVENT_ATTR(chan_3_rw_bytes, CHAN3_COUNTER_ID), 274 + AML_DDR_PMU_EVENT_ATTR(chan_4_rw_bytes, CHAN4_COUNTER_ID), 275 + AML_DDR_PMU_EVENT_ATTR(chan_5_rw_bytes, CHAN5_COUNTER_ID), 276 + AML_DDR_PMU_EVENT_ATTR(chan_6_rw_bytes, CHAN6_COUNTER_ID), 277 + AML_DDR_PMU_EVENT_ATTR(chan_7_rw_bytes, CHAN7_COUNTER_ID), 278 + AML_DDR_PMU_EVENT_ATTR(chan_8_rw_bytes, CHAN8_COUNTER_ID), 279 + }; 280 + 281 + /* three attrs are combined an event */ 282 + static struct attribute *ddr_perf_events_attrs[COUNTER_MAX_ID * 3]; 283 + 284 + static struct attribute_group ddr_perf_events_attr_group = { 285 + .name = "events", 286 + .attrs = ddr_perf_events_attrs, 287 + }; 288 + 289 + static umode_t meson_ddr_perf_format_attr_visible(struct kobject *kobj, 290 + struct attribute *attr, 291 + int n) 292 + { 293 + struct pmu *pmu = dev_get_drvdata(kobj_to_dev(kobj)); 294 + struct ddr_pmu *ddr_pmu = to_ddr_pmu(pmu); 295 + const u64 *capability = ddr_pmu->info.hw_info->capability; 296 + struct device_attribute *dev_attr; 297 + int id; 298 + char value[20]; // config1:xxx, 20 is enough 299 + 300 + dev_attr = container_of(attr, struct device_attribute, attr); 301 + dev_attr->show(NULL, NULL, value); 302 + 303 + if (sscanf(value, "config1:%d", &id) == 1) 304 + return capability[0] & (1ULL << id) ? attr->mode : 0; 305 + 306 + if (sscanf(value, "config2:%d", &id) == 1) 307 + return capability[1] & (1ULL << id) ? attr->mode : 0; 308 + 309 + return attr->mode; 310 + } 311 + 312 + static struct attribute_group ddr_perf_format_attr_group = { 313 + .name = "format", 314 + .is_visible = meson_ddr_perf_format_attr_visible, 315 + }; 316 + 317 + static ssize_t meson_ddr_perf_identifier_show(struct device *dev, 318 + struct device_attribute *attr, 319 + char *page) 320 + { 321 + struct ddr_pmu *pmu = dev_get_drvdata(dev); 322 + 323 + return sysfs_emit(page, "%s\n", pmu->name); 324 + } 325 + 326 + static struct device_attribute meson_ddr_perf_identifier_attr = 327 + __ATTR(identifier, 0444, meson_ddr_perf_identifier_show, NULL); 328 + 329 + static struct attribute *meson_ddr_perf_identifier_attrs[] = { 330 + &meson_ddr_perf_identifier_attr.attr, 331 + NULL, 332 + }; 333 + 334 + static const struct attribute_group ddr_perf_identifier_attr_group = { 335 + .attrs = meson_ddr_perf_identifier_attrs, 336 + }; 337 + 338 + static const struct attribute_group *attr_groups[] = { 339 + &ddr_perf_events_attr_group, 340 + &ddr_perf_format_attr_group, 341 + &ddr_perf_cpumask_attr_group, 342 + &ddr_perf_identifier_attr_group, 343 + NULL, 344 + }; 345 + 346 + static irqreturn_t dmc_irq_handler(int irq, void *dev_id) 347 + { 348 + struct dmc_info *info = dev_id; 349 + struct ddr_pmu *pmu; 350 + struct dmc_counter counters, *sum_cnter; 351 + int i; 352 + 353 + pmu = dmc_info_to_pmu(info); 354 + 355 + if (info->hw_info->irq_handler(info, &counters) != 0) 356 + goto out; 357 + 358 + sum_cnter = &pmu->counters; 359 + sum_cnter->all_cnt += counters.all_cnt; 360 + sum_cnter->all_req += counters.all_req; 361 + 362 + for (i = 0; i < pmu->info.hw_info->chann_nr; i++) 363 + sum_cnter->channel_cnt[i] += counters.channel_cnt[i]; 364 + 365 + if (pmu->pmu_enabled) 366 + /* 367 + * the timer interrupt only supprt 368 + * one shot mode, we have to re-enable 369 + * it in ISR to support continue mode. 370 + */ 371 + info->hw_info->enable(info); 372 + 373 + dev_dbg(pmu->dev, "counts: %llu %llu %llu, %llu, %llu, %llu\t\t" 374 + "sum: %llu %llu %llu, %llu, %llu, %llu\n", 375 + counters.all_req, 376 + counters.all_cnt, 377 + counters.channel_cnt[0], 378 + counters.channel_cnt[1], 379 + counters.channel_cnt[2], 380 + counters.channel_cnt[3], 381 + 382 + pmu->counters.all_req, 383 + pmu->counters.all_cnt, 384 + pmu->counters.channel_cnt[0], 385 + pmu->counters.channel_cnt[1], 386 + pmu->counters.channel_cnt[2], 387 + pmu->counters.channel_cnt[3]); 388 + out: 389 + return IRQ_HANDLED; 390 + } 391 + 392 + static int ddr_perf_offline_cpu(unsigned int cpu, struct hlist_node *node) 393 + { 394 + struct ddr_pmu *pmu = hlist_entry_safe(node, struct ddr_pmu, node); 395 + int target; 396 + 397 + if (cpu != pmu->cpu) 398 + return 0; 399 + 400 + target = cpumask_any_but(cpu_online_mask, cpu); 401 + if (target >= nr_cpu_ids) 402 + return 0; 403 + 404 + perf_pmu_migrate_context(&pmu->pmu, cpu, target); 405 + pmu->cpu = target; 406 + 407 + WARN_ON(irq_set_affinity(pmu->info.irq_num, cpumask_of(pmu->cpu))); 408 + 409 + return 0; 410 + } 411 + 412 + static void fill_event_attr(struct ddr_pmu *pmu) 413 + { 414 + int i, j, k; 415 + struct attribute **dst = ddr_perf_events_attrs; 416 + 417 + j = 0; 418 + k = 0; 419 + 420 + /* fill ALL_CHAN_COUNTER_ID event */ 421 + dst[j++] = &event_attrs[k].attr.attr; 422 + dst[j++] = &event_unit_attrs[k].attr; 423 + dst[j++] = &event_scale_attrs[k].attr; 424 + 425 + k++; 426 + 427 + /* fill each channel event */ 428 + for (i = 0; i < pmu->info.hw_info->chann_nr; i++, k++) { 429 + dst[j++] = &event_attrs[k].attr.attr; 430 + dst[j++] = &event_unit_attrs[k].attr; 431 + dst[j++] = &event_scale_attrs[k].attr; 432 + } 433 + 434 + dst[j] = NULL; /* mark end */ 435 + } 436 + 437 + static void fmt_attr_fill(struct attribute **fmt_attr) 438 + { 439 + ddr_perf_format_attr_group.attrs = fmt_attr; 440 + } 441 + 442 + static int ddr_pmu_parse_dt(struct platform_device *pdev, 443 + struct dmc_info *info) 444 + { 445 + void __iomem *base; 446 + int i, ret; 447 + 448 + info->hw_info = of_device_get_match_data(&pdev->dev); 449 + 450 + for (i = 0; i < info->hw_info->dmc_nr; i++) { 451 + /* resource 0 for ddr register base */ 452 + base = devm_platform_ioremap_resource(pdev, i); 453 + if (IS_ERR(base)) 454 + return PTR_ERR(base); 455 + 456 + info->ddr_reg[i] = base; 457 + } 458 + 459 + /* resource i for pll register base */ 460 + base = devm_platform_ioremap_resource(pdev, i); 461 + if (IS_ERR(base)) 462 + return PTR_ERR(base); 463 + 464 + info->pll_reg = base; 465 + 466 + ret = platform_get_irq(pdev, 0); 467 + if (ret < 0) 468 + return ret; 469 + 470 + info->irq_num = ret; 471 + 472 + ret = devm_request_irq(&pdev->dev, info->irq_num, dmc_irq_handler, 473 + IRQF_NOBALANCING, dev_name(&pdev->dev), 474 + (void *)info); 475 + if (ret < 0) 476 + return ret; 477 + 478 + return 0; 479 + } 480 + 481 + int meson_ddr_pmu_create(struct platform_device *pdev) 482 + { 483 + int ret; 484 + char *name; 485 + struct ddr_pmu *pmu; 486 + 487 + pmu = devm_kzalloc(&pdev->dev, sizeof(struct ddr_pmu), GFP_KERNEL); 488 + if (!pmu) 489 + return -ENOMEM; 490 + 491 + *pmu = (struct ddr_pmu) { 492 + .pmu = { 493 + .module = THIS_MODULE, 494 + .capabilities = PERF_PMU_CAP_NO_EXCLUDE, 495 + .task_ctx_nr = perf_invalid_context, 496 + .attr_groups = attr_groups, 497 + .event_init = meson_ddr_perf_event_init, 498 + .add = meson_ddr_perf_event_add, 499 + .del = meson_ddr_perf_event_del, 500 + .start = meson_ddr_perf_event_start, 501 + .stop = meson_ddr_perf_event_stop, 502 + .read = meson_ddr_perf_event_update, 503 + }, 504 + }; 505 + 506 + ret = ddr_pmu_parse_dt(pdev, &pmu->info); 507 + if (ret < 0) 508 + return ret; 509 + 510 + fmt_attr_fill(pmu->info.hw_info->fmt_attr); 511 + 512 + pmu->cpu = smp_processor_id(); 513 + 514 + name = devm_kasprintf(&pdev->dev, GFP_KERNEL, DDR_PERF_DEV_NAME); 515 + if (!name) 516 + return -ENOMEM; 517 + 518 + ret = cpuhp_setup_state_multi(CPUHP_AP_ONLINE_DYN, name, NULL, 519 + ddr_perf_offline_cpu); 520 + if (ret < 0) 521 + return ret; 522 + 523 + pmu->cpuhp_state = ret; 524 + 525 + /* Register the pmu instance for cpu hotplug */ 526 + ret = cpuhp_state_add_instance_nocalls(pmu->cpuhp_state, &pmu->node); 527 + if (ret) 528 + goto cpuhp_instance_err; 529 + 530 + fill_event_attr(pmu); 531 + 532 + ret = perf_pmu_register(&pmu->pmu, name, -1); 533 + if (ret) 534 + goto pmu_register_err; 535 + 536 + pmu->name = name; 537 + pmu->dev = &pdev->dev; 538 + pmu->pmu_enabled = false; 539 + 540 + platform_set_drvdata(pdev, pmu); 541 + 542 + return 0; 543 + 544 + pmu_register_err: 545 + cpuhp_state_remove_instance_nocalls(pmu->cpuhp_state, &pmu->node); 546 + 547 + cpuhp_instance_err: 548 + cpuhp_remove_state(pmu->cpuhp_state); 549 + 550 + return ret; 551 + } 552 + 553 + int meson_ddr_pmu_remove(struct platform_device *pdev) 554 + { 555 + struct ddr_pmu *pmu = platform_get_drvdata(pdev); 556 + 557 + perf_pmu_unregister(&pmu->pmu); 558 + cpuhp_state_remove_instance_nocalls(pmu->cpuhp_state, &pmu->node); 559 + cpuhp_remove_state(pmu->cpuhp_state); 560 + 561 + return 0; 562 + }
+394
drivers/perf/amlogic/meson_g12_ddr_pmu.c
··· 1 + // SPDX-License-Identifier: GPL-2.0 2 + /* 3 + * Copyright (c) 2022 Amlogic, Inc. All rights reserved. 4 + */ 5 + 6 + #include <linux/err.h> 7 + #include <linux/kernel.h> 8 + #include <linux/module.h> 9 + #include <linux/of.h> 10 + #include <linux/perf_event.h> 11 + #include <linux/platform_device.h> 12 + #include <linux/printk.h> 13 + #include <linux/types.h> 14 + #include <linux/version.h> 15 + 16 + #include <soc/amlogic/meson_ddr_pmu.h> 17 + 18 + #define PORT_MAJOR 32 19 + #define DEFAULT_XTAL_FREQ 24000000UL 20 + 21 + #define DMC_QOS_IRQ BIT(30) 22 + 23 + /* DMC bandwidth monitor register address offset */ 24 + #define DMC_MON_G12_CTRL0 (0x20 << 2) 25 + #define DMC_MON_G12_CTRL1 (0x21 << 2) 26 + #define DMC_MON_G12_CTRL2 (0x22 << 2) 27 + #define DMC_MON_G12_CTRL3 (0x23 << 2) 28 + #define DMC_MON_G12_CTRL4 (0x24 << 2) 29 + #define DMC_MON_G12_CTRL5 (0x25 << 2) 30 + #define DMC_MON_G12_CTRL6 (0x26 << 2) 31 + #define DMC_MON_G12_CTRL7 (0x27 << 2) 32 + #define DMC_MON_G12_CTRL8 (0x28 << 2) 33 + 34 + #define DMC_MON_G12_ALL_REQ_CNT (0x29 << 2) 35 + #define DMC_MON_G12_ALL_GRANT_CNT (0x2a << 2) 36 + #define DMC_MON_G12_ONE_GRANT_CNT (0x2b << 2) 37 + #define DMC_MON_G12_SEC_GRANT_CNT (0x2c << 2) 38 + #define DMC_MON_G12_THD_GRANT_CNT (0x2d << 2) 39 + #define DMC_MON_G12_FOR_GRANT_CNT (0x2e << 2) 40 + #define DMC_MON_G12_TIMER (0x2f << 2) 41 + 42 + /* Each bit represent a axi line */ 43 + PMU_FORMAT_ATTR(event, "config:0-7"); 44 + PMU_FORMAT_ATTR(arm, "config1:0"); 45 + PMU_FORMAT_ATTR(gpu, "config1:1"); 46 + PMU_FORMAT_ATTR(pcie, "config1:2"); 47 + PMU_FORMAT_ATTR(hdcp, "config1:3"); 48 + PMU_FORMAT_ATTR(hevc_front, "config1:4"); 49 + PMU_FORMAT_ATTR(usb3_0, "config1:6"); 50 + PMU_FORMAT_ATTR(device, "config1:7"); 51 + PMU_FORMAT_ATTR(hevc_back, "config1:8"); 52 + PMU_FORMAT_ATTR(h265enc, "config1:9"); 53 + PMU_FORMAT_ATTR(vpu_read1, "config1:16"); 54 + PMU_FORMAT_ATTR(vpu_read2, "config1:17"); 55 + PMU_FORMAT_ATTR(vpu_read3, "config1:18"); 56 + PMU_FORMAT_ATTR(vpu_write1, "config1:19"); 57 + PMU_FORMAT_ATTR(vpu_write2, "config1:20"); 58 + PMU_FORMAT_ATTR(vdec, "config1:21"); 59 + PMU_FORMAT_ATTR(hcodec, "config1:22"); 60 + PMU_FORMAT_ATTR(ge2d, "config1:23"); 61 + 62 + PMU_FORMAT_ATTR(spicc1, "config1:32"); 63 + PMU_FORMAT_ATTR(usb0, "config1:33"); 64 + PMU_FORMAT_ATTR(dma, "config1:34"); 65 + PMU_FORMAT_ATTR(arb0, "config1:35"); 66 + PMU_FORMAT_ATTR(sd_emmc_b, "config1:36"); 67 + PMU_FORMAT_ATTR(usb1, "config1:37"); 68 + PMU_FORMAT_ATTR(audio, "config1:38"); 69 + PMU_FORMAT_ATTR(aififo, "config1:39"); 70 + PMU_FORMAT_ATTR(parser, "config1:41"); 71 + PMU_FORMAT_ATTR(ao_cpu, "config1:42"); 72 + PMU_FORMAT_ATTR(sd_emmc_c, "config1:43"); 73 + PMU_FORMAT_ATTR(spicc2, "config1:44"); 74 + PMU_FORMAT_ATTR(ethernet, "config1:45"); 75 + PMU_FORMAT_ATTR(sana, "config1:46"); 76 + 77 + /* for sm1 and g12b */ 78 + PMU_FORMAT_ATTR(nna, "config1:10"); 79 + 80 + /* for g12b only */ 81 + PMU_FORMAT_ATTR(gdc, "config1:11"); 82 + PMU_FORMAT_ATTR(mipi_isp, "config1:12"); 83 + PMU_FORMAT_ATTR(arm1, "config1:13"); 84 + PMU_FORMAT_ATTR(sd_emmc_a, "config1:40"); 85 + 86 + static struct attribute *g12_pmu_format_attrs[] = { 87 + &format_attr_event.attr, 88 + &format_attr_arm.attr, 89 + &format_attr_gpu.attr, 90 + &format_attr_nna.attr, 91 + &format_attr_gdc.attr, 92 + &format_attr_arm1.attr, 93 + &format_attr_mipi_isp.attr, 94 + &format_attr_sd_emmc_a.attr, 95 + &format_attr_pcie.attr, 96 + &format_attr_hdcp.attr, 97 + &format_attr_hevc_front.attr, 98 + &format_attr_usb3_0.attr, 99 + &format_attr_device.attr, 100 + &format_attr_hevc_back.attr, 101 + &format_attr_h265enc.attr, 102 + &format_attr_vpu_read1.attr, 103 + &format_attr_vpu_read2.attr, 104 + &format_attr_vpu_read3.attr, 105 + &format_attr_vpu_write1.attr, 106 + &format_attr_vpu_write2.attr, 107 + &format_attr_vdec.attr, 108 + &format_attr_hcodec.attr, 109 + &format_attr_ge2d.attr, 110 + &format_attr_spicc1.attr, 111 + &format_attr_usb0.attr, 112 + &format_attr_dma.attr, 113 + &format_attr_arb0.attr, 114 + &format_attr_sd_emmc_b.attr, 115 + &format_attr_usb1.attr, 116 + &format_attr_audio.attr, 117 + &format_attr_aififo.attr, 118 + &format_attr_parser.attr, 119 + &format_attr_ao_cpu.attr, 120 + &format_attr_sd_emmc_c.attr, 121 + &format_attr_spicc2.attr, 122 + &format_attr_ethernet.attr, 123 + &format_attr_sana.attr, 124 + NULL, 125 + }; 126 + 127 + /* calculate ddr clock */ 128 + static unsigned long dmc_g12_get_freq_quick(struct dmc_info *info) 129 + { 130 + unsigned int val; 131 + unsigned int n, m, od1; 132 + unsigned int od_div = 0xfff; 133 + unsigned long freq = 0; 134 + 135 + val = readl(info->pll_reg); 136 + val = val & 0xfffff; 137 + switch ((val >> 16) & 7) { 138 + case 0: 139 + od_div = 2; 140 + break; 141 + 142 + case 1: 143 + od_div = 3; 144 + break; 145 + 146 + case 2: 147 + od_div = 4; 148 + break; 149 + 150 + case 3: 151 + od_div = 6; 152 + break; 153 + 154 + case 4: 155 + od_div = 8; 156 + break; 157 + 158 + default: 159 + break; 160 + } 161 + 162 + m = val & 0x1ff; 163 + n = ((val >> 10) & 0x1f); 164 + od1 = (((val >> 19) & 0x1)) == 1 ? 2 : 1; 165 + freq = DEFAULT_XTAL_FREQ / 1000; /* avoid overflow */ 166 + if (n) 167 + freq = ((((freq * m) / n) >> od1) / od_div) * 1000; 168 + 169 + return freq; 170 + } 171 + 172 + #ifdef DEBUG 173 + static void g12_dump_reg(struct dmc_info *db) 174 + { 175 + int s = 0, i; 176 + unsigned int r; 177 + 178 + for (i = 0; i < 9; i++) { 179 + r = readl(db->ddr_reg[0] + (DMC_MON_G12_CTRL0 + (i << 2))); 180 + pr_notice("DMC_MON_CTRL%d: %08x\n", i, r); 181 + } 182 + r = readl(db->ddr_reg[0] + DMC_MON_G12_ALL_REQ_CNT); 183 + pr_notice("DMC_MON_ALL_REQ_CNT: %08x\n", r); 184 + r = readl(db->ddr_reg[0] + DMC_MON_G12_ALL_GRANT_CNT); 185 + pr_notice("DMC_MON_ALL_GRANT_CNT:%08x\n", r); 186 + r = readl(db->ddr_reg[0] + DMC_MON_G12_ONE_GRANT_CNT); 187 + pr_notice("DMC_MON_ONE_GRANT_CNT:%08x\n", r); 188 + r = readl(db->ddr_reg[0] + DMC_MON_G12_SEC_GRANT_CNT); 189 + pr_notice("DMC_MON_SEC_GRANT_CNT:%08x\n", r); 190 + r = readl(db->ddr_reg[0] + DMC_MON_G12_THD_GRANT_CNT); 191 + pr_notice("DMC_MON_THD_GRANT_CNT:%08x\n", r); 192 + r = readl(db->ddr_reg[0] + DMC_MON_G12_FOR_GRANT_CNT); 193 + pr_notice("DMC_MON_FOR_GRANT_CNT:%08x\n", r); 194 + r = readl(db->ddr_reg[0] + DMC_MON_G12_TIMER); 195 + pr_notice("DMC_MON_TIMER: %08x\n", r); 196 + } 197 + #endif 198 + 199 + static void dmc_g12_counter_enable(struct dmc_info *info) 200 + { 201 + unsigned int val; 202 + unsigned long clock_count = dmc_g12_get_freq_quick(info) / 10; /* 100ms */ 203 + 204 + writel(clock_count, info->ddr_reg[0] + DMC_MON_G12_TIMER); 205 + 206 + val = readl(info->ddr_reg[0] + DMC_MON_G12_CTRL0); 207 + 208 + /* enable all channel */ 209 + val = BIT(31) | /* enable bit */ 210 + BIT(20) | /* use timer */ 211 + 0x0f; /* 4 channels */ 212 + 213 + writel(val, info->ddr_reg[0] + DMC_MON_G12_CTRL0); 214 + 215 + #ifdef DEBUG 216 + g12_dump_reg(info); 217 + #endif 218 + } 219 + 220 + static void dmc_g12_config_fiter(struct dmc_info *info, 221 + int port, int channel) 222 + { 223 + u32 val; 224 + u32 rp[MAX_CHANNEL_NUM] = {DMC_MON_G12_CTRL1, DMC_MON_G12_CTRL3, 225 + DMC_MON_G12_CTRL5, DMC_MON_G12_CTRL7}; 226 + u32 rs[MAX_CHANNEL_NUM] = {DMC_MON_G12_CTRL2, DMC_MON_G12_CTRL4, 227 + DMC_MON_G12_CTRL6, DMC_MON_G12_CTRL8}; 228 + int subport = -1; 229 + 230 + /* clear all port mask */ 231 + if (port < 0) { 232 + writel(0, info->ddr_reg[0] + rp[channel]); 233 + writel(0, info->ddr_reg[0] + rs[channel]); 234 + return; 235 + } 236 + 237 + if (port >= PORT_MAJOR) 238 + subport = port - PORT_MAJOR; 239 + 240 + if (subport < 0) { 241 + val = readl(info->ddr_reg[0] + rp[channel]); 242 + val |= (1 << port); 243 + writel(val, info->ddr_reg[0] + rp[channel]); 244 + val = 0xffff; 245 + writel(val, info->ddr_reg[0] + rs[channel]); 246 + } else { 247 + val = BIT(23); /* select device */ 248 + writel(val, info->ddr_reg[0] + rp[channel]); 249 + val = readl(info->ddr_reg[0] + rs[channel]); 250 + val |= (1 << subport); 251 + writel(val, info->ddr_reg[0] + rs[channel]); 252 + } 253 + } 254 + 255 + static void dmc_g12_set_axi_filter(struct dmc_info *info, int axi_id, int channel) 256 + { 257 + if (channel > info->hw_info->chann_nr) 258 + return; 259 + 260 + dmc_g12_config_fiter(info, axi_id, channel); 261 + } 262 + 263 + static void dmc_g12_counter_disable(struct dmc_info *info) 264 + { 265 + int i; 266 + 267 + /* clear timer */ 268 + writel(0, info->ddr_reg[0] + DMC_MON_G12_CTRL0); 269 + writel(0, info->ddr_reg[0] + DMC_MON_G12_TIMER); 270 + 271 + writel(0, info->ddr_reg[0] + DMC_MON_G12_ALL_REQ_CNT); 272 + writel(0, info->ddr_reg[0] + DMC_MON_G12_ALL_GRANT_CNT); 273 + writel(0, info->ddr_reg[0] + DMC_MON_G12_ONE_GRANT_CNT); 274 + writel(0, info->ddr_reg[0] + DMC_MON_G12_SEC_GRANT_CNT); 275 + writel(0, info->ddr_reg[0] + DMC_MON_G12_THD_GRANT_CNT); 276 + writel(0, info->ddr_reg[0] + DMC_MON_G12_FOR_GRANT_CNT); 277 + 278 + /* clear port channel mapping */ 279 + for (i = 0; i < info->hw_info->chann_nr; i++) 280 + dmc_g12_config_fiter(info, -1, i); 281 + } 282 + 283 + static void dmc_g12_get_counters(struct dmc_info *info, 284 + struct dmc_counter *counter) 285 + { 286 + int i; 287 + unsigned int reg; 288 + 289 + counter->all_cnt = readl(info->ddr_reg[0] + DMC_MON_G12_ALL_GRANT_CNT); 290 + counter->all_req = readl(info->ddr_reg[0] + DMC_MON_G12_ALL_REQ_CNT); 291 + 292 + for (i = 0; i < info->hw_info->chann_nr; i++) { 293 + reg = DMC_MON_G12_ONE_GRANT_CNT + (i << 2); 294 + counter->channel_cnt[i] = readl(info->ddr_reg[0] + reg); 295 + } 296 + } 297 + 298 + static int dmc_g12_irq_handler(struct dmc_info *info, 299 + struct dmc_counter *counter) 300 + { 301 + unsigned int val; 302 + int ret = -EINVAL; 303 + 304 + val = readl(info->ddr_reg[0] + DMC_MON_G12_CTRL0); 305 + if (val & DMC_QOS_IRQ) { 306 + dmc_g12_get_counters(info, counter); 307 + /* clear irq flags */ 308 + writel(val, info->ddr_reg[0] + DMC_MON_G12_CTRL0); 309 + ret = 0; 310 + } 311 + return ret; 312 + } 313 + 314 + static const struct dmc_hw_info g12a_dmc_info = { 315 + .enable = dmc_g12_counter_enable, 316 + .disable = dmc_g12_counter_disable, 317 + .irq_handler = dmc_g12_irq_handler, 318 + .get_counters = dmc_g12_get_counters, 319 + .set_axi_filter = dmc_g12_set_axi_filter, 320 + 321 + .dmc_nr = 1, 322 + .chann_nr = 4, 323 + .capability = {0X7EFF00FF03DF, 0}, 324 + .fmt_attr = g12_pmu_format_attrs, 325 + }; 326 + 327 + static const struct dmc_hw_info g12b_dmc_info = { 328 + .enable = dmc_g12_counter_enable, 329 + .disable = dmc_g12_counter_disable, 330 + .irq_handler = dmc_g12_irq_handler, 331 + .get_counters = dmc_g12_get_counters, 332 + .set_axi_filter = dmc_g12_set_axi_filter, 333 + 334 + .dmc_nr = 1, 335 + .chann_nr = 4, 336 + .capability = {0X7FFF00FF3FDF, 0}, 337 + .fmt_attr = g12_pmu_format_attrs, 338 + }; 339 + 340 + static const struct dmc_hw_info sm1_dmc_info = { 341 + .enable = dmc_g12_counter_enable, 342 + .disable = dmc_g12_counter_disable, 343 + .irq_handler = dmc_g12_irq_handler, 344 + .get_counters = dmc_g12_get_counters, 345 + .set_axi_filter = dmc_g12_set_axi_filter, 346 + 347 + .dmc_nr = 1, 348 + .chann_nr = 4, 349 + .capability = {0X7EFF00FF07DF, 0}, 350 + .fmt_attr = g12_pmu_format_attrs, 351 + }; 352 + 353 + static int g12_ddr_pmu_probe(struct platform_device *pdev) 354 + { 355 + return meson_ddr_pmu_create(pdev); 356 + } 357 + 358 + static int g12_ddr_pmu_remove(struct platform_device *pdev) 359 + { 360 + meson_ddr_pmu_remove(pdev); 361 + 362 + return 0; 363 + } 364 + 365 + static const struct of_device_id meson_ddr_pmu_dt_match[] = { 366 + { 367 + .compatible = "amlogic,g12a-ddr-pmu", 368 + .data = &g12a_dmc_info, 369 + }, 370 + { 371 + .compatible = "amlogic,g12b-ddr-pmu", 372 + .data = &g12b_dmc_info, 373 + }, 374 + { 375 + .compatible = "amlogic,sm1-ddr-pmu", 376 + .data = &sm1_dmc_info, 377 + }, 378 + {} 379 + }; 380 + 381 + static struct platform_driver g12_ddr_pmu_driver = { 382 + .probe = g12_ddr_pmu_probe, 383 + .remove = g12_ddr_pmu_remove, 384 + 385 + .driver = { 386 + .name = "meson-g12-ddr-pmu", 387 + .of_match_table = meson_ddr_pmu_dt_match, 388 + }, 389 + }; 390 + 391 + module_platform_driver(g12_ddr_pmu_driver); 392 + MODULE_AUTHOR("Jiucheng Xu"); 393 + MODULE_LICENSE("GPL"); 394 + MODULE_DESCRIPTION("Amlogic G12 series SoC DDR PMU");
+66
include/soc/amlogic/meson_ddr_pmu.h
··· 1 + /* SPDX-License-Identifier: GPL-2.0 */ 2 + /* 3 + * Copyright (c) 2022 Amlogic, Inc. All rights reserved. 4 + */ 5 + 6 + #ifndef __MESON_DDR_PMU_H__ 7 + #define __MESON_DDR_PMU_H__ 8 + 9 + #define MAX_CHANNEL_NUM 8 10 + 11 + enum { 12 + ALL_CHAN_COUNTER_ID, 13 + CHAN1_COUNTER_ID, 14 + CHAN2_COUNTER_ID, 15 + CHAN3_COUNTER_ID, 16 + CHAN4_COUNTER_ID, 17 + CHAN5_COUNTER_ID, 18 + CHAN6_COUNTER_ID, 19 + CHAN7_COUNTER_ID, 20 + CHAN8_COUNTER_ID, 21 + COUNTER_MAX_ID, 22 + }; 23 + 24 + struct dmc_info; 25 + 26 + struct dmc_counter { 27 + u64 all_cnt; /* The count of all requests come in/out ddr controller */ 28 + union { 29 + u64 all_req; 30 + struct { 31 + u64 all_idle_cnt; 32 + u64 all_16bit_cnt; 33 + }; 34 + }; 35 + u64 channel_cnt[MAX_CHANNEL_NUM]; /* To save a DMC bandwidth-monitor channel counter */ 36 + }; 37 + 38 + struct dmc_hw_info { 39 + void (*enable)(struct dmc_info *info); 40 + void (*disable)(struct dmc_info *info); 41 + /* Bind an axi line to a bandwidth-monitor channel */ 42 + void (*set_axi_filter)(struct dmc_info *info, int axi_id, int chann); 43 + int (*irq_handler)(struct dmc_info *info, 44 + struct dmc_counter *counter); 45 + void (*get_counters)(struct dmc_info *info, 46 + struct dmc_counter *counter); 47 + 48 + int dmc_nr; /* The number of dmc controller */ 49 + int chann_nr; /* The number of dmc bandwidth monitor channels */ 50 + struct attribute **fmt_attr; 51 + const u64 capability[2]; 52 + }; 53 + 54 + struct dmc_info { 55 + const struct dmc_hw_info *hw_info; 56 + 57 + void __iomem *ddr_reg[4]; 58 + unsigned long timer_value; /* Timer value in TIMER register */ 59 + void __iomem *pll_reg; 60 + int irq_num; /* irq vector number */ 61 + }; 62 + 63 + int meson_ddr_pmu_create(struct platform_device *pdev); 64 + int meson_ddr_pmu_remove(struct platform_device *pdev); 65 + 66 + #endif /* __MESON_DDR_PMU_H__ */