Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

C6X: cache control

Original port to early 2.6 kernel using TI COFF toolchain.
Brought up to date by Mark Salter <msalter@redhat.com>

Signed-off-by: Aurelien Jacquiot <a-jacquiot@ti.com>
Signed-off-by: Mark Salter <msalter@redhat.com>
Acked-by: Arnd Bergmann <arnd@arndb.de>

authored by

Aurelien Jacquiot and committed by
Mark Salter
784bdcd0 81ec9889

+600
+90
arch/c6x/include/asm/cache.h
··· 1 + /* 2 + * Port on Texas Instruments TMS320C6x architecture 3 + * 4 + * Copyright (C) 2005, 2006, 2009, 2010 Texas Instruments Incorporated 5 + * Author: Aurelien Jacquiot (aurelien.jacquiot@jaluna.com) 6 + * 7 + * This program is free software; you can redistribute it and/or modify 8 + * it under the terms of the GNU General Public License version 2 as 9 + * published by the Free Software Foundation. 10 + */ 11 + #ifndef _ASM_C6X_CACHE_H 12 + #define _ASM_C6X_CACHE_H 13 + 14 + #include <linux/irqflags.h> 15 + 16 + /* 17 + * Cache line size 18 + */ 19 + #define L1D_CACHE_BYTES 64 20 + #define L1P_CACHE_BYTES 32 21 + #define L2_CACHE_BYTES 128 22 + 23 + /* 24 + * L2 used as cache 25 + */ 26 + #define L2MODE_SIZE L2MODE_256K_CACHE 27 + 28 + /* 29 + * For practical reasons the L1_CACHE_BYTES defines should not be smaller than 30 + * the L2 line size 31 + */ 32 + #define L1_CACHE_BYTES L2_CACHE_BYTES 33 + 34 + #define L2_CACHE_ALIGN_LOW(x) \ 35 + (((x) & ~(L2_CACHE_BYTES - 1))) 36 + #define L2_CACHE_ALIGN_UP(x) \ 37 + (((x) + (L2_CACHE_BYTES - 1)) & ~(L2_CACHE_BYTES - 1)) 38 + #define L2_CACHE_ALIGN_CNT(x) \ 39 + (((x) + (sizeof(int) - 1)) & ~(sizeof(int) - 1)) 40 + 41 + #define ARCH_DMA_MINALIGN L1_CACHE_BYTES 42 + #define ARCH_SLAB_MINALIGN L1_CACHE_BYTES 43 + 44 + /* 45 + * This is the granularity of hardware cacheability control. 46 + */ 47 + #define CACHEABILITY_ALIGN 0x01000000 48 + 49 + /* 50 + * Align a physical address to MAR regions 51 + */ 52 + #define CACHE_REGION_START(v) \ 53 + (((u32) (v)) & ~(CACHEABILITY_ALIGN - 1)) 54 + #define CACHE_REGION_END(v) \ 55 + (((u32) (v) + (CACHEABILITY_ALIGN - 1)) & ~(CACHEABILITY_ALIGN - 1)) 56 + 57 + extern void __init c6x_cache_init(void); 58 + 59 + extern void enable_caching(unsigned long start, unsigned long end); 60 + extern void disable_caching(unsigned long start, unsigned long end); 61 + 62 + extern void L1_cache_off(void); 63 + extern void L1_cache_on(void); 64 + 65 + extern void L1P_cache_global_invalidate(void); 66 + extern void L1D_cache_global_invalidate(void); 67 + extern void L1D_cache_global_writeback(void); 68 + extern void L1D_cache_global_writeback_invalidate(void); 69 + extern void L2_cache_set_mode(unsigned int mode); 70 + extern void L2_cache_global_writeback_invalidate(void); 71 + extern void L2_cache_global_writeback(void); 72 + 73 + extern void L1P_cache_block_invalidate(unsigned int start, unsigned int end); 74 + extern void L1D_cache_block_invalidate(unsigned int start, unsigned int end); 75 + extern void L1D_cache_block_writeback_invalidate(unsigned int start, 76 + unsigned int end); 77 + extern void L1D_cache_block_writeback(unsigned int start, unsigned int end); 78 + extern void L2_cache_block_invalidate(unsigned int start, unsigned int end); 79 + extern void L2_cache_block_writeback(unsigned int start, unsigned int end); 80 + extern void L2_cache_block_writeback_invalidate(unsigned int start, 81 + unsigned int end); 82 + extern void L2_cache_block_invalidate_nowait(unsigned int start, 83 + unsigned int end); 84 + extern void L2_cache_block_writeback_nowait(unsigned int start, 85 + unsigned int end); 86 + 87 + extern void L2_cache_block_writeback_invalidate_nowait(unsigned int start, 88 + unsigned int end); 89 + 90 + #endif /* _ASM_C6X_CACHE_H */
+65
arch/c6x/include/asm/cacheflush.h
··· 1 + /* 2 + * Port on Texas Instruments TMS320C6x architecture 3 + * 4 + * Copyright (C) 2004, 2009, 2010 Texas Instruments Incorporated 5 + * Author: Aurelien Jacquiot (aurelien.jacquiot@jaluna.com) 6 + * 7 + * This program is free software; you can redistribute it and/or modify 8 + * it under the terms of the GNU General Public License version 2 as 9 + * published by the Free Software Foundation. 10 + */ 11 + #ifndef _ASM_C6X_CACHEFLUSH_H 12 + #define _ASM_C6X_CACHEFLUSH_H 13 + 14 + #include <linux/spinlock.h> 15 + 16 + #include <asm/setup.h> 17 + #include <asm/cache.h> 18 + #include <asm/mman.h> 19 + #include <asm/page.h> 20 + #include <asm/string.h> 21 + 22 + /* 23 + * virtually-indexed cache management (our cache is physically indexed) 24 + */ 25 + #define flush_cache_all() do {} while (0) 26 + #define flush_cache_mm(mm) do {} while (0) 27 + #define flush_cache_dup_mm(mm) do {} while (0) 28 + #define flush_cache_range(mm, start, end) do {} while (0) 29 + #define flush_cache_page(vma, vmaddr, pfn) do {} while (0) 30 + #define flush_cache_vmap(start, end) do {} while (0) 31 + #define flush_cache_vunmap(start, end) do {} while (0) 32 + #define ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE 0 33 + #define flush_dcache_page(page) do {} while (0) 34 + #define flush_dcache_mmap_lock(mapping) do {} while (0) 35 + #define flush_dcache_mmap_unlock(mapping) do {} while (0) 36 + 37 + /* 38 + * physically-indexed cache management 39 + */ 40 + #define flush_icache_range(s, e) \ 41 + do { \ 42 + L1D_cache_block_writeback((s), (e)); \ 43 + L1P_cache_block_invalidate((s), (e)); \ 44 + } while (0) 45 + 46 + #define flush_icache_page(vma, page) \ 47 + do { \ 48 + if ((vma)->vm_flags & PROT_EXEC) \ 49 + L1D_cache_block_writeback_invalidate(page_address(page), \ 50 + (unsigned long) page_address(page) + PAGE_SIZE)); \ 51 + L1P_cache_block_invalidate(page_address(page), \ 52 + (unsigned long) page_address(page) + PAGE_SIZE)); \ 53 + } while (0) 54 + 55 + 56 + #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 57 + do { \ 58 + memcpy(dst, src, len); \ 59 + flush_icache_range((unsigned) (dst), (unsigned) (dst) + (len)); \ 60 + } while (0) 61 + 62 + #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ 63 + memcpy(dst, src, len) 64 + 65 + #endif /* _ASM_C6X_CACHEFLUSH_H */
+445
arch/c6x/platforms/cache.c
··· 1 + /* 2 + * Copyright (C) 2011 Texas Instruments Incorporated 3 + * Author: Mark Salter <msalter@redhat.com> 4 + * 5 + * This program is free software; you can redistribute it and/or modify 6 + * it under the terms of the GNU General Public License version 2 as 7 + * published by the Free Software Foundation. 8 + */ 9 + #include <linux/of.h> 10 + #include <linux/of_address.h> 11 + #include <linux/io.h> 12 + 13 + #include <asm/cache.h> 14 + #include <asm/soc.h> 15 + 16 + /* 17 + * Internal Memory Control Registers for caches 18 + */ 19 + #define IMCR_CCFG 0x0000 20 + #define IMCR_L1PCFG 0x0020 21 + #define IMCR_L1PCC 0x0024 22 + #define IMCR_L1DCFG 0x0040 23 + #define IMCR_L1DCC 0x0044 24 + #define IMCR_L2ALLOC0 0x2000 25 + #define IMCR_L2ALLOC1 0x2004 26 + #define IMCR_L2ALLOC2 0x2008 27 + #define IMCR_L2ALLOC3 0x200c 28 + #define IMCR_L2WBAR 0x4000 29 + #define IMCR_L2WWC 0x4004 30 + #define IMCR_L2WIBAR 0x4010 31 + #define IMCR_L2WIWC 0x4014 32 + #define IMCR_L2IBAR 0x4018 33 + #define IMCR_L2IWC 0x401c 34 + #define IMCR_L1PIBAR 0x4020 35 + #define IMCR_L1PIWC 0x4024 36 + #define IMCR_L1DWIBAR 0x4030 37 + #define IMCR_L1DWIWC 0x4034 38 + #define IMCR_L1DWBAR 0x4040 39 + #define IMCR_L1DWWC 0x4044 40 + #define IMCR_L1DIBAR 0x4048 41 + #define IMCR_L1DIWC 0x404c 42 + #define IMCR_L2WB 0x5000 43 + #define IMCR_L2WBINV 0x5004 44 + #define IMCR_L2INV 0x5008 45 + #define IMCR_L1PINV 0x5028 46 + #define IMCR_L1DWB 0x5040 47 + #define IMCR_L1DWBINV 0x5044 48 + #define IMCR_L1DINV 0x5048 49 + #define IMCR_MAR_BASE 0x8000 50 + #define IMCR_MAR96_111 0x8180 51 + #define IMCR_MAR128_191 0x8200 52 + #define IMCR_MAR224_239 0x8380 53 + #define IMCR_L2MPFAR 0xa000 54 + #define IMCR_L2MPFSR 0xa004 55 + #define IMCR_L2MPFCR 0xa008 56 + #define IMCR_L2MPLK0 0xa100 57 + #define IMCR_L2MPLK1 0xa104 58 + #define IMCR_L2MPLK2 0xa108 59 + #define IMCR_L2MPLK3 0xa10c 60 + #define IMCR_L2MPLKCMD 0xa110 61 + #define IMCR_L2MPLKSTAT 0xa114 62 + #define IMCR_L2MPPA_BASE 0xa200 63 + #define IMCR_L1PMPFAR 0xa400 64 + #define IMCR_L1PMPFSR 0xa404 65 + #define IMCR_L1PMPFCR 0xa408 66 + #define IMCR_L1PMPLK0 0xa500 67 + #define IMCR_L1PMPLK1 0xa504 68 + #define IMCR_L1PMPLK2 0xa508 69 + #define IMCR_L1PMPLK3 0xa50c 70 + #define IMCR_L1PMPLKCMD 0xa510 71 + #define IMCR_L1PMPLKSTAT 0xa514 72 + #define IMCR_L1PMPPA_BASE 0xa600 73 + #define IMCR_L1DMPFAR 0xac00 74 + #define IMCR_L1DMPFSR 0xac04 75 + #define IMCR_L1DMPFCR 0xac08 76 + #define IMCR_L1DMPLK0 0xad00 77 + #define IMCR_L1DMPLK1 0xad04 78 + #define IMCR_L1DMPLK2 0xad08 79 + #define IMCR_L1DMPLK3 0xad0c 80 + #define IMCR_L1DMPLKCMD 0xad10 81 + #define IMCR_L1DMPLKSTAT 0xad14 82 + #define IMCR_L1DMPPA_BASE 0xae00 83 + #define IMCR_L2PDWAKE0 0xc040 84 + #define IMCR_L2PDWAKE1 0xc044 85 + #define IMCR_L2PDSLEEP0 0xc050 86 + #define IMCR_L2PDSLEEP1 0xc054 87 + #define IMCR_L2PDSTAT0 0xc060 88 + #define IMCR_L2PDSTAT1 0xc064 89 + 90 + /* 91 + * CCFG register values and bits 92 + */ 93 + #define L2MODE_0K_CACHE 0x0 94 + #define L2MODE_32K_CACHE 0x1 95 + #define L2MODE_64K_CACHE 0x2 96 + #define L2MODE_128K_CACHE 0x3 97 + #define L2MODE_256K_CACHE 0x7 98 + 99 + #define L2PRIO_URGENT 0x0 100 + #define L2PRIO_HIGH 0x1 101 + #define L2PRIO_MEDIUM 0x2 102 + #define L2PRIO_LOW 0x3 103 + 104 + #define CCFG_ID 0x100 /* Invalidate L1P bit */ 105 + #define CCFG_IP 0x200 /* Invalidate L1D bit */ 106 + 107 + static void __iomem *cache_base; 108 + 109 + /* 110 + * L1 & L2 caches generic functions 111 + */ 112 + #define imcr_get(reg) soc_readl(cache_base + (reg)) 113 + #define imcr_set(reg, value) \ 114 + do { \ 115 + soc_writel((value), cache_base + (reg)); \ 116 + soc_readl(cache_base + (reg)); \ 117 + } while (0) 118 + 119 + static void cache_block_operation_wait(unsigned int wc_reg) 120 + { 121 + /* Wait for completion */ 122 + while (imcr_get(wc_reg)) 123 + cpu_relax(); 124 + } 125 + 126 + static DEFINE_SPINLOCK(cache_lock); 127 + 128 + /* 129 + * Generic function to perform a block cache operation as 130 + * invalidate or writeback/invalidate 131 + */ 132 + static void cache_block_operation(unsigned int *start, 133 + unsigned int *end, 134 + unsigned int bar_reg, 135 + unsigned int wc_reg) 136 + { 137 + unsigned long flags; 138 + unsigned int wcnt = 139 + (L2_CACHE_ALIGN_CNT((unsigned int) end) 140 + - L2_CACHE_ALIGN_LOW((unsigned int) start)) >> 2; 141 + unsigned int wc = 0; 142 + 143 + for (; wcnt; wcnt -= wc, start += wc) { 144 + loop: 145 + spin_lock_irqsave(&cache_lock, flags); 146 + 147 + /* 148 + * If another cache operation is occuring 149 + */ 150 + if (unlikely(imcr_get(wc_reg))) { 151 + spin_unlock_irqrestore(&cache_lock, flags); 152 + 153 + /* Wait for previous operation completion */ 154 + cache_block_operation_wait(wc_reg); 155 + 156 + /* Try again */ 157 + goto loop; 158 + } 159 + 160 + imcr_set(bar_reg, L2_CACHE_ALIGN_LOW((unsigned int) start)); 161 + 162 + if (wcnt > 0xffff) 163 + wc = 0xffff; 164 + else 165 + wc = wcnt; 166 + 167 + /* Set word count value in the WC register */ 168 + imcr_set(wc_reg, wc & 0xffff); 169 + 170 + spin_unlock_irqrestore(&cache_lock, flags); 171 + 172 + /* Wait for completion */ 173 + cache_block_operation_wait(wc_reg); 174 + } 175 + } 176 + 177 + static void cache_block_operation_nowait(unsigned int *start, 178 + unsigned int *end, 179 + unsigned int bar_reg, 180 + unsigned int wc_reg) 181 + { 182 + unsigned long flags; 183 + unsigned int wcnt = 184 + (L2_CACHE_ALIGN_CNT((unsigned int) end) 185 + - L2_CACHE_ALIGN_LOW((unsigned int) start)) >> 2; 186 + unsigned int wc = 0; 187 + 188 + for (; wcnt; wcnt -= wc, start += wc) { 189 + 190 + spin_lock_irqsave(&cache_lock, flags); 191 + 192 + imcr_set(bar_reg, L2_CACHE_ALIGN_LOW((unsigned int) start)); 193 + 194 + if (wcnt > 0xffff) 195 + wc = 0xffff; 196 + else 197 + wc = wcnt; 198 + 199 + /* Set word count value in the WC register */ 200 + imcr_set(wc_reg, wc & 0xffff); 201 + 202 + spin_unlock_irqrestore(&cache_lock, flags); 203 + 204 + /* Don't wait for completion on last cache operation */ 205 + if (wcnt > 0xffff) 206 + cache_block_operation_wait(wc_reg); 207 + } 208 + } 209 + 210 + /* 211 + * L1 caches management 212 + */ 213 + 214 + /* 215 + * Disable L1 caches 216 + */ 217 + void L1_cache_off(void) 218 + { 219 + unsigned int dummy; 220 + 221 + imcr_set(IMCR_L1PCFG, 0); 222 + dummy = imcr_get(IMCR_L1PCFG); 223 + 224 + imcr_set(IMCR_L1DCFG, 0); 225 + dummy = imcr_get(IMCR_L1DCFG); 226 + } 227 + 228 + /* 229 + * Enable L1 caches 230 + */ 231 + void L1_cache_on(void) 232 + { 233 + unsigned int dummy; 234 + 235 + imcr_set(IMCR_L1PCFG, 7); 236 + dummy = imcr_get(IMCR_L1PCFG); 237 + 238 + imcr_set(IMCR_L1DCFG, 7); 239 + dummy = imcr_get(IMCR_L1DCFG); 240 + } 241 + 242 + /* 243 + * L1P global-invalidate all 244 + */ 245 + void L1P_cache_global_invalidate(void) 246 + { 247 + unsigned int set = 1; 248 + imcr_set(IMCR_L1PINV, set); 249 + while (imcr_get(IMCR_L1PINV) & 1) 250 + cpu_relax(); 251 + } 252 + 253 + /* 254 + * L1D global-invalidate all 255 + * 256 + * Warning: this operation causes all updated data in L1D to 257 + * be discarded rather than written back to the lower levels of 258 + * memory 259 + */ 260 + void L1D_cache_global_invalidate(void) 261 + { 262 + unsigned int set = 1; 263 + imcr_set(IMCR_L1DINV, set); 264 + while (imcr_get(IMCR_L1DINV) & 1) 265 + cpu_relax(); 266 + } 267 + 268 + void L1D_cache_global_writeback(void) 269 + { 270 + unsigned int set = 1; 271 + imcr_set(IMCR_L1DWB, set); 272 + while (imcr_get(IMCR_L1DWB) & 1) 273 + cpu_relax(); 274 + } 275 + 276 + void L1D_cache_global_writeback_invalidate(void) 277 + { 278 + unsigned int set = 1; 279 + imcr_set(IMCR_L1DWBINV, set); 280 + while (imcr_get(IMCR_L1DWBINV) & 1) 281 + cpu_relax(); 282 + } 283 + 284 + /* 285 + * L2 caches management 286 + */ 287 + 288 + /* 289 + * Set L2 operation mode 290 + */ 291 + void L2_cache_set_mode(unsigned int mode) 292 + { 293 + unsigned int ccfg = imcr_get(IMCR_CCFG); 294 + 295 + /* Clear and set the L2MODE bits in CCFG */ 296 + ccfg &= ~7; 297 + ccfg |= (mode & 7); 298 + imcr_set(IMCR_CCFG, ccfg); 299 + ccfg = imcr_get(IMCR_CCFG); 300 + } 301 + 302 + /* 303 + * L2 global-writeback and global-invalidate all 304 + */ 305 + void L2_cache_global_writeback_invalidate(void) 306 + { 307 + imcr_set(IMCR_L2WBINV, 1); 308 + while (imcr_get(IMCR_L2WBINV)) 309 + cpu_relax(); 310 + } 311 + 312 + /* 313 + * L2 global-writeback all 314 + */ 315 + void L2_cache_global_writeback(void) 316 + { 317 + imcr_set(IMCR_L2WB, 1); 318 + while (imcr_get(IMCR_L2WB)) 319 + cpu_relax(); 320 + } 321 + 322 + /* 323 + * Cacheability controls 324 + */ 325 + void enable_caching(unsigned long start, unsigned long end) 326 + { 327 + unsigned int mar = IMCR_MAR_BASE + ((start >> 24) << 2); 328 + unsigned int mar_e = IMCR_MAR_BASE + ((end >> 24) << 2); 329 + 330 + for (; mar <= mar_e; mar += 4) 331 + imcr_set(mar, imcr_get(mar) | 1); 332 + } 333 + 334 + void disable_caching(unsigned long start, unsigned long end) 335 + { 336 + unsigned int mar = IMCR_MAR_BASE + ((start >> 24) << 2); 337 + unsigned int mar_e = IMCR_MAR_BASE + ((end >> 24) << 2); 338 + 339 + for (; mar <= mar_e; mar += 4) 340 + imcr_set(mar, imcr_get(mar) & ~1); 341 + } 342 + 343 + 344 + /* 345 + * L1 block operations 346 + */ 347 + void L1P_cache_block_invalidate(unsigned int start, unsigned int end) 348 + { 349 + cache_block_operation((unsigned int *) start, 350 + (unsigned int *) end, 351 + IMCR_L1PIBAR, IMCR_L1PIWC); 352 + } 353 + 354 + void L1D_cache_block_invalidate(unsigned int start, unsigned int end) 355 + { 356 + cache_block_operation((unsigned int *) start, 357 + (unsigned int *) end, 358 + IMCR_L1DIBAR, IMCR_L1DIWC); 359 + } 360 + 361 + void L1D_cache_block_writeback_invalidate(unsigned int start, unsigned int end) 362 + { 363 + cache_block_operation((unsigned int *) start, 364 + (unsigned int *) end, 365 + IMCR_L1DWIBAR, IMCR_L1DWIWC); 366 + } 367 + 368 + void L1D_cache_block_writeback(unsigned int start, unsigned int end) 369 + { 370 + cache_block_operation((unsigned int *) start, 371 + (unsigned int *) end, 372 + IMCR_L1DWBAR, IMCR_L1DWWC); 373 + } 374 + 375 + /* 376 + * L2 block operations 377 + */ 378 + void L2_cache_block_invalidate(unsigned int start, unsigned int end) 379 + { 380 + cache_block_operation((unsigned int *) start, 381 + (unsigned int *) end, 382 + IMCR_L2IBAR, IMCR_L2IWC); 383 + } 384 + 385 + void L2_cache_block_writeback(unsigned int start, unsigned int end) 386 + { 387 + cache_block_operation((unsigned int *) start, 388 + (unsigned int *) end, 389 + IMCR_L2WBAR, IMCR_L2WWC); 390 + } 391 + 392 + void L2_cache_block_writeback_invalidate(unsigned int start, unsigned int end) 393 + { 394 + cache_block_operation((unsigned int *) start, 395 + (unsigned int *) end, 396 + IMCR_L2WIBAR, IMCR_L2WIWC); 397 + } 398 + 399 + void L2_cache_block_invalidate_nowait(unsigned int start, unsigned int end) 400 + { 401 + cache_block_operation_nowait((unsigned int *) start, 402 + (unsigned int *) end, 403 + IMCR_L2IBAR, IMCR_L2IWC); 404 + } 405 + 406 + void L2_cache_block_writeback_nowait(unsigned int start, unsigned int end) 407 + { 408 + cache_block_operation_nowait((unsigned int *) start, 409 + (unsigned int *) end, 410 + IMCR_L2WBAR, IMCR_L2WWC); 411 + } 412 + 413 + void L2_cache_block_writeback_invalidate_nowait(unsigned int start, 414 + unsigned int end) 415 + { 416 + cache_block_operation_nowait((unsigned int *) start, 417 + (unsigned int *) end, 418 + IMCR_L2WIBAR, IMCR_L2WIWC); 419 + } 420 + 421 + 422 + /* 423 + * L1 and L2 caches configuration 424 + */ 425 + void __init c6x_cache_init(void) 426 + { 427 + struct device_node *node; 428 + 429 + node = of_find_compatible_node(NULL, NULL, "ti,c64x+cache"); 430 + if (!node) 431 + return; 432 + 433 + cache_base = of_iomap(node, 0); 434 + 435 + of_node_put(node); 436 + 437 + if (!cache_base) 438 + return; 439 + 440 + /* Set L2 caches on the the whole L2 SRAM memory */ 441 + L2_cache_set_mode(L2MODE_SIZE); 442 + 443 + /* Enable L1 */ 444 + L1_cache_on(); 445 + }