at v6.3 377 lines 8.7 kB view raw
1/* SPDX-License-Identifier: GPL-2.0-only */ 2/* 3 * Copyright (C) 2012 Regents of the University of California 4 */ 5 6#ifndef _ASM_RISCV_PGTABLE_64_H 7#define _ASM_RISCV_PGTABLE_64_H 8 9#include <linux/bits.h> 10#include <linux/const.h> 11#include <asm/errata_list.h> 12 13extern bool pgtable_l4_enabled; 14extern bool pgtable_l5_enabled; 15 16#define PGDIR_SHIFT_L3 30 17#define PGDIR_SHIFT_L4 39 18#define PGDIR_SHIFT_L5 48 19#define PGDIR_SIZE_L3 (_AC(1, UL) << PGDIR_SHIFT_L3) 20 21#define PGDIR_SHIFT (pgtable_l5_enabled ? PGDIR_SHIFT_L5 : \ 22 (pgtable_l4_enabled ? PGDIR_SHIFT_L4 : PGDIR_SHIFT_L3)) 23/* Size of region mapped by a page global directory */ 24#define PGDIR_SIZE (_AC(1, UL) << PGDIR_SHIFT) 25#define PGDIR_MASK (~(PGDIR_SIZE - 1)) 26 27/* p4d is folded into pgd in case of 4-level page table */ 28#define P4D_SHIFT_L3 30 29#define P4D_SHIFT_L4 39 30#define P4D_SHIFT_L5 39 31#define P4D_SHIFT (pgtable_l5_enabled ? P4D_SHIFT_L5 : \ 32 (pgtable_l4_enabled ? P4D_SHIFT_L4 : P4D_SHIFT_L3)) 33#define P4D_SIZE (_AC(1, UL) << P4D_SHIFT) 34#define P4D_MASK (~(P4D_SIZE - 1)) 35 36/* pud is folded into pgd in case of 3-level page table */ 37#define PUD_SHIFT 30 38#define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) 39#define PUD_MASK (~(PUD_SIZE - 1)) 40 41#define PMD_SHIFT 21 42/* Size of region mapped by a page middle directory */ 43#define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) 44#define PMD_MASK (~(PMD_SIZE - 1)) 45 46/* Page 4th Directory entry */ 47typedef struct { 48 unsigned long p4d; 49} p4d_t; 50 51#define p4d_val(x) ((x).p4d) 52#define __p4d(x) ((p4d_t) { (x) }) 53#define PTRS_PER_P4D (PAGE_SIZE / sizeof(p4d_t)) 54 55/* Page Upper Directory entry */ 56typedef struct { 57 unsigned long pud; 58} pud_t; 59 60#define pud_val(x) ((x).pud) 61#define __pud(x) ((pud_t) { (x) }) 62#define PTRS_PER_PUD (PAGE_SIZE / sizeof(pud_t)) 63 64/* Page Middle Directory entry */ 65typedef struct { 66 unsigned long pmd; 67} pmd_t; 68 69#define pmd_val(x) ((x).pmd) 70#define __pmd(x) ((pmd_t) { (x) }) 71 72#define PTRS_PER_PMD (PAGE_SIZE / sizeof(pmd_t)) 73 74/* 75 * rv64 PTE format: 76 * | 63 | 62 61 | 60 54 | 53 10 | 9 8 | 7 | 6 | 5 | 4 | 3 | 2 | 1 | 0 77 * N MT RSV PFN reserved for SW D A G U X W R V 78 */ 79#define _PAGE_PFN_MASK GENMASK(53, 10) 80 81/* 82 * [62:61] Svpbmt Memory Type definitions: 83 * 84 * 00 - PMA Normal Cacheable, No change to implied PMA memory type 85 * 01 - NC Non-cacheable, idempotent, weakly-ordered Main Memory 86 * 10 - IO Non-cacheable, non-idempotent, strongly-ordered I/O memory 87 * 11 - Rsvd Reserved for future standard use 88 */ 89#define _PAGE_NOCACHE_SVPBMT (1UL << 61) 90#define _PAGE_IO_SVPBMT (1UL << 62) 91#define _PAGE_MTMASK_SVPBMT (_PAGE_NOCACHE_SVPBMT | _PAGE_IO_SVPBMT) 92 93/* 94 * [63:59] T-Head Memory Type definitions: 95 * 96 * 00000 - NC Weakly-ordered, Non-cacheable, Non-bufferable, Non-shareable, Non-trustable 97 * 01110 - PMA Weakly-ordered, Cacheable, Bufferable, Shareable, Non-trustable 98 * 10000 - IO Strongly-ordered, Non-cacheable, Non-bufferable, Non-shareable, Non-trustable 99 */ 100#define _PAGE_PMA_THEAD ((1UL << 62) | (1UL << 61) | (1UL << 60)) 101#define _PAGE_NOCACHE_THEAD 0UL 102#define _PAGE_IO_THEAD (1UL << 63) 103#define _PAGE_MTMASK_THEAD (_PAGE_PMA_THEAD | _PAGE_IO_THEAD | (1UL << 59)) 104 105static inline u64 riscv_page_mtmask(void) 106{ 107 u64 val; 108 109 ALT_SVPBMT(val, _PAGE_MTMASK); 110 return val; 111} 112 113static inline u64 riscv_page_nocache(void) 114{ 115 u64 val; 116 117 ALT_SVPBMT(val, _PAGE_NOCACHE); 118 return val; 119} 120 121static inline u64 riscv_page_io(void) 122{ 123 u64 val; 124 125 ALT_SVPBMT(val, _PAGE_IO); 126 return val; 127} 128 129#define _PAGE_NOCACHE riscv_page_nocache() 130#define _PAGE_IO riscv_page_io() 131#define _PAGE_MTMASK riscv_page_mtmask() 132 133/* Set of bits to preserve across pte_modify() */ 134#define _PAGE_CHG_MASK (~(unsigned long)(_PAGE_PRESENT | _PAGE_READ | \ 135 _PAGE_WRITE | _PAGE_EXEC | \ 136 _PAGE_USER | _PAGE_GLOBAL | \ 137 _PAGE_MTMASK)) 138 139static inline int pud_present(pud_t pud) 140{ 141 return (pud_val(pud) & _PAGE_PRESENT); 142} 143 144static inline int pud_none(pud_t pud) 145{ 146 return (pud_val(pud) == 0); 147} 148 149static inline int pud_bad(pud_t pud) 150{ 151 return !pud_present(pud); 152} 153 154#define pud_leaf pud_leaf 155static inline int pud_leaf(pud_t pud) 156{ 157 return pud_present(pud) && (pud_val(pud) & _PAGE_LEAF); 158} 159 160static inline int pud_user(pud_t pud) 161{ 162 return pud_val(pud) & _PAGE_USER; 163} 164 165static inline void set_pud(pud_t *pudp, pud_t pud) 166{ 167 *pudp = pud; 168} 169 170static inline void pud_clear(pud_t *pudp) 171{ 172 set_pud(pudp, __pud(0)); 173} 174 175static inline pud_t pfn_pud(unsigned long pfn, pgprot_t prot) 176{ 177 return __pud((pfn << _PAGE_PFN_SHIFT) | pgprot_val(prot)); 178} 179 180static inline unsigned long _pud_pfn(pud_t pud) 181{ 182 return __page_val_to_pfn(pud_val(pud)); 183} 184 185static inline pmd_t *pud_pgtable(pud_t pud) 186{ 187 return (pmd_t *)pfn_to_virt(__page_val_to_pfn(pud_val(pud))); 188} 189 190static inline struct page *pud_page(pud_t pud) 191{ 192 return pfn_to_page(__page_val_to_pfn(pud_val(pud))); 193} 194 195#define mm_p4d_folded mm_p4d_folded 196static inline bool mm_p4d_folded(struct mm_struct *mm) 197{ 198 if (pgtable_l5_enabled) 199 return false; 200 201 return true; 202} 203 204#define mm_pud_folded mm_pud_folded 205static inline bool mm_pud_folded(struct mm_struct *mm) 206{ 207 if (pgtable_l4_enabled) 208 return false; 209 210 return true; 211} 212 213#define pmd_index(addr) (((addr) >> PMD_SHIFT) & (PTRS_PER_PMD - 1)) 214 215static inline pmd_t pfn_pmd(unsigned long pfn, pgprot_t prot) 216{ 217 unsigned long prot_val = pgprot_val(prot); 218 219 ALT_THEAD_PMA(prot_val); 220 221 return __pmd((pfn << _PAGE_PFN_SHIFT) | prot_val); 222} 223 224static inline unsigned long _pmd_pfn(pmd_t pmd) 225{ 226 return __page_val_to_pfn(pmd_val(pmd)); 227} 228 229#define mk_pmd(page, prot) pfn_pmd(page_to_pfn(page), prot) 230 231#define pmd_ERROR(e) \ 232 pr_err("%s:%d: bad pmd %016lx.\n", __FILE__, __LINE__, pmd_val(e)) 233 234#define pud_ERROR(e) \ 235 pr_err("%s:%d: bad pud %016lx.\n", __FILE__, __LINE__, pud_val(e)) 236 237#define p4d_ERROR(e) \ 238 pr_err("%s:%d: bad p4d %016lx.\n", __FILE__, __LINE__, p4d_val(e)) 239 240static inline void set_p4d(p4d_t *p4dp, p4d_t p4d) 241{ 242 if (pgtable_l4_enabled) 243 *p4dp = p4d; 244 else 245 set_pud((pud_t *)p4dp, (pud_t){ p4d_val(p4d) }); 246} 247 248static inline int p4d_none(p4d_t p4d) 249{ 250 if (pgtable_l4_enabled) 251 return (p4d_val(p4d) == 0); 252 253 return 0; 254} 255 256static inline int p4d_present(p4d_t p4d) 257{ 258 if (pgtable_l4_enabled) 259 return (p4d_val(p4d) & _PAGE_PRESENT); 260 261 return 1; 262} 263 264static inline int p4d_bad(p4d_t p4d) 265{ 266 if (pgtable_l4_enabled) 267 return !p4d_present(p4d); 268 269 return 0; 270} 271 272static inline void p4d_clear(p4d_t *p4d) 273{ 274 if (pgtable_l4_enabled) 275 set_p4d(p4d, __p4d(0)); 276} 277 278static inline p4d_t pfn_p4d(unsigned long pfn, pgprot_t prot) 279{ 280 return __p4d((pfn << _PAGE_PFN_SHIFT) | pgprot_val(prot)); 281} 282 283static inline unsigned long _p4d_pfn(p4d_t p4d) 284{ 285 return __page_val_to_pfn(p4d_val(p4d)); 286} 287 288static inline pud_t *p4d_pgtable(p4d_t p4d) 289{ 290 if (pgtable_l4_enabled) 291 return (pud_t *)pfn_to_virt(__page_val_to_pfn(p4d_val(p4d))); 292 293 return (pud_t *)pud_pgtable((pud_t) { p4d_val(p4d) }); 294} 295#define p4d_page_vaddr(p4d) ((unsigned long)p4d_pgtable(p4d)) 296 297static inline struct page *p4d_page(p4d_t p4d) 298{ 299 return pfn_to_page(__page_val_to_pfn(p4d_val(p4d))); 300} 301 302#define pud_index(addr) (((addr) >> PUD_SHIFT) & (PTRS_PER_PUD - 1)) 303 304#define pud_offset pud_offset 305static inline pud_t *pud_offset(p4d_t *p4d, unsigned long address) 306{ 307 if (pgtable_l4_enabled) 308 return p4d_pgtable(*p4d) + pud_index(address); 309 310 return (pud_t *)p4d; 311} 312 313static inline void set_pgd(pgd_t *pgdp, pgd_t pgd) 314{ 315 if (pgtable_l5_enabled) 316 *pgdp = pgd; 317 else 318 set_p4d((p4d_t *)pgdp, (p4d_t){ pgd_val(pgd) }); 319} 320 321static inline int pgd_none(pgd_t pgd) 322{ 323 if (pgtable_l5_enabled) 324 return (pgd_val(pgd) == 0); 325 326 return 0; 327} 328 329static inline int pgd_present(pgd_t pgd) 330{ 331 if (pgtable_l5_enabled) 332 return (pgd_val(pgd) & _PAGE_PRESENT); 333 334 return 1; 335} 336 337static inline int pgd_bad(pgd_t pgd) 338{ 339 if (pgtable_l5_enabled) 340 return !pgd_present(pgd); 341 342 return 0; 343} 344 345static inline void pgd_clear(pgd_t *pgd) 346{ 347 if (pgtable_l5_enabled) 348 set_pgd(pgd, __pgd(0)); 349} 350 351static inline p4d_t *pgd_pgtable(pgd_t pgd) 352{ 353 if (pgtable_l5_enabled) 354 return (p4d_t *)pfn_to_virt(__page_val_to_pfn(pgd_val(pgd))); 355 356 return (p4d_t *)p4d_pgtable((p4d_t) { pgd_val(pgd) }); 357} 358#define pgd_page_vaddr(pgd) ((unsigned long)pgd_pgtable(pgd)) 359 360static inline struct page *pgd_page(pgd_t pgd) 361{ 362 return pfn_to_page(__page_val_to_pfn(pgd_val(pgd))); 363} 364#define pgd_page(pgd) pgd_page(pgd) 365 366#define p4d_index(addr) (((addr) >> P4D_SHIFT) & (PTRS_PER_P4D - 1)) 367 368#define p4d_offset p4d_offset 369static inline p4d_t *p4d_offset(pgd_t *pgd, unsigned long address) 370{ 371 if (pgtable_l5_enabled) 372 return pgd_pgtable(*pgd) + p4d_index(address); 373 374 return (p4d_t *)pgd; 375} 376 377#endif /* _ASM_RISCV_PGTABLE_64_H */