Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

consolidate io_remap_pfn_range definitions

Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>

Al Viro 40d158e6 605c912b

+7 -105
-3
arch/alpha/include/asm/pgtable.h
··· 354 354 #define kern_addr_valid(addr) (1) 355 355 #endif 356 356 357 - #define io_remap_pfn_range(vma, start, pfn, size, prot) \ 358 - remap_pfn_range(vma, start, pfn, size, prot) 359 - 360 357 #define pte_ERROR(e) \ 361 358 printk("%s:%d: bad pte %016lx.\n", __FILE__, __LINE__, pte_val(e)) 362 359 #define pmd_ERROR(e) \
-3
arch/arc/include/asm/pgtable.h
··· 394 394 * remap a physical page `pfn' of size `size' with page protection `prot' 395 395 * into virtual address `from' 396 396 */ 397 - #define io_remap_pfn_range(vma, from, pfn, size, prot) \ 398 - remap_pfn_range(vma, from, pfn, size, prot) 399 - 400 397 #include <asm-generic/pgtable.h> 401 398 402 399 /* to cope with aliasing VIPT cache */
-2
arch/arm/include/asm/pgtable-nommu.h
··· 79 79 * No page table caches to initialise. 80 80 */ 81 81 #define pgtable_cache_init() do { } while (0) 82 - #define io_remap_pfn_range remap_pfn_range 83 - 84 82 85 83 /* 86 84 * All 32bit addresses are effectively valid for vmalloc...
-7
arch/arm/include/asm/pgtable.h
··· 318 318 #define HAVE_ARCH_UNMAPPED_AREA 319 319 #define HAVE_ARCH_UNMAPPED_AREA_TOPDOWN 320 320 321 - /* 322 - * remap a physical page `pfn' of size `size' with page protection `prot' 323 - * into virtual address `from' 324 - */ 325 - #define io_remap_pfn_range(vma,from,pfn,size,prot) \ 326 - remap_pfn_range(vma, from, pfn, size, prot) 327 - 328 321 #define pgtable_cache_init() do { } while (0) 329 322 330 323 #endif /* !__ASSEMBLY__ */
-7
arch/arm64/include/asm/pgtable.h
··· 320 320 321 321 #include <asm-generic/pgtable.h> 322 322 323 - /* 324 - * remap a physical page `pfn' of size `size' with page protection `prot' 325 - * into virtual address `from' 326 - */ 327 - #define io_remap_pfn_range(vma,from,pfn,size,prot) \ 328 - remap_pfn_range(vma, from, pfn, size, prot) 329 - 330 323 #define pgtable_cache_init() do { } while (0) 331 324 332 325 #endif /* !__ASSEMBLY__ */
-3
arch/avr32/include/asm/pgtable.h
··· 362 362 363 363 #define kern_addr_valid(addr) (1) 364 364 365 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 366 - remap_pfn_range(vma, vaddr, pfn, size, prot) 367 - 368 365 /* No page table caches to initialize (?) */ 369 366 #define pgtable_cache_init() do { } while(0) 370 367
-1
arch/blackfin/include/asm/pgtable.h
··· 88 88 * No page table caches to initialise. 89 89 */ 90 90 #define pgtable_cache_init() do { } while (0) 91 - #define io_remap_pfn_range remap_pfn_range 92 91 93 92 /* 94 93 * All 32bit addresses are effectively valid for vmalloc...
-1
arch/c6x/include/asm/pgtable.h
··· 71 71 * No page table caches to initialise 72 72 */ 73 73 #define pgtable_cache_init() do { } while (0) 74 - #define io_remap_pfn_range remap_pfn_range 75 74 76 75 #include <asm-generic/pgtable.h> 77 76
-3
arch/cris/include/asm/pgtable.h
··· 258 258 #define pgd_ERROR(e) \ 259 259 printk("%s:%d: bad pgd %p(%08lx).\n", __FILE__, __LINE__, &(e), pgd_val(e)) 260 260 261 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 262 - remap_pfn_range(vma, vaddr, pfn, size, prot) 263 - 264 261 265 262 extern pgd_t swapper_pg_dir[PTRS_PER_PGD]; /* defined in head.S */ 266 263
-3
arch/frv/include/asm/pgtable.h
··· 488 488 #define PageSkip(page) (0) 489 489 #define kern_addr_valid(addr) (1) 490 490 491 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 492 - remap_pfn_range(vma, vaddr, pfn, size, prot) 493 - 494 491 #define __HAVE_ARCH_PTEP_TEST_AND_CLEAR_YOUNG 495 492 #define __HAVE_ARCH_PTEP_GET_AND_CLEAR 496 493 #define __HAVE_ARCH_PTEP_SET_WRPROTECT
-3
arch/h8300/include/asm/pgtable.h
··· 52 52 */ 53 53 #define pgtable_cache_init() do { } while (0) 54 54 55 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 56 - remap_pfn_range(vma, vaddr, pfn, size, prot) 57 - 58 55 /* 59 56 * All 32bit addresses are effectively valid for vmalloc... 60 57 * Sort of meaningless for non-VM targets.
-4
arch/hexagon/include/asm/pgtable.h
··· 452 452 453 453 #define __pte_offset(address) (((address) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1)) 454 454 455 - /* Nothing special about IO remapping at this point */ 456 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 457 - remap_pfn_range(vma, vaddr, pfn, size, prot) 458 - 459 455 /* I think this is in case we have page table caches; needed by init/main.c */ 460 456 #define pgtable_cache_init() do { } while (0) 461 457
-3
arch/ia64/include/asm/pgtable.h
··· 493 493 #define pte_to_pgoff(pte) ((pte_val(pte) << 1) >> 3) 494 494 #define pgoff_to_pte(off) ((pte_t) { ((off) << 2) | _PAGE_FILE }) 495 495 496 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 497 - remap_pfn_range(vma, vaddr, pfn, size, prot) 498 - 499 496 /* 500 497 * ZERO_PAGE is a global shared page that is always zero: used 501 498 * for zero-mapped memory areas etc..
-3
arch/m32r/include/asm/pgtable.h
··· 347 347 /* Needs to be defined here and not in linux/mm.h, as it is arch dependent */ 348 348 #define kern_addr_valid(addr) (1) 349 349 350 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 351 - remap_pfn_range(vma, vaddr, pfn, size, prot) 352 - 353 350 #define __HAVE_ARCH_PTEP_TEST_AND_CLEAR_YOUNG 354 351 #define __HAVE_ARCH_PTEP_GET_AND_CLEAR 355 352 #define __HAVE_ARCH_PTEP_SET_WRPROTECT
-3
arch/m68k/include/asm/pgtable_mm.h
··· 135 135 136 136 #define kern_addr_valid(addr) (1) 137 137 138 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 139 - remap_pfn_range(vma, vaddr, pfn, size, prot) 140 - 141 138 /* MMU-specific headers */ 142 139 143 140 #ifdef CONFIG_SUN3
-3
arch/m68k/include/asm/pgtable_no.h
··· 55 55 */ 56 56 #define pgtable_cache_init() do { } while (0) 57 57 58 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 59 - remap_pfn_range(vma, vaddr, pfn, size, prot) 60 - 61 58 /* 62 59 * All 32bit addresses are effectively valid for vmalloc... 63 60 * Sort of meaningless for non-VM targets.
-3
arch/metag/include/asm/pgtable.h
··· 333 333 334 334 #define kern_addr_valid(addr) (1) 335 335 336 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 337 - remap_pfn_range(vma, vaddr, pfn, size, prot) 338 - 339 336 /* 340 337 * No page table caches to initialise 341 338 */
-3
arch/microblaze/include/asm/pgtable.h
··· 13 13 14 14 #include <asm/setup.h> 15 15 16 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 17 - remap_pfn_range(vma, vaddr, pfn, size, prot) 18 - 19 16 #ifndef __ASSEMBLY__ 20 17 extern int mem_init_done; 21 18 #endif
+1 -3
arch/mips/include/asm/pgtable.h
··· 394 394 phys_t phys_addr_high = fixup_bigphys_addr(pfn << PAGE_SHIFT, size); 395 395 return remap_pfn_range(vma, vaddr, phys_addr_high >> PAGE_SHIFT, size, prot); 396 396 } 397 - #else 398 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 399 - remap_pfn_range(vma, vaddr, pfn, size, prot) 397 + #define io_remap_pfn_range io_remap_pfn_range 400 398 #endif 401 399 402 400 #ifdef CONFIG_TRANSPARENT_HUGEPAGE
-3
arch/mn10300/include/asm/pgtable.h
··· 486 486 487 487 #define kern_addr_valid(addr) (1) 488 488 489 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 490 - remap_pfn_range((vma), (vaddr), (pfn), (size), (prot)) 491 - 492 489 #define MK_IOSPACE_PFN(space, pfn) (pfn) 493 490 #define GET_IOSPACE(pfn) 0 494 491 #define GET_PFN(pfn) (pfn)
-3
arch/openrisc/include/asm/pgtable.h
··· 446 446 447 447 #define kern_addr_valid(addr) (1) 448 448 449 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 450 - remap_pfn_range(vma, vaddr, pfn, size, prot) 451 - 452 449 #include <asm-generic/pgtable.h> 453 450 454 451 /*
-3
arch/parisc/include/asm/pgtable.h
··· 506 506 #endif 507 507 508 508 509 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 510 - remap_pfn_range(vma, vaddr, pfn, size, prot) 511 - 512 509 #define pgprot_noncached(prot) __pgprot(pgprot_val(prot) | _PAGE_NO_CACHE) 513 510 514 511 /* We provide our own get_unmapped_area to provide cache coherency */
-3
arch/powerpc/include/asm/pgtable.h
··· 198 198 */ 199 199 #define kern_addr_valid(addr) (1) 200 200 201 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 202 - remap_pfn_range(vma, vaddr, pfn, size, prot) 203 - 204 201 #include <asm-generic/pgtable.h> 205 202 206 203
-3
arch/s390/include/asm/pgtable.h
··· 58 58 #define __HAVE_COLOR_ZERO_PAGE 59 59 60 60 /* TODO: s390 cannot support io_remap_pfn_range... */ 61 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 62 - remap_pfn_range(vma, vaddr, pfn, size, prot) 63 - 64 61 #endif /* !__ASSEMBLY__ */ 65 62 66 63 /*
-3
arch/score/include/asm/pgtable.h
··· 113 113 #define pte_clear(mm, addr, xp) \ 114 114 do { set_pte_at(mm, addr, xp, __pte(0)); } while (0) 115 115 116 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 117 - remap_pfn_range(vma, vaddr, pfn, size, prot) 118 - 119 116 /* 120 117 * The "pgd_xxx()" functions here are trivial for a folded two-level 121 118 * setup: the pgd is never bad, and a pmd always exists (as it's folded
-3
arch/sh/include/asm/pgtable.h
··· 124 124 125 125 #define kern_addr_valid(addr) (1) 126 126 127 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 128 - remap_pfn_range(vma, vaddr, pfn, size, prot) 129 - 130 127 #define pte_pfn(x) ((unsigned long)(((x).pte_low >> PAGE_SHIFT))) 131 128 132 129 /*
+1
arch/sparc/include/asm/pgtable_32.h
··· 443 443 444 444 return remap_pfn_range(vma, from, phys_base >> PAGE_SHIFT, size, prot); 445 445 } 446 + #define io_remap_pfn_range io_remap_pfn_range 446 447 447 448 #define __HAVE_ARCH_PTEP_SET_ACCESS_FLAGS 448 449 #define ptep_set_access_flags(__vma, __address, __ptep, __entry, __dirty) \
+1
arch/sparc/include/asm/pgtable_64.h
··· 914 914 915 915 return remap_pfn_range(vma, from, phys_base >> PAGE_SHIFT, size, prot); 916 916 } 917 + #define io_remap_pfn_range io_remap_pfn_range 917 918 918 919 #include <asm/tlbflush.h> 919 920 #include <asm-generic/pgtable.h>
-3
arch/tile/include/asm/pgtable.h
··· 362 362 #define kern_addr_valid(addr) (1) 363 363 #endif /* CONFIG_FLATMEM */ 364 364 365 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 366 - remap_pfn_range(vma, vaddr, pfn, size, prot) 367 - 368 365 extern void vmalloc_sync_all(void); 369 366 370 367 #endif /* !__ASSEMBLY__ */
-2
arch/um/include/asm/pgtable.h
··· 69 69 #define PAGE_KERNEL __pgprot(_PAGE_PRESENT | _PAGE_RW | _PAGE_DIRTY | _PAGE_ACCESSED) 70 70 #define PAGE_KERNEL_EXEC __pgprot(__PAGE_KERNEL_EXEC) 71 71 72 - #define io_remap_pfn_range remap_pfn_range 73 - 74 72 /* 75 73 * The i386 can't do page protection for execute, and considers that the same 76 74 * are read.
-7
arch/unicore32/include/asm/pgtable.h
··· 303 303 304 304 #include <asm-generic/pgtable.h> 305 305 306 - /* 307 - * remap a physical page `pfn' of size `size' with page protection `prot' 308 - * into virtual address `from' 309 - */ 310 - #define io_remap_pfn_range(vma, from, pfn, size, prot) \ 311 - remap_pfn_range(vma, from, pfn, size, prot) 312 - 313 306 #define pgtable_cache_init() do { } while (0) 314 307 315 308 #endif /* !__ASSEMBLY__ */
-3
arch/x86/include/asm/pgtable.h
··· 506 506 return npg >> (20 - PAGE_SHIFT); 507 507 } 508 508 509 - #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 510 - remap_pfn_range(vma, vaddr, pfn, size, prot) 511 - 512 509 #if PAGETABLE_LEVELS > 2 513 510 static inline int pud_none(pud_t pud) 514 511 {
-8
arch/xtensa/include/asm/pgtable.h
··· 393 393 extern void update_mmu_cache(struct vm_area_struct * vma, 394 394 unsigned long address, pte_t *ptep); 395 395 396 - /* 397 - * remap a physical page `pfn' of size `size' with page protection `prot' 398 - * into virtual address `from' 399 - */ 400 - 401 - #define io_remap_pfn_range(vma,from,pfn,size,prot) \ 402 - remap_pfn_range(vma, from, pfn, size, prot) 403 - 404 396 typedef pte_t *pte_addr_t; 405 397 406 398 #endif /* !defined (__ASSEMBLY__) */
+4
include/asm-generic/pgtable.h
··· 692 692 693 693 #endif /* !__ASSEMBLY__ */ 694 694 695 + #ifndef io_remap_pfn_range 696 + #define io_remap_pfn_range remap_pfn_range 697 + #endif 698 + 695 699 #endif /* _ASM_GENERIC_PGTABLE_H */