Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

Merge branch 'upstream' of git://ftp.linux-mips.org/pub/scm/upstream-linus

* 'upstream' of git://ftp.linux-mips.org/pub/scm/upstream-linus:
[MIPS] IP27: Build fix
[MIPS] Wire up ioprio_set and ioprio_get.
[MIPS] Fix __raw_read_trylock() to allow multiple readers
[MIPS] Export __copy_user_inatomic.
[MIPS] R2 bitops compile fix for gcc < 4.0.
[MIPS] TX39: Remove redundant tx39_blast_icache() calls
[MIPS] Cobalt: Fix early printk
[MIPS] SMTC: De-obscure Malta hooks.
[MIPS] SMTC: Add fordward declarations for mm_struct and task_struct.
[MIPS] SMTC: <asm/mips_mt.h> must include <linux/cpumask.h>
[MIPS] SMTC: <asm/smtc_ipi.h> must include <linux/spinlock.h>
[MIPS] Atlas, Malta: Fix build warning.

+84 -87
+1 -1
arch/mips/Kconfig
··· 167 167 select IRQ_CPU 168 168 select MIPS_GT64111 169 169 select SYS_HAS_CPU_NEVADA 170 + select SYS_HAS_EARLY_PRINTK 170 171 select SYS_SUPPORTS_32BIT_KERNEL 171 172 select SYS_SUPPORTS_64BIT_KERNEL if EXPERIMENTAL 172 173 select SYS_SUPPORTS_LITTLE_ENDIAN ··· 838 837 source "arch/mips/tx4938/Kconfig" 839 838 source "arch/mips/vr41xx/Kconfig" 840 839 source "arch/mips/philips/pnx8550/common/Kconfig" 841 - source "arch/mips/cobalt/Kconfig" 842 840 843 841 endmenu 844 842
-7
arch/mips/cobalt/Kconfig
··· 1 - config EARLY_PRINTK 2 - bool "Early console support" 3 - depends on MIPS_COBALT 4 - help 5 - Provide early console support by direct access to the 6 - on board UART. The UART must have been previously 7 - initialised by the boot loader.
+1 -4
arch/mips/cobalt/console.c
··· 9 9 #include <asm/addrspace.h> 10 10 #include <asm/mach-cobalt/cobalt.h> 11 11 12 - static void putchar(int c) 12 + void prom_putchar(char c) 13 13 { 14 - if(c == '\n') 15 - putchar('\r'); 16 - 17 14 while(!(COBALT_UART[UART_LSR] & UART_LSR_THRE)) 18 15 ; 19 16
+1
arch/mips/kernel/mips_ksyms.c
··· 37 37 * Userspace access stuff. 38 38 */ 39 39 EXPORT_SYMBOL(__copy_user); 40 + EXPORT_SYMBOL(__copy_user_inatomic); 40 41 EXPORT_SYMBOL(__bzero); 41 42 EXPORT_SYMBOL(__strncpy_from_user_nocheck_asm); 42 43 EXPORT_SYMBOL(__strncpy_from_user_asm);
+2
arch/mips/kernel/scall32-o32.S
··· 656 656 sys sys_kexec_load 4 657 657 sys sys_getcpu 3 658 658 sys sys_epoll_pwait 6 659 + sys sys_ioprio_set 3 660 + sys sys_ioprio_get 2 659 661 .endm 660 662 661 663 /* We pre-compute the number of _instruction_ bytes needed to
+3
arch/mips/kernel/scall64-64.S
··· 471 471 PTR sys_kexec_load /* 5270 */ 472 472 PTR sys_getcpu 473 473 PTR sys_epoll_pwait 474 + PTR sys_ioprio_set 475 + PTR sys_ioprio_get 476 + .size sys_call_table,.-sys_call_table
+4 -1
arch/mips/kernel/scall64-n32.S
··· 395 395 PTR compat_sys_set_robust_list 396 396 PTR compat_sys_get_robust_list 397 397 PTR compat_sys_kexec_load 398 - PTR sys_getcpu 398 + PTR sys_getcpu /* 6275 */ 399 399 PTR compat_sys_epoll_pwait 400 + PTR sys_ioprio_set 401 + PTR sys_ioprio_get 402 + .size sysn32_call_table,.-sysn32_call_table
+2
arch/mips/kernel/scall64-o32.S
··· 519 519 PTR compat_sys_kexec_load 520 520 PTR sys_getcpu 521 521 PTR compat_sys_epoll_pwait 522 + PTR sys_ioprio_set 523 + PTR sys_ioprio_get /* 4315 */ 522 524 .size sys_call_table,.-sys_call_table
+2 -2
arch/mips/mips-boards/generic/init.c
··· 251 251 252 252 void __init prom_init(void) 253 253 { 254 - u32 start, map, mask, data; 255 - 256 254 prom_argc = fw_arg0; 257 255 _prom_argv = (int *) fw_arg1; 258 256 _prom_envp = (int *) fw_arg2; ··· 276 278 mips_revision_corid = MIPS_REVISION_CORID_CORE_EMUL_MSC; 277 279 } 278 280 switch(mips_revision_corid) { 281 + u32 start, map, mask, data; 282 + 279 283 case MIPS_REVISION_CORID_QED_RM5261: 280 284 case MIPS_REVISION_CORID_CORE_LV: 281 285 case MIPS_REVISION_CORID_CORE_FPGA:
+1 -1
arch/mips/mips-boards/malta/Makefile
··· 21 21 22 22 obj-y := malta_int.o malta_setup.o 23 23 obj-$(CONFIG_MTD) += malta_mtd.o 24 - obj-$(CONFIG_SMP) += malta_smp.o 24 + obj-$(CONFIG_MIPS_MT_SMTC) += malta_smtc.o
+5 -26
arch/mips/mips-boards/malta/malta_smp.c arch/mips/mips-boards/malta/malta_smtc.c
··· 1 1 /* 2 2 * Malta Platform-specific hooks for SMP operation 3 3 */ 4 + #include <linux/init.h> 4 5 5 - #include <linux/kernel.h> 6 - #include <linux/sched.h> 7 - #include <linux/cpumask.h> 8 - #include <linux/interrupt.h> 9 - 10 - #include <asm/atomic.h> 11 - #include <asm/cpu.h> 12 - #include <asm/processor.h> 13 - #include <asm/system.h> 14 - #include <asm/hardirq.h> 15 - #include <asm/mmu_context.h> 16 - #include <asm/smp.h> 17 - #ifdef CONFIG_MIPS_MT_SMTC 6 + #include <asm/mipsregs.h> 7 + #include <asm/mipsmtregs.h> 8 + #include <asm/smtc.h> 18 9 #include <asm/smtc_ipi.h> 19 - #endif /* CONFIG_MIPS_MT_SMTC */ 20 10 21 11 /* VPE/SMP Prototype implements platform interfaces directly */ 22 - #if !defined(CONFIG_MIPS_MT_SMP) 23 12 24 13 /* 25 14 * Cause the specified action to be performed on a targeted "CPU" ··· 16 27 17 28 void core_send_ipi(int cpu, unsigned int action) 18 29 { 19 - /* "CPU" may be TC of same VPE, VPE of same CPU, or different CPU */ 20 - #ifdef CONFIG_MIPS_MT_SMTC 30 + /* "CPU" may be TC of same VPE, VPE of same CPU, or different CPU */ 21 31 smtc_send_ipi(cpu, LINUX_SMP_IPI, action); 22 - #endif /* CONFIG_MIPS_MT_SMTC */ 23 32 } 24 33 25 34 /* ··· 26 39 27 40 void prom_boot_secondary(int cpu, struct task_struct *idle) 28 41 { 29 - #ifdef CONFIG_MIPS_MT_SMTC 30 42 smtc_boot_secondary(cpu, idle); 31 - #endif /* CONFIG_MIPS_MT_SMTC */ 32 43 } 33 44 34 45 /* ··· 35 50 36 51 void prom_init_secondary(void) 37 52 { 38 - #ifdef CONFIG_MIPS_MT_SMTC 39 53 void smtc_init_secondary(void); 40 54 int myvpe; 41 55 ··· 49 65 } 50 66 51 67 smtc_init_secondary(); 52 - #endif /* CONFIG_MIPS_MT_SMTC */ 53 68 } 54 69 55 70 /* ··· 76 93 77 94 void prom_smp_finish(void) 78 95 { 79 - #ifdef CONFIG_MIPS_MT_SMTC 80 96 smtc_smp_finish(); 81 - #endif /* CONFIG_MIPS_MT_SMTC */ 82 97 } 83 98 84 99 /* ··· 86 105 void prom_cpus_done(void) 87 106 { 88 107 } 89 - 90 - #endif /* CONFIG_MIPS32R2_MT_SMP */
+6 -12
arch/mips/mm/c-tx39.c
··· 128 128 return; 129 129 130 130 tx39_blast_dcache(); 131 - tx39_blast_icache(); 132 131 } 133 132 134 133 static inline void tx39___flush_cache_all(void) ··· 141 142 if (!cpu_has_dc_aliases) 142 143 return; 143 144 144 - if (cpu_context(smp_processor_id(), mm) != 0) { 145 - tx39_flush_cache_all(); 146 - } 145 + if (cpu_context(smp_processor_id(), mm) != 0) 146 + tx39_blast_dcache(); 147 147 } 148 148 149 149 static void tx39_flush_cache_range(struct vm_area_struct *vma, 150 150 unsigned long start, unsigned long end) 151 151 { 152 - int exec; 153 - 152 + if (!cpu_has_dc_aliases) 153 + return; 154 154 if (!(cpu_context(smp_processor_id(), vma->vm_mm))) 155 155 return; 156 156 157 - exec = vma->vm_flags & VM_EXEC; 158 - if (cpu_has_dc_aliases || exec) 159 - tx39_blast_dcache(); 160 - if (exec) 161 - tx39_blast_icache(); 157 + tx39_blast_dcache(); 162 158 } 163 159 164 160 static void tx39_flush_cache_page(struct vm_area_struct *vma, unsigned long page, unsigned long pfn) ··· 212 218 213 219 static void local_tx39_flush_data_cache_page(void * addr) 214 220 { 215 - tx39_blast_dcache_page(addr); 221 + tx39_blast_dcache_page((unsigned long)addr); 216 222 } 217 223 218 224 static void tx39_flush_data_cache_page(unsigned long addr)
-2
arch/mips/sgi-ip27/ip27-init.c
··· 191 191 ioc3->eier = 0; 192 192 } 193 193 194 - extern void ip27_setup_console(void); 195 194 extern void ip27_time_init(void); 196 195 extern void ip27_reboot_setup(void); 197 196 ··· 199 200 hubreg_t p, e, n_mode; 200 201 nasid_t nid; 201 202 202 - ip27_setup_console(); 203 203 ip27_reboot_setup(); 204 204 205 205 /*
+33 -23
include/asm-mips/bitops.h
··· 54 54 static inline void set_bit(unsigned long nr, volatile unsigned long *addr) 55 55 { 56 56 unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); 57 + unsigned short bit = nr & SZLONG_MASK; 57 58 unsigned long temp; 58 59 59 60 if (cpu_has_llsc && R10000_LLSC_WAR) { ··· 66 65 " beqzl %0, 1b \n" 67 66 " .set mips0 \n" 68 67 : "=&r" (temp), "=m" (*m) 69 - : "ir" (1UL << (nr & SZLONG_MASK)), "m" (*m)); 68 + : "ir" (1UL << bit), "m" (*m)); 70 69 #ifdef CONFIG_CPU_MIPSR2 71 - } else if (__builtin_constant_p(nr)) { 70 + } else if (__builtin_constant_p(bit)) { 72 71 __asm__ __volatile__( 73 72 "1: " __LL "%0, %1 # set_bit \n" 74 73 " " __INS "%0, %4, %2, 1 \n" ··· 78 77 "2: b 1b \n" 79 78 " .previous \n" 80 79 : "=&r" (temp), "=m" (*m) 81 - : "ir" (nr & SZLONG_MASK), "m" (*m), "r" (~0)); 80 + : "ir" (bit), "m" (*m), "r" (~0)); 82 81 #endif /* CONFIG_CPU_MIPSR2 */ 83 82 } else if (cpu_has_llsc) { 84 83 __asm__ __volatile__( ··· 92 91 " .previous \n" 93 92 " .set mips0 \n" 94 93 : "=&r" (temp), "=m" (*m) 95 - : "ir" (1UL << (nr & SZLONG_MASK)), "m" (*m)); 94 + : "ir" (1UL << bit), "m" (*m)); 96 95 } else { 97 96 volatile unsigned long *a = addr; 98 97 unsigned long mask; 99 98 unsigned long flags; 100 99 101 100 a += nr >> SZLONG_LOG; 102 - mask = 1UL << (nr & SZLONG_MASK); 101 + mask = 1UL << bit; 103 102 local_irq_save(flags); 104 103 *a |= mask; 105 104 local_irq_restore(flags); ··· 119 118 static inline void clear_bit(unsigned long nr, volatile unsigned long *addr) 120 119 { 121 120 unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); 121 + unsigned short bit = nr & SZLONG_MASK; 122 122 unsigned long temp; 123 123 124 124 if (cpu_has_llsc && R10000_LLSC_WAR) { ··· 131 129 " beqzl %0, 1b \n" 132 130 " .set mips0 \n" 133 131 : "=&r" (temp), "=m" (*m) 134 - : "ir" (~(1UL << (nr & SZLONG_MASK))), "m" (*m)); 132 + : "ir" (~(1UL << bit)), "m" (*m)); 135 133 #ifdef CONFIG_CPU_MIPSR2 136 - } else if (__builtin_constant_p(nr)) { 134 + } else if (__builtin_constant_p(bit)) { 137 135 __asm__ __volatile__( 138 136 "1: " __LL "%0, %1 # clear_bit \n" 139 137 " " __INS "%0, $0, %2, 1 \n" ··· 143 141 "2: b 1b \n" 144 142 " .previous \n" 145 143 : "=&r" (temp), "=m" (*m) 146 - : "ir" (nr & SZLONG_MASK), "m" (*m)); 144 + : "ir" (bit), "m" (*m)); 147 145 #endif /* CONFIG_CPU_MIPSR2 */ 148 146 } else if (cpu_has_llsc) { 149 147 __asm__ __volatile__( ··· 157 155 " .previous \n" 158 156 " .set mips0 \n" 159 157 : "=&r" (temp), "=m" (*m) 160 - : "ir" (~(1UL << (nr & SZLONG_MASK))), "m" (*m)); 158 + : "ir" (~(1UL << bit)), "m" (*m)); 161 159 } else { 162 160 volatile unsigned long *a = addr; 163 161 unsigned long mask; 164 162 unsigned long flags; 165 163 166 164 a += nr >> SZLONG_LOG; 167 - mask = 1UL << (nr & SZLONG_MASK); 165 + mask = 1UL << bit; 168 166 local_irq_save(flags); 169 167 *a &= ~mask; 170 168 local_irq_restore(flags); ··· 182 180 */ 183 181 static inline void change_bit(unsigned long nr, volatile unsigned long *addr) 184 182 { 183 + unsigned short bit = nr & SZLONG_MASK; 184 + 185 185 if (cpu_has_llsc && R10000_LLSC_WAR) { 186 186 unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); 187 187 unsigned long temp; ··· 196 192 " beqzl %0, 1b \n" 197 193 " .set mips0 \n" 198 194 : "=&r" (temp), "=m" (*m) 199 - : "ir" (1UL << (nr & SZLONG_MASK)), "m" (*m)); 195 + : "ir" (1UL << bit), "m" (*m)); 200 196 } else if (cpu_has_llsc) { 201 197 unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); 202 198 unsigned long temp; ··· 212 208 " .previous \n" 213 209 " .set mips0 \n" 214 210 : "=&r" (temp), "=m" (*m) 215 - : "ir" (1UL << (nr & SZLONG_MASK)), "m" (*m)); 211 + : "ir" (1UL << bit), "m" (*m)); 216 212 } else { 217 213 volatile unsigned long *a = addr; 218 214 unsigned long mask; 219 215 unsigned long flags; 220 216 221 217 a += nr >> SZLONG_LOG; 222 - mask = 1UL << (nr & SZLONG_MASK); 218 + mask = 1UL << bit; 223 219 local_irq_save(flags); 224 220 *a ^= mask; 225 221 local_irq_restore(flags); ··· 237 233 static inline int test_and_set_bit(unsigned long nr, 238 234 volatile unsigned long *addr) 239 235 { 236 + unsigned short bit = nr & SZLONG_MASK; 237 + 240 238 if (cpu_has_llsc && R10000_LLSC_WAR) { 241 239 unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); 242 240 unsigned long temp, res; ··· 252 246 " and %2, %0, %3 \n" 253 247 " .set mips0 \n" 254 248 : "=&r" (temp), "=m" (*m), "=&r" (res) 255 - : "r" (1UL << (nr & SZLONG_MASK)), "m" (*m) 249 + : "r" (1UL << bit), "m" (*m) 256 250 : "memory"); 257 251 258 252 return res != 0; ··· 275 269 " .previous \n" 276 270 " .set pop \n" 277 271 : "=&r" (temp), "=m" (*m), "=&r" (res) 278 - : "r" (1UL << (nr & SZLONG_MASK)), "m" (*m) 272 + : "r" (1UL << bit), "m" (*m) 279 273 : "memory"); 280 274 281 275 return res != 0; ··· 286 280 unsigned long flags; 287 281 288 282 a += nr >> SZLONG_LOG; 289 - mask = 1UL << (nr & SZLONG_MASK); 283 + mask = 1UL << bit; 290 284 local_irq_save(flags); 291 285 retval = (mask & *a) != 0; 292 286 *a |= mask; ··· 309 303 static inline int test_and_clear_bit(unsigned long nr, 310 304 volatile unsigned long *addr) 311 305 { 306 + unsigned short bit = nr & SZLONG_MASK; 307 + 312 308 if (cpu_has_llsc && R10000_LLSC_WAR) { 313 309 unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); 314 310 unsigned long temp, res; ··· 325 317 " and %2, %0, %3 \n" 326 318 " .set mips0 \n" 327 319 : "=&r" (temp), "=m" (*m), "=&r" (res) 328 - : "r" (1UL << (nr & SZLONG_MASK)), "m" (*m) 320 + : "r" (1UL << bit), "m" (*m) 329 321 : "memory"); 330 322 331 323 return res != 0; ··· 344 336 "2: b 1b \n" 345 337 " .previous \n" 346 338 : "=&r" (temp), "=m" (*m), "=&r" (res) 347 - : "ri" (nr & SZLONG_MASK), "m" (*m) 339 + : "ri" (bit), "m" (*m) 348 340 : "memory"); 349 341 350 342 return res; ··· 369 361 " .previous \n" 370 362 " .set pop \n" 371 363 : "=&r" (temp), "=m" (*m), "=&r" (res) 372 - : "r" (1UL << (nr & SZLONG_MASK)), "m" (*m) 364 + : "r" (1UL << bit), "m" (*m) 373 365 : "memory"); 374 366 375 367 return res != 0; ··· 380 372 unsigned long flags; 381 373 382 374 a += nr >> SZLONG_LOG; 383 - mask = 1UL << (nr & SZLONG_MASK); 375 + mask = 1UL << bit; 384 376 local_irq_save(flags); 385 377 retval = (mask & *a) != 0; 386 378 *a &= ~mask; ··· 403 395 static inline int test_and_change_bit(unsigned long nr, 404 396 volatile unsigned long *addr) 405 397 { 398 + unsigned short bit = nr & SZLONG_MASK; 399 + 406 400 if (cpu_has_llsc && R10000_LLSC_WAR) { 407 401 unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); 408 402 unsigned long temp, res; ··· 418 408 " and %2, %0, %3 \n" 419 409 " .set mips0 \n" 420 410 : "=&r" (temp), "=m" (*m), "=&r" (res) 421 - : "r" (1UL << (nr & SZLONG_MASK)), "m" (*m) 411 + : "r" (1UL << bit), "m" (*m) 422 412 : "memory"); 423 413 424 414 return res != 0; ··· 441 431 " .previous \n" 442 432 " .set pop \n" 443 433 : "=&r" (temp), "=m" (*m), "=&r" (res) 444 - : "r" (1UL << (nr & SZLONG_MASK)), "m" (*m) 434 + : "r" (1UL << bit), "m" (*m) 445 435 : "memory"); 446 436 447 437 return res != 0; ··· 451 441 unsigned long flags; 452 442 453 443 a += nr >> SZLONG_LOG; 454 - mask = 1UL << (nr & SZLONG_MASK); 444 + mask = 1UL << bit; 455 445 local_irq_save(flags); 456 446 retval = (mask & *a) != 0; 457 447 *a ^= mask;
+2
include/asm-mips/mips_mt.h
··· 6 6 #ifndef __ASM_MIPS_MT_H 7 7 #define __ASM_MIPS_MT_H 8 8 9 + #include <linux/cpumask.h> 10 + 9 11 extern cpumask_t mt_fpu_cpumask; 10 12 extern unsigned long mt_fpemul_threshold; 11 13
+3
include/asm-mips/smtc.h
··· 34 34 35 35 extern asiduse smtc_live_asid[MAX_SMTC_TLBS][MAX_SMTC_ASIDS]; 36 36 37 + struct mm_struct; 38 + struct task_struct; 39 + 37 40 void smtc_get_new_mmu_context(struct mm_struct *mm, unsigned long cpu); 38 41 39 42 void smtc_flush_tlb_asid(unsigned long asid);
+2
include/asm-mips/smtc_ipi.h
··· 4 4 #ifndef __ASM_SMTC_IPI_H 5 5 #define __ASM_SMTC_IPI_H 6 6 7 + #include <linux/spinlock.h> 8 + 7 9 //#define SMTC_IPI_DEBUG 8 10 9 11 #ifdef SMTC_IPI_DEBUG
+2 -2
include/asm-mips/spinlock.h
··· 287 287 " .set noreorder # __raw_read_trylock \n" 288 288 " li %2, 0 \n" 289 289 "1: ll %1, %3 \n" 290 - " bnez %1, 2f \n" 290 + " bltz %1, 2f \n" 291 291 " addu %1, 1 \n" 292 292 " sc %1, %0 \n" 293 293 " .set reorder \n" ··· 304 304 " .set noreorder # __raw_read_trylock \n" 305 305 " li %2, 0 \n" 306 306 "1: ll %1, %3 \n" 307 - " bnez %1, 2f \n" 307 + " bltz %1, 2f \n" 308 308 " addu %1, 1 \n" 309 309 " sc %1, %0 \n" 310 310 " beqz %1, 1b \n"
+2
include/asm-mips/uaccess.h
··· 435 435 __cu_len; \ 436 436 }) 437 437 438 + extern size_t __copy_user_inatomic(void *__to, const void *__from, size_t __n); 439 + 438 440 #define __copy_to_user_inatomic(to,from,n) \ 439 441 ({ \ 440 442 void __user *__cu_to; \
+12 -6
include/asm-mips/unistd.h
··· 334 334 #define __NR_kexec_load (__NR_Linux + 311) 335 335 #define __NR_getcpu (__NR_Linux + 312) 336 336 #define __NR_epoll_pwait (__NR_Linux + 313) 337 + #define __NR_ioprio_set (__NR_Linux + 314) 338 + #define __NR_ioprio_get (__NR_Linux + 315) 337 339 338 340 /* 339 341 * Offset of the last Linux o32 flavoured syscall 340 342 */ 341 - #define __NR_Linux_syscalls 313 343 + #define __NR_Linux_syscalls 315 342 344 343 345 #endif /* _MIPS_SIM == _MIPS_SIM_ABI32 */ 344 346 345 347 #define __NR_O32_Linux 4000 346 - #define __NR_O32_Linux_syscalls 313 348 + #define __NR_O32_Linux_syscalls 315 347 349 348 350 #if _MIPS_SIM == _MIPS_SIM_ABI64 349 351 ··· 626 624 #define __NR_kexec_load (__NR_Linux + 270) 627 625 #define __NR_getcpu (__NR_Linux + 271) 628 626 #define __NR_epoll_pwait (__NR_Linux + 272) 627 + #define __NR_ioprio_set (__NR_Linux + 273) 628 + #define __NR_ioprio_get (__NR_Linux + 274) 629 629 630 630 /* 631 631 * Offset of the last Linux 64-bit flavoured syscall 632 632 */ 633 - #define __NR_Linux_syscalls 272 633 + #define __NR_Linux_syscalls 274 634 634 635 635 #endif /* _MIPS_SIM == _MIPS_SIM_ABI64 */ 636 636 637 637 #define __NR_64_Linux 5000 638 - #define __NR_64_Linux_syscalls 272 638 + #define __NR_64_Linux_syscalls 274 639 639 640 640 #if _MIPS_SIM == _MIPS_SIM_NABI32 641 641 ··· 922 918 #define __NR_kexec_load (__NR_Linux + 274) 923 919 #define __NR_getcpu (__NR_Linux + 275) 924 920 #define __NR_epoll_pwait (__NR_Linux + 276) 921 + #define __NR_ioprio_set (__NR_Linux + 277) 922 + #define __NR_ioprio_get (__NR_Linux + 278) 925 923 926 924 /* 927 925 * Offset of the last N32 flavoured syscall 928 926 */ 929 - #define __NR_Linux_syscalls 276 927 + #define __NR_Linux_syscalls 278 930 928 931 929 #endif /* _MIPS_SIM == _MIPS_SIM_NABI32 */ 932 930 933 931 #define __NR_N32_Linux 6000 934 - #define __NR_N32_Linux_syscalls 276 932 + #define __NR_N32_Linux_syscalls 278 935 933 936 934 #ifdef __KERNEL__ 937 935