Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

locking: Cleanup the name space completely

Make the name space hierarchy of locking functions consistent:
raw_spin* -> _raw_spin* -> __raw_spin*

No functional change.

Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Acked-by: Peter Zijlstra <peterz@infradead.org>
Acked-by: Ingo Molnar <mingo@elte.hu>

+272 -259
+24 -24
include/linux/rwlock.h
··· 38 38 extern int do_raw_write_trylock(rwlock_t *lock); 39 39 extern void do_raw_write_unlock(rwlock_t *lock); 40 40 #else 41 - # define do_raw_read_lock(rwlock) arch_read_lock(&(rwlock)->raw_lock) 41 + # define do_raw_read_lock(rwlock) arch_read_lock(&(rwlock)->raw_lock) 42 42 # define do_raw_read_lock_flags(lock, flags) \ 43 43 arch_read_lock_flags(&(lock)->raw_lock, *(flags)) 44 44 # define do_raw_read_trylock(rwlock) arch_read_trylock(&(rwlock)->raw_lock) ··· 58 58 * regardless of whether CONFIG_SMP or CONFIG_PREEMPT are set. The various 59 59 * methods are defined as nops in the case they are not required. 60 60 */ 61 - #define read_trylock(lock) __cond_lock(lock, _read_trylock(lock)) 62 - #define write_trylock(lock) __cond_lock(lock, _write_trylock(lock)) 61 + #define read_trylock(lock) __cond_lock(lock, _raw_read_trylock(lock)) 62 + #define write_trylock(lock) __cond_lock(lock, _raw_write_trylock(lock)) 63 63 64 - #define write_lock(lock) _write_lock(lock) 65 - #define read_lock(lock) _read_lock(lock) 64 + #define write_lock(lock) _raw_write_lock(lock) 65 + #define read_lock(lock) _raw_read_lock(lock) 66 66 67 67 #if defined(CONFIG_SMP) || defined(CONFIG_DEBUG_SPINLOCK) 68 68 69 69 #define read_lock_irqsave(lock, flags) \ 70 70 do { \ 71 71 typecheck(unsigned long, flags); \ 72 - flags = _read_lock_irqsave(lock); \ 72 + flags = _raw_read_lock_irqsave(lock); \ 73 73 } while (0) 74 74 #define write_lock_irqsave(lock, flags) \ 75 75 do { \ 76 76 typecheck(unsigned long, flags); \ 77 - flags = _write_lock_irqsave(lock); \ 77 + flags = _raw_write_lock_irqsave(lock); \ 78 78 } while (0) 79 79 80 80 #else ··· 82 82 #define read_lock_irqsave(lock, flags) \ 83 83 do { \ 84 84 typecheck(unsigned long, flags); \ 85 - _read_lock_irqsave(lock, flags); \ 85 + _raw_read_lock_irqsave(lock, flags); \ 86 86 } while (0) 87 87 #define write_lock_irqsave(lock, flags) \ 88 88 do { \ 89 89 typecheck(unsigned long, flags); \ 90 - _write_lock_irqsave(lock, flags); \ 90 + _raw_write_lock_irqsave(lock, flags); \ 91 91 } while (0) 92 92 93 93 #endif 94 94 95 - #define read_lock_irq(lock) _read_lock_irq(lock) 96 - #define read_lock_bh(lock) _read_lock_bh(lock) 97 - #define write_lock_irq(lock) _write_lock_irq(lock) 98 - #define write_lock_bh(lock) _write_lock_bh(lock) 99 - #define read_unlock(lock) _read_unlock(lock) 100 - #define write_unlock(lock) _write_unlock(lock) 101 - #define read_unlock_irq(lock) _read_unlock_irq(lock) 102 - #define write_unlock_irq(lock) _write_unlock_irq(lock) 95 + #define read_lock_irq(lock) _raw_read_lock_irq(lock) 96 + #define read_lock_bh(lock) _raw_read_lock_bh(lock) 97 + #define write_lock_irq(lock) _raw_write_lock_irq(lock) 98 + #define write_lock_bh(lock) _raw_write_lock_bh(lock) 99 + #define read_unlock(lock) _raw_read_unlock(lock) 100 + #define write_unlock(lock) _raw_write_unlock(lock) 101 + #define read_unlock_irq(lock) _raw_read_unlock_irq(lock) 102 + #define write_unlock_irq(lock) _raw_write_unlock_irq(lock) 103 103 104 - #define read_unlock_irqrestore(lock, flags) \ 105 - do { \ 106 - typecheck(unsigned long, flags); \ 107 - _read_unlock_irqrestore(lock, flags); \ 104 + #define read_unlock_irqrestore(lock, flags) \ 105 + do { \ 106 + typecheck(unsigned long, flags); \ 107 + _raw_read_unlock_irqrestore(lock, flags); \ 108 108 } while (0) 109 - #define read_unlock_bh(lock) _read_unlock_bh(lock) 109 + #define read_unlock_bh(lock) _raw_read_unlock_bh(lock) 110 110 111 111 #define write_unlock_irqrestore(lock, flags) \ 112 112 do { \ 113 113 typecheck(unsigned long, flags); \ 114 - _write_unlock_irqrestore(lock, flags); \ 114 + _raw_write_unlock_irqrestore(lock, flags); \ 115 115 } while (0) 116 - #define write_unlock_bh(lock) _write_unlock_bh(lock) 116 + #define write_unlock_bh(lock) _raw_write_unlock_bh(lock) 117 117 118 118 #define write_trylock_irqsave(lock, flags) \ 119 119 ({ \
+59 -54
include/linux/rwlock_api_smp.h
··· 15 15 * Released under the General Public License (GPL). 16 16 */ 17 17 18 - void __lockfunc _read_lock(rwlock_t *lock) __acquires(lock); 19 - void __lockfunc _write_lock(rwlock_t *lock) __acquires(lock); 20 - void __lockfunc _read_lock_bh(rwlock_t *lock) __acquires(lock); 21 - void __lockfunc _write_lock_bh(rwlock_t *lock) __acquires(lock); 22 - void __lockfunc _read_lock_irq(rwlock_t *lock) __acquires(lock); 23 - void __lockfunc _write_lock_irq(rwlock_t *lock) __acquires(lock); 24 - unsigned long __lockfunc _read_lock_irqsave(rwlock_t *lock) 18 + void __lockfunc _raw_read_lock(rwlock_t *lock) __acquires(lock); 19 + void __lockfunc _raw_write_lock(rwlock_t *lock) __acquires(lock); 20 + void __lockfunc _raw_read_lock_bh(rwlock_t *lock) __acquires(lock); 21 + void __lockfunc _raw_write_lock_bh(rwlock_t *lock) __acquires(lock); 22 + void __lockfunc _raw_read_lock_irq(rwlock_t *lock) __acquires(lock); 23 + void __lockfunc _raw_write_lock_irq(rwlock_t *lock) __acquires(lock); 24 + unsigned long __lockfunc _raw_read_lock_irqsave(rwlock_t *lock) 25 25 __acquires(lock); 26 - unsigned long __lockfunc _write_lock_irqsave(rwlock_t *lock) 26 + unsigned long __lockfunc _raw_write_lock_irqsave(rwlock_t *lock) 27 27 __acquires(lock); 28 - int __lockfunc _read_trylock(rwlock_t *lock); 29 - int __lockfunc _write_trylock(rwlock_t *lock); 30 - void __lockfunc _read_unlock(rwlock_t *lock) __releases(lock); 31 - void __lockfunc _write_unlock(rwlock_t *lock) __releases(lock); 32 - void __lockfunc _read_unlock_bh(rwlock_t *lock) __releases(lock); 33 - void __lockfunc _write_unlock_bh(rwlock_t *lock) __releases(lock); 34 - void __lockfunc _read_unlock_irq(rwlock_t *lock) __releases(lock); 35 - void __lockfunc _write_unlock_irq(rwlock_t *lock) __releases(lock); 36 - void __lockfunc _read_unlock_irqrestore(rwlock_t *lock, unsigned long flags) 28 + int __lockfunc _raw_read_trylock(rwlock_t *lock); 29 + int __lockfunc _raw_write_trylock(rwlock_t *lock); 30 + void __lockfunc _raw_read_unlock(rwlock_t *lock) __releases(lock); 31 + void __lockfunc _raw_write_unlock(rwlock_t *lock) __releases(lock); 32 + void __lockfunc _raw_read_unlock_bh(rwlock_t *lock) __releases(lock); 33 + void __lockfunc _raw_write_unlock_bh(rwlock_t *lock) __releases(lock); 34 + void __lockfunc _raw_read_unlock_irq(rwlock_t *lock) __releases(lock); 35 + void __lockfunc _raw_write_unlock_irq(rwlock_t *lock) __releases(lock); 36 + void __lockfunc 37 + _raw_read_unlock_irqrestore(rwlock_t *lock, unsigned long flags) 37 38 __releases(lock); 38 - void __lockfunc _write_unlock_irqrestore(rwlock_t *lock, unsigned long flags) 39 + void __lockfunc 40 + _raw_write_unlock_irqrestore(rwlock_t *lock, unsigned long flags) 39 41 __releases(lock); 40 42 41 43 #ifdef CONFIG_INLINE_READ_LOCK 42 - #define _read_lock(lock) __read_lock(lock) 44 + #define _raw_read_lock(lock) __raw_read_lock(lock) 43 45 #endif 44 46 45 47 #ifdef CONFIG_INLINE_WRITE_LOCK 46 - #define _write_lock(lock) __write_lock(lock) 48 + #define _raw_write_lock(lock) __raw_write_lock(lock) 47 49 #endif 48 50 49 51 #ifdef CONFIG_INLINE_READ_LOCK_BH 50 - #define _read_lock_bh(lock) __read_lock_bh(lock) 52 + #define _raw_read_lock_bh(lock) __raw_read_lock_bh(lock) 51 53 #endif 52 54 53 55 #ifdef CONFIG_INLINE_WRITE_LOCK_BH 54 - #define _write_lock_bh(lock) __write_lock_bh(lock) 56 + #define _raw_write_lock_bh(lock) __raw_write_lock_bh(lock) 55 57 #endif 56 58 57 59 #ifdef CONFIG_INLINE_READ_LOCK_IRQ 58 - #define _read_lock_irq(lock) __read_lock_irq(lock) 60 + #define _raw_read_lock_irq(lock) __raw_read_lock_irq(lock) 59 61 #endif 60 62 61 63 #ifdef CONFIG_INLINE_WRITE_LOCK_IRQ 62 - #define _write_lock_irq(lock) __write_lock_irq(lock) 64 + #define _raw_write_lock_irq(lock) __raw_write_lock_irq(lock) 63 65 #endif 64 66 65 67 #ifdef CONFIG_INLINE_READ_LOCK_IRQSAVE 66 - #define _read_lock_irqsave(lock) __read_lock_irqsave(lock) 68 + #define _raw_read_lock_irqsave(lock) __raw_read_lock_irqsave(lock) 67 69 #endif 68 70 69 71 #ifdef CONFIG_INLINE_WRITE_LOCK_IRQSAVE 70 - #define _write_lock_irqsave(lock) __write_lock_irqsave(lock) 72 + #define _raw_write_lock_irqsave(lock) __raw_write_lock_irqsave(lock) 71 73 #endif 72 74 73 75 #ifdef CONFIG_INLINE_READ_TRYLOCK 74 - #define _read_trylock(lock) __read_trylock(lock) 76 + #define _raw_read_trylock(lock) __raw_read_trylock(lock) 75 77 #endif 76 78 77 79 #ifdef CONFIG_INLINE_WRITE_TRYLOCK 78 - #define _write_trylock(lock) __write_trylock(lock) 80 + #define _raw_write_trylock(lock) __raw_write_trylock(lock) 79 81 #endif 80 82 81 83 #ifdef CONFIG_INLINE_READ_UNLOCK 82 - #define _read_unlock(lock) __read_unlock(lock) 84 + #define _raw_read_unlock(lock) __raw_read_unlock(lock) 83 85 #endif 84 86 85 87 #ifdef CONFIG_INLINE_WRITE_UNLOCK 86 - #define _write_unlock(lock) __write_unlock(lock) 88 + #define _raw_write_unlock(lock) __raw_write_unlock(lock) 87 89 #endif 88 90 89 91 #ifdef CONFIG_INLINE_READ_UNLOCK_BH 90 - #define _read_unlock_bh(lock) __read_unlock_bh(lock) 92 + #define _raw_read_unlock_bh(lock) __raw_read_unlock_bh(lock) 91 93 #endif 92 94 93 95 #ifdef CONFIG_INLINE_WRITE_UNLOCK_BH 94 - #define _write_unlock_bh(lock) __write_unlock_bh(lock) 96 + #define _raw_write_unlock_bh(lock) __raw_write_unlock_bh(lock) 95 97 #endif 96 98 97 99 #ifdef CONFIG_INLINE_READ_UNLOCK_IRQ 98 - #define _read_unlock_irq(lock) __read_unlock_irq(lock) 100 + #define _raw_read_unlock_irq(lock) __raw_read_unlock_irq(lock) 99 101 #endif 100 102 101 103 #ifdef CONFIG_INLINE_WRITE_UNLOCK_IRQ 102 - #define _write_unlock_irq(lock) __write_unlock_irq(lock) 104 + #define _raw_write_unlock_irq(lock) __raw_write_unlock_irq(lock) 103 105 #endif 104 106 105 107 #ifdef CONFIG_INLINE_READ_UNLOCK_IRQRESTORE 106 - #define _read_unlock_irqrestore(lock, flags) __read_unlock_irqrestore(lock, flags) 108 + #define _raw_read_unlock_irqrestore(lock, flags) \ 109 + __raw_read_unlock_irqrestore(lock, flags) 107 110 #endif 108 111 109 112 #ifdef CONFIG_INLINE_WRITE_UNLOCK_IRQRESTORE 110 - #define _write_unlock_irqrestore(lock, flags) __write_unlock_irqrestore(lock, flags) 113 + #define _raw_write_unlock_irqrestore(lock, flags) \ 114 + __raw_write_unlock_irqrestore(lock, flags) 111 115 #endif 112 116 113 - static inline int __read_trylock(rwlock_t *lock) 117 + static inline int __raw_read_trylock(rwlock_t *lock) 114 118 { 115 119 preempt_disable(); 116 120 if (do_raw_read_trylock(lock)) { ··· 125 121 return 0; 126 122 } 127 123 128 - static inline int __write_trylock(rwlock_t *lock) 124 + static inline int __raw_write_trylock(rwlock_t *lock) 129 125 { 130 126 preempt_disable(); 131 127 if (do_raw_write_trylock(lock)) { ··· 143 139 */ 144 140 #if !defined(CONFIG_GENERIC_LOCKBREAK) || defined(CONFIG_DEBUG_LOCK_ALLOC) 145 141 146 - static inline void __read_lock(rwlock_t *lock) 142 + static inline void __raw_read_lock(rwlock_t *lock) 147 143 { 148 144 preempt_disable(); 149 145 rwlock_acquire_read(&lock->dep_map, 0, 0, _RET_IP_); 150 146 LOCK_CONTENDED(lock, do_raw_read_trylock, do_raw_read_lock); 151 147 } 152 148 153 - static inline unsigned long __read_lock_irqsave(rwlock_t *lock) 149 + static inline unsigned long __raw_read_lock_irqsave(rwlock_t *lock) 154 150 { 155 151 unsigned long flags; 156 152 ··· 162 158 return flags; 163 159 } 164 160 165 - static inline void __read_lock_irq(rwlock_t *lock) 161 + static inline void __raw_read_lock_irq(rwlock_t *lock) 166 162 { 167 163 local_irq_disable(); 168 164 preempt_disable(); ··· 170 166 LOCK_CONTENDED(lock, do_raw_read_trylock, do_raw_read_lock); 171 167 } 172 168 173 - static inline void __read_lock_bh(rwlock_t *lock) 169 + static inline void __raw_read_lock_bh(rwlock_t *lock) 174 170 { 175 171 local_bh_disable(); 176 172 preempt_disable(); ··· 178 174 LOCK_CONTENDED(lock, do_raw_read_trylock, do_raw_read_lock); 179 175 } 180 176 181 - static inline unsigned long __write_lock_irqsave(rwlock_t *lock) 177 + static inline unsigned long __raw_write_lock_irqsave(rwlock_t *lock) 182 178 { 183 179 unsigned long flags; 184 180 ··· 190 186 return flags; 191 187 } 192 188 193 - static inline void __write_lock_irq(rwlock_t *lock) 189 + static inline void __raw_write_lock_irq(rwlock_t *lock) 194 190 { 195 191 local_irq_disable(); 196 192 preempt_disable(); ··· 198 194 LOCK_CONTENDED(lock, do_raw_write_trylock, do_raw_write_lock); 199 195 } 200 196 201 - static inline void __write_lock_bh(rwlock_t *lock) 197 + static inline void __raw_write_lock_bh(rwlock_t *lock) 202 198 { 203 199 local_bh_disable(); 204 200 preempt_disable(); ··· 206 202 LOCK_CONTENDED(lock, do_raw_write_trylock, do_raw_write_lock); 207 203 } 208 204 209 - static inline void __write_lock(rwlock_t *lock) 205 + static inline void __raw_write_lock(rwlock_t *lock) 210 206 { 211 207 preempt_disable(); 212 208 rwlock_acquire(&lock->dep_map, 0, 0, _RET_IP_); ··· 215 211 216 212 #endif /* CONFIG_PREEMPT */ 217 213 218 - static inline void __write_unlock(rwlock_t *lock) 214 + static inline void __raw_write_unlock(rwlock_t *lock) 219 215 { 220 216 rwlock_release(&lock->dep_map, 1, _RET_IP_); 221 217 do_raw_write_unlock(lock); 222 218 preempt_enable(); 223 219 } 224 220 225 - static inline void __read_unlock(rwlock_t *lock) 221 + static inline void __raw_read_unlock(rwlock_t *lock) 226 222 { 227 223 rwlock_release(&lock->dep_map, 1, _RET_IP_); 228 224 do_raw_read_unlock(lock); 229 225 preempt_enable(); 230 226 } 231 227 232 - static inline void __read_unlock_irqrestore(rwlock_t *lock, unsigned long flags) 228 + static inline void 229 + __raw_read_unlock_irqrestore(rwlock_t *lock, unsigned long flags) 233 230 { 234 231 rwlock_release(&lock->dep_map, 1, _RET_IP_); 235 232 do_raw_read_unlock(lock); ··· 238 233 preempt_enable(); 239 234 } 240 235 241 - static inline void __read_unlock_irq(rwlock_t *lock) 236 + static inline void __raw_read_unlock_irq(rwlock_t *lock) 242 237 { 243 238 rwlock_release(&lock->dep_map, 1, _RET_IP_); 244 239 do_raw_read_unlock(lock); ··· 246 241 preempt_enable(); 247 242 } 248 243 249 - static inline void __read_unlock_bh(rwlock_t *lock) 244 + static inline void __raw_read_unlock_bh(rwlock_t *lock) 250 245 { 251 246 rwlock_release(&lock->dep_map, 1, _RET_IP_); 252 247 do_raw_read_unlock(lock); ··· 254 249 local_bh_enable_ip((unsigned long)__builtin_return_address(0)); 255 250 } 256 251 257 - static inline void __write_unlock_irqrestore(rwlock_t *lock, 252 + static inline void __raw_write_unlock_irqrestore(rwlock_t *lock, 258 253 unsigned long flags) 259 254 { 260 255 rwlock_release(&lock->dep_map, 1, _RET_IP_); ··· 263 258 preempt_enable(); 264 259 } 265 260 266 - static inline void __write_unlock_irq(rwlock_t *lock) 261 + static inline void __raw_write_unlock_irq(rwlock_t *lock) 267 262 { 268 263 rwlock_release(&lock->dep_map, 1, _RET_IP_); 269 264 do_raw_write_unlock(lock); ··· 271 266 preempt_enable(); 272 267 } 273 268 274 - static inline void __write_unlock_bh(rwlock_t *lock) 269 + static inline void __raw_write_unlock_bh(rwlock_t *lock) 275 270 { 276 271 rwlock_release(&lock->dep_map, 1, _RET_IP_); 277 272 do_raw_write_unlock(lock);
+20 -17
include/linux/spinlock.h
··· 161 161 * various methods are defined as nops in the case they are not 162 162 * required. 163 163 */ 164 - #define raw_spin_trylock(lock) __cond_lock(lock, _spin_trylock(lock)) 164 + #define raw_spin_trylock(lock) __cond_lock(lock, _raw_spin_trylock(lock)) 165 165 166 - #define raw_spin_lock(lock) _spin_lock(lock) 166 + #define raw_spin_lock(lock) _raw_spin_lock(lock) 167 167 168 168 #ifdef CONFIG_DEBUG_LOCK_ALLOC 169 - # define raw_spin_lock_nested(lock, subclass) _spin_lock_nested(lock, subclass) 169 + # define raw_spin_lock_nested(lock, subclass) \ 170 + _raw_spin_lock_nested(lock, subclass) 171 + 170 172 # define raw_spin_lock_nest_lock(lock, nest_lock) \ 171 173 do { \ 172 174 typecheck(struct lockdep_map *, &(nest_lock)->dep_map);\ 173 - _spin_lock_nest_lock(lock, &(nest_lock)->dep_map); \ 175 + _raw_spin_lock_nest_lock(lock, &(nest_lock)->dep_map); \ 174 176 } while (0) 175 177 #else 176 - # define raw_spin_lock_nested(lock, subclass) _spin_lock(lock) 177 - # define raw_spin_lock_nest_lock(lock, nest_lock) _spin_lock(lock) 178 + # define raw_spin_lock_nested(lock, subclass) _raw_spin_lock(lock) 179 + # define raw_spin_lock_nest_lock(lock, nest_lock) _raw_spin_lock(lock) 178 180 #endif 179 181 180 182 #if defined(CONFIG_SMP) || defined(CONFIG_DEBUG_SPINLOCK) ··· 184 182 #define raw_spin_lock_irqsave(lock, flags) \ 185 183 do { \ 186 184 typecheck(unsigned long, flags); \ 187 - flags = _spin_lock_irqsave(lock); \ 185 + flags = _raw_spin_lock_irqsave(lock); \ 188 186 } while (0) 189 187 190 188 #ifdef CONFIG_DEBUG_LOCK_ALLOC 191 189 #define raw_spin_lock_irqsave_nested(lock, flags, subclass) \ 192 190 do { \ 193 191 typecheck(unsigned long, flags); \ 194 - flags = _spin_lock_irqsave_nested(lock, subclass); \ 192 + flags = _raw_spin_lock_irqsave_nested(lock, subclass); \ 195 193 } while (0) 196 194 #else 197 195 #define raw_spin_lock_irqsave_nested(lock, flags, subclass) \ 198 196 do { \ 199 197 typecheck(unsigned long, flags); \ 200 - flags = _spin_lock_irqsave(lock); \ 198 + flags = _raw_spin_lock_irqsave(lock); \ 201 199 } while (0) 202 200 #endif 203 201 ··· 206 204 #define raw_spin_lock_irqsave(lock, flags) \ 207 205 do { \ 208 206 typecheck(unsigned long, flags); \ 209 - _spin_lock_irqsave(lock, flags); \ 207 + _raw_spin_lock_irqsave(lock, flags); \ 210 208 } while (0) 211 209 212 210 #define raw_spin_lock_irqsave_nested(lock, flags, subclass) \ ··· 214 212 215 213 #endif 216 214 217 - #define raw_spin_lock_irq(lock) _spin_lock_irq(lock) 218 - #define raw_spin_lock_bh(lock) _spin_lock_bh(lock) 219 - #define raw_spin_unlock(lock) _spin_unlock(lock) 220 - #define raw_spin_unlock_irq(lock) _spin_unlock_irq(lock) 215 + #define raw_spin_lock_irq(lock) _raw_spin_lock_irq(lock) 216 + #define raw_spin_lock_bh(lock) _raw_spin_lock_bh(lock) 217 + #define raw_spin_unlock(lock) _raw_spin_unlock(lock) 218 + #define raw_spin_unlock_irq(lock) _raw_spin_unlock_irq(lock) 221 219 222 220 #define raw_spin_unlock_irqrestore(lock, flags) \ 223 221 do { \ 224 222 typecheck(unsigned long, flags); \ 225 - _spin_unlock_irqrestore(lock, flags); \ 223 + _raw_spin_unlock_irqrestore(lock, flags); \ 226 224 } while (0) 227 - #define raw_spin_unlock_bh(lock) _spin_unlock_bh(lock) 225 + #define raw_spin_unlock_bh(lock) _raw_spin_unlock_bh(lock) 228 226 229 - #define raw_spin_trylock_bh(lock) __cond_lock(lock, _spin_trylock_bh(lock)) 227 + #define raw_spin_trylock_bh(lock) \ 228 + __cond_lock(lock, _raw_spin_trylock_bh(lock)) 230 229 231 230 #define raw_spin_trylock_irq(lock) \ 232 231 ({ \
+39 -38
include/linux/spinlock_api_smp.h
··· 19 19 20 20 #define assert_raw_spin_locked(x) BUG_ON(!raw_spin_is_locked(x)) 21 21 22 - void __lockfunc _spin_lock(raw_spinlock_t *lock) __acquires(lock); 23 - void __lockfunc _spin_lock_nested(raw_spinlock_t *lock, int subclass) 24 - __acquires(lock); 22 + void __lockfunc _raw_spin_lock(raw_spinlock_t *lock) __acquires(lock); 23 + void __lockfunc _raw_spin_lock_nested(raw_spinlock_t *lock, int subclass) 24 + __acquires(lock); 25 25 void __lockfunc 26 - _spin_lock_nest_lock(raw_spinlock_t *lock, struct lockdep_map *map) 27 - __acquires(lock); 28 - void __lockfunc _spin_lock_bh(raw_spinlock_t *lock) __acquires(lock); 29 - void __lockfunc _spin_lock_irq(raw_spinlock_t *lock) __acquires(lock); 26 + _raw_spin_lock_nest_lock(raw_spinlock_t *lock, struct lockdep_map *map) 27 + __acquires(lock); 28 + void __lockfunc _raw_spin_lock_bh(raw_spinlock_t *lock) __acquires(lock); 29 + void __lockfunc _raw_spin_lock_irq(raw_spinlock_t *lock) 30 + __acquires(lock); 30 31 31 - unsigned long __lockfunc _spin_lock_irqsave(raw_spinlock_t *lock) 32 - __acquires(lock); 32 + unsigned long __lockfunc _raw_spin_lock_irqsave(raw_spinlock_t *lock) 33 + __acquires(lock); 33 34 unsigned long __lockfunc 34 - _spin_lock_irqsave_nested(raw_spinlock_t *lock, int subclass) 35 - __acquires(lock); 36 - int __lockfunc _spin_trylock(raw_spinlock_t *lock); 37 - int __lockfunc _spin_trylock_bh(raw_spinlock_t *lock); 38 - void __lockfunc _spin_unlock(raw_spinlock_t *lock) __releases(lock); 39 - void __lockfunc _spin_unlock_bh(raw_spinlock_t *lock) __releases(lock); 40 - void __lockfunc _spin_unlock_irq(raw_spinlock_t *lock) __releases(lock); 35 + _raw_spin_lock_irqsave_nested(raw_spinlock_t *lock, int subclass) 36 + __acquires(lock); 37 + int __lockfunc _raw_spin_trylock(raw_spinlock_t *lock); 38 + int __lockfunc _raw_spin_trylock_bh(raw_spinlock_t *lock); 39 + void __lockfunc _raw_spin_unlock(raw_spinlock_t *lock) __releases(lock); 40 + void __lockfunc _raw_spin_unlock_bh(raw_spinlock_t *lock) __releases(lock); 41 + void __lockfunc _raw_spin_unlock_irq(raw_spinlock_t *lock) __releases(lock); 41 42 void __lockfunc 42 - _spin_unlock_irqrestore(raw_spinlock_t *lock, unsigned long flags) 43 - __releases(lock); 43 + _raw_spin_unlock_irqrestore(raw_spinlock_t *lock, unsigned long flags) 44 + __releases(lock); 44 45 45 46 #ifdef CONFIG_INLINE_SPIN_LOCK 46 - #define _spin_lock(lock) __spin_lock(lock) 47 + #define _raw_spin_lock(lock) __raw_spin_lock(lock) 47 48 #endif 48 49 49 50 #ifdef CONFIG_INLINE_SPIN_LOCK_BH 50 - #define _spin_lock_bh(lock) __spin_lock_bh(lock) 51 + #define _raw_spin_lock_bh(lock) __raw_spin_lock_bh(lock) 51 52 #endif 52 53 53 54 #ifdef CONFIG_INLINE_SPIN_LOCK_IRQ 54 - #define _spin_lock_irq(lock) __spin_lock_irq(lock) 55 + #define _raw_spin_lock_irq(lock) __raw_spin_lock_irq(lock) 55 56 #endif 56 57 57 58 #ifdef CONFIG_INLINE_SPIN_LOCK_IRQSAVE 58 - #define _spin_lock_irqsave(lock) __spin_lock_irqsave(lock) 59 + #define _raw_spin_lock_irqsave(lock) __raw_spin_lock_irqsave(lock) 59 60 #endif 60 61 61 62 #ifdef CONFIG_INLINE_SPIN_TRYLOCK 62 - #define _spin_trylock(lock) __spin_trylock(lock) 63 + #define _raw_spin_trylock(lock) __raw_spin_trylock(lock) 63 64 #endif 64 65 65 66 #ifdef CONFIG_INLINE_SPIN_TRYLOCK_BH 66 - #define _spin_trylock_bh(lock) __spin_trylock_bh(lock) 67 + #define _raw_spin_trylock_bh(lock) __raw_spin_trylock_bh(lock) 67 68 #endif 68 69 69 70 #ifdef CONFIG_INLINE_SPIN_UNLOCK 70 - #define _spin_unlock(lock) __spin_unlock(lock) 71 + #define _raw_spin_unlock(lock) __raw_spin_unlock(lock) 71 72 #endif 72 73 73 74 #ifdef CONFIG_INLINE_SPIN_UNLOCK_BH 74 - #define _spin_unlock_bh(lock) __spin_unlock_bh(lock) 75 + #define _raw_spin_unlock_bh(lock) __raw_spin_unlock_bh(lock) 75 76 #endif 76 77 77 78 #ifdef CONFIG_INLINE_SPIN_UNLOCK_IRQ 78 - #define _spin_unlock_irq(lock) __spin_unlock_irq(lock) 79 + #define _raw_spin_unlock_irq(lock) __raw_spin_unlock_irq(lock) 79 80 #endif 80 81 81 82 #ifdef CONFIG_INLINE_SPIN_UNLOCK_IRQRESTORE 82 - #define _spin_unlock_irqrestore(lock, flags) __spin_unlock_irqrestore(lock, flags) 83 + #define _raw_spin_unlock_irqrestore(lock, flags) __raw_spin_unlock_irqrestore(lock, flags) 83 84 #endif 84 85 85 - static inline int __spin_trylock(raw_spinlock_t *lock) 86 + static inline int __raw_spin_trylock(raw_spinlock_t *lock) 86 87 { 87 88 preempt_disable(); 88 89 if (do_raw_spin_trylock(lock)) { ··· 101 100 */ 102 101 #if !defined(CONFIG_GENERIC_LOCKBREAK) || defined(CONFIG_DEBUG_LOCK_ALLOC) 103 102 104 - static inline unsigned long __spin_lock_irqsave(raw_spinlock_t *lock) 103 + static inline unsigned long __raw_spin_lock_irqsave(raw_spinlock_t *lock) 105 104 { 106 105 unsigned long flags; 107 106 ··· 121 120 return flags; 122 121 } 123 122 124 - static inline void __spin_lock_irq(raw_spinlock_t *lock) 123 + static inline void __raw_spin_lock_irq(raw_spinlock_t *lock) 125 124 { 126 125 local_irq_disable(); 127 126 preempt_disable(); ··· 129 128 LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock); 130 129 } 131 130 132 - static inline void __spin_lock_bh(raw_spinlock_t *lock) 131 + static inline void __raw_spin_lock_bh(raw_spinlock_t *lock) 133 132 { 134 133 local_bh_disable(); 135 134 preempt_disable(); ··· 137 136 LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock); 138 137 } 139 138 140 - static inline void __spin_lock(raw_spinlock_t *lock) 139 + static inline void __raw_spin_lock(raw_spinlock_t *lock) 141 140 { 142 141 preempt_disable(); 143 142 spin_acquire(&lock->dep_map, 0, 0, _RET_IP_); ··· 146 145 147 146 #endif /* CONFIG_PREEMPT */ 148 147 149 - static inline void __spin_unlock(raw_spinlock_t *lock) 148 + static inline void __raw_spin_unlock(raw_spinlock_t *lock) 150 149 { 151 150 spin_release(&lock->dep_map, 1, _RET_IP_); 152 151 do_raw_spin_unlock(lock); 153 152 preempt_enable(); 154 153 } 155 154 156 - static inline void __spin_unlock_irqrestore(raw_spinlock_t *lock, 155 + static inline void __raw_spin_unlock_irqrestore(raw_spinlock_t *lock, 157 156 unsigned long flags) 158 157 { 159 158 spin_release(&lock->dep_map, 1, _RET_IP_); ··· 162 161 preempt_enable(); 163 162 } 164 163 165 - static inline void __spin_unlock_irq(raw_spinlock_t *lock) 164 + static inline void __raw_spin_unlock_irq(raw_spinlock_t *lock) 166 165 { 167 166 spin_release(&lock->dep_map, 1, _RET_IP_); 168 167 do_raw_spin_unlock(lock); ··· 170 169 preempt_enable(); 171 170 } 172 171 173 - static inline void __spin_unlock_bh(raw_spinlock_t *lock) 172 + static inline void __raw_spin_unlock_bh(raw_spinlock_t *lock) 174 173 { 175 174 spin_release(&lock->dep_map, 1, _RET_IP_); 176 175 do_raw_spin_unlock(lock); ··· 178 177 local_bh_enable_ip((unsigned long)__builtin_return_address(0)); 179 178 } 180 179 181 - static inline int __spin_trylock_bh(raw_spinlock_t *lock) 180 + static inline int __raw_spin_trylock_bh(raw_spinlock_t *lock) 182 181 { 183 182 local_bh_disable(); 184 183 preempt_disable();
+34 -30
include/linux/spinlock_api_up.h
··· 40 40 do { preempt_enable(); __release(lock); (void)(lock); } while (0) 41 41 42 42 #define __UNLOCK_BH(lock) \ 43 - do { preempt_enable_no_resched(); local_bh_enable(); __release(lock); (void)(lock); } while (0) 43 + do { preempt_enable_no_resched(); local_bh_enable(); \ 44 + __release(lock); (void)(lock); } while (0) 44 45 45 46 #define __UNLOCK_IRQ(lock) \ 46 47 do { local_irq_enable(); __UNLOCK(lock); } while (0) ··· 49 48 #define __UNLOCK_IRQRESTORE(lock, flags) \ 50 49 do { local_irq_restore(flags); __UNLOCK(lock); } while (0) 51 50 52 - #define _spin_lock(lock) __LOCK(lock) 53 - #define _spin_lock_nested(lock, subclass) __LOCK(lock) 54 - #define _read_lock(lock) __LOCK(lock) 55 - #define _write_lock(lock) __LOCK(lock) 56 - #define _spin_lock_bh(lock) __LOCK_BH(lock) 57 - #define _read_lock_bh(lock) __LOCK_BH(lock) 58 - #define _write_lock_bh(lock) __LOCK_BH(lock) 59 - #define _spin_lock_irq(lock) __LOCK_IRQ(lock) 60 - #define _read_lock_irq(lock) __LOCK_IRQ(lock) 61 - #define _write_lock_irq(lock) __LOCK_IRQ(lock) 62 - #define _spin_lock_irqsave(lock, flags) __LOCK_IRQSAVE(lock, flags) 63 - #define _read_lock_irqsave(lock, flags) __LOCK_IRQSAVE(lock, flags) 64 - #define _write_lock_irqsave(lock, flags) __LOCK_IRQSAVE(lock, flags) 65 - #define _spin_trylock(lock) ({ __LOCK(lock); 1; }) 66 - #define _read_trylock(lock) ({ __LOCK(lock); 1; }) 67 - #define _write_trylock(lock) ({ __LOCK(lock); 1; }) 68 - #define _spin_trylock_bh(lock) ({ __LOCK_BH(lock); 1; }) 69 - #define _spin_unlock(lock) __UNLOCK(lock) 70 - #define _read_unlock(lock) __UNLOCK(lock) 71 - #define _write_unlock(lock) __UNLOCK(lock) 72 - #define _spin_unlock_bh(lock) __UNLOCK_BH(lock) 73 - #define _write_unlock_bh(lock) __UNLOCK_BH(lock) 74 - #define _read_unlock_bh(lock) __UNLOCK_BH(lock) 75 - #define _spin_unlock_irq(lock) __UNLOCK_IRQ(lock) 76 - #define _read_unlock_irq(lock) __UNLOCK_IRQ(lock) 77 - #define _write_unlock_irq(lock) __UNLOCK_IRQ(lock) 78 - #define _spin_unlock_irqrestore(lock, flags) __UNLOCK_IRQRESTORE(lock, flags) 79 - #define _read_unlock_irqrestore(lock, flags) __UNLOCK_IRQRESTORE(lock, flags) 80 - #define _write_unlock_irqrestore(lock, flags) __UNLOCK_IRQRESTORE(lock, flags) 51 + #define _raw_spin_lock(lock) __LOCK(lock) 52 + #define _raw_spin_lock_nested(lock, subclass) __LOCK(lock) 53 + #define _raw_read_lock(lock) __LOCK(lock) 54 + #define _raw_write_lock(lock) __LOCK(lock) 55 + #define _raw_spin_lock_bh(lock) __LOCK_BH(lock) 56 + #define _raw_read_lock_bh(lock) __LOCK_BH(lock) 57 + #define _raw_write_lock_bh(lock) __LOCK_BH(lock) 58 + #define _raw_spin_lock_irq(lock) __LOCK_IRQ(lock) 59 + #define _raw_read_lock_irq(lock) __LOCK_IRQ(lock) 60 + #define _raw_write_lock_irq(lock) __LOCK_IRQ(lock) 61 + #define _raw_spin_lock_irqsave(lock, flags) __LOCK_IRQSAVE(lock, flags) 62 + #define _raw_read_lock_irqsave(lock, flags) __LOCK_IRQSAVE(lock, flags) 63 + #define _raw_write_lock_irqsave(lock, flags) __LOCK_IRQSAVE(lock, flags) 64 + #define _raw_spin_trylock(lock) ({ __LOCK(lock); 1; }) 65 + #define _raw_read_trylock(lock) ({ __LOCK(lock); 1; }) 66 + #define _raw_write_trylock(lock) ({ __LOCK(lock); 1; }) 67 + #define _raw_spin_trylock_bh(lock) ({ __LOCK_BH(lock); 1; }) 68 + #define _raw_spin_unlock(lock) __UNLOCK(lock) 69 + #define _raw_read_unlock(lock) __UNLOCK(lock) 70 + #define _raw_write_unlock(lock) __UNLOCK(lock) 71 + #define _raw_spin_unlock_bh(lock) __UNLOCK_BH(lock) 72 + #define _raw_write_unlock_bh(lock) __UNLOCK_BH(lock) 73 + #define _raw_read_unlock_bh(lock) __UNLOCK_BH(lock) 74 + #define _raw_spin_unlock_irq(lock) __UNLOCK_IRQ(lock) 75 + #define _raw_read_unlock_irq(lock) __UNLOCK_IRQ(lock) 76 + #define _raw_write_unlock_irq(lock) __UNLOCK_IRQ(lock) 77 + #define _raw_spin_unlock_irqrestore(lock, flags) \ 78 + __UNLOCK_IRQRESTORE(lock, flags) 79 + #define _raw_read_unlock_irqrestore(lock, flags) \ 80 + __UNLOCK_IRQRESTORE(lock, flags) 81 + #define _raw_write_unlock_irqrestore(lock, flags) \ 82 + __UNLOCK_IRQRESTORE(lock, flags) 81 83 82 84 #endif /* __LINUX_SPINLOCK_API_UP_H */
+96 -96
kernel/spinlock.c
··· 44 44 * towards that other CPU that it should break the lock ASAP. 45 45 */ 46 46 #define BUILD_LOCK_OPS(op, locktype) \ 47 - void __lockfunc __##op##_lock(locktype##_t *lock) \ 47 + void __lockfunc __raw_##op##_lock(locktype##_t *lock) \ 48 48 { \ 49 49 for (;;) { \ 50 50 preempt_disable(); \ ··· 60 60 (lock)->break_lock = 0; \ 61 61 } \ 62 62 \ 63 - unsigned long __lockfunc __##op##_lock_irqsave(locktype##_t *lock) \ 63 + unsigned long __lockfunc __raw_##op##_lock_irqsave(locktype##_t *lock) \ 64 64 { \ 65 65 unsigned long flags; \ 66 66 \ ··· 81 81 return flags; \ 82 82 } \ 83 83 \ 84 - void __lockfunc __##op##_lock_irq(locktype##_t *lock) \ 84 + void __lockfunc __raw_##op##_lock_irq(locktype##_t *lock) \ 85 85 { \ 86 - _##op##_lock_irqsave(lock); \ 86 + _raw_##op##_lock_irqsave(lock); \ 87 87 } \ 88 88 \ 89 - void __lockfunc __##op##_lock_bh(locktype##_t *lock) \ 89 + void __lockfunc __raw_##op##_lock_bh(locktype##_t *lock) \ 90 90 { \ 91 91 unsigned long flags; \ 92 92 \ ··· 95 95 /* irq-disabling. We use the generic preemption-aware */ \ 96 96 /* function: */ \ 97 97 /**/ \ 98 - flags = _##op##_lock_irqsave(lock); \ 98 + flags = _raw_##op##_lock_irqsave(lock); \ 99 99 local_bh_disable(); \ 100 100 local_irq_restore(flags); \ 101 101 } \ ··· 116 116 #endif 117 117 118 118 #ifndef CONFIG_INLINE_SPIN_TRYLOCK 119 - int __lockfunc _spin_trylock(raw_spinlock_t *lock) 119 + int __lockfunc _raw_spin_trylock(raw_spinlock_t *lock) 120 120 { 121 - return __spin_trylock(lock); 121 + return __raw_spin_trylock(lock); 122 122 } 123 - EXPORT_SYMBOL(_spin_trylock); 123 + EXPORT_SYMBOL(_raw_spin_trylock); 124 124 #endif 125 125 126 126 #ifndef CONFIG_INLINE_SPIN_TRYLOCK_BH 127 - int __lockfunc _spin_trylock_bh(raw_spinlock_t *lock) 127 + int __lockfunc _raw_spin_trylock_bh(raw_spinlock_t *lock) 128 128 { 129 - return __spin_trylock_bh(lock); 129 + return __raw_spin_trylock_bh(lock); 130 130 } 131 - EXPORT_SYMBOL(_spin_trylock_bh); 131 + EXPORT_SYMBOL(_raw_spin_trylock_bh); 132 132 #endif 133 133 134 134 #ifndef CONFIG_INLINE_SPIN_LOCK 135 - void __lockfunc _spin_lock(raw_spinlock_t *lock) 135 + void __lockfunc _raw_spin_lock(raw_spinlock_t *lock) 136 136 { 137 - __spin_lock(lock); 137 + __raw_spin_lock(lock); 138 138 } 139 - EXPORT_SYMBOL(_spin_lock); 139 + EXPORT_SYMBOL(_raw_spin_lock); 140 140 #endif 141 141 142 142 #ifndef CONFIG_INLINE_SPIN_LOCK_IRQSAVE 143 - unsigned long __lockfunc _spin_lock_irqsave(raw_spinlock_t *lock) 143 + unsigned long __lockfunc _raw_spin_lock_irqsave(raw_spinlock_t *lock) 144 144 { 145 - return __spin_lock_irqsave(lock); 145 + return __raw_spin_lock_irqsave(lock); 146 146 } 147 - EXPORT_SYMBOL(_spin_lock_irqsave); 147 + EXPORT_SYMBOL(_raw_spin_lock_irqsave); 148 148 #endif 149 149 150 150 #ifndef CONFIG_INLINE_SPIN_LOCK_IRQ 151 - void __lockfunc _spin_lock_irq(raw_spinlock_t *lock) 151 + void __lockfunc _raw_spin_lock_irq(raw_spinlock_t *lock) 152 152 { 153 - __spin_lock_irq(lock); 153 + __raw_spin_lock_irq(lock); 154 154 } 155 - EXPORT_SYMBOL(_spin_lock_irq); 155 + EXPORT_SYMBOL(_raw_spin_lock_irq); 156 156 #endif 157 157 158 158 #ifndef CONFIG_INLINE_SPIN_LOCK_BH 159 - void __lockfunc _spin_lock_bh(raw_spinlock_t *lock) 159 + void __lockfunc _raw_spin_lock_bh(raw_spinlock_t *lock) 160 160 { 161 - __spin_lock_bh(lock); 161 + __raw_spin_lock_bh(lock); 162 162 } 163 - EXPORT_SYMBOL(_spin_lock_bh); 163 + EXPORT_SYMBOL(_raw_spin_lock_bh); 164 164 #endif 165 165 166 166 #ifndef CONFIG_INLINE_SPIN_UNLOCK 167 - void __lockfunc _spin_unlock(raw_spinlock_t *lock) 167 + void __lockfunc _raw_spin_unlock(raw_spinlock_t *lock) 168 168 { 169 - __spin_unlock(lock); 169 + __raw_spin_unlock(lock); 170 170 } 171 - EXPORT_SYMBOL(_spin_unlock); 171 + EXPORT_SYMBOL(_raw_spin_unlock); 172 172 #endif 173 173 174 174 #ifndef CONFIG_INLINE_SPIN_UNLOCK_IRQRESTORE 175 - void __lockfunc _spin_unlock_irqrestore(raw_spinlock_t *lock, unsigned long flags) 175 + void __lockfunc _raw_spin_unlock_irqrestore(raw_spinlock_t *lock, unsigned long flags) 176 176 { 177 - __spin_unlock_irqrestore(lock, flags); 177 + __raw_spin_unlock_irqrestore(lock, flags); 178 178 } 179 - EXPORT_SYMBOL(_spin_unlock_irqrestore); 179 + EXPORT_SYMBOL(_raw_spin_unlock_irqrestore); 180 180 #endif 181 181 182 182 #ifndef CONFIG_INLINE_SPIN_UNLOCK_IRQ 183 - void __lockfunc _spin_unlock_irq(raw_spinlock_t *lock) 183 + void __lockfunc _raw_spin_unlock_irq(raw_spinlock_t *lock) 184 184 { 185 - __spin_unlock_irq(lock); 185 + __raw_spin_unlock_irq(lock); 186 186 } 187 - EXPORT_SYMBOL(_spin_unlock_irq); 187 + EXPORT_SYMBOL(_raw_spin_unlock_irq); 188 188 #endif 189 189 190 190 #ifndef CONFIG_INLINE_SPIN_UNLOCK_BH 191 - void __lockfunc _spin_unlock_bh(raw_spinlock_t *lock) 191 + void __lockfunc _raw_spin_unlock_bh(raw_spinlock_t *lock) 192 192 { 193 - __spin_unlock_bh(lock); 193 + __raw_spin_unlock_bh(lock); 194 194 } 195 - EXPORT_SYMBOL(_spin_unlock_bh); 195 + EXPORT_SYMBOL(_raw_spin_unlock_bh); 196 196 #endif 197 197 198 198 #ifndef CONFIG_INLINE_READ_TRYLOCK 199 - int __lockfunc _read_trylock(rwlock_t *lock) 199 + int __lockfunc _raw_read_trylock(rwlock_t *lock) 200 200 { 201 - return __read_trylock(lock); 201 + return __raw_read_trylock(lock); 202 202 } 203 - EXPORT_SYMBOL(_read_trylock); 203 + EXPORT_SYMBOL(_raw_read_trylock); 204 204 #endif 205 205 206 206 #ifndef CONFIG_INLINE_READ_LOCK 207 - void __lockfunc _read_lock(rwlock_t *lock) 207 + void __lockfunc _raw_read_lock(rwlock_t *lock) 208 208 { 209 - __read_lock(lock); 209 + __raw_read_lock(lock); 210 210 } 211 - EXPORT_SYMBOL(_read_lock); 211 + EXPORT_SYMBOL(_raw_read_lock); 212 212 #endif 213 213 214 214 #ifndef CONFIG_INLINE_READ_LOCK_IRQSAVE 215 - unsigned long __lockfunc _read_lock_irqsave(rwlock_t *lock) 215 + unsigned long __lockfunc _raw_read_lock_irqsave(rwlock_t *lock) 216 216 { 217 - return __read_lock_irqsave(lock); 217 + return __raw_read_lock_irqsave(lock); 218 218 } 219 - EXPORT_SYMBOL(_read_lock_irqsave); 219 + EXPORT_SYMBOL(_raw_read_lock_irqsave); 220 220 #endif 221 221 222 222 #ifndef CONFIG_INLINE_READ_LOCK_IRQ 223 - void __lockfunc _read_lock_irq(rwlock_t *lock) 223 + void __lockfunc _raw_read_lock_irq(rwlock_t *lock) 224 224 { 225 - __read_lock_irq(lock); 225 + __raw_read_lock_irq(lock); 226 226 } 227 - EXPORT_SYMBOL(_read_lock_irq); 227 + EXPORT_SYMBOL(_raw_read_lock_irq); 228 228 #endif 229 229 230 230 #ifndef CONFIG_INLINE_READ_LOCK_BH 231 - void __lockfunc _read_lock_bh(rwlock_t *lock) 231 + void __lockfunc _raw_read_lock_bh(rwlock_t *lock) 232 232 { 233 - __read_lock_bh(lock); 233 + __raw_read_lock_bh(lock); 234 234 } 235 - EXPORT_SYMBOL(_read_lock_bh); 235 + EXPORT_SYMBOL(_raw_read_lock_bh); 236 236 #endif 237 237 238 238 #ifndef CONFIG_INLINE_READ_UNLOCK 239 - void __lockfunc _read_unlock(rwlock_t *lock) 239 + void __lockfunc _raw_read_unlock(rwlock_t *lock) 240 240 { 241 - __read_unlock(lock); 241 + __raw_read_unlock(lock); 242 242 } 243 - EXPORT_SYMBOL(_read_unlock); 243 + EXPORT_SYMBOL(_raw_read_unlock); 244 244 #endif 245 245 246 246 #ifndef CONFIG_INLINE_READ_UNLOCK_IRQRESTORE 247 - void __lockfunc _read_unlock_irqrestore(rwlock_t *lock, unsigned long flags) 247 + void __lockfunc _raw_read_unlock_irqrestore(rwlock_t *lock, unsigned long flags) 248 248 { 249 - __read_unlock_irqrestore(lock, flags); 249 + __raw_read_unlock_irqrestore(lock, flags); 250 250 } 251 - EXPORT_SYMBOL(_read_unlock_irqrestore); 251 + EXPORT_SYMBOL(_raw_read_unlock_irqrestore); 252 252 #endif 253 253 254 254 #ifndef CONFIG_INLINE_READ_UNLOCK_IRQ 255 - void __lockfunc _read_unlock_irq(rwlock_t *lock) 255 + void __lockfunc _raw_read_unlock_irq(rwlock_t *lock) 256 256 { 257 - __read_unlock_irq(lock); 257 + __raw_read_unlock_irq(lock); 258 258 } 259 - EXPORT_SYMBOL(_read_unlock_irq); 259 + EXPORT_SYMBOL(_raw_read_unlock_irq); 260 260 #endif 261 261 262 262 #ifndef CONFIG_INLINE_READ_UNLOCK_BH 263 - void __lockfunc _read_unlock_bh(rwlock_t *lock) 263 + void __lockfunc _raw_read_unlock_bh(rwlock_t *lock) 264 264 { 265 - __read_unlock_bh(lock); 265 + __raw_read_unlock_bh(lock); 266 266 } 267 - EXPORT_SYMBOL(_read_unlock_bh); 267 + EXPORT_SYMBOL(_raw_read_unlock_bh); 268 268 #endif 269 269 270 270 #ifndef CONFIG_INLINE_WRITE_TRYLOCK 271 - int __lockfunc _write_trylock(rwlock_t *lock) 271 + int __lockfunc _raw_write_trylock(rwlock_t *lock) 272 272 { 273 - return __write_trylock(lock); 273 + return __raw_write_trylock(lock); 274 274 } 275 - EXPORT_SYMBOL(_write_trylock); 275 + EXPORT_SYMBOL(_raw_write_trylock); 276 276 #endif 277 277 278 278 #ifndef CONFIG_INLINE_WRITE_LOCK 279 - void __lockfunc _write_lock(rwlock_t *lock) 279 + void __lockfunc _raw_write_lock(rwlock_t *lock) 280 280 { 281 - __write_lock(lock); 281 + __raw_write_lock(lock); 282 282 } 283 - EXPORT_SYMBOL(_write_lock); 283 + EXPORT_SYMBOL(_raw_write_lock); 284 284 #endif 285 285 286 286 #ifndef CONFIG_INLINE_WRITE_LOCK_IRQSAVE 287 - unsigned long __lockfunc _write_lock_irqsave(rwlock_t *lock) 287 + unsigned long __lockfunc _raw_write_lock_irqsave(rwlock_t *lock) 288 288 { 289 - return __write_lock_irqsave(lock); 289 + return __raw_write_lock_irqsave(lock); 290 290 } 291 - EXPORT_SYMBOL(_write_lock_irqsave); 291 + EXPORT_SYMBOL(_raw_write_lock_irqsave); 292 292 #endif 293 293 294 294 #ifndef CONFIG_INLINE_WRITE_LOCK_IRQ 295 - void __lockfunc _write_lock_irq(rwlock_t *lock) 295 + void __lockfunc _raw_write_lock_irq(rwlock_t *lock) 296 296 { 297 - __write_lock_irq(lock); 297 + __raw_write_lock_irq(lock); 298 298 } 299 - EXPORT_SYMBOL(_write_lock_irq); 299 + EXPORT_SYMBOL(_raw_write_lock_irq); 300 300 #endif 301 301 302 302 #ifndef CONFIG_INLINE_WRITE_LOCK_BH 303 - void __lockfunc _write_lock_bh(rwlock_t *lock) 303 + void __lockfunc _raw_write_lock_bh(rwlock_t *lock) 304 304 { 305 - __write_lock_bh(lock); 305 + __raw_write_lock_bh(lock); 306 306 } 307 - EXPORT_SYMBOL(_write_lock_bh); 307 + EXPORT_SYMBOL(_raw_write_lock_bh); 308 308 #endif 309 309 310 310 #ifndef CONFIG_INLINE_WRITE_UNLOCK 311 - void __lockfunc _write_unlock(rwlock_t *lock) 311 + void __lockfunc _raw_write_unlock(rwlock_t *lock) 312 312 { 313 - __write_unlock(lock); 313 + __raw_write_unlock(lock); 314 314 } 315 - EXPORT_SYMBOL(_write_unlock); 315 + EXPORT_SYMBOL(_raw_write_unlock); 316 316 #endif 317 317 318 318 #ifndef CONFIG_INLINE_WRITE_UNLOCK_IRQRESTORE 319 - void __lockfunc _write_unlock_irqrestore(rwlock_t *lock, unsigned long flags) 319 + void __lockfunc _raw_write_unlock_irqrestore(rwlock_t *lock, unsigned long flags) 320 320 { 321 - __write_unlock_irqrestore(lock, flags); 321 + __raw_write_unlock_irqrestore(lock, flags); 322 322 } 323 - EXPORT_SYMBOL(_write_unlock_irqrestore); 323 + EXPORT_SYMBOL(_raw_write_unlock_irqrestore); 324 324 #endif 325 325 326 326 #ifndef CONFIG_INLINE_WRITE_UNLOCK_IRQ 327 - void __lockfunc _write_unlock_irq(rwlock_t *lock) 327 + void __lockfunc _raw_write_unlock_irq(rwlock_t *lock) 328 328 { 329 - __write_unlock_irq(lock); 329 + __raw_write_unlock_irq(lock); 330 330 } 331 - EXPORT_SYMBOL(_write_unlock_irq); 331 + EXPORT_SYMBOL(_raw_write_unlock_irq); 332 332 #endif 333 333 334 334 #ifndef CONFIG_INLINE_WRITE_UNLOCK_BH 335 - void __lockfunc _write_unlock_bh(rwlock_t *lock) 335 + void __lockfunc _raw_write_unlock_bh(rwlock_t *lock) 336 336 { 337 - __write_unlock_bh(lock); 337 + __raw_write_unlock_bh(lock); 338 338 } 339 - EXPORT_SYMBOL(_write_unlock_bh); 339 + EXPORT_SYMBOL(_raw_write_unlock_bh); 340 340 #endif 341 341 342 342 #ifdef CONFIG_DEBUG_LOCK_ALLOC 343 343 344 - void __lockfunc _spin_lock_nested(raw_spinlock_t *lock, int subclass) 344 + void __lockfunc _raw_spin_lock_nested(raw_spinlock_t *lock, int subclass) 345 345 { 346 346 preempt_disable(); 347 347 spin_acquire(&lock->dep_map, subclass, 0, _RET_IP_); 348 348 LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock); 349 349 } 350 - EXPORT_SYMBOL(_spin_lock_nested); 350 + EXPORT_SYMBOL(_raw_spin_lock_nested); 351 351 352 - unsigned long __lockfunc _spin_lock_irqsave_nested(raw_spinlock_t *lock, 352 + unsigned long __lockfunc _raw_spin_lock_irqsave_nested(raw_spinlock_t *lock, 353 353 int subclass) 354 354 { 355 355 unsigned long flags; ··· 361 361 do_raw_spin_lock_flags, &flags); 362 362 return flags; 363 363 } 364 - EXPORT_SYMBOL(_spin_lock_irqsave_nested); 364 + EXPORT_SYMBOL(_raw_spin_lock_irqsave_nested); 365 365 366 - void __lockfunc _spin_lock_nest_lock(raw_spinlock_t *lock, 366 + void __lockfunc _raw_spin_lock_nest_lock(raw_spinlock_t *lock, 367 367 struct lockdep_map *nest_lock) 368 368 { 369 369 preempt_disable(); 370 370 spin_acquire_nest(&lock->dep_map, 0, 0, nest_lock, _RET_IP_); 371 371 LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock); 372 372 } 373 - EXPORT_SYMBOL(_spin_lock_nest_lock); 373 + EXPORT_SYMBOL(_raw_spin_lock_nest_lock); 374 374 375 375 #endif 376 376