Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

locking: Further name space cleanups

The name space hierarchy for the internal lock functions is now a bit
backwards. raw_spin* functions map to _spin* which use __spin*, while
we would like to have _raw_spin* and __raw_spin*.

_raw_spin* is already used by lock debugging, so rename those funtions
to do_raw_spin* to free up the _raw_spin* name space.

No functional change.

Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Acked-by: Peter Zijlstra <peterz@infradead.org>
Acked-by: Ingo Molnar <mingo@elte.hu>

+81 -81
+16 -16
include/linux/rwlock.h
··· 29 29 #endif 30 30 31 31 #ifdef CONFIG_DEBUG_SPINLOCK 32 - extern void _raw_read_lock(rwlock_t *lock); 33 - #define _raw_read_lock_flags(lock, flags) _raw_read_lock(lock) 34 - extern int _raw_read_trylock(rwlock_t *lock); 35 - extern void _raw_read_unlock(rwlock_t *lock); 36 - extern void _raw_write_lock(rwlock_t *lock); 37 - #define _raw_write_lock_flags(lock, flags) _raw_write_lock(lock) 38 - extern int _raw_write_trylock(rwlock_t *lock); 39 - extern void _raw_write_unlock(rwlock_t *lock); 32 + extern void do_raw_read_lock(rwlock_t *lock); 33 + #define do_raw_read_lock_flags(lock, flags) do_raw_read_lock(lock) 34 + extern int do_raw_read_trylock(rwlock_t *lock); 35 + extern void do_raw_read_unlock(rwlock_t *lock); 36 + extern void do_raw_write_lock(rwlock_t *lock); 37 + #define do_raw_write_lock_flags(lock, flags) do_raw_write_lock(lock) 38 + extern int do_raw_write_trylock(rwlock_t *lock); 39 + extern void do_raw_write_unlock(rwlock_t *lock); 40 40 #else 41 - # define _raw_read_lock(rwlock) arch_read_lock(&(rwlock)->raw_lock) 42 - # define _raw_read_lock_flags(lock, flags) \ 41 + # define do_raw_read_lock(rwlock) arch_read_lock(&(rwlock)->raw_lock) 42 + # define do_raw_read_lock_flags(lock, flags) \ 43 43 arch_read_lock_flags(&(lock)->raw_lock, *(flags)) 44 - # define _raw_read_trylock(rwlock) arch_read_trylock(&(rwlock)->raw_lock) 45 - # define _raw_read_unlock(rwlock) arch_read_unlock(&(rwlock)->raw_lock) 46 - # define _raw_write_lock(rwlock) arch_write_lock(&(rwlock)->raw_lock) 47 - # define _raw_write_lock_flags(lock, flags) \ 44 + # define do_raw_read_trylock(rwlock) arch_read_trylock(&(rwlock)->raw_lock) 45 + # define do_raw_read_unlock(rwlock) arch_read_unlock(&(rwlock)->raw_lock) 46 + # define do_raw_write_lock(rwlock) arch_write_lock(&(rwlock)->raw_lock) 47 + # define do_raw_write_lock_flags(lock, flags) \ 48 48 arch_write_lock_flags(&(lock)->raw_lock, *(flags)) 49 - # define _raw_write_trylock(rwlock) arch_write_trylock(&(rwlock)->raw_lock) 50 - # define _raw_write_unlock(rwlock) arch_write_unlock(&(rwlock)->raw_lock) 49 + # define do_raw_write_trylock(rwlock) arch_write_trylock(&(rwlock)->raw_lock) 50 + # define do_raw_write_unlock(rwlock) arch_write_unlock(&(rwlock)->raw_lock) 51 51 #endif 52 52 53 53 #define read_can_lock(rwlock) arch_read_can_lock(&(rwlock)->raw_lock)
+20 -20
include/linux/rwlock_api_smp.h
··· 113 113 static inline int __read_trylock(rwlock_t *lock) 114 114 { 115 115 preempt_disable(); 116 - if (_raw_read_trylock(lock)) { 116 + if (do_raw_read_trylock(lock)) { 117 117 rwlock_acquire_read(&lock->dep_map, 0, 1, _RET_IP_); 118 118 return 1; 119 119 } ··· 124 124 static inline int __write_trylock(rwlock_t *lock) 125 125 { 126 126 preempt_disable(); 127 - if (_raw_write_trylock(lock)) { 127 + if (do_raw_write_trylock(lock)) { 128 128 rwlock_acquire(&lock->dep_map, 0, 1, _RET_IP_); 129 129 return 1; 130 130 } ··· 143 143 { 144 144 preempt_disable(); 145 145 rwlock_acquire_read(&lock->dep_map, 0, 0, _RET_IP_); 146 - LOCK_CONTENDED(lock, _raw_read_trylock, _raw_read_lock); 146 + LOCK_CONTENDED(lock, do_raw_read_trylock, do_raw_read_lock); 147 147 } 148 148 149 149 static inline unsigned long __read_lock_irqsave(rwlock_t *lock) ··· 153 153 local_irq_save(flags); 154 154 preempt_disable(); 155 155 rwlock_acquire_read(&lock->dep_map, 0, 0, _RET_IP_); 156 - LOCK_CONTENDED_FLAGS(lock, _raw_read_trylock, _raw_read_lock, 157 - _raw_read_lock_flags, &flags); 156 + LOCK_CONTENDED_FLAGS(lock, do_raw_read_trylock, do_raw_read_lock, 157 + do_raw_read_lock_flags, &flags); 158 158 return flags; 159 159 } 160 160 ··· 163 163 local_irq_disable(); 164 164 preempt_disable(); 165 165 rwlock_acquire_read(&lock->dep_map, 0, 0, _RET_IP_); 166 - LOCK_CONTENDED(lock, _raw_read_trylock, _raw_read_lock); 166 + LOCK_CONTENDED(lock, do_raw_read_trylock, do_raw_read_lock); 167 167 } 168 168 169 169 static inline void __read_lock_bh(rwlock_t *lock) ··· 171 171 local_bh_disable(); 172 172 preempt_disable(); 173 173 rwlock_acquire_read(&lock->dep_map, 0, 0, _RET_IP_); 174 - LOCK_CONTENDED(lock, _raw_read_trylock, _raw_read_lock); 174 + LOCK_CONTENDED(lock, do_raw_read_trylock, do_raw_read_lock); 175 175 } 176 176 177 177 static inline unsigned long __write_lock_irqsave(rwlock_t *lock) ··· 181 181 local_irq_save(flags); 182 182 preempt_disable(); 183 183 rwlock_acquire(&lock->dep_map, 0, 0, _RET_IP_); 184 - LOCK_CONTENDED_FLAGS(lock, _raw_write_trylock, _raw_write_lock, 185 - _raw_write_lock_flags, &flags); 184 + LOCK_CONTENDED_FLAGS(lock, do_raw_write_trylock, do_raw_write_lock, 185 + do_raw_write_lock_flags, &flags); 186 186 return flags; 187 187 } 188 188 ··· 191 191 local_irq_disable(); 192 192 preempt_disable(); 193 193 rwlock_acquire(&lock->dep_map, 0, 0, _RET_IP_); 194 - LOCK_CONTENDED(lock, _raw_write_trylock, _raw_write_lock); 194 + LOCK_CONTENDED(lock, do_raw_write_trylock, do_raw_write_lock); 195 195 } 196 196 197 197 static inline void __write_lock_bh(rwlock_t *lock) ··· 199 199 local_bh_disable(); 200 200 preempt_disable(); 201 201 rwlock_acquire(&lock->dep_map, 0, 0, _RET_IP_); 202 - LOCK_CONTENDED(lock, _raw_write_trylock, _raw_write_lock); 202 + LOCK_CONTENDED(lock, do_raw_write_trylock, do_raw_write_lock); 203 203 } 204 204 205 205 static inline void __write_lock(rwlock_t *lock) 206 206 { 207 207 preempt_disable(); 208 208 rwlock_acquire(&lock->dep_map, 0, 0, _RET_IP_); 209 - LOCK_CONTENDED(lock, _raw_write_trylock, _raw_write_lock); 209 + LOCK_CONTENDED(lock, do_raw_write_trylock, do_raw_write_lock); 210 210 } 211 211 212 212 #endif /* CONFIG_PREEMPT */ ··· 214 214 static inline void __write_unlock(rwlock_t *lock) 215 215 { 216 216 rwlock_release(&lock->dep_map, 1, _RET_IP_); 217 - _raw_write_unlock(lock); 217 + do_raw_write_unlock(lock); 218 218 preempt_enable(); 219 219 } 220 220 221 221 static inline void __read_unlock(rwlock_t *lock) 222 222 { 223 223 rwlock_release(&lock->dep_map, 1, _RET_IP_); 224 - _raw_read_unlock(lock); 224 + do_raw_read_unlock(lock); 225 225 preempt_enable(); 226 226 } 227 227 228 228 static inline void __read_unlock_irqrestore(rwlock_t *lock, unsigned long flags) 229 229 { 230 230 rwlock_release(&lock->dep_map, 1, _RET_IP_); 231 - _raw_read_unlock(lock); 231 + do_raw_read_unlock(lock); 232 232 local_irq_restore(flags); 233 233 preempt_enable(); 234 234 } ··· 236 236 static inline void __read_unlock_irq(rwlock_t *lock) 237 237 { 238 238 rwlock_release(&lock->dep_map, 1, _RET_IP_); 239 - _raw_read_unlock(lock); 239 + do_raw_read_unlock(lock); 240 240 local_irq_enable(); 241 241 preempt_enable(); 242 242 } ··· 244 244 static inline void __read_unlock_bh(rwlock_t *lock) 245 245 { 246 246 rwlock_release(&lock->dep_map, 1, _RET_IP_); 247 - _raw_read_unlock(lock); 247 + do_raw_read_unlock(lock); 248 248 preempt_enable_no_resched(); 249 249 local_bh_enable_ip((unsigned long)__builtin_return_address(0)); 250 250 } ··· 253 253 unsigned long flags) 254 254 { 255 255 rwlock_release(&lock->dep_map, 1, _RET_IP_); 256 - _raw_write_unlock(lock); 256 + do_raw_write_unlock(lock); 257 257 local_irq_restore(flags); 258 258 preempt_enable(); 259 259 } ··· 261 261 static inline void __write_unlock_irq(rwlock_t *lock) 262 262 { 263 263 rwlock_release(&lock->dep_map, 1, _RET_IP_); 264 - _raw_write_unlock(lock); 264 + do_raw_write_unlock(lock); 265 265 local_irq_enable(); 266 266 preempt_enable(); 267 267 } ··· 269 269 static inline void __write_unlock_bh(rwlock_t *lock) 270 270 { 271 271 rwlock_release(&lock->dep_map, 1, _RET_IP_); 272 - _raw_write_unlock(lock); 272 + do_raw_write_unlock(lock); 273 273 preempt_enable_no_resched(); 274 274 local_bh_enable_ip((unsigned long)__builtin_return_address(0)); 275 275 }
+8 -8
include/linux/spinlock.h
··· 128 128 #define raw_spin_unlock_wait(lock) arch_spin_unlock_wait(&(lock)->raw_lock) 129 129 130 130 #ifdef CONFIG_DEBUG_SPINLOCK 131 - extern void _raw_spin_lock(raw_spinlock_t *lock); 132 - #define _raw_spin_lock_flags(lock, flags) _raw_spin_lock(lock) 133 - extern int _raw_spin_trylock(raw_spinlock_t *lock); 134 - extern void _raw_spin_unlock(raw_spinlock_t *lock); 131 + extern void do_raw_spin_lock(raw_spinlock_t *lock); 132 + #define do_raw_spin_lock_flags(lock, flags) do_raw_spin_lock(lock) 133 + extern int do_raw_spin_trylock(raw_spinlock_t *lock); 134 + extern void do_raw_spin_unlock(raw_spinlock_t *lock); 135 135 #else 136 - static inline void _raw_spin_lock(raw_spinlock_t *lock) 136 + static inline void do_raw_spin_lock(raw_spinlock_t *lock) 137 137 { 138 138 arch_spin_lock(&lock->raw_lock); 139 139 } 140 140 141 141 static inline void 142 - _raw_spin_lock_flags(raw_spinlock_t *lock, unsigned long *flags) 142 + do_raw_spin_lock_flags(raw_spinlock_t *lock, unsigned long *flags) 143 143 { 144 144 arch_spin_lock_flags(&lock->raw_lock, *flags); 145 145 } 146 146 147 - static inline int _raw_spin_trylock(raw_spinlock_t *lock) 147 + static inline int do_raw_spin_trylock(raw_spinlock_t *lock) 148 148 { 149 149 return arch_spin_trylock(&(lock)->raw_lock); 150 150 } 151 151 152 - static inline void _raw_spin_unlock(raw_spinlock_t *lock) 152 + static inline void do_raw_spin_unlock(raw_spinlock_t *lock) 153 153 { 154 154 arch_spin_unlock(&lock->raw_lock); 155 155 }
+12 -12
include/linux/spinlock_api_smp.h
··· 85 85 static inline int __spin_trylock(raw_spinlock_t *lock) 86 86 { 87 87 preempt_disable(); 88 - if (_raw_spin_trylock(lock)) { 88 + if (do_raw_spin_trylock(lock)) { 89 89 spin_acquire(&lock->dep_map, 0, 1, _RET_IP_); 90 90 return 1; 91 91 } ··· 109 109 spin_acquire(&lock->dep_map, 0, 0, _RET_IP_); 110 110 /* 111 111 * On lockdep we dont want the hand-coded irq-enable of 112 - * _raw_spin_lock_flags() code, because lockdep assumes 112 + * do_raw_spin_lock_flags() code, because lockdep assumes 113 113 * that interrupts are not re-enabled during lock-acquire: 114 114 */ 115 115 #ifdef CONFIG_LOCKDEP 116 - LOCK_CONTENDED(lock, _raw_spin_trylock, _raw_spin_lock); 116 + LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock); 117 117 #else 118 - _raw_spin_lock_flags(lock, &flags); 118 + do_raw_spin_lock_flags(lock, &flags); 119 119 #endif 120 120 return flags; 121 121 } ··· 125 125 local_irq_disable(); 126 126 preempt_disable(); 127 127 spin_acquire(&lock->dep_map, 0, 0, _RET_IP_); 128 - LOCK_CONTENDED(lock, _raw_spin_trylock, _raw_spin_lock); 128 + LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock); 129 129 } 130 130 131 131 static inline void __spin_lock_bh(raw_spinlock_t *lock) ··· 133 133 local_bh_disable(); 134 134 preempt_disable(); 135 135 spin_acquire(&lock->dep_map, 0, 0, _RET_IP_); 136 - LOCK_CONTENDED(lock, _raw_spin_trylock, _raw_spin_lock); 136 + LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock); 137 137 } 138 138 139 139 static inline void __spin_lock(raw_spinlock_t *lock) 140 140 { 141 141 preempt_disable(); 142 142 spin_acquire(&lock->dep_map, 0, 0, _RET_IP_); 143 - LOCK_CONTENDED(lock, _raw_spin_trylock, _raw_spin_lock); 143 + LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock); 144 144 } 145 145 146 146 #endif /* CONFIG_PREEMPT */ ··· 148 148 static inline void __spin_unlock(raw_spinlock_t *lock) 149 149 { 150 150 spin_release(&lock->dep_map, 1, _RET_IP_); 151 - _raw_spin_unlock(lock); 151 + do_raw_spin_unlock(lock); 152 152 preempt_enable(); 153 153 } 154 154 ··· 156 156 unsigned long flags) 157 157 { 158 158 spin_release(&lock->dep_map, 1, _RET_IP_); 159 - _raw_spin_unlock(lock); 159 + do_raw_spin_unlock(lock); 160 160 local_irq_restore(flags); 161 161 preempt_enable(); 162 162 } ··· 164 164 static inline void __spin_unlock_irq(raw_spinlock_t *lock) 165 165 { 166 166 spin_release(&lock->dep_map, 1, _RET_IP_); 167 - _raw_spin_unlock(lock); 167 + do_raw_spin_unlock(lock); 168 168 local_irq_enable(); 169 169 preempt_enable(); 170 170 } ··· 172 172 static inline void __spin_unlock_bh(raw_spinlock_t *lock) 173 173 { 174 174 spin_release(&lock->dep_map, 1, _RET_IP_); 175 - _raw_spin_unlock(lock); 175 + do_raw_spin_unlock(lock); 176 176 preempt_enable_no_resched(); 177 177 local_bh_enable_ip((unsigned long)__builtin_return_address(0)); 178 178 } ··· 181 181 { 182 182 local_bh_disable(); 183 183 preempt_disable(); 184 - if (_raw_spin_trylock(lock)) { 184 + if (do_raw_spin_trylock(lock)) { 185 185 spin_acquire(&lock->dep_map, 0, 1, _RET_IP_); 186 186 return 1; 187 187 }
+1 -1
kernel/sched.c
··· 6684 6684 */ 6685 6685 __release(rq->lock); 6686 6686 spin_release(&rq->lock.dep_map, 1, _THIS_IP_); 6687 - _raw_spin_unlock(&rq->lock); 6687 + do_raw_spin_unlock(&rq->lock); 6688 6688 preempt_enable_no_resched(); 6689 6689 6690 6690 schedule();
+6 -6
kernel/spinlock.c
··· 48 48 { \ 49 49 for (;;) { \ 50 50 preempt_disable(); \ 51 - if (likely(_raw_##op##_trylock(lock))) \ 51 + if (likely(do_raw_##op##_trylock(lock))) \ 52 52 break; \ 53 53 preempt_enable(); \ 54 54 \ ··· 67 67 for (;;) { \ 68 68 preempt_disable(); \ 69 69 local_irq_save(flags); \ 70 - if (likely(_raw_##op##_trylock(lock))) \ 70 + if (likely(do_raw_##op##_trylock(lock))) \ 71 71 break; \ 72 72 local_irq_restore(flags); \ 73 73 preempt_enable(); \ ··· 345 345 { 346 346 preempt_disable(); 347 347 spin_acquire(&lock->dep_map, subclass, 0, _RET_IP_); 348 - LOCK_CONTENDED(lock, _raw_spin_trylock, _raw_spin_lock); 348 + LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock); 349 349 } 350 350 EXPORT_SYMBOL(_spin_lock_nested); 351 351 ··· 357 357 local_irq_save(flags); 358 358 preempt_disable(); 359 359 spin_acquire(&lock->dep_map, subclass, 0, _RET_IP_); 360 - LOCK_CONTENDED_FLAGS(lock, _raw_spin_trylock, _raw_spin_lock, 361 - _raw_spin_lock_flags, &flags); 360 + LOCK_CONTENDED_FLAGS(lock, do_raw_spin_trylock, do_raw_spin_lock, 361 + do_raw_spin_lock_flags, &flags); 362 362 return flags; 363 363 } 364 364 EXPORT_SYMBOL(_spin_lock_irqsave_nested); ··· 368 368 { 369 369 preempt_disable(); 370 370 spin_acquire_nest(&lock->dep_map, 0, 0, nest_lock, _RET_IP_); 371 - LOCK_CONTENDED(lock, _raw_spin_trylock, _raw_spin_lock); 371 + LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock); 372 372 } 373 373 EXPORT_SYMBOL(_spin_lock_nest_lock); 374 374
+9 -9
lib/kernel_lock.c
··· 36 36 * If it successfully gets the lock, it should increment 37 37 * the preemption count like any spinlock does. 38 38 * 39 - * (This works on UP too - _raw_spin_trylock will never 39 + * (This works on UP too - do_raw_spin_trylock will never 40 40 * return false in that case) 41 41 */ 42 42 int __lockfunc __reacquire_kernel_lock(void) 43 43 { 44 - while (!_raw_spin_trylock(&kernel_flag)) { 44 + while (!do_raw_spin_trylock(&kernel_flag)) { 45 45 if (need_resched()) 46 46 return -EAGAIN; 47 47 cpu_relax(); ··· 52 52 53 53 void __lockfunc __release_kernel_lock(void) 54 54 { 55 - _raw_spin_unlock(&kernel_flag); 55 + do_raw_spin_unlock(&kernel_flag); 56 56 preempt_enable_no_resched(); 57 57 } 58 58 59 59 /* 60 60 * These are the BKL spinlocks - we try to be polite about preemption. 61 61 * If SMP is not on (ie UP preemption), this all goes away because the 62 - * _raw_spin_trylock() will always succeed. 62 + * do_raw_spin_trylock() will always succeed. 63 63 */ 64 64 #ifdef CONFIG_PREEMPT 65 65 static inline void __lock_kernel(void) 66 66 { 67 67 preempt_disable(); 68 - if (unlikely(!_raw_spin_trylock(&kernel_flag))) { 68 + if (unlikely(!do_raw_spin_trylock(&kernel_flag))) { 69 69 /* 70 70 * If preemption was disabled even before this 71 71 * was called, there's nothing we can be polite 72 72 * about - just spin. 73 73 */ 74 74 if (preempt_count() > 1) { 75 - _raw_spin_lock(&kernel_flag); 75 + do_raw_spin_lock(&kernel_flag); 76 76 return; 77 77 } 78 78 ··· 85 85 while (spin_is_locked(&kernel_flag)) 86 86 cpu_relax(); 87 87 preempt_disable(); 88 - } while (!_raw_spin_trylock(&kernel_flag)); 88 + } while (!do_raw_spin_trylock(&kernel_flag)); 89 89 } 90 90 } 91 91 ··· 96 96 */ 97 97 static inline void __lock_kernel(void) 98 98 { 99 - _raw_spin_lock(&kernel_flag); 99 + do_raw_spin_lock(&kernel_flag); 100 100 } 101 101 #endif 102 102 ··· 106 106 * the BKL is not covered by lockdep, so we open-code the 107 107 * unlocking sequence (and thus avoid the dep-chain ops): 108 108 */ 109 - _raw_spin_unlock(&kernel_flag); 109 + do_raw_spin_unlock(&kernel_flag); 110 110 preempt_enable(); 111 111 } 112 112
+9 -9
lib/spinlock_debug.c
··· 125 125 } 126 126 } 127 127 128 - void _raw_spin_lock(raw_spinlock_t *lock) 128 + void do_raw_spin_lock(raw_spinlock_t *lock) 129 129 { 130 130 debug_spin_lock_before(lock); 131 131 if (unlikely(!arch_spin_trylock(&lock->raw_lock))) ··· 133 133 debug_spin_lock_after(lock); 134 134 } 135 135 136 - int _raw_spin_trylock(raw_spinlock_t *lock) 136 + int do_raw_spin_trylock(raw_spinlock_t *lock) 137 137 { 138 138 int ret = arch_spin_trylock(&lock->raw_lock); 139 139 ··· 148 148 return ret; 149 149 } 150 150 151 - void _raw_spin_unlock(raw_spinlock_t *lock) 151 + void do_raw_spin_unlock(raw_spinlock_t *lock) 152 152 { 153 153 debug_spin_unlock(lock); 154 154 arch_spin_unlock(&lock->raw_lock); ··· 193 193 } 194 194 #endif 195 195 196 - void _raw_read_lock(rwlock_t *lock) 196 + void do_raw_read_lock(rwlock_t *lock) 197 197 { 198 198 RWLOCK_BUG_ON(lock->magic != RWLOCK_MAGIC, lock, "bad magic"); 199 199 arch_read_lock(&lock->raw_lock); 200 200 } 201 201 202 - int _raw_read_trylock(rwlock_t *lock) 202 + int do_raw_read_trylock(rwlock_t *lock) 203 203 { 204 204 int ret = arch_read_trylock(&lock->raw_lock); 205 205 ··· 212 212 return ret; 213 213 } 214 214 215 - void _raw_read_unlock(rwlock_t *lock) 215 + void do_raw_read_unlock(rwlock_t *lock) 216 216 { 217 217 RWLOCK_BUG_ON(lock->magic != RWLOCK_MAGIC, lock, "bad magic"); 218 218 arch_read_unlock(&lock->raw_lock); ··· 268 268 } 269 269 #endif 270 270 271 - void _raw_write_lock(rwlock_t *lock) 271 + void do_raw_write_lock(rwlock_t *lock) 272 272 { 273 273 debug_write_lock_before(lock); 274 274 arch_write_lock(&lock->raw_lock); 275 275 debug_write_lock_after(lock); 276 276 } 277 277 278 - int _raw_write_trylock(rwlock_t *lock) 278 + int do_raw_write_trylock(rwlock_t *lock) 279 279 { 280 280 int ret = arch_write_trylock(&lock->raw_lock); 281 281 ··· 290 290 return ret; 291 291 } 292 292 293 - void _raw_write_unlock(rwlock_t *lock) 293 + void do_raw_write_unlock(rwlock_t *lock) 294 294 { 295 295 debug_write_unlock(lock); 296 296 arch_write_unlock(&lock->raw_lock);