Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
at v2.6.16-rc4 147 lines 3.1 kB view raw
1#ifndef __ASM_SPINLOCK_H 2#define __ASM_SPINLOCK_H 3 4#include <asm/system.h> 5#include <asm/processor.h> 6#include <asm/spinlock_types.h> 7 8static inline int __raw_spin_is_locked(raw_spinlock_t *x) 9{ 10 volatile unsigned int *a = __ldcw_align(x); 11 return *a == 0; 12} 13 14#define __raw_spin_lock(lock) __raw_spin_lock_flags(lock, 0) 15#define __raw_spin_unlock_wait(x) \ 16 do { cpu_relax(); } while (__raw_spin_is_locked(x)) 17 18static inline void __raw_spin_lock_flags(raw_spinlock_t *x, 19 unsigned long flags) 20{ 21 volatile unsigned int *a; 22 23 mb(); 24 a = __ldcw_align(x); 25 while (__ldcw(a) == 0) 26 while (*a == 0) 27 if (flags & PSW_SM_I) { 28 local_irq_enable(); 29 cpu_relax(); 30 local_irq_disable(); 31 } else 32 cpu_relax(); 33 mb(); 34} 35 36static inline void __raw_spin_unlock(raw_spinlock_t *x) 37{ 38 volatile unsigned int *a; 39 mb(); 40 a = __ldcw_align(x); 41 *a = 1; 42 mb(); 43} 44 45static inline int __raw_spin_trylock(raw_spinlock_t *x) 46{ 47 volatile unsigned int *a; 48 int ret; 49 50 mb(); 51 a = __ldcw_align(x); 52 ret = __ldcw(a) != 0; 53 mb(); 54 55 return ret; 56} 57 58/* 59 * Read-write spinlocks, allowing multiple readers 60 * but only one writer. 61 */ 62 63#define __raw_read_trylock(lock) generic__raw_read_trylock(lock) 64 65/* read_lock, read_unlock are pretty straightforward. Of course it somehow 66 * sucks we end up saving/restoring flags twice for read_lock_irqsave aso. */ 67 68static __inline__ void __raw_read_lock(raw_rwlock_t *rw) 69{ 70 __raw_spin_lock(&rw->lock); 71 72 rw->counter++; 73 74 __raw_spin_unlock(&rw->lock); 75} 76 77static __inline__ void __raw_read_unlock(raw_rwlock_t *rw) 78{ 79 __raw_spin_lock(&rw->lock); 80 81 rw->counter--; 82 83 __raw_spin_unlock(&rw->lock); 84} 85 86/* write_lock is less trivial. We optimistically grab the lock and check 87 * if we surprised any readers. If so we release the lock and wait till 88 * they're all gone before trying again 89 * 90 * Also note that we don't use the _irqsave / _irqrestore suffixes here. 91 * If we're called with interrupts enabled and we've got readers (or other 92 * writers) in interrupt handlers someone fucked up and we'd dead-lock 93 * sooner or later anyway. prumpf */ 94 95static __inline__ void __raw_write_lock(raw_rwlock_t *rw) 96{ 97retry: 98 __raw_spin_lock(&rw->lock); 99 100 if(rw->counter != 0) { 101 /* this basically never happens */ 102 __raw_spin_unlock(&rw->lock); 103 104 while (rw->counter != 0) 105 cpu_relax(); 106 107 goto retry; 108 } 109 110 /* got it. now leave without unlocking */ 111 rw->counter = -1; /* remember we are locked */ 112} 113 114/* write_unlock is absolutely trivial - we don't have to wait for anything */ 115 116static __inline__ void __raw_write_unlock(raw_rwlock_t *rw) 117{ 118 rw->counter = 0; 119 __raw_spin_unlock(&rw->lock); 120} 121 122static __inline__ int __raw_write_trylock(raw_rwlock_t *rw) 123{ 124 __raw_spin_lock(&rw->lock); 125 if (rw->counter != 0) { 126 /* this basically never happens */ 127 __raw_spin_unlock(&rw->lock); 128 129 return 0; 130 } 131 132 /* got it. now leave without unlocking */ 133 rw->counter = -1; /* remember we are locked */ 134 return 1; 135} 136 137static __inline__ int __raw_is_read_locked(raw_rwlock_t *rw) 138{ 139 return rw->counter > 0; 140} 141 142static __inline__ int __raw_is_write_locked(raw_rwlock_t *rw) 143{ 144 return rw->counter < 0; 145} 146 147#endif /* __ASM_SPINLOCK_H */