diff options
Diffstat (limited to 'arch/m32r/include/asm/spinlock.h')
| -rw-r--r-- | arch/m32r/include/asm/spinlock.h | 48 | 
1 files changed, 24 insertions, 24 deletions
| diff --git a/arch/m32r/include/asm/spinlock.h b/arch/m32r/include/asm/spinlock.h index dded923883b2..179a06489b10 100644 --- a/arch/m32r/include/asm/spinlock.h +++ b/arch/m32r/include/asm/spinlock.h @@ -24,19 +24,19 @@   * We make no fairness assumptions. They have a cost.   */ -#define __raw_spin_is_locked(x)		(*(volatile int *)(&(x)->slock) <= 0) -#define __raw_spin_lock_flags(lock, flags) __raw_spin_lock(lock) -#define __raw_spin_unlock_wait(x) \ -		do { cpu_relax(); } while (__raw_spin_is_locked(x)) +#define arch_spin_is_locked(x)		(*(volatile int *)(&(x)->slock) <= 0) +#define arch_spin_lock_flags(lock, flags) arch_spin_lock(lock) +#define arch_spin_unlock_wait(x) \ +		do { cpu_relax(); } while (arch_spin_is_locked(x))  /** - * __raw_spin_trylock - Try spin lock and return a result + * arch_spin_trylock - Try spin lock and return a result   * @lock: Pointer to the lock variable   * - * __raw_spin_trylock() tries to get the lock and returns a result. + * arch_spin_trylock() tries to get the lock and returns a result.   * On the m32r, the result value is 1 (= Success) or 0 (= Failure).   */ -static inline int __raw_spin_trylock(raw_spinlock_t *lock) +static inline int arch_spin_trylock(arch_spinlock_t *lock)  {  	int oldval;  	unsigned long tmp1, tmp2; @@ -50,7 +50,7 @@ static inline int __raw_spin_trylock(raw_spinlock_t *lock)  	 * }  	 */  	__asm__ __volatile__ ( -		"# __raw_spin_trylock		\n\t" +		"# arch_spin_trylock		\n\t"  		"ldi	%1, #0;			\n\t"  		"mvfc	%2, psw;		\n\t"  		"clrpsw	#0x40 -> nop;		\n\t" @@ -69,7 +69,7 @@ static inline int __raw_spin_trylock(raw_spinlock_t *lock)  	return (oldval > 0);  } -static inline void __raw_spin_lock(raw_spinlock_t *lock) +static inline void arch_spin_lock(arch_spinlock_t *lock)  {  	unsigned long tmp0, tmp1; @@ -84,7 +84,7 @@ static inline void __raw_spin_lock(raw_spinlock_t *lock)  	 * }  	 */  	__asm__ __volatile__ ( -		"# __raw_spin_lock		\n\t" +		"# arch_spin_lock		\n\t"  		".fillinsn			\n"  		"1:				\n\t"  		"mvfc	%1, psw;		\n\t" @@ -111,7 +111,7 @@ static inline void __raw_spin_lock(raw_spinlock_t *lock)  	);  } -static inline void __raw_spin_unlock(raw_spinlock_t *lock) +static inline void arch_spin_unlock(arch_spinlock_t *lock)  {  	mb();  	lock->slock = 1; @@ -140,15 +140,15 @@ static inline void __raw_spin_unlock(raw_spinlock_t *lock)   * read_can_lock - would read_trylock() succeed?   * @lock: the rwlock in question.   */ -#define __raw_read_can_lock(x) ((int)(x)->lock > 0) +#define arch_read_can_lock(x) ((int)(x)->lock > 0)  /**   * write_can_lock - would write_trylock() succeed?   * @lock: the rwlock in question.   */ -#define __raw_write_can_lock(x) ((x)->lock == RW_LOCK_BIAS) +#define arch_write_can_lock(x) ((x)->lock == RW_LOCK_BIAS) -static inline void __raw_read_lock(raw_rwlock_t *rw) +static inline void arch_read_lock(arch_rwlock_t *rw)  {  	unsigned long tmp0, tmp1; @@ -199,7 +199,7 @@ static inline void __raw_read_lock(raw_rwlock_t *rw)  	);  } -static inline void __raw_write_lock(raw_rwlock_t *rw) +static inline void arch_write_lock(arch_rwlock_t *rw)  {  	unsigned long tmp0, tmp1, tmp2; @@ -252,7 +252,7 @@ static inline void __raw_write_lock(raw_rwlock_t *rw)  	);  } -static inline void __raw_read_unlock(raw_rwlock_t *rw) +static inline void arch_read_unlock(arch_rwlock_t *rw)  {  	unsigned long tmp0, tmp1; @@ -274,7 +274,7 @@ static inline void __raw_read_unlock(raw_rwlock_t *rw)  	);  } -static inline void __raw_write_unlock(raw_rwlock_t *rw) +static inline void arch_write_unlock(arch_rwlock_t *rw)  {  	unsigned long tmp0, tmp1, tmp2; @@ -298,7 +298,7 @@ static inline void __raw_write_unlock(raw_rwlock_t *rw)  	);  } -static inline int __raw_read_trylock(raw_rwlock_t *lock) +static inline int arch_read_trylock(arch_rwlock_t *lock)  {  	atomic_t *count = (atomic_t*)lock;  	if (atomic_dec_return(count) >= 0) @@ -307,7 +307,7 @@ static inline int __raw_read_trylock(raw_rwlock_t *lock)  	return 0;  } -static inline int __raw_write_trylock(raw_rwlock_t *lock) +static inline int arch_write_trylock(arch_rwlock_t *lock)  {  	atomic_t *count = (atomic_t *)lock;  	if (atomic_sub_and_test(RW_LOCK_BIAS, count)) @@ -316,11 +316,11 @@ static inline int __raw_write_trylock(raw_rwlock_t *lock)  	return 0;  } -#define __raw_read_lock_flags(lock, flags) __raw_read_lock(lock) -#define __raw_write_lock_flags(lock, flags) __raw_write_lock(lock) +#define arch_read_lock_flags(lock, flags) arch_read_lock(lock) +#define arch_write_lock_flags(lock, flags) arch_write_lock(lock) -#define _raw_spin_relax(lock)	cpu_relax() -#define _raw_read_relax(lock)	cpu_relax() -#define _raw_write_relax(lock)	cpu_relax() +#define arch_spin_relax(lock)	cpu_relax() +#define arch_read_relax(lock)	cpu_relax() +#define arch_write_relax(lock)	cpu_relax()  #endif	/* _ASM_M32R_SPINLOCK_H */ | 
