| #ifndef _ASM_POWERPC_SEMAPHORE_H | 
 | #define _ASM_POWERPC_SEMAPHORE_H | 
 |  | 
 | /* | 
 |  * Remove spinlock-based RW semaphores; RW semaphore definitions are | 
 |  * now in rwsem.h and we use the generic lib/rwsem.c implementation. | 
 |  * Rework semaphores to use atomic_dec_if_positive. | 
 |  * -- Paul Mackerras (paulus@samba.org) | 
 |  */ | 
 |  | 
 | #ifdef __KERNEL__ | 
 |  | 
 | #include <asm/atomic.h> | 
 | #include <asm/system.h> | 
 | #include <linux/wait.h> | 
 | #include <linux/rwsem.h> | 
 |  | 
 | struct semaphore { | 
 | 	/* | 
 | 	 * Note that any negative value of count is equivalent to 0, | 
 | 	 * but additionally indicates that some process(es) might be | 
 | 	 * sleeping on `wait'. | 
 | 	 */ | 
 | 	atomic_t count; | 
 | 	wait_queue_head_t wait; | 
 | }; | 
 |  | 
 | #define __SEMAPHORE_INITIALIZER(name, n)				\ | 
 | {									\ | 
 | 	.count		= ATOMIC_INIT(n),				\ | 
 | 	.wait		= __WAIT_QUEUE_HEAD_INITIALIZER((name).wait)	\ | 
 | } | 
 |  | 
 | #define __DECLARE_SEMAPHORE_GENERIC(name, count) \ | 
 | 	struct semaphore name = __SEMAPHORE_INITIALIZER(name,count) | 
 |  | 
 | #define DECLARE_MUTEX(name)		__DECLARE_SEMAPHORE_GENERIC(name, 1) | 
 | #define DECLARE_MUTEX_LOCKED(name)	__DECLARE_SEMAPHORE_GENERIC(name, 0) | 
 |  | 
 | static inline void sema_init (struct semaphore *sem, int val) | 
 | { | 
 | 	atomic_set(&sem->count, val); | 
 | 	init_waitqueue_head(&sem->wait); | 
 | } | 
 |  | 
 | static inline void init_MUTEX (struct semaphore *sem) | 
 | { | 
 | 	sema_init(sem, 1); | 
 | } | 
 |  | 
 | static inline void init_MUTEX_LOCKED (struct semaphore *sem) | 
 | { | 
 | 	sema_init(sem, 0); | 
 | } | 
 |  | 
 | extern void __down(struct semaphore * sem); | 
 | extern int  __down_interruptible(struct semaphore * sem); | 
 | extern void __up(struct semaphore * sem); | 
 |  | 
 | static inline void down(struct semaphore * sem) | 
 | { | 
 | 	might_sleep(); | 
 |  | 
 | 	/* | 
 | 	 * Try to get the semaphore, take the slow path if we fail. | 
 | 	 */ | 
 | 	if (unlikely(atomic_dec_return(&sem->count) < 0)) | 
 | 		__down(sem); | 
 | } | 
 |  | 
 | static inline int down_interruptible(struct semaphore * sem) | 
 | { | 
 | 	int ret = 0; | 
 |  | 
 | 	might_sleep(); | 
 |  | 
 | 	if (unlikely(atomic_dec_return(&sem->count) < 0)) | 
 | 		ret = __down_interruptible(sem); | 
 | 	return ret; | 
 | } | 
 |  | 
 | static inline int down_trylock(struct semaphore * sem) | 
 | { | 
 | 	return atomic_dec_if_positive(&sem->count) < 0; | 
 | } | 
 |  | 
 | static inline void up(struct semaphore * sem) | 
 | { | 
 | 	if (unlikely(atomic_inc_return(&sem->count) <= 0)) | 
 | 		__up(sem); | 
 | } | 
 |  | 
 | #endif /* __KERNEL__ */ | 
 |  | 
 | #endif /* _ASM_POWERPC_SEMAPHORE_H */ |