| #ifndef _M68K_SEMAPHORE_HELPER_H | 
 | #define _M68K_SEMAPHORE_HELPER_H | 
 |  | 
 | /* | 
 |  * SMP- and interrupt-safe semaphores helper functions. | 
 |  * | 
 |  * (C) Copyright 1996 Linus Torvalds | 
 |  * | 
 |  * m68k version by Andreas Schwab | 
 |  */ | 
 |  | 
 | #include <linux/errno.h> | 
 |  | 
 | /* | 
 |  * These two _must_ execute atomically wrt each other. | 
 |  */ | 
 | static inline void wake_one_more(struct semaphore * sem) | 
 | { | 
 | 	atomic_inc(&sem->waking); | 
 | } | 
 |  | 
 | #ifndef CONFIG_RMW_INSNS | 
 | extern spinlock_t semaphore_wake_lock; | 
 | #endif | 
 |  | 
 | static inline int waking_non_zero(struct semaphore *sem) | 
 | { | 
 | 	int ret; | 
 | #ifndef CONFIG_RMW_INSNS | 
 | 	unsigned long flags; | 
 |  | 
 | 	spin_lock_irqsave(&semaphore_wake_lock, flags); | 
 | 	ret = 0; | 
 | 	if (atomic_read(&sem->waking) > 0) { | 
 | 		atomic_dec(&sem->waking); | 
 | 		ret = 1; | 
 | 	} | 
 | 	spin_unlock_irqrestore(&semaphore_wake_lock, flags); | 
 | #else | 
 | 	int tmp1, tmp2; | 
 |  | 
 | 	__asm__ __volatile__ | 
 | 	  ("1:	movel	%1,%2\n" | 
 | 	   "    jle	2f\n" | 
 | 	   "	subql	#1,%2\n" | 
 | 	   "	casl	%1,%2,%3\n" | 
 | 	   "	jne	1b\n" | 
 | 	   "	moveq	#1,%0\n" | 
 | 	   "2:" | 
 | 	   : "=d" (ret), "=d" (tmp1), "=d" (tmp2) | 
 | 	   : "m" (sem->waking), "0" (0), "1" (sem->waking)); | 
 | #endif | 
 |  | 
 | 	return ret; | 
 | } | 
 |  | 
 | /* | 
 |  * waking_non_zero_interruptible: | 
 |  *	1	got the lock | 
 |  *	0	go to sleep | 
 |  *	-EINTR	interrupted | 
 |  */ | 
 | static inline int waking_non_zero_interruptible(struct semaphore *sem, | 
 | 						struct task_struct *tsk) | 
 | { | 
 | 	int ret; | 
 | #ifndef CONFIG_RMW_INSNS | 
 | 	unsigned long flags; | 
 |  | 
 | 	spin_lock_irqsave(&semaphore_wake_lock, flags); | 
 | 	ret = 0; | 
 | 	if (atomic_read(&sem->waking) > 0) { | 
 | 		atomic_dec(&sem->waking); | 
 | 		ret = 1; | 
 | 	} else if (signal_pending(tsk)) { | 
 | 		atomic_inc(&sem->count); | 
 | 		ret = -EINTR; | 
 | 	} | 
 | 	spin_unlock_irqrestore(&semaphore_wake_lock, flags); | 
 | #else | 
 | 	int tmp1, tmp2; | 
 |  | 
 | 	__asm__ __volatile__ | 
 | 	  ("1:	movel	%1,%2\n" | 
 | 	   "	jle	2f\n" | 
 | 	   "	subql	#1,%2\n" | 
 | 	   "	casl	%1,%2,%3\n" | 
 | 	   "	jne	1b\n" | 
 | 	   "	moveq	#1,%0\n" | 
 | 	   "	jra	%a4\n" | 
 | 	   "2:" | 
 | 	   : "=d" (ret), "=d" (tmp1), "=d" (tmp2) | 
 | 	   : "m" (sem->waking), "i" (&&next), "0" (0), "1" (sem->waking)); | 
 | 	if (signal_pending(tsk)) { | 
 | 		atomic_inc(&sem->count); | 
 | 		ret = -EINTR; | 
 | 	} | 
 | next: | 
 | #endif | 
 |  | 
 | 	return ret; | 
 | } | 
 |  | 
 | /* | 
 |  * waking_non_zero_trylock: | 
 |  *	1	failed to lock | 
 |  *	0	got the lock | 
 |  */ | 
 | static inline int waking_non_zero_trylock(struct semaphore *sem) | 
 | { | 
 | 	int ret; | 
 | #ifndef CONFIG_RMW_INSNS | 
 | 	unsigned long flags; | 
 |  | 
 | 	spin_lock_irqsave(&semaphore_wake_lock, flags); | 
 | 	ret = 1; | 
 | 	if (atomic_read(&sem->waking) > 0) { | 
 | 		atomic_dec(&sem->waking); | 
 | 		ret = 0; | 
 | 	} else | 
 | 		atomic_inc(&sem->count); | 
 | 	spin_unlock_irqrestore(&semaphore_wake_lock, flags); | 
 | #else | 
 | 	int tmp1, tmp2; | 
 |  | 
 | 	__asm__ __volatile__ | 
 | 	  ("1:	movel	%1,%2\n" | 
 | 	   "    jle	2f\n" | 
 | 	   "	subql	#1,%2\n" | 
 | 	   "	casl	%1,%2,%3\n" | 
 | 	   "	jne	1b\n" | 
 | 	   "	moveq	#0,%0\n" | 
 | 	   "2:" | 
 | 	   : "=d" (ret), "=d" (tmp1), "=d" (tmp2) | 
 | 	   : "m" (sem->waking), "0" (1), "1" (sem->waking)); | 
 | 	if (ret) | 
 | 		atomic_inc(&sem->count); | 
 | #endif | 
 | 	return ret; | 
 | } | 
 |  | 
 | #endif |