| /* | 
 |  * MIPS-specific semaphore code. | 
 |  * | 
 |  * Copyright (C) 1999 Cort Dougan <cort@cs.nmt.edu> | 
 |  * Copyright (C) 2004 Ralf Baechle <ralf@linux-mips.org> | 
 |  * | 
 |  * This program is free software; you can redistribute it and/or | 
 |  * modify it under the terms of the GNU General Public License | 
 |  * as published by the Free Software Foundation; either version | 
 |  * 2 of the License, or (at your option) any later version. | 
 |  * | 
 |  * April 2001 - Reworked by Paul Mackerras <paulus@samba.org> | 
 |  * to eliminate the SMP races in the old version between the updates | 
 |  * of `count' and `waking'.  Now we use negative `count' values to | 
 |  * indicate that some process(es) are waiting for the semaphore. | 
 |  */ | 
 |  | 
 | #include <linux/module.h> | 
 | #include <linux/sched.h> | 
 | #include <linux/init.h> | 
 | #include <asm/atomic.h> | 
 | #include <asm/cpu-features.h> | 
 | #include <asm/errno.h> | 
 | #include <asm/semaphore.h> | 
 | #include <asm/war.h> | 
 | /* | 
 |  * Atomically update sem->count. | 
 |  * This does the equivalent of the following: | 
 |  * | 
 |  *	old_count = sem->count; | 
 |  *	tmp = MAX(old_count, 0) + incr; | 
 |  *	sem->count = tmp; | 
 |  *	return old_count; | 
 |  * | 
 |  * On machines without lld/scd we need a spinlock to make the manipulation of | 
 |  * sem->count and sem->waking atomic.  Scalability isn't an issue because | 
 |  * this lock is used on UP only so it's just an empty variable. | 
 |  */ | 
 | static inline int __sem_update_count(struct semaphore *sem, int incr) | 
 | { | 
 | 	int old_count, tmp; | 
 |  | 
 | 	if (cpu_has_llsc && R10000_LLSC_WAR) { | 
 | 		__asm__ __volatile__( | 
 | 		"	.set	mips3					\n" | 
 | 		"1:	ll	%0, %2		# __sem_update_count	\n" | 
 | 		"	sra	%1, %0, 31				\n" | 
 | 		"	not	%1					\n" | 
 | 		"	and	%1, %0, %1				\n" | 
 | 		"	addu	%1, %1, %3				\n" | 
 | 		"	sc	%1, %2					\n" | 
 | 		"	beqzl	%1, 1b					\n" | 
 | 		"	.set	mips0					\n" | 
 | 		: "=&r" (old_count), "=&r" (tmp), "=m" (sem->count) | 
 | 		: "r" (incr), "m" (sem->count)); | 
 | 	} else if (cpu_has_llsc) { | 
 | 		__asm__ __volatile__( | 
 | 		"	.set	mips3					\n" | 
 | 		"1:	ll	%0, %2		# __sem_update_count	\n" | 
 | 		"	sra	%1, %0, 31				\n" | 
 | 		"	not	%1					\n" | 
 | 		"	and	%1, %0, %1				\n" | 
 | 		"	addu	%1, %1, %3				\n" | 
 | 		"	sc	%1, %2					\n" | 
 | 		"	beqz	%1, 1b					\n" | 
 | 		"	.set	mips0					\n" | 
 | 		: "=&r" (old_count), "=&r" (tmp), "=m" (sem->count) | 
 | 		: "r" (incr), "m" (sem->count)); | 
 | 	} else { | 
 | 		static DEFINE_SPINLOCK(semaphore_lock); | 
 | 		unsigned long flags; | 
 |  | 
 | 		spin_lock_irqsave(&semaphore_lock, flags); | 
 | 		old_count = atomic_read(&sem->count); | 
 | 		tmp = max_t(int, old_count, 0) + incr; | 
 | 		atomic_set(&sem->count, tmp); | 
 | 		spin_unlock_irqrestore(&semaphore_lock, flags); | 
 | 	} | 
 |  | 
 | 	return old_count; | 
 | } | 
 |  | 
 | void __up(struct semaphore *sem) | 
 | { | 
 | 	/* | 
 | 	 * Note that we incremented count in up() before we came here, | 
 | 	 * but that was ineffective since the result was <= 0, and | 
 | 	 * any negative value of count is equivalent to 0. | 
 | 	 * This ends up setting count to 1, unless count is now > 0 | 
 | 	 * (i.e. because some other cpu has called up() in the meantime), | 
 | 	 * in which case we just increment count. | 
 | 	 */ | 
 | 	__sem_update_count(sem, 1); | 
 | 	wake_up(&sem->wait); | 
 | } | 
 |  | 
 | EXPORT_SYMBOL(__up); | 
 |  | 
 | /* | 
 |  * Note that when we come in to __down or __down_interruptible, | 
 |  * we have already decremented count, but that decrement was | 
 |  * ineffective since the result was < 0, and any negative value | 
 |  * of count is equivalent to 0. | 
 |  * Thus it is only when we decrement count from some value > 0 | 
 |  * that we have actually got the semaphore. | 
 |  */ | 
 | void __sched __down(struct semaphore *sem) | 
 | { | 
 | 	struct task_struct *tsk = current; | 
 | 	DECLARE_WAITQUEUE(wait, tsk); | 
 |  | 
 | 	__set_task_state(tsk, TASK_UNINTERRUPTIBLE); | 
 | 	add_wait_queue_exclusive(&sem->wait, &wait); | 
 |  | 
 | 	/* | 
 | 	 * Try to get the semaphore.  If the count is > 0, then we've | 
 | 	 * got the semaphore; we decrement count and exit the loop. | 
 | 	 * If the count is 0 or negative, we set it to -1, indicating | 
 | 	 * that we are asleep, and then sleep. | 
 | 	 */ | 
 | 	while (__sem_update_count(sem, -1) <= 0) { | 
 | 		schedule(); | 
 | 		set_task_state(tsk, TASK_UNINTERRUPTIBLE); | 
 | 	} | 
 | 	remove_wait_queue(&sem->wait, &wait); | 
 | 	__set_task_state(tsk, TASK_RUNNING); | 
 |  | 
 | 	/* | 
 | 	 * If there are any more sleepers, wake one of them up so | 
 | 	 * that it can either get the semaphore, or set count to -1 | 
 | 	 * indicating that there are still processes sleeping. | 
 | 	 */ | 
 | 	wake_up(&sem->wait); | 
 | } | 
 |  | 
 | EXPORT_SYMBOL(__down); | 
 |  | 
 | int __sched __down_interruptible(struct semaphore * sem) | 
 | { | 
 | 	int retval = 0; | 
 | 	struct task_struct *tsk = current; | 
 | 	DECLARE_WAITQUEUE(wait, tsk); | 
 |  | 
 | 	__set_task_state(tsk, TASK_INTERRUPTIBLE); | 
 | 	add_wait_queue_exclusive(&sem->wait, &wait); | 
 |  | 
 | 	while (__sem_update_count(sem, -1) <= 0) { | 
 | 		if (signal_pending(current)) { | 
 | 			/* | 
 | 			 * A signal is pending - give up trying. | 
 | 			 * Set sem->count to 0 if it is negative, | 
 | 			 * since we are no longer sleeping. | 
 | 			 */ | 
 | 			__sem_update_count(sem, 0); | 
 | 			retval = -EINTR; | 
 | 			break; | 
 | 		} | 
 | 		schedule(); | 
 | 		set_task_state(tsk, TASK_INTERRUPTIBLE); | 
 | 	} | 
 | 	remove_wait_queue(&sem->wait, &wait); | 
 | 	__set_task_state(tsk, TASK_RUNNING); | 
 |  | 
 | 	wake_up(&sem->wait); | 
 | 	return retval; | 
 | } | 
 |  | 
 | EXPORT_SYMBOL(__down_interruptible); |