/* * MIPS-specific semaphore code. * * Copyright (C) 1999 Cort Dougan * Copyright (C) 2004 Ralf Baechle * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public License * as published by the Free Software Foundation; either version * 2 of the License, or (at your option) any later version. * * April 2001 - Reworked by Paul Mackerras * to eliminate the SMP races in the old version between the updates * of `count' and `waking'. Now we use negative `count' values to * indicate that some process(es) are waiting for the semaphore. */ #include #include #include #include #include #include #include #include /* * Atomically update sem->count. * This does the equivalent of the following: * * old_count = sem->count; * tmp = MAX(old_count, 0) + incr; * sem->count = tmp; * return old_count; * * On machines without lld/scd we need a spinlock to make the manipulation of * sem->count and sem->waking atomic. Scalability isn't an issue because * this lock is used on UP only so it's just an empty variable. */ static inline int __sem_update_count(struct semaphore *sem, int incr) { int old_count, tmp; if (cpu_has_llsc && R10000_LLSC_WAR) { __asm__ __volatile__( " .set mips3 \n" "1: ll %0, %2 # __sem_update_count \n" " sra %1, %0, 31 \n" " not %1 \n" " and %1, %0, %1 \n" " addu %1, %1, %3 \n" " sc %1, %2 \n" " beqzl %1, 1b \n" " .set mips0 \n" : "=&r" (old_count), "=&r" (tmp), "=m" (sem->count) : "r" (incr), "m" (sem->count)); } else if (cpu_has_llsc) { __asm__ __volatile__( " .set mips3 \n" "1: ll %0, %2 # __sem_update_count \n" " sra %1, %0, 31 \n" " not %1 \n" " and %1, %0, %1 \n" " addu %1, %1, %3 \n" " sc %1, %2 \n" " beqz %1, 1b \n" " .set mips0 \n" : "=&r" (old_count), "=&r" (tmp), "=m" (sem->count) : "r" (incr), "m" (sem->count)); } else { static DEFINE_SPINLOCK(semaphore_lock); unsigned long flags; spin_lock_irqsave(&semaphore_lock, flags); old_count = atomic_read(&sem->count); tmp = max_t(int, old_count, 0) + incr; atomic_set(&sem->count, tmp); spin_unlock_irqrestore(&semaphore_lock, flags); } return old_count; } void __up(struct semaphore *sem) { /* * Note that we incremented count in up() before we came here, * but that was ineffective since the result was <= 0, and * any negative value of count is equivalent to 0. * This ends up setting count to 1, unless count is now > 0 * (i.e. because some other cpu has called up() in the meantime), * in which case we just increment count. */ __sem_update_count(sem, 1); wake_up(&sem->wait); } EXPORT_SYMBOL(__up); /* * Note that when we come in to __down or __down_interruptible, * we have already decremented count, but that decrement was * ineffective since the result was < 0, and any negative value * of count is equivalent to 0. * Thus it is only when we decrement count from some value > 0 * that we have actually got the semaphore. */ void __sched __down(struct semaphore *sem) { struct task_struct *tsk = current; DECLARE_WAITQUEUE(wait, tsk); __set_task_state(tsk, TASK_UNINTERRUPTIBLE); add_wait_queue_exclusive(&sem->wait, &wait); /* * Try to get the semaphore. If the count is > 0, then we've * got the semaphore; we decrement count and exit the loop. * If the count is 0 or negative, we set it to -1, indicating * that we are asleep, and then sleep. */ while (__sem_update_count(sem, -1) <= 0) { schedule(); set_task_state(tsk, TASK_UNINTERRUPTIBLE); } remove_wait_queue(&sem->wait, &wait); __set_task_state(tsk, TASK_RUNNING); /* * If there are any more sleepers, wake one of them up so * that it can either get the semaphore, or set count to -1 * indicating that there are still processes sleeping. */ wake_up(&sem->wait); } EXPORT_SYMBOL(__down); int __sched __down_interruptible(struct semaphore * sem) { int retval = 0; struct task_struct *tsk = current; DECLARE_WAITQUEUE(wait, tsk); __set_task_state(tsk, TASK_INTERRUPTIBLE); add_wait_queue_exclusive(&sem->wait, &wait); while (__sem_update_count(sem, -1) <= 0) { if (signal_pending(current)) { /* * A signal is pending - give up trying. * Set sem->count to 0 if it is negative, * since we are no longer sleeping. */ __sem_update_count(sem, 0); retval = -EINTR; break; } schedule(); set_task_state(tsk, TASK_INTERRUPTIBLE); } remove_wait_queue(&sem->wait, &wait); __set_task_state(tsk, TASK_RUNNING); wake_up(&sem->wait); return retval; } EXPORT_SYMBOL(__down_interruptible);