git://git.onelab.eu
/
linux-2.6.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Merge to Fedora kernel-2.6.18-1.2224_FC5 patched with stable patch-2.6.18.1-vs2.0...
[linux-2.6.git]
/
lib
/
rwsem-spinlock.c
diff --git
a/lib/rwsem-spinlock.c
b/lib/rwsem-spinlock.c
index
96255f4
..
c4cfd6c
100644
(file)
--- a/
lib/rwsem-spinlock.c
+++ b/
lib/rwsem-spinlock.c
@@
-17,27
+17,22
@@
struct rwsem_waiter {
#define RWSEM_WAITING_FOR_WRITE 0x00000002
};
#define RWSEM_WAITING_FOR_WRITE 0x00000002
};
-#if RWSEM_DEBUG
-void rwsemtrace(struct rw_semaphore *sem, const char *str)
-{
- if (sem->debug)
- printk("[%d] %s({%d,%d})\n",
- current->pid, str, sem->activity,
- list_empty(&sem->wait_list) ? 0 : 1);
-}
-#endif
-
/*
* initialise the semaphore
*/
/*
* initialise the semaphore
*/
-void fastcall init_rwsem(struct rw_semaphore *sem)
+void __init_rwsem(struct rw_semaphore *sem, const char *name,
+ struct lock_class_key *key)
{
{
+#ifdef CONFIG_DEBUG_LOCK_ALLOC
+ /*
+ * Make sure we are not reinitializing a held semaphore:
+ */
+ debug_check_no_locks_freed((void *)sem, sizeof(*sem));
+ lockdep_init_map(&sem->dep_map, name, key, 0);
+#endif
sem->activity = 0;
spin_lock_init(&sem->wait_lock);
INIT_LIST_HEAD(&sem->wait_list);
sem->activity = 0;
spin_lock_init(&sem->wait_lock);
INIT_LIST_HEAD(&sem->wait_list);
-#if RWSEM_DEBUG
- sem->debug = 0;
-#endif
}
/*
}
/*
@@
-56,8
+51,6
@@
__rwsem_do_wake(struct rw_semaphore *sem, int wakewrite)
struct task_struct *tsk;
int woken;
struct task_struct *tsk;
int woken;
- rwsemtrace(sem, "Entering __rwsem_do_wake");
-
waiter = list_entry(sem->wait_list.next, struct rwsem_waiter, list);
if (!wakewrite) {
waiter = list_entry(sem->wait_list.next, struct rwsem_waiter, list);
if (!wakewrite) {
@@
-76,7
+69,7
@@
__rwsem_do_wake(struct rw_semaphore *sem, int wakewrite)
list_del(&waiter->list);
tsk = waiter->task;
/* Don't touch waiter after ->task has been NULLed */
list_del(&waiter->list);
tsk = waiter->task;
/* Don't touch waiter after ->task has been NULLed */
- mb();
+
smp_
mb();
waiter->task = NULL;
wake_up_process(tsk);
put_task_struct(tsk);
waiter->task = NULL;
wake_up_process(tsk);
put_task_struct(tsk);
@@
-91,7
+84,7
@@
__rwsem_do_wake(struct rw_semaphore *sem, int wakewrite)
list_del(&waiter->list);
tsk = waiter->task;
list_del(&waiter->list);
tsk = waiter->task;
- mb();
+
smp_
mb();
waiter->task = NULL;
wake_up_process(tsk);
put_task_struct(tsk);
waiter->task = NULL;
wake_up_process(tsk);
put_task_struct(tsk);
@@
-104,7
+97,6
@@
__rwsem_do_wake(struct rw_semaphore *sem, int wakewrite)
sem->activity += woken;
out:
sem->activity += woken;
out:
- rwsemtrace(sem, "Leaving __rwsem_do_wake");
return sem;
}
return sem;
}
@@
-123,7
+115,7
@@
__rwsem_wake_one_writer(struct rw_semaphore *sem)
list_del(&waiter->list);
tsk = waiter->task;
list_del(&waiter->list);
tsk = waiter->task;
- mb();
+
smp_
mb();
waiter->task = NULL;
wake_up_process(tsk);
put_task_struct(tsk);
waiter->task = NULL;
wake_up_process(tsk);
put_task_struct(tsk);
@@
-138,14
+130,12
@@
void fastcall __sched __down_read(struct rw_semaphore *sem)
struct rwsem_waiter waiter;
struct task_struct *tsk;
struct rwsem_waiter waiter;
struct task_struct *tsk;
- rwsemtrace(sem, "Entering __down_read");
-
- spin_lock(&sem->wait_lock);
+ spin_lock_irq(&sem->wait_lock);
if (sem->activity >= 0 && list_empty(&sem->wait_list)) {
/* granted */
sem->activity++;
if (sem->activity >= 0 && list_empty(&sem->wait_list)) {
/* granted */
sem->activity++;
- spin_unlock(&sem->wait_lock);
+ spin_unlock
_irq
(&sem->wait_lock);
goto out;
}
goto out;
}
@@
-160,7
+150,7
@@
void fastcall __sched __down_read(struct rw_semaphore *sem)
list_add_tail(&waiter.list, &sem->wait_list);
/* we don't need to touch the semaphore struct anymore */
list_add_tail(&waiter.list, &sem->wait_list);
/* we don't need to touch the semaphore struct anymore */
- spin_unlock(&sem->wait_lock);
+ spin_unlock
_irq
(&sem->wait_lock);
/* wait to be given the lock */
for (;;) {
/* wait to be given the lock */
for (;;) {
@@
-171,9
+161,8
@@
void fastcall __sched __down_read(struct rw_semaphore *sem)
}
tsk->state = TASK_RUNNING;
}
tsk->state = TASK_RUNNING;
-
out:
out:
-
rwsemtrace(sem, "Leaving __down_read")
;
+ ;
}
/*
}
/*
@@
-181,10
+170,11
@@
void fastcall __sched __down_read(struct rw_semaphore *sem)
*/
int fastcall __down_read_trylock(struct rw_semaphore *sem)
{
*/
int fastcall __down_read_trylock(struct rw_semaphore *sem)
{
+ unsigned long flags;
int ret = 0;
int ret = 0;
- rwsemtrace(sem, "Entering __down_read_trylock");
- spin_lock(&sem->wait_lock);
+
+ spin_lock_irqsave(&sem->wait_lock, flags);
if (sem->activity >= 0 && list_empty(&sem->wait_list)) {
/* granted */
if (sem->activity >= 0 && list_empty(&sem->wait_list)) {
/* granted */
@@
-192,9
+182,8
@@
int fastcall __down_read_trylock(struct rw_semaphore *sem)
ret = 1;
}
ret = 1;
}
- spin_unlock
(&sem->wait_lock
);
+ spin_unlock
_irqrestore(&sem->wait_lock, flags
);
- rwsemtrace(sem, "Leaving __down_read_trylock");
return ret;
}
return ret;
}
@@
-202,19
+191,17
@@
int fastcall __down_read_trylock(struct rw_semaphore *sem)
* get a write lock on the semaphore
* - we increment the waiting count anyway to indicate an exclusive lock
*/
* get a write lock on the semaphore
* - we increment the waiting count anyway to indicate an exclusive lock
*/
-void fastcall __sched __down_write
(struct rw_semaphore *sem
)
+void fastcall __sched __down_write
_nested(struct rw_semaphore *sem, int subclass
)
{
struct rwsem_waiter waiter;
struct task_struct *tsk;
{
struct rwsem_waiter waiter;
struct task_struct *tsk;
- rwsemtrace(sem, "Entering __down_write");
-
- spin_lock(&sem->wait_lock);
+ spin_lock_irq(&sem->wait_lock);
if (sem->activity == 0 && list_empty(&sem->wait_list)) {
/* granted */
sem->activity = -1;
if (sem->activity == 0 && list_empty(&sem->wait_list)) {
/* granted */
sem->activity = -1;
- spin_unlock(&sem->wait_lock);
+ spin_unlock
_irq
(&sem->wait_lock);
goto out;
}
goto out;
}
@@
-229,7
+216,7
@@
void fastcall __sched __down_write(struct rw_semaphore *sem)
list_add_tail(&waiter.list, &sem->wait_list);
/* we don't need to touch the semaphore struct anymore */
list_add_tail(&waiter.list, &sem->wait_list);
/* we don't need to touch the semaphore struct anymore */
- spin_unlock(&sem->wait_lock);
+ spin_unlock
_irq
(&sem->wait_lock);
/* wait to be given the lock */
for (;;) {
/* wait to be given the lock */
for (;;) {
@@
-240,9
+227,13
@@
void fastcall __sched __down_write(struct rw_semaphore *sem)
}
tsk->state = TASK_RUNNING;
}
tsk->state = TASK_RUNNING;
-
out:
out:
- rwsemtrace(sem, "Leaving __down_write");
+ ;
+}
+
+void fastcall __sched __down_write(struct rw_semaphore *sem)
+{
+ __down_write_nested(sem, 0);
}
/*
}
/*
@@
-250,10
+241,10
@@
void fastcall __sched __down_write(struct rw_semaphore *sem)
*/
int fastcall __down_write_trylock(struct rw_semaphore *sem)
{
*/
int fastcall __down_write_trylock(struct rw_semaphore *sem)
{
+ unsigned long flags;
int ret = 0;
int ret = 0;
- rwsemtrace(sem, "Entering __down_write_trylock");
- spin_lock
(&sem->wait_lock
);
+ spin_lock
_irqsave(&sem->wait_lock, flags
);
if (sem->activity == 0 && list_empty(&sem->wait_list)) {
/* granted */
if (sem->activity == 0 && list_empty(&sem->wait_list)) {
/* granted */
@@
-261,9
+252,8
@@
int fastcall __down_write_trylock(struct rw_semaphore *sem)
ret = 1;
}
ret = 1;
}
- spin_unlock
(&sem->wait_lock
);
+ spin_unlock
_irqrestore(&sem->wait_lock, flags
);
- rwsemtrace(sem, "Leaving __down_write_trylock");
return ret;
}
return ret;
}
@@
-272,16
+262,14
@@
int fastcall __down_write_trylock(struct rw_semaphore *sem)
*/
void fastcall __up_read(struct rw_semaphore *sem)
{
*/
void fastcall __up_read(struct rw_semaphore *sem)
{
-
rwsemtrace(sem, "Entering __up_read")
;
+
unsigned long flags
;
- spin_lock
(&sem->wait_lock
);
+ spin_lock
_irqsave(&sem->wait_lock, flags
);
if (--sem->activity == 0 && !list_empty(&sem->wait_list))
sem = __rwsem_wake_one_writer(sem);
if (--sem->activity == 0 && !list_empty(&sem->wait_list))
sem = __rwsem_wake_one_writer(sem);
- spin_unlock(&sem->wait_lock);
-
- rwsemtrace(sem, "Leaving __up_read");
+ spin_unlock_irqrestore(&sem->wait_lock, flags);
}
/*
}
/*
@@
-289,17
+277,15
@@
void fastcall __up_read(struct rw_semaphore *sem)
*/
void fastcall __up_write(struct rw_semaphore *sem)
{
*/
void fastcall __up_write(struct rw_semaphore *sem)
{
-
rwsemtrace(sem, "Entering __up_write")
;
+
unsigned long flags
;
- spin_lock
(&sem->wait_lock
);
+ spin_lock
_irqsave(&sem->wait_lock, flags
);
sem->activity = 0;
if (!list_empty(&sem->wait_list))
sem = __rwsem_do_wake(sem, 1);
sem->activity = 0;
if (!list_empty(&sem->wait_list))
sem = __rwsem_do_wake(sem, 1);
- spin_unlock(&sem->wait_lock);
-
- rwsemtrace(sem, "Leaving __up_write");
+ spin_unlock_irqrestore(&sem->wait_lock, flags);
}
/*
}
/*
@@
-308,27
+294,23
@@
void fastcall __up_write(struct rw_semaphore *sem)
*/
void fastcall __downgrade_write(struct rw_semaphore *sem)
{
*/
void fastcall __downgrade_write(struct rw_semaphore *sem)
{
-
rwsemtrace(sem, "Entering __downgrade_write")
;
+
unsigned long flags
;
- spin_lock
(&sem->wait_lock
);
+ spin_lock
_irqsave(&sem->wait_lock, flags
);
sem->activity = 1;
if (!list_empty(&sem->wait_list))
sem = __rwsem_do_wake(sem, 0);
sem->activity = 1;
if (!list_empty(&sem->wait_list))
sem = __rwsem_do_wake(sem, 0);
- spin_unlock(&sem->wait_lock);
-
- rwsemtrace(sem, "Leaving __downgrade_write");
+ spin_unlock_irqrestore(&sem->wait_lock, flags);
}
}
-EXPORT_SYMBOL(init_rwsem);
+EXPORT_SYMBOL(
__
init_rwsem);
EXPORT_SYMBOL(__down_read);
EXPORT_SYMBOL(__down_read_trylock);
EXPORT_SYMBOL(__down_read);
EXPORT_SYMBOL(__down_read_trylock);
+EXPORT_SYMBOL(__down_write_nested);
EXPORT_SYMBOL(__down_write);
EXPORT_SYMBOL(__down_write_trylock);
EXPORT_SYMBOL(__up_read);
EXPORT_SYMBOL(__up_write);
EXPORT_SYMBOL(__downgrade_write);
EXPORT_SYMBOL(__down_write);
EXPORT_SYMBOL(__down_write_trylock);
EXPORT_SYMBOL(__up_read);
EXPORT_SYMBOL(__up_write);
EXPORT_SYMBOL(__downgrade_write);
-#if RWSEM_DEBUG
-EXPORT_SYMBOL(rwsemtrace);
-#endif