Update to 3.4-final.
[linux-flexiantxendom0-3.2.10.git] / lib / rwsem.c
index 318d435..8337e1b 100644 (file)
@@ -6,7 +6,7 @@
 #include <linux/rwsem.h>
 #include <linux/sched.h>
 #include <linux/init.h>
-#include <linux/module.h>
+#include <linux/export.h>
 
 /*
  * Initialize an rwsem:
@@ -22,7 +22,7 @@ void __init_rwsem(struct rw_semaphore *sem, const char *name,
        lockdep_init_map(&sem->dep_map, name, key, 0);
 #endif
        sem->count = RWSEM_UNLOCKED_VALUE;
-       spin_lock_init(&sem->wait_lock);
+       raw_spin_lock_init(&sem->wait_lock);
        INIT_LIST_HEAD(&sem->wait_list);
 }
 
@@ -171,21 +171,23 @@ __rwsem_do_wake(struct rw_semaphore *sem, int wake_type)
  */
 static struct rw_semaphore __sched *
 rwsem_down_failed_common(struct rw_semaphore *sem,
-                       struct rwsem_waiter *waiter, signed long adjustment)
+                        unsigned int flags, signed long adjustment)
 {
+       struct rwsem_waiter waiter;
        struct task_struct *tsk = current;
        signed long count;
 
        set_task_state(tsk, TASK_UNINTERRUPTIBLE);
 
        /* set up my own style of waitqueue */
-       spin_lock_irq(&sem->wait_lock);
-       waiter->task = tsk;
+       raw_spin_lock_irq(&sem->wait_lock);
+       waiter.task = tsk;
+       waiter.flags = flags;
        get_task_struct(tsk);
 
        if (list_empty(&sem->wait_list))
                adjustment += RWSEM_WAITING_BIAS;
-       list_add_tail(&waiter->list, &sem->wait_list);
+       list_add_tail(&waiter.list, &sem->wait_list);
 
        /* we're now waiting on the lock, but no longer actively locking */
        count = rwsem_atomic_update(adjustment, sem);
@@ -202,11 +204,11 @@ rwsem_down_failed_common(struct rw_semaphore *sem,
                 adjustment == -RWSEM_ACTIVE_WRITE_BIAS)
                sem = __rwsem_do_wake(sem, RWSEM_WAKE_READ_OWNED);
 
-       spin_unlock_irq(&sem->wait_lock);
+       raw_spin_unlock_irq(&sem->wait_lock);
 
        /* wait to be given the lock */
        for (;;) {
-               if (!waiter->task)
+               if (!waiter.task)
                        break;
                schedule();
                set_task_state(tsk, TASK_UNINTERRUPTIBLE);
@@ -220,45 +222,36 @@ rwsem_down_failed_common(struct rw_semaphore *sem,
 /*
  * wait for the read lock to be granted
  */
-asmregparm struct rw_semaphore __sched *
-rwsem_down_read_failed(struct rw_semaphore *sem)
+struct rw_semaphore __sched *rwsem_down_read_failed(struct rw_semaphore *sem)
 {
-       struct rwsem_waiter waiter;
-
-       waiter.flags = RWSEM_WAITING_FOR_READ;
-       rwsem_down_failed_common(sem, &waiter, -RWSEM_ACTIVE_READ_BIAS);
-       return sem;
+       return rwsem_down_failed_common(sem, RWSEM_WAITING_FOR_READ,
+                                       -RWSEM_ACTIVE_READ_BIAS);
 }
 
 /*
  * wait for the write lock to be granted
  */
-asmregparm struct rw_semaphore __sched *
-rwsem_down_write_failed(struct rw_semaphore *sem)
+struct rw_semaphore __sched *rwsem_down_write_failed(struct rw_semaphore *sem)
 {
-       struct rwsem_waiter waiter;
-
-       waiter.flags = RWSEM_WAITING_FOR_WRITE;
-       rwsem_down_failed_common(sem, &waiter, -RWSEM_ACTIVE_WRITE_BIAS);
-
-       return sem;
+       return rwsem_down_failed_common(sem, RWSEM_WAITING_FOR_WRITE,
+                                       -RWSEM_ACTIVE_WRITE_BIAS);
 }
 
 /*
  * handle waking up a waiter on the semaphore
  * - up_read/up_write has decremented the active part of count if we come here
  */
-asmregparm struct rw_semaphore *rwsem_wake(struct rw_semaphore *sem)
+struct rw_semaphore *rwsem_wake(struct rw_semaphore *sem)
 {
        unsigned long flags;
 
-       spin_lock_irqsave(&sem->wait_lock, flags);
+       raw_spin_lock_irqsave(&sem->wait_lock, flags);
 
        /* do nothing if list empty */
        if (!list_empty(&sem->wait_list))
                sem = __rwsem_do_wake(sem, RWSEM_WAKE_ANY);
 
-       spin_unlock_irqrestore(&sem->wait_lock, flags);
+       raw_spin_unlock_irqrestore(&sem->wait_lock, flags);
 
        return sem;
 }
@@ -268,17 +261,17 @@ asmregparm struct rw_semaphore *rwsem_wake(struct rw_semaphore *sem)
  * - caller incremented waiting part of count and discovered it still negative
  * - just wake up any readers at the front of the queue
  */
-asmregparm struct rw_semaphore *rwsem_downgrade_wake(struct rw_semaphore *sem)
+struct rw_semaphore *rwsem_downgrade_wake(struct rw_semaphore *sem)
 {
        unsigned long flags;
 
-       spin_lock_irqsave(&sem->wait_lock, flags);
+       raw_spin_lock_irqsave(&sem->wait_lock, flags);
 
        /* do nothing if list empty */
        if (!list_empty(&sem->wait_list))
                sem = __rwsem_do_wake(sem, RWSEM_WAKE_READ_OWNED);
 
-       spin_unlock_irqrestore(&sem->wait_lock, flags);
+       raw_spin_unlock_irqrestore(&sem->wait_lock, flags);
 
        return sem;
 }