| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1 | /* rwsem-spinlock.c: R/W semaphores: contention handling functions for | 
|  | 2 | * generic spinlock implementation | 
|  | 3 | * | 
|  | 4 | * Copyright (c) 2001   David Howells (dhowells@redhat.com). | 
|  | 5 | * - Derived partially from idea by Andrea Arcangeli <andrea@suse.de> | 
|  | 6 | * - Derived also from comments by Linus | 
|  | 7 | */ | 
|  | 8 | #include <linux/rwsem.h> | 
|  | 9 | #include <linux/sched.h> | 
|  | 10 | #include <linux/module.h> | 
|  | 11 |  | 
|  | 12 | struct rwsem_waiter { | 
|  | 13 | struct list_head list; | 
|  | 14 | struct task_struct *task; | 
|  | 15 | unsigned int flags; | 
|  | 16 | #define RWSEM_WAITING_FOR_READ	0x00000001 | 
|  | 17 | #define RWSEM_WAITING_FOR_WRITE	0x00000002 | 
|  | 18 | }; | 
|  | 19 |  | 
| Amerigo Wang | 29671f2 | 2009-12-14 18:00:21 -0800 | [diff] [blame] | 20 | int rwsem_is_locked(struct rw_semaphore *sem) | 
|  | 21 | { | 
|  | 22 | int ret = 1; | 
|  | 23 | unsigned long flags; | 
|  | 24 |  | 
|  | 25 | if (spin_trylock_irqsave(&sem->wait_lock, flags)) { | 
|  | 26 | ret = (sem->activity != 0); | 
|  | 27 | spin_unlock_irqrestore(&sem->wait_lock, flags); | 
|  | 28 | } | 
|  | 29 | return ret; | 
|  | 30 | } | 
|  | 31 | EXPORT_SYMBOL(rwsem_is_locked); | 
|  | 32 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 33 | /* | 
|  | 34 | * initialise the semaphore | 
|  | 35 | */ | 
| Ingo Molnar | 4ea2176 | 2006-07-03 00:24:53 -0700 | [diff] [blame] | 36 | void __init_rwsem(struct rw_semaphore *sem, const char *name, | 
|  | 37 | struct lock_class_key *key) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 38 | { | 
| Ingo Molnar | 4ea2176 | 2006-07-03 00:24:53 -0700 | [diff] [blame] | 39 | #ifdef CONFIG_DEBUG_LOCK_ALLOC | 
|  | 40 | /* | 
|  | 41 | * Make sure we are not reinitializing a held semaphore: | 
|  | 42 | */ | 
|  | 43 | debug_check_no_locks_freed((void *)sem, sizeof(*sem)); | 
| Peter Zijlstra | 4dfbb9d | 2006-10-11 01:45:14 -0400 | [diff] [blame] | 44 | lockdep_init_map(&sem->dep_map, name, key, 0); | 
| Ingo Molnar | 4ea2176 | 2006-07-03 00:24:53 -0700 | [diff] [blame] | 45 | #endif | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 46 | sem->activity = 0; | 
|  | 47 | spin_lock_init(&sem->wait_lock); | 
|  | 48 | INIT_LIST_HEAD(&sem->wait_list); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 49 | } | 
| Amerigo Wang | 118d52d | 2009-12-14 18:00:20 -0800 | [diff] [blame] | 50 | EXPORT_SYMBOL(__init_rwsem); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 51 |  | 
|  | 52 | /* | 
|  | 53 | * handle the lock release when processes blocked on it that can now run | 
|  | 54 | * - if we come here, then: | 
|  | 55 | *   - the 'active count' _reached_ zero | 
|  | 56 | *   - the 'waiting count' is non-zero | 
|  | 57 | * - the spinlock must be held by the caller | 
|  | 58 | * - woken process blocks are discarded from the list after having task zeroed | 
|  | 59 | * - writers are only woken if wakewrite is non-zero | 
|  | 60 | */ | 
|  | 61 | static inline struct rw_semaphore * | 
|  | 62 | __rwsem_do_wake(struct rw_semaphore *sem, int wakewrite) | 
|  | 63 | { | 
|  | 64 | struct rwsem_waiter *waiter; | 
|  | 65 | struct task_struct *tsk; | 
|  | 66 | int woken; | 
|  | 67 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 68 | waiter = list_entry(sem->wait_list.next, struct rwsem_waiter, list); | 
|  | 69 |  | 
|  | 70 | if (!wakewrite) { | 
|  | 71 | if (waiter->flags & RWSEM_WAITING_FOR_WRITE) | 
|  | 72 | goto out; | 
|  | 73 | goto dont_wake_writers; | 
|  | 74 | } | 
|  | 75 |  | 
|  | 76 | /* if we are allowed to wake writers try to grant a single write lock | 
|  | 77 | * if there's a writer at the front of the queue | 
|  | 78 | * - we leave the 'waiting count' incremented to signify potential | 
|  | 79 | *   contention | 
|  | 80 | */ | 
|  | 81 | if (waiter->flags & RWSEM_WAITING_FOR_WRITE) { | 
|  | 82 | sem->activity = -1; | 
|  | 83 | list_del(&waiter->list); | 
|  | 84 | tsk = waiter->task; | 
|  | 85 | /* Don't touch waiter after ->task has been NULLed */ | 
| akpm@osdl.org | d59dd46 | 2005-05-01 08:58:47 -0700 | [diff] [blame] | 86 | smp_mb(); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 87 | waiter->task = NULL; | 
|  | 88 | wake_up_process(tsk); | 
|  | 89 | put_task_struct(tsk); | 
|  | 90 | goto out; | 
|  | 91 | } | 
|  | 92 |  | 
|  | 93 | /* grant an infinite number of read locks to the front of the queue */ | 
|  | 94 | dont_wake_writers: | 
|  | 95 | woken = 0; | 
|  | 96 | while (waiter->flags & RWSEM_WAITING_FOR_READ) { | 
|  | 97 | struct list_head *next = waiter->list.next; | 
|  | 98 |  | 
|  | 99 | list_del(&waiter->list); | 
|  | 100 | tsk = waiter->task; | 
| akpm@osdl.org | d59dd46 | 2005-05-01 08:58:47 -0700 | [diff] [blame] | 101 | smp_mb(); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 102 | waiter->task = NULL; | 
|  | 103 | wake_up_process(tsk); | 
|  | 104 | put_task_struct(tsk); | 
|  | 105 | woken++; | 
|  | 106 | if (list_empty(&sem->wait_list)) | 
|  | 107 | break; | 
|  | 108 | waiter = list_entry(next, struct rwsem_waiter, list); | 
|  | 109 | } | 
|  | 110 |  | 
|  | 111 | sem->activity += woken; | 
|  | 112 |  | 
|  | 113 | out: | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 114 | return sem; | 
|  | 115 | } | 
|  | 116 |  | 
|  | 117 | /* | 
|  | 118 | * wake a single writer | 
|  | 119 | */ | 
|  | 120 | static inline struct rw_semaphore * | 
|  | 121 | __rwsem_wake_one_writer(struct rw_semaphore *sem) | 
|  | 122 | { | 
|  | 123 | struct rwsem_waiter *waiter; | 
|  | 124 | struct task_struct *tsk; | 
|  | 125 |  | 
|  | 126 | sem->activity = -1; | 
|  | 127 |  | 
|  | 128 | waiter = list_entry(sem->wait_list.next, struct rwsem_waiter, list); | 
|  | 129 | list_del(&waiter->list); | 
|  | 130 |  | 
|  | 131 | tsk = waiter->task; | 
| akpm@osdl.org | d59dd46 | 2005-05-01 08:58:47 -0700 | [diff] [blame] | 132 | smp_mb(); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 133 | waiter->task = NULL; | 
|  | 134 | wake_up_process(tsk); | 
|  | 135 | put_task_struct(tsk); | 
|  | 136 | return sem; | 
|  | 137 | } | 
|  | 138 |  | 
|  | 139 | /* | 
|  | 140 | * get a read lock on the semaphore | 
|  | 141 | */ | 
| Harvey Harrison | 9f741cb | 2008-02-08 04:19:55 -0800 | [diff] [blame] | 142 | void __sched __down_read(struct rw_semaphore *sem) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 143 | { | 
|  | 144 | struct rwsem_waiter waiter; | 
|  | 145 | struct task_struct *tsk; | 
| Kevin Hilman | 3eac4ab | 2010-04-07 11:52:46 -0700 | [diff] [blame] | 146 | unsigned long flags; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 147 |  | 
| Kevin Hilman | 3eac4ab | 2010-04-07 11:52:46 -0700 | [diff] [blame] | 148 | spin_lock_irqsave(&sem->wait_lock, flags); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 149 |  | 
|  | 150 | if (sem->activity >= 0 && list_empty(&sem->wait_list)) { | 
|  | 151 | /* granted */ | 
|  | 152 | sem->activity++; | 
| Kevin Hilman | 3eac4ab | 2010-04-07 11:52:46 -0700 | [diff] [blame] | 153 | spin_unlock_irqrestore(&sem->wait_lock, flags); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 154 | goto out; | 
|  | 155 | } | 
|  | 156 |  | 
|  | 157 | tsk = current; | 
|  | 158 | set_task_state(tsk, TASK_UNINTERRUPTIBLE); | 
|  | 159 |  | 
|  | 160 | /* set up my own style of waitqueue */ | 
|  | 161 | waiter.task = tsk; | 
|  | 162 | waiter.flags = RWSEM_WAITING_FOR_READ; | 
|  | 163 | get_task_struct(tsk); | 
|  | 164 |  | 
|  | 165 | list_add_tail(&waiter.list, &sem->wait_list); | 
|  | 166 |  | 
|  | 167 | /* we don't need to touch the semaphore struct anymore */ | 
| Kevin Hilman | 3eac4ab | 2010-04-07 11:52:46 -0700 | [diff] [blame] | 168 | spin_unlock_irqrestore(&sem->wait_lock, flags); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 169 |  | 
|  | 170 | /* wait to be given the lock */ | 
|  | 171 | for (;;) { | 
|  | 172 | if (!waiter.task) | 
|  | 173 | break; | 
|  | 174 | schedule(); | 
|  | 175 | set_task_state(tsk, TASK_UNINTERRUPTIBLE); | 
|  | 176 | } | 
|  | 177 |  | 
|  | 178 | tsk->state = TASK_RUNNING; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 179 | out: | 
| Ingo Molnar | c4e0511 | 2006-07-03 00:24:29 -0700 | [diff] [blame] | 180 | ; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 181 | } | 
|  | 182 |  | 
|  | 183 | /* | 
|  | 184 | * trylock for reading -- returns 1 if successful, 0 if contention | 
|  | 185 | */ | 
| Harvey Harrison | 9f741cb | 2008-02-08 04:19:55 -0800 | [diff] [blame] | 186 | int __down_read_trylock(struct rw_semaphore *sem) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 187 | { | 
|  | 188 | unsigned long flags; | 
|  | 189 | int ret = 0; | 
|  | 190 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 191 |  | 
|  | 192 | spin_lock_irqsave(&sem->wait_lock, flags); | 
|  | 193 |  | 
|  | 194 | if (sem->activity >= 0 && list_empty(&sem->wait_list)) { | 
|  | 195 | /* granted */ | 
|  | 196 | sem->activity++; | 
|  | 197 | ret = 1; | 
|  | 198 | } | 
|  | 199 |  | 
|  | 200 | spin_unlock_irqrestore(&sem->wait_lock, flags); | 
|  | 201 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 202 | return ret; | 
|  | 203 | } | 
|  | 204 |  | 
|  | 205 | /* | 
|  | 206 | * get a write lock on the semaphore | 
|  | 207 | * - we increment the waiting count anyway to indicate an exclusive lock | 
|  | 208 | */ | 
| Harvey Harrison | 9f741cb | 2008-02-08 04:19:55 -0800 | [diff] [blame] | 209 | void __sched __down_write_nested(struct rw_semaphore *sem, int subclass) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 210 | { | 
|  | 211 | struct rwsem_waiter waiter; | 
|  | 212 | struct task_struct *tsk; | 
| Kevin Hilman | 3eac4ab | 2010-04-07 11:52:46 -0700 | [diff] [blame] | 213 | unsigned long flags; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 214 |  | 
| Kevin Hilman | 3eac4ab | 2010-04-07 11:52:46 -0700 | [diff] [blame] | 215 | spin_lock_irqsave(&sem->wait_lock, flags); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 216 |  | 
|  | 217 | if (sem->activity == 0 && list_empty(&sem->wait_list)) { | 
|  | 218 | /* granted */ | 
|  | 219 | sem->activity = -1; | 
| Kevin Hilman | 3eac4ab | 2010-04-07 11:52:46 -0700 | [diff] [blame] | 220 | spin_unlock_irqrestore(&sem->wait_lock, flags); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 221 | goto out; | 
|  | 222 | } | 
|  | 223 |  | 
|  | 224 | tsk = current; | 
|  | 225 | set_task_state(tsk, TASK_UNINTERRUPTIBLE); | 
|  | 226 |  | 
|  | 227 | /* set up my own style of waitqueue */ | 
|  | 228 | waiter.task = tsk; | 
|  | 229 | waiter.flags = RWSEM_WAITING_FOR_WRITE; | 
|  | 230 | get_task_struct(tsk); | 
|  | 231 |  | 
|  | 232 | list_add_tail(&waiter.list, &sem->wait_list); | 
|  | 233 |  | 
|  | 234 | /* we don't need to touch the semaphore struct anymore */ | 
| Kevin Hilman | 3eac4ab | 2010-04-07 11:52:46 -0700 | [diff] [blame] | 235 | spin_unlock_irqrestore(&sem->wait_lock, flags); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 236 |  | 
|  | 237 | /* wait to be given the lock */ | 
|  | 238 | for (;;) { | 
|  | 239 | if (!waiter.task) | 
|  | 240 | break; | 
|  | 241 | schedule(); | 
|  | 242 | set_task_state(tsk, TASK_UNINTERRUPTIBLE); | 
|  | 243 | } | 
|  | 244 |  | 
|  | 245 | tsk->state = TASK_RUNNING; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 246 | out: | 
| Ingo Molnar | c4e0511 | 2006-07-03 00:24:29 -0700 | [diff] [blame] | 247 | ; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 248 | } | 
|  | 249 |  | 
| Harvey Harrison | 9f741cb | 2008-02-08 04:19:55 -0800 | [diff] [blame] | 250 | void __sched __down_write(struct rw_semaphore *sem) | 
| Ingo Molnar | 4ea2176 | 2006-07-03 00:24:53 -0700 | [diff] [blame] | 251 | { | 
|  | 252 | __down_write_nested(sem, 0); | 
|  | 253 | } | 
|  | 254 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 255 | /* | 
|  | 256 | * trylock for writing -- returns 1 if successful, 0 if contention | 
|  | 257 | */ | 
| Harvey Harrison | 9f741cb | 2008-02-08 04:19:55 -0800 | [diff] [blame] | 258 | int __down_write_trylock(struct rw_semaphore *sem) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 259 | { | 
|  | 260 | unsigned long flags; | 
|  | 261 | int ret = 0; | 
|  | 262 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 263 | spin_lock_irqsave(&sem->wait_lock, flags); | 
|  | 264 |  | 
|  | 265 | if (sem->activity == 0 && list_empty(&sem->wait_list)) { | 
|  | 266 | /* granted */ | 
|  | 267 | sem->activity = -1; | 
|  | 268 | ret = 1; | 
|  | 269 | } | 
|  | 270 |  | 
|  | 271 | spin_unlock_irqrestore(&sem->wait_lock, flags); | 
|  | 272 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 273 | return ret; | 
|  | 274 | } | 
|  | 275 |  | 
|  | 276 | /* | 
|  | 277 | * release a read lock on the semaphore | 
|  | 278 | */ | 
| Harvey Harrison | 9f741cb | 2008-02-08 04:19:55 -0800 | [diff] [blame] | 279 | void __up_read(struct rw_semaphore *sem) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 280 | { | 
|  | 281 | unsigned long flags; | 
|  | 282 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 283 | spin_lock_irqsave(&sem->wait_lock, flags); | 
|  | 284 |  | 
|  | 285 | if (--sem->activity == 0 && !list_empty(&sem->wait_list)) | 
|  | 286 | sem = __rwsem_wake_one_writer(sem); | 
|  | 287 |  | 
|  | 288 | spin_unlock_irqrestore(&sem->wait_lock, flags); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 289 | } | 
|  | 290 |  | 
|  | 291 | /* | 
|  | 292 | * release a write lock on the semaphore | 
|  | 293 | */ | 
| Harvey Harrison | 9f741cb | 2008-02-08 04:19:55 -0800 | [diff] [blame] | 294 | void __up_write(struct rw_semaphore *sem) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 295 | { | 
|  | 296 | unsigned long flags; | 
|  | 297 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 298 | spin_lock_irqsave(&sem->wait_lock, flags); | 
|  | 299 |  | 
|  | 300 | sem->activity = 0; | 
|  | 301 | if (!list_empty(&sem->wait_list)) | 
|  | 302 | sem = __rwsem_do_wake(sem, 1); | 
|  | 303 |  | 
|  | 304 | spin_unlock_irqrestore(&sem->wait_lock, flags); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 305 | } | 
|  | 306 |  | 
|  | 307 | /* | 
|  | 308 | * downgrade a write lock into a read lock | 
|  | 309 | * - just wake up any readers at the front of the queue | 
|  | 310 | */ | 
| Harvey Harrison | 9f741cb | 2008-02-08 04:19:55 -0800 | [diff] [blame] | 311 | void __downgrade_write(struct rw_semaphore *sem) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 312 | { | 
|  | 313 | unsigned long flags; | 
|  | 314 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 315 | spin_lock_irqsave(&sem->wait_lock, flags); | 
|  | 316 |  | 
|  | 317 | sem->activity = 1; | 
|  | 318 | if (!list_empty(&sem->wait_list)) | 
|  | 319 | sem = __rwsem_do_wake(sem, 0); | 
|  | 320 |  | 
|  | 321 | spin_unlock_irqrestore(&sem->wait_lock, flags); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 322 | } | 
|  | 323 |  |