| Steffen Klassert | 16295be | 2010-01-06 19:47:10 +1100 | [diff] [blame] | 1 | /* | 
 | 2 |  * padata.c - generic interface to process data streams in parallel | 
 | 3 |  * | 
 | 4 |  * Copyright (C) 2008, 2009 secunet Security Networks AG | 
 | 5 |  * Copyright (C) 2008, 2009 Steffen Klassert <steffen.klassert@secunet.com> | 
 | 6 |  * | 
 | 7 |  * This program is free software; you can redistribute it and/or modify it | 
 | 8 |  * under the terms and conditions of the GNU General Public License, | 
 | 9 |  * version 2, as published by the Free Software Foundation. | 
 | 10 |  * | 
 | 11 |  * This program is distributed in the hope it will be useful, but WITHOUT | 
 | 12 |  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | 
 | 13 |  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for | 
 | 14 |  * more details. | 
 | 15 |  * | 
 | 16 |  * You should have received a copy of the GNU General Public License along with | 
 | 17 |  * this program; if not, write to the Free Software Foundation, Inc., | 
 | 18 |  * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA. | 
 | 19 |  */ | 
 | 20 |  | 
 | 21 | #include <linux/module.h> | 
 | 22 | #include <linux/cpumask.h> | 
 | 23 | #include <linux/err.h> | 
 | 24 | #include <linux/cpu.h> | 
 | 25 | #include <linux/padata.h> | 
 | 26 | #include <linux/mutex.h> | 
 | 27 | #include <linux/sched.h> | 
 | 28 | #include <linux/rcupdate.h> | 
 | 29 |  | 
 | 30 | #define MAX_SEQ_NR INT_MAX - NR_CPUS | 
 | 31 | #define MAX_OBJ_NUM 10000 * NR_CPUS | 
 | 32 |  | 
 | 33 | static int padata_index_to_cpu(struct parallel_data *pd, int cpu_index) | 
 | 34 | { | 
 | 35 | 	int cpu, target_cpu; | 
 | 36 |  | 
 | 37 | 	target_cpu = cpumask_first(pd->cpumask); | 
 | 38 | 	for (cpu = 0; cpu < cpu_index; cpu++) | 
 | 39 | 		target_cpu = cpumask_next(target_cpu, pd->cpumask); | 
 | 40 |  | 
 | 41 | 	return target_cpu; | 
 | 42 | } | 
 | 43 |  | 
 | 44 | static int padata_cpu_hash(struct padata_priv *padata) | 
 | 45 | { | 
 | 46 | 	int cpu_index; | 
 | 47 | 	struct parallel_data *pd; | 
 | 48 |  | 
 | 49 | 	pd =  padata->pd; | 
 | 50 |  | 
 | 51 | 	/* | 
 | 52 | 	 * Hash the sequence numbers to the cpus by taking | 
 | 53 | 	 * seq_nr mod. number of cpus in use. | 
 | 54 | 	 */ | 
 | 55 | 	cpu_index =  padata->seq_nr % cpumask_weight(pd->cpumask); | 
 | 56 |  | 
 | 57 | 	return padata_index_to_cpu(pd, cpu_index); | 
 | 58 | } | 
 | 59 |  | 
 | 60 | static void padata_parallel_worker(struct work_struct *work) | 
 | 61 | { | 
 | 62 | 	struct padata_queue *queue; | 
 | 63 | 	struct parallel_data *pd; | 
 | 64 | 	struct padata_instance *pinst; | 
 | 65 | 	LIST_HEAD(local_list); | 
 | 66 |  | 
 | 67 | 	local_bh_disable(); | 
 | 68 | 	queue = container_of(work, struct padata_queue, pwork); | 
 | 69 | 	pd = queue->pd; | 
 | 70 | 	pinst = pd->pinst; | 
 | 71 |  | 
 | 72 | 	spin_lock(&queue->parallel.lock); | 
 | 73 | 	list_replace_init(&queue->parallel.list, &local_list); | 
 | 74 | 	spin_unlock(&queue->parallel.lock); | 
 | 75 |  | 
 | 76 | 	while (!list_empty(&local_list)) { | 
 | 77 | 		struct padata_priv *padata; | 
 | 78 |  | 
 | 79 | 		padata = list_entry(local_list.next, | 
 | 80 | 				    struct padata_priv, list); | 
 | 81 |  | 
 | 82 | 		list_del_init(&padata->list); | 
 | 83 |  | 
 | 84 | 		padata->parallel(padata); | 
 | 85 | 	} | 
 | 86 |  | 
 | 87 | 	local_bh_enable(); | 
 | 88 | } | 
 | 89 |  | 
 | 90 | /* | 
 | 91 |  * padata_do_parallel - padata parallelization function | 
 | 92 |  * | 
 | 93 |  * @pinst: padata instance | 
 | 94 |  * @padata: object to be parallelized | 
 | 95 |  * @cb_cpu: cpu the serialization callback function will run on, | 
 | 96 |  *          must be in the cpumask of padata. | 
 | 97 |  * | 
 | 98 |  * The parallelization callback function will run with BHs off. | 
 | 99 |  * Note: Every object which is parallelized by padata_do_parallel | 
 | 100 |  * must be seen by padata_do_serial. | 
 | 101 |  */ | 
 | 102 | int padata_do_parallel(struct padata_instance *pinst, | 
 | 103 | 		       struct padata_priv *padata, int cb_cpu) | 
 | 104 | { | 
 | 105 | 	int target_cpu, err; | 
 | 106 | 	struct padata_queue *queue; | 
 | 107 | 	struct parallel_data *pd; | 
 | 108 |  | 
 | 109 | 	rcu_read_lock_bh(); | 
 | 110 |  | 
 | 111 | 	pd = rcu_dereference(pinst->pd); | 
 | 112 |  | 
 | 113 | 	err = 0; | 
 | 114 | 	if (!(pinst->flags & PADATA_INIT)) | 
 | 115 | 		goto out; | 
 | 116 |  | 
 | 117 | 	err =  -EBUSY; | 
 | 118 | 	if ((pinst->flags & PADATA_RESET)) | 
 | 119 | 		goto out; | 
 | 120 |  | 
 | 121 | 	if (atomic_read(&pd->refcnt) >= MAX_OBJ_NUM) | 
 | 122 | 		goto out; | 
 | 123 |  | 
 | 124 | 	err = -EINVAL; | 
 | 125 | 	if (!cpumask_test_cpu(cb_cpu, pd->cpumask)) | 
 | 126 | 		goto out; | 
 | 127 |  | 
 | 128 | 	err = -EINPROGRESS; | 
 | 129 | 	atomic_inc(&pd->refcnt); | 
 | 130 | 	padata->pd = pd; | 
 | 131 | 	padata->cb_cpu = cb_cpu; | 
 | 132 |  | 
 | 133 | 	if (unlikely(atomic_read(&pd->seq_nr) == pd->max_seq_nr)) | 
 | 134 | 		atomic_set(&pd->seq_nr, -1); | 
 | 135 |  | 
 | 136 | 	padata->seq_nr = atomic_inc_return(&pd->seq_nr); | 
 | 137 |  | 
 | 138 | 	target_cpu = padata_cpu_hash(padata); | 
 | 139 | 	queue = per_cpu_ptr(pd->queue, target_cpu); | 
 | 140 |  | 
 | 141 | 	spin_lock(&queue->parallel.lock); | 
 | 142 | 	list_add_tail(&padata->list, &queue->parallel.list); | 
 | 143 | 	spin_unlock(&queue->parallel.lock); | 
 | 144 |  | 
 | 145 | 	queue_work_on(target_cpu, pinst->wq, &queue->pwork); | 
 | 146 |  | 
 | 147 | out: | 
 | 148 | 	rcu_read_unlock_bh(); | 
 | 149 |  | 
 | 150 | 	return err; | 
 | 151 | } | 
 | 152 | EXPORT_SYMBOL(padata_do_parallel); | 
 | 153 |  | 
 | 154 | static struct padata_priv *padata_get_next(struct parallel_data *pd) | 
 | 155 | { | 
 | 156 | 	int cpu, num_cpus, empty, calc_seq_nr; | 
 | 157 | 	int seq_nr, next_nr, overrun, next_overrun; | 
 | 158 | 	struct padata_queue *queue, *next_queue; | 
 | 159 | 	struct padata_priv *padata; | 
 | 160 | 	struct padata_list *reorder; | 
 | 161 |  | 
 | 162 | 	empty = 0; | 
 | 163 | 	next_nr = -1; | 
 | 164 | 	next_overrun = 0; | 
 | 165 | 	next_queue = NULL; | 
 | 166 |  | 
 | 167 | 	num_cpus = cpumask_weight(pd->cpumask); | 
 | 168 |  | 
 | 169 | 	for_each_cpu(cpu, pd->cpumask) { | 
 | 170 | 		queue = per_cpu_ptr(pd->queue, cpu); | 
 | 171 | 		reorder = &queue->reorder; | 
 | 172 |  | 
 | 173 | 		/* | 
 | 174 | 		 * Calculate the seq_nr of the object that should be | 
 | 175 | 		 * next in this queue. | 
 | 176 | 		 */ | 
 | 177 | 		overrun = 0; | 
 | 178 | 		calc_seq_nr = (atomic_read(&queue->num_obj) * num_cpus) | 
 | 179 | 			       + queue->cpu_index; | 
 | 180 |  | 
 | 181 | 		if (unlikely(calc_seq_nr > pd->max_seq_nr)) { | 
 | 182 | 			calc_seq_nr = calc_seq_nr - pd->max_seq_nr - 1; | 
 | 183 | 			overrun = 1; | 
 | 184 | 		} | 
 | 185 |  | 
 | 186 | 		if (!list_empty(&reorder->list)) { | 
 | 187 | 			padata = list_entry(reorder->list.next, | 
 | 188 | 					    struct padata_priv, list); | 
 | 189 |  | 
 | 190 | 			seq_nr  = padata->seq_nr; | 
 | 191 | 			BUG_ON(calc_seq_nr != seq_nr); | 
 | 192 | 		} else { | 
 | 193 | 			seq_nr = calc_seq_nr; | 
 | 194 | 			empty++; | 
 | 195 | 		} | 
 | 196 |  | 
 | 197 | 		if (next_nr < 0 || seq_nr < next_nr | 
 | 198 | 		    || (next_overrun && !overrun)) { | 
 | 199 | 			next_nr = seq_nr; | 
 | 200 | 			next_overrun = overrun; | 
 | 201 | 			next_queue = queue; | 
 | 202 | 		} | 
 | 203 | 	} | 
 | 204 |  | 
 | 205 | 	padata = NULL; | 
 | 206 |  | 
 | 207 | 	if (empty == num_cpus) | 
 | 208 | 		goto out; | 
 | 209 |  | 
 | 210 | 	reorder = &next_queue->reorder; | 
 | 211 |  | 
 | 212 | 	if (!list_empty(&reorder->list)) { | 
 | 213 | 		padata = list_entry(reorder->list.next, | 
 | 214 | 				    struct padata_priv, list); | 
 | 215 |  | 
 | 216 | 		if (unlikely(next_overrun)) { | 
 | 217 | 			for_each_cpu(cpu, pd->cpumask) { | 
 | 218 | 				queue = per_cpu_ptr(pd->queue, cpu); | 
 | 219 | 				atomic_set(&queue->num_obj, 0); | 
 | 220 | 			} | 
 | 221 | 		} | 
 | 222 |  | 
 | 223 | 		spin_lock(&reorder->lock); | 
 | 224 | 		list_del_init(&padata->list); | 
 | 225 | 		atomic_dec(&pd->reorder_objects); | 
 | 226 | 		spin_unlock(&reorder->lock); | 
 | 227 |  | 
 | 228 | 		atomic_inc(&next_queue->num_obj); | 
 | 229 |  | 
 | 230 | 		goto out; | 
 | 231 | 	} | 
 | 232 |  | 
 | 233 | 	if (next_nr % num_cpus == next_queue->cpu_index) { | 
 | 234 | 		padata = ERR_PTR(-ENODATA); | 
 | 235 | 		goto out; | 
 | 236 | 	} | 
 | 237 |  | 
 | 238 | 	padata = ERR_PTR(-EINPROGRESS); | 
 | 239 | out: | 
 | 240 | 	return padata; | 
 | 241 | } | 
 | 242 |  | 
 | 243 | static void padata_reorder(struct parallel_data *pd) | 
 | 244 | { | 
 | 245 | 	struct padata_priv *padata; | 
 | 246 | 	struct padata_queue *queue; | 
 | 247 | 	struct padata_instance *pinst = pd->pinst; | 
 | 248 |  | 
 | 249 | try_again: | 
 | 250 | 	if (!spin_trylock_bh(&pd->lock)) | 
 | 251 | 		goto out; | 
 | 252 |  | 
 | 253 | 	while (1) { | 
 | 254 | 		padata = padata_get_next(pd); | 
 | 255 |  | 
 | 256 | 		if (!padata || PTR_ERR(padata) == -EINPROGRESS) | 
 | 257 | 			break; | 
 | 258 |  | 
 | 259 | 		if (PTR_ERR(padata) == -ENODATA) { | 
 | 260 | 			spin_unlock_bh(&pd->lock); | 
 | 261 | 			goto out; | 
 | 262 | 		} | 
 | 263 |  | 
 | 264 | 		queue = per_cpu_ptr(pd->queue, padata->cb_cpu); | 
 | 265 |  | 
 | 266 | 		spin_lock(&queue->serial.lock); | 
 | 267 | 		list_add_tail(&padata->list, &queue->serial.list); | 
 | 268 | 		spin_unlock(&queue->serial.lock); | 
 | 269 |  | 
 | 270 | 		queue_work_on(padata->cb_cpu, pinst->wq, &queue->swork); | 
 | 271 | 	} | 
 | 272 |  | 
 | 273 | 	spin_unlock_bh(&pd->lock); | 
 | 274 |  | 
 | 275 | 	if (atomic_read(&pd->reorder_objects)) | 
 | 276 | 		goto try_again; | 
 | 277 |  | 
 | 278 | out: | 
 | 279 | 	return; | 
 | 280 | } | 
 | 281 |  | 
 | 282 | static void padata_serial_worker(struct work_struct *work) | 
 | 283 | { | 
 | 284 | 	struct padata_queue *queue; | 
 | 285 | 	struct parallel_data *pd; | 
 | 286 | 	LIST_HEAD(local_list); | 
 | 287 |  | 
 | 288 | 	local_bh_disable(); | 
 | 289 | 	queue = container_of(work, struct padata_queue, swork); | 
 | 290 | 	pd = queue->pd; | 
 | 291 |  | 
 | 292 | 	spin_lock(&queue->serial.lock); | 
 | 293 | 	list_replace_init(&queue->serial.list, &local_list); | 
 | 294 | 	spin_unlock(&queue->serial.lock); | 
 | 295 |  | 
 | 296 | 	while (!list_empty(&local_list)) { | 
 | 297 | 		struct padata_priv *padata; | 
 | 298 |  | 
 | 299 | 		padata = list_entry(local_list.next, | 
 | 300 | 				    struct padata_priv, list); | 
 | 301 |  | 
 | 302 | 		list_del_init(&padata->list); | 
 | 303 |  | 
 | 304 | 		padata->serial(padata); | 
 | 305 | 		atomic_dec(&pd->refcnt); | 
 | 306 | 	} | 
 | 307 | 	local_bh_enable(); | 
 | 308 | } | 
 | 309 |  | 
 | 310 | /* | 
 | 311 |  * padata_do_serial - padata serialization function | 
 | 312 |  * | 
 | 313 |  * @padata: object to be serialized. | 
 | 314 |  * | 
 | 315 |  * padata_do_serial must be called for every parallelized object. | 
 | 316 |  * The serialization callback function will run with BHs off. | 
 | 317 |  */ | 
 | 318 | void padata_do_serial(struct padata_priv *padata) | 
 | 319 | { | 
 | 320 | 	int cpu; | 
 | 321 | 	struct padata_queue *queue; | 
 | 322 | 	struct parallel_data *pd; | 
 | 323 |  | 
 | 324 | 	pd = padata->pd; | 
 | 325 |  | 
 | 326 | 	cpu = get_cpu(); | 
 | 327 | 	queue = per_cpu_ptr(pd->queue, cpu); | 
 | 328 |  | 
 | 329 | 	spin_lock(&queue->reorder.lock); | 
 | 330 | 	atomic_inc(&pd->reorder_objects); | 
 | 331 | 	list_add_tail(&padata->list, &queue->reorder.list); | 
 | 332 | 	spin_unlock(&queue->reorder.lock); | 
 | 333 |  | 
 | 334 | 	put_cpu(); | 
 | 335 |  | 
 | 336 | 	padata_reorder(pd); | 
 | 337 | } | 
 | 338 | EXPORT_SYMBOL(padata_do_serial); | 
 | 339 |  | 
 | 340 | static struct parallel_data *padata_alloc_pd(struct padata_instance *pinst, | 
 | 341 | 					     const struct cpumask *cpumask) | 
 | 342 | { | 
 | 343 | 	int cpu, cpu_index, num_cpus; | 
 | 344 | 	struct padata_queue *queue; | 
 | 345 | 	struct parallel_data *pd; | 
 | 346 |  | 
 | 347 | 	cpu_index = 0; | 
 | 348 |  | 
 | 349 | 	pd = kzalloc(sizeof(struct parallel_data), GFP_KERNEL); | 
 | 350 | 	if (!pd) | 
 | 351 | 		goto err; | 
 | 352 |  | 
 | 353 | 	pd->queue = alloc_percpu(struct padata_queue); | 
 | 354 | 	if (!pd->queue) | 
 | 355 | 		goto err_free_pd; | 
 | 356 |  | 
 | 357 | 	if (!alloc_cpumask_var(&pd->cpumask, GFP_KERNEL)) | 
 | 358 | 		goto err_free_queue; | 
 | 359 |  | 
 | 360 | 	for_each_possible_cpu(cpu) { | 
 | 361 | 		queue = per_cpu_ptr(pd->queue, cpu); | 
 | 362 |  | 
 | 363 | 		queue->pd = pd; | 
 | 364 |  | 
 | 365 | 		if (cpumask_test_cpu(cpu, cpumask) | 
 | 366 | 		    && cpumask_test_cpu(cpu, cpu_active_mask)) { | 
 | 367 | 			queue->cpu_index = cpu_index; | 
 | 368 | 			cpu_index++; | 
 | 369 | 		} else | 
 | 370 | 			queue->cpu_index = -1; | 
 | 371 |  | 
 | 372 | 		INIT_LIST_HEAD(&queue->reorder.list); | 
 | 373 | 		INIT_LIST_HEAD(&queue->parallel.list); | 
 | 374 | 		INIT_LIST_HEAD(&queue->serial.list); | 
 | 375 | 		spin_lock_init(&queue->reorder.lock); | 
 | 376 | 		spin_lock_init(&queue->parallel.lock); | 
 | 377 | 		spin_lock_init(&queue->serial.lock); | 
 | 378 |  | 
 | 379 | 		INIT_WORK(&queue->pwork, padata_parallel_worker); | 
 | 380 | 		INIT_WORK(&queue->swork, padata_serial_worker); | 
 | 381 | 		atomic_set(&queue->num_obj, 0); | 
 | 382 | 	} | 
 | 383 |  | 
 | 384 | 	cpumask_and(pd->cpumask, cpumask, cpu_active_mask); | 
 | 385 |  | 
 | 386 | 	num_cpus = cpumask_weight(pd->cpumask); | 
 | 387 | 	pd->max_seq_nr = (MAX_SEQ_NR / num_cpus) * num_cpus - 1; | 
 | 388 |  | 
 | 389 | 	atomic_set(&pd->seq_nr, -1); | 
 | 390 | 	atomic_set(&pd->reorder_objects, 0); | 
 | 391 | 	atomic_set(&pd->refcnt, 0); | 
 | 392 | 	pd->pinst = pinst; | 
 | 393 | 	spin_lock_init(&pd->lock); | 
 | 394 |  | 
 | 395 | 	return pd; | 
 | 396 |  | 
 | 397 | err_free_queue: | 
 | 398 | 	free_percpu(pd->queue); | 
 | 399 | err_free_pd: | 
 | 400 | 	kfree(pd); | 
 | 401 | err: | 
 | 402 | 	return NULL; | 
 | 403 | } | 
 | 404 |  | 
 | 405 | static void padata_free_pd(struct parallel_data *pd) | 
 | 406 | { | 
 | 407 | 	free_cpumask_var(pd->cpumask); | 
 | 408 | 	free_percpu(pd->queue); | 
 | 409 | 	kfree(pd); | 
 | 410 | } | 
 | 411 |  | 
 | 412 | static void padata_replace(struct padata_instance *pinst, | 
 | 413 | 			   struct parallel_data *pd_new) | 
 | 414 | { | 
 | 415 | 	struct parallel_data *pd_old = pinst->pd; | 
 | 416 |  | 
 | 417 | 	pinst->flags |= PADATA_RESET; | 
 | 418 |  | 
 | 419 | 	rcu_assign_pointer(pinst->pd, pd_new); | 
 | 420 |  | 
 | 421 | 	synchronize_rcu(); | 
 | 422 |  | 
 | 423 | 	while (atomic_read(&pd_old->refcnt) != 0) | 
 | 424 | 		yield(); | 
 | 425 |  | 
 | 426 | 	flush_workqueue(pinst->wq); | 
 | 427 |  | 
 | 428 | 	padata_free_pd(pd_old); | 
 | 429 |  | 
 | 430 | 	pinst->flags &= ~PADATA_RESET; | 
 | 431 | } | 
 | 432 |  | 
 | 433 | /* | 
 | 434 |  * padata_set_cpumask - set the cpumask that padata should use | 
 | 435 |  * | 
 | 436 |  * @pinst: padata instance | 
 | 437 |  * @cpumask: the cpumask to use | 
 | 438 |  */ | 
 | 439 | int padata_set_cpumask(struct padata_instance *pinst, | 
 | 440 | 			cpumask_var_t cpumask) | 
 | 441 | { | 
 | 442 | 	struct parallel_data *pd; | 
 | 443 | 	int err = 0; | 
 | 444 |  | 
 | 445 | 	might_sleep(); | 
 | 446 |  | 
 | 447 | 	mutex_lock(&pinst->lock); | 
 | 448 |  | 
 | 449 | 	pd = padata_alloc_pd(pinst, cpumask); | 
 | 450 | 	if (!pd) { | 
 | 451 | 		err = -ENOMEM; | 
 | 452 | 		goto out; | 
 | 453 | 	} | 
 | 454 |  | 
 | 455 | 	cpumask_copy(pinst->cpumask, cpumask); | 
 | 456 |  | 
 | 457 | 	padata_replace(pinst, pd); | 
 | 458 |  | 
 | 459 | out: | 
 | 460 | 	mutex_unlock(&pinst->lock); | 
 | 461 |  | 
 | 462 | 	return err; | 
 | 463 | } | 
 | 464 | EXPORT_SYMBOL(padata_set_cpumask); | 
 | 465 |  | 
 | 466 | static int __padata_add_cpu(struct padata_instance *pinst, int cpu) | 
 | 467 | { | 
 | 468 | 	struct parallel_data *pd; | 
 | 469 |  | 
 | 470 | 	if (cpumask_test_cpu(cpu, cpu_active_mask)) { | 
 | 471 | 		pd = padata_alloc_pd(pinst, pinst->cpumask); | 
 | 472 | 		if (!pd) | 
 | 473 | 			return -ENOMEM; | 
 | 474 |  | 
 | 475 | 		padata_replace(pinst, pd); | 
 | 476 | 	} | 
 | 477 |  | 
 | 478 | 	return 0; | 
 | 479 | } | 
 | 480 |  | 
 | 481 | /* | 
 | 482 |  * padata_add_cpu - add a cpu to the padata cpumask | 
 | 483 |  * | 
 | 484 |  * @pinst: padata instance | 
 | 485 |  * @cpu: cpu to add | 
 | 486 |  */ | 
 | 487 | int padata_add_cpu(struct padata_instance *pinst, int cpu) | 
 | 488 | { | 
 | 489 | 	int err; | 
 | 490 |  | 
 | 491 | 	might_sleep(); | 
 | 492 |  | 
 | 493 | 	mutex_lock(&pinst->lock); | 
 | 494 |  | 
 | 495 | 	cpumask_set_cpu(cpu, pinst->cpumask); | 
 | 496 | 	err = __padata_add_cpu(pinst, cpu); | 
 | 497 |  | 
 | 498 | 	mutex_unlock(&pinst->lock); | 
 | 499 |  | 
 | 500 | 	return err; | 
 | 501 | } | 
 | 502 | EXPORT_SYMBOL(padata_add_cpu); | 
 | 503 |  | 
 | 504 | static int __padata_remove_cpu(struct padata_instance *pinst, int cpu) | 
 | 505 | { | 
 | 506 | 	struct parallel_data *pd; | 
 | 507 |  | 
 | 508 | 	if (cpumask_test_cpu(cpu, cpu_online_mask)) { | 
 | 509 | 		pd = padata_alloc_pd(pinst, pinst->cpumask); | 
 | 510 | 		if (!pd) | 
 | 511 | 			return -ENOMEM; | 
 | 512 |  | 
 | 513 | 		padata_replace(pinst, pd); | 
 | 514 | 	} | 
 | 515 |  | 
 | 516 | 	return 0; | 
 | 517 | } | 
 | 518 |  | 
 | 519 | /* | 
 | 520 |  * padata_remove_cpu - remove a cpu from the padata cpumask | 
 | 521 |  * | 
 | 522 |  * @pinst: padata instance | 
 | 523 |  * @cpu: cpu to remove | 
 | 524 |  */ | 
 | 525 | int padata_remove_cpu(struct padata_instance *pinst, int cpu) | 
 | 526 | { | 
 | 527 | 	int err; | 
 | 528 |  | 
 | 529 | 	might_sleep(); | 
 | 530 |  | 
 | 531 | 	mutex_lock(&pinst->lock); | 
 | 532 |  | 
 | 533 | 	cpumask_clear_cpu(cpu, pinst->cpumask); | 
 | 534 | 	err = __padata_remove_cpu(pinst, cpu); | 
 | 535 |  | 
 | 536 | 	mutex_unlock(&pinst->lock); | 
 | 537 |  | 
 | 538 | 	return err; | 
 | 539 | } | 
 | 540 | EXPORT_SYMBOL(padata_remove_cpu); | 
 | 541 |  | 
 | 542 | /* | 
 | 543 |  * padata_start - start the parallel processing | 
 | 544 |  * | 
 | 545 |  * @pinst: padata instance to start | 
 | 546 |  */ | 
 | 547 | void padata_start(struct padata_instance *pinst) | 
 | 548 | { | 
 | 549 | 	might_sleep(); | 
 | 550 |  | 
 | 551 | 	mutex_lock(&pinst->lock); | 
 | 552 | 	pinst->flags |= PADATA_INIT; | 
 | 553 | 	mutex_unlock(&pinst->lock); | 
 | 554 | } | 
 | 555 | EXPORT_SYMBOL(padata_start); | 
 | 556 |  | 
 | 557 | /* | 
 | 558 |  * padata_stop - stop the parallel processing | 
 | 559 |  * | 
 | 560 |  * @pinst: padata instance to stop | 
 | 561 |  */ | 
 | 562 | void padata_stop(struct padata_instance *pinst) | 
 | 563 | { | 
 | 564 | 	might_sleep(); | 
 | 565 |  | 
 | 566 | 	mutex_lock(&pinst->lock); | 
 | 567 | 	pinst->flags &= ~PADATA_INIT; | 
 | 568 | 	mutex_unlock(&pinst->lock); | 
 | 569 | } | 
 | 570 | EXPORT_SYMBOL(padata_stop); | 
 | 571 |  | 
 | 572 | static int __cpuinit padata_cpu_callback(struct notifier_block *nfb, | 
 | 573 | 					 unsigned long action, void *hcpu) | 
 | 574 | { | 
 | 575 | 	int err; | 
 | 576 | 	struct padata_instance *pinst; | 
 | 577 | 	int cpu = (unsigned long)hcpu; | 
 | 578 |  | 
 | 579 | 	pinst = container_of(nfb, struct padata_instance, cpu_notifier); | 
 | 580 |  | 
 | 581 | 	switch (action) { | 
 | 582 | 	case CPU_ONLINE: | 
 | 583 | 	case CPU_ONLINE_FROZEN: | 
 | 584 | 		if (!cpumask_test_cpu(cpu, pinst->cpumask)) | 
 | 585 | 			break; | 
 | 586 | 		mutex_lock(&pinst->lock); | 
 | 587 | 		err = __padata_add_cpu(pinst, cpu); | 
 | 588 | 		mutex_unlock(&pinst->lock); | 
 | 589 | 		if (err) | 
 | 590 | 			return NOTIFY_BAD; | 
 | 591 | 		break; | 
 | 592 |  | 
 | 593 | 	case CPU_DOWN_PREPARE: | 
 | 594 | 	case CPU_DOWN_PREPARE_FROZEN: | 
 | 595 | 		if (!cpumask_test_cpu(cpu, pinst->cpumask)) | 
 | 596 | 			break; | 
 | 597 | 		mutex_lock(&pinst->lock); | 
 | 598 | 		err = __padata_remove_cpu(pinst, cpu); | 
 | 599 | 		mutex_unlock(&pinst->lock); | 
 | 600 | 		if (err) | 
 | 601 | 			return NOTIFY_BAD; | 
 | 602 | 		break; | 
 | 603 |  | 
 | 604 | 	case CPU_UP_CANCELED: | 
 | 605 | 	case CPU_UP_CANCELED_FROZEN: | 
 | 606 | 		if (!cpumask_test_cpu(cpu, pinst->cpumask)) | 
 | 607 | 			break; | 
 | 608 | 		mutex_lock(&pinst->lock); | 
 | 609 | 		__padata_remove_cpu(pinst, cpu); | 
 | 610 | 		mutex_unlock(&pinst->lock); | 
 | 611 |  | 
 | 612 | 	case CPU_DOWN_FAILED: | 
 | 613 | 	case CPU_DOWN_FAILED_FROZEN: | 
 | 614 | 		if (!cpumask_test_cpu(cpu, pinst->cpumask)) | 
 | 615 | 			break; | 
 | 616 | 		mutex_lock(&pinst->lock); | 
 | 617 | 		__padata_add_cpu(pinst, cpu); | 
 | 618 | 		mutex_unlock(&pinst->lock); | 
 | 619 | 	} | 
 | 620 |  | 
 | 621 | 	return NOTIFY_OK; | 
 | 622 | } | 
 | 623 |  | 
 | 624 | /* | 
 | 625 |  * padata_alloc - allocate and initialize a padata instance | 
 | 626 |  * | 
 | 627 |  * @cpumask: cpumask that padata uses for parallelization | 
 | 628 |  * @wq: workqueue to use for the allocated padata instance | 
 | 629 |  */ | 
 | 630 | struct padata_instance *padata_alloc(const struct cpumask *cpumask, | 
 | 631 | 				     struct workqueue_struct *wq) | 
 | 632 | { | 
 | 633 | 	int err; | 
 | 634 | 	struct padata_instance *pinst; | 
 | 635 | 	struct parallel_data *pd; | 
 | 636 |  | 
 | 637 | 	pinst = kzalloc(sizeof(struct padata_instance), GFP_KERNEL); | 
 | 638 | 	if (!pinst) | 
 | 639 | 		goto err; | 
 | 640 |  | 
 | 641 | 	pd = padata_alloc_pd(pinst, cpumask); | 
 | 642 | 	if (!pd) | 
 | 643 | 		goto err_free_inst; | 
 | 644 |  | 
| Steffen Klassert | 7478138 | 2010-03-04 13:30:22 +0800 | [diff] [blame] | 645 | 	if (!alloc_cpumask_var(&pinst->cpumask, GFP_KERNEL)) | 
 | 646 | 		goto err_free_pd; | 
 | 647 |  | 
| Steffen Klassert | 16295be | 2010-01-06 19:47:10 +1100 | [diff] [blame] | 648 | 	rcu_assign_pointer(pinst->pd, pd); | 
 | 649 |  | 
 | 650 | 	pinst->wq = wq; | 
 | 651 |  | 
 | 652 | 	cpumask_copy(pinst->cpumask, cpumask); | 
 | 653 |  | 
 | 654 | 	pinst->flags = 0; | 
 | 655 |  | 
 | 656 | 	pinst->cpu_notifier.notifier_call = padata_cpu_callback; | 
 | 657 | 	pinst->cpu_notifier.priority = 0; | 
 | 658 | 	err = register_hotcpu_notifier(&pinst->cpu_notifier); | 
 | 659 | 	if (err) | 
| Steffen Klassert | 7478138 | 2010-03-04 13:30:22 +0800 | [diff] [blame] | 660 | 		goto err_free_cpumask; | 
| Steffen Klassert | 16295be | 2010-01-06 19:47:10 +1100 | [diff] [blame] | 661 |  | 
 | 662 | 	mutex_init(&pinst->lock); | 
 | 663 |  | 
 | 664 | 	return pinst; | 
 | 665 |  | 
| Steffen Klassert | 7478138 | 2010-03-04 13:30:22 +0800 | [diff] [blame] | 666 | err_free_cpumask: | 
 | 667 | 	free_cpumask_var(pinst->cpumask); | 
| Steffen Klassert | 16295be | 2010-01-06 19:47:10 +1100 | [diff] [blame] | 668 | err_free_pd: | 
 | 669 | 	padata_free_pd(pd); | 
 | 670 | err_free_inst: | 
 | 671 | 	kfree(pinst); | 
 | 672 | err: | 
 | 673 | 	return NULL; | 
 | 674 | } | 
 | 675 | EXPORT_SYMBOL(padata_alloc); | 
 | 676 |  | 
 | 677 | /* | 
 | 678 |  * padata_free - free a padata instance | 
 | 679 |  * | 
 | 680 |  * @ padata_inst: padata instance to free | 
 | 681 |  */ | 
 | 682 | void padata_free(struct padata_instance *pinst) | 
 | 683 | { | 
 | 684 | 	padata_stop(pinst); | 
 | 685 |  | 
 | 686 | 	synchronize_rcu(); | 
 | 687 |  | 
 | 688 | 	while (atomic_read(&pinst->pd->refcnt) != 0) | 
 | 689 | 		yield(); | 
 | 690 |  | 
 | 691 | 	unregister_hotcpu_notifier(&pinst->cpu_notifier); | 
 | 692 | 	padata_free_pd(pinst->pd); | 
| Steffen Klassert | 7478138 | 2010-03-04 13:30:22 +0800 | [diff] [blame] | 693 | 	free_cpumask_var(pinst->cpumask); | 
| Steffen Klassert | 16295be | 2010-01-06 19:47:10 +1100 | [diff] [blame] | 694 | 	kfree(pinst); | 
 | 695 | } | 
 | 696 | EXPORT_SYMBOL(padata_free); |