| Christoph Lameter | 6225e93 | 2007-05-06 14:49:50 -0700 | [diff] [blame] | 1 | /* | 
 | 2 |  * Quicklist support. | 
 | 3 |  * | 
 | 4 |  * Quicklists are light weight lists of pages that have a defined state | 
 | 5 |  * on alloc and free. Pages must be in the quicklist specific defined state | 
 | 6 |  * (zero by default) when the page is freed. It seems that the initial idea | 
 | 7 |  * for such lists first came from Dave Miller and then various other people | 
 | 8 |  * improved on it. | 
 | 9 |  * | 
 | 10 |  * Copyright (C) 2007 SGI, | 
 | 11 |  * 	Christoph Lameter <clameter@sgi.com> | 
 | 12 |  * 		Generalized, added support for multiple lists and | 
 | 13 |  * 		constructors / destructors. | 
 | 14 |  */ | 
 | 15 | #include <linux/kernel.h> | 
 | 16 |  | 
 | 17 | #include <linux/mm.h> | 
 | 18 | #include <linux/mmzone.h> | 
 | 19 | #include <linux/module.h> | 
 | 20 | #include <linux/quicklist.h> | 
 | 21 |  | 
 | 22 | DEFINE_PER_CPU(struct quicklist, quicklist)[CONFIG_NR_QUICK]; | 
 | 23 |  | 
 | 24 | #define FRACTION_OF_NODE_MEM	16 | 
 | 25 |  | 
 | 26 | static unsigned long max_pages(unsigned long min_pages) | 
 | 27 | { | 
 | 28 | 	unsigned long node_free_pages, max; | 
| Christoph Lameter | 96990a4 | 2008-01-14 00:55:14 -0800 | [diff] [blame] | 29 | 	struct zone *zones = NODE_DATA(numa_node_id())->node_zones; | 
| Christoph Lameter | 6225e93 | 2007-05-06 14:49:50 -0700 | [diff] [blame] | 30 |  | 
| Christoph Lameter | 96990a4 | 2008-01-14 00:55:14 -0800 | [diff] [blame] | 31 | 	node_free_pages = | 
 | 32 | #ifdef CONFIG_ZONE_DMA | 
 | 33 | 		zone_page_state(&zones[ZONE_DMA], NR_FREE_PAGES) + | 
 | 34 | #endif | 
 | 35 | #ifdef CONFIG_ZONE_DMA32 | 
 | 36 | 		zone_page_state(&zones[ZONE_DMA32], NR_FREE_PAGES) + | 
 | 37 | #endif | 
 | 38 | 		zone_page_state(&zones[ZONE_NORMAL], NR_FREE_PAGES); | 
 | 39 |  | 
| Christoph Lameter | 6225e93 | 2007-05-06 14:49:50 -0700 | [diff] [blame] | 40 | 	max = node_free_pages / FRACTION_OF_NODE_MEM; | 
 | 41 | 	return max(max, min_pages); | 
 | 42 | } | 
 | 43 |  | 
 | 44 | static long min_pages_to_free(struct quicklist *q, | 
 | 45 | 	unsigned long min_pages, long max_free) | 
 | 46 | { | 
 | 47 | 	long pages_to_free; | 
 | 48 |  | 
 | 49 | 	pages_to_free = q->nr_pages - max_pages(min_pages); | 
 | 50 |  | 
 | 51 | 	return min(pages_to_free, max_free); | 
 | 52 | } | 
 | 53 |  | 
 | 54 | /* | 
 | 55 |  * Trim down the number of pages in the quicklist | 
 | 56 |  */ | 
 | 57 | void quicklist_trim(int nr, void (*dtor)(void *), | 
 | 58 | 	unsigned long min_pages, unsigned long max_free) | 
 | 59 | { | 
 | 60 | 	long pages_to_free; | 
 | 61 | 	struct quicklist *q; | 
 | 62 |  | 
 | 63 | 	q = &get_cpu_var(quicklist)[nr]; | 
 | 64 | 	if (q->nr_pages > min_pages) { | 
 | 65 | 		pages_to_free = min_pages_to_free(q, min_pages, max_free); | 
 | 66 |  | 
 | 67 | 		while (pages_to_free > 0) { | 
 | 68 | 			/* | 
 | 69 | 			 * We pass a gfp_t of 0 to quicklist_alloc here | 
 | 70 | 			 * because we will never call into the page allocator. | 
 | 71 | 			 */ | 
 | 72 | 			void *p = quicklist_alloc(nr, 0, NULL); | 
 | 73 |  | 
 | 74 | 			if (dtor) | 
 | 75 | 				dtor(p); | 
 | 76 | 			free_page((unsigned long)p); | 
 | 77 | 			pages_to_free--; | 
 | 78 | 		} | 
 | 79 | 	} | 
 | 80 | 	put_cpu_var(quicklist); | 
 | 81 | } | 
 | 82 |  | 
 | 83 | unsigned long quicklist_total_size(void) | 
 | 84 | { | 
 | 85 | 	unsigned long count = 0; | 
 | 86 | 	int cpu; | 
 | 87 | 	struct quicklist *ql, *q; | 
 | 88 |  | 
 | 89 | 	for_each_online_cpu(cpu) { | 
 | 90 | 		ql = per_cpu(quicklist, cpu); | 
 | 91 | 		for (q = ql; q < ql + CONFIG_NR_QUICK; q++) | 
 | 92 | 			count += q->nr_pages; | 
 | 93 | 	} | 
 | 94 | 	return count; | 
 | 95 | } | 
 | 96 |  |