blob: 5df82186fc93ea12ab9f33c85e4665e5a4022b8e [file] [log] [blame]
Linus Torvalds1da177e2005-04-16 15:20:36 -07001/*
2 * PPC64 (POWER4) Huge TLB Page Support for Kernel.
3 *
4 * Copyright (C) 2003 David Gibson, IBM Corporation.
5 *
6 * Based on the IA-32 version:
7 * Copyright (C) 2002, Rohit Seth <rohit.seth@intel.com>
8 */
9
10#include <linux/init.h>
11#include <linux/fs.h>
12#include <linux/mm.h>
13#include <linux/hugetlb.h>
14#include <linux/pagemap.h>
Linus Torvalds1da177e2005-04-16 15:20:36 -070015#include <linux/slab.h>
16#include <linux/err.h>
17#include <linux/sysctl.h>
18#include <asm/mman.h>
19#include <asm/pgalloc.h>
20#include <asm/tlb.h>
21#include <asm/tlbflush.h>
22#include <asm/mmu_context.h>
23#include <asm/machdep.h>
24#include <asm/cputable.h>
Benjamin Herrenschmidt94b2a432007-03-10 00:05:37 +010025#include <asm/spu.h>
Linus Torvalds1da177e2005-04-16 15:20:36 -070026
Jon Tollefson4ec161c2008-01-04 09:59:50 +110027#define HPAGE_SHIFT_64K 16
28#define HPAGE_SHIFT_16M 24
29
David Gibsonc594ada2005-08-11 16:55:21 +100030#define NUM_LOW_AREAS (0x100000000UL >> SID_SHIFT)
31#define NUM_HIGH_AREAS (PGTABLE_RANGE >> HTLB_AREA_SHIFT)
Jon Tollefsonec4b2c02008-07-23 21:27:53 -070032#define MAX_NUMBER_GPAGES 1024
33
34/* Tracks the 16G pages after the device tree is scanned and before the
35 * huge_boot_pages list is ready. */
36static unsigned long gpage_freearray[MAX_NUMBER_GPAGES];
37static unsigned nr_gpages;
David Gibsonc594ada2005-08-11 16:55:21 +100038
Jon Tollefson4ec161c2008-01-04 09:59:50 +110039unsigned int hugepte_shift;
40#define PTRS_PER_HUGEPTE (1 << hugepte_shift)
41#define HUGEPTE_TABLE_SIZE (sizeof(pte_t) << hugepte_shift)
David Gibsonf10a04c2006-04-28 15:02:51 +100042
Jon Tollefson4ec161c2008-01-04 09:59:50 +110043#define HUGEPD_SHIFT (HPAGE_SHIFT + hugepte_shift)
David Gibsonf10a04c2006-04-28 15:02:51 +100044#define HUGEPD_SIZE (1UL << HUGEPD_SHIFT)
45#define HUGEPD_MASK (~(HUGEPD_SIZE-1))
46
47#define huge_pgtable_cache (pgtable_cache[HUGEPTE_CACHE_NUM])
48
49/* Flag to mark huge PD pointers. This means pmd_bad() and pud_bad()
50 * will choke on pointers to hugepte tables, which is handy for
51 * catching screwups early. */
52#define HUGEPD_OK 0x1
53
54typedef struct { unsigned long pd; } hugepd_t;
55
56#define hugepd_none(hpd) ((hpd).pd == 0)
57
58static inline pte_t *hugepd_page(hugepd_t hpd)
59{
60 BUG_ON(!(hpd.pd & HUGEPD_OK));
61 return (pte_t *)(hpd.pd & ~HUGEPD_OK);
62}
63
64static inline pte_t *hugepte_offset(hugepd_t *hpdp, unsigned long addr)
65{
66 unsigned long idx = ((addr >> HPAGE_SHIFT) & (PTRS_PER_HUGEPTE-1));
67 pte_t *dir = hugepd_page(*hpdp);
68
69 return dir + idx;
70}
71
72static int __hugepte_alloc(struct mm_struct *mm, hugepd_t *hpdp,
73 unsigned long address)
74{
75 pte_t *new = kmem_cache_alloc(huge_pgtable_cache,
76 GFP_KERNEL|__GFP_REPEAT);
77
78 if (! new)
79 return -ENOMEM;
80
81 spin_lock(&mm->page_table_lock);
82 if (!hugepd_none(*hpdp))
83 kmem_cache_free(huge_pgtable_cache, new);
84 else
85 hpdp->pd = (unsigned long)new | HUGEPD_OK;
86 spin_unlock(&mm->page_table_lock);
87 return 0;
88}
89
Jon Tollefson4ec161c2008-01-04 09:59:50 +110090/* Base page size affects how we walk hugetlb page tables */
91#ifdef CONFIG_PPC_64K_PAGES
92#define hpmd_offset(pud, addr) pmd_offset(pud, addr)
93#define hpmd_alloc(mm, pud, addr) pmd_alloc(mm, pud, addr)
94#else
95static inline
96pmd_t *hpmd_offset(pud_t *pud, unsigned long addr)
97{
98 if (HPAGE_SHIFT == HPAGE_SHIFT_64K)
99 return pmd_offset(pud, addr);
100 else
101 return (pmd_t *) pud;
102}
103static inline
104pmd_t *hpmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long addr)
105{
106 if (HPAGE_SHIFT == HPAGE_SHIFT_64K)
107 return pmd_alloc(mm, pud, addr);
108 else
109 return (pmd_t *) pud;
110}
111#endif
112
Jon Tollefsonec4b2c02008-07-23 21:27:53 -0700113/* Moves the gigantic page addresses from the temporary list to the
114 * huge_boot_pages list. */
115int alloc_bootmem_huge_page(struct hstate *h)
116{
117 struct huge_bootmem_page *m;
118 if (nr_gpages == 0)
119 return 0;
120 m = phys_to_virt(gpage_freearray[--nr_gpages]);
121 gpage_freearray[nr_gpages] = 0;
122 list_add(&m->list, &huge_boot_pages);
123 m->hstate = h;
124 return 1;
125}
126
127
David Gibsone28f7fa2005-08-05 19:39:06 +1000128/* Modelled after find_linux_pte() */
129pte_t *huge_pte_offset(struct mm_struct *mm, unsigned long addr)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700130{
David Gibsone28f7fa2005-08-05 19:39:06 +1000131 pgd_t *pg;
132 pud_t *pu;
Jon Tollefson4ec161c2008-01-04 09:59:50 +1100133 pmd_t *pm;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700134
Benjamin Herrenschmidtd0f13e32007-05-08 16:27:27 +1000135 BUG_ON(get_slice_psize(mm, addr) != mmu_huge_psize);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700136
David Gibsone28f7fa2005-08-05 19:39:06 +1000137 addr &= HPAGE_MASK;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700138
David Gibsone28f7fa2005-08-05 19:39:06 +1000139 pg = pgd_offset(mm, addr);
140 if (!pgd_none(*pg)) {
141 pu = pud_offset(pg, addr);
142 if (!pud_none(*pu)) {
Jon Tollefson4ec161c2008-01-04 09:59:50 +1100143 pm = hpmd_offset(pu, addr);
David Gibsonf10a04c2006-04-28 15:02:51 +1000144 if (!pmd_none(*pm))
145 return hugepte_offset((hugepd_t *)pm, addr);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700146 }
147 }
148
David Gibsone28f7fa2005-08-05 19:39:06 +1000149 return NULL;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700150}
151
Andi Kleena5516432008-07-23 21:27:41 -0700152pte_t *huge_pte_alloc(struct mm_struct *mm,
153 unsigned long addr, unsigned long sz)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700154{
David Gibsone28f7fa2005-08-05 19:39:06 +1000155 pgd_t *pg;
156 pud_t *pu;
Jon Tollefson4ec161c2008-01-04 09:59:50 +1100157 pmd_t *pm;
David Gibsonf10a04c2006-04-28 15:02:51 +1000158 hugepd_t *hpdp = NULL;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700159
Benjamin Herrenschmidtd0f13e32007-05-08 16:27:27 +1000160 BUG_ON(get_slice_psize(mm, addr) != mmu_huge_psize);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700161
David Gibsone28f7fa2005-08-05 19:39:06 +1000162 addr &= HPAGE_MASK;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700163
David Gibsone28f7fa2005-08-05 19:39:06 +1000164 pg = pgd_offset(mm, addr);
165 pu = pud_alloc(mm, pg, addr);
166
167 if (pu) {
Jon Tollefson4ec161c2008-01-04 09:59:50 +1100168 pm = hpmd_alloc(mm, pu, addr);
David Gibsonf10a04c2006-04-28 15:02:51 +1000169 if (pm)
170 hpdp = (hugepd_t *)pm;
David Gibsone28f7fa2005-08-05 19:39:06 +1000171 }
172
David Gibsonf10a04c2006-04-28 15:02:51 +1000173 if (! hpdp)
174 return NULL;
175
176 if (hugepd_none(*hpdp) && __hugepte_alloc(mm, hpdp, addr))
177 return NULL;
178
179 return hugepte_offset(hpdp, addr);
180}
181
Chen, Kenneth W39dde652006-12-06 20:32:03 -0800182int huge_pmd_unshare(struct mm_struct *mm, unsigned long *addr, pte_t *ptep)
183{
184 return 0;
185}
186
David Gibsonf10a04c2006-04-28 15:02:51 +1000187static void free_hugepte_range(struct mmu_gather *tlb, hugepd_t *hpdp)
188{
189 pte_t *hugepte = hugepd_page(*hpdp);
190
191 hpdp->pd = 0;
192 tlb->need_flush = 1;
193 pgtable_free_tlb(tlb, pgtable_free_cache(hugepte, HUGEPTE_CACHE_NUM,
Adam Litkec9169f82006-08-18 11:22:21 -0700194 PGF_CACHENUM_MASK));
David Gibsonf10a04c2006-04-28 15:02:51 +1000195}
196
David Gibsonf10a04c2006-04-28 15:02:51 +1000197static void hugetlb_free_pmd_range(struct mmu_gather *tlb, pud_t *pud,
198 unsigned long addr, unsigned long end,
199 unsigned long floor, unsigned long ceiling)
200{
201 pmd_t *pmd;
202 unsigned long next;
203 unsigned long start;
204
205 start = addr;
206 pmd = pmd_offset(pud, addr);
207 do {
208 next = pmd_addr_end(addr, end);
209 if (pmd_none(*pmd))
210 continue;
211 free_hugepte_range(tlb, (hugepd_t *)pmd);
212 } while (pmd++, addr = next, addr != end);
213
214 start &= PUD_MASK;
215 if (start < floor)
216 return;
217 if (ceiling) {
218 ceiling &= PUD_MASK;
219 if (!ceiling)
220 return;
221 }
222 if (end - 1 > ceiling - 1)
223 return;
224
225 pmd = pmd_offset(pud, start);
226 pud_clear(pud);
227 pmd_free_tlb(tlb, pmd);
228}
David Gibsonf10a04c2006-04-28 15:02:51 +1000229
230static void hugetlb_free_pud_range(struct mmu_gather *tlb, pgd_t *pgd,
231 unsigned long addr, unsigned long end,
232 unsigned long floor, unsigned long ceiling)
233{
234 pud_t *pud;
235 unsigned long next;
236 unsigned long start;
237
238 start = addr;
239 pud = pud_offset(pgd, addr);
240 do {
241 next = pud_addr_end(addr, end);
242#ifdef CONFIG_PPC_64K_PAGES
243 if (pud_none_or_clear_bad(pud))
244 continue;
245 hugetlb_free_pmd_range(tlb, pud, addr, next, floor, ceiling);
246#else
Jon Tollefson4ec161c2008-01-04 09:59:50 +1100247 if (HPAGE_SHIFT == HPAGE_SHIFT_64K) {
248 if (pud_none_or_clear_bad(pud))
249 continue;
250 hugetlb_free_pmd_range(tlb, pud, addr, next, floor, ceiling);
251 } else {
252 if (pud_none(*pud))
253 continue;
254 free_hugepte_range(tlb, (hugepd_t *)pud);
255 }
David Gibsonf10a04c2006-04-28 15:02:51 +1000256#endif
257 } while (pud++, addr = next, addr != end);
258
259 start &= PGDIR_MASK;
260 if (start < floor)
261 return;
262 if (ceiling) {
263 ceiling &= PGDIR_MASK;
264 if (!ceiling)
265 return;
266 }
267 if (end - 1 > ceiling - 1)
268 return;
269
270 pud = pud_offset(pgd, start);
271 pgd_clear(pgd);
272 pud_free_tlb(tlb, pud);
273}
274
275/*
276 * This function frees user-level page tables of a process.
277 *
278 * Must be called with pagetable lock held.
279 */
Jan Beulich42b77722008-07-23 21:27:10 -0700280void hugetlb_free_pgd_range(struct mmu_gather *tlb,
David Gibsonf10a04c2006-04-28 15:02:51 +1000281 unsigned long addr, unsigned long end,
282 unsigned long floor, unsigned long ceiling)
283{
284 pgd_t *pgd;
285 unsigned long next;
286 unsigned long start;
287
288 /*
289 * Comments below take from the normal free_pgd_range(). They
290 * apply here too. The tests against HUGEPD_MASK below are
291 * essential, because we *don't* test for this at the bottom
292 * level. Without them we'll attempt to free a hugepte table
293 * when we unmap just part of it, even if there are other
294 * active mappings using it.
295 *
296 * The next few lines have given us lots of grief...
297 *
298 * Why are we testing HUGEPD* at this top level? Because
299 * often there will be no work to do at all, and we'd prefer
300 * not to go all the way down to the bottom just to discover
301 * that.
302 *
303 * Why all these "- 1"s? Because 0 represents both the bottom
304 * of the address space and the top of it (using -1 for the
305 * top wouldn't help much: the masks would do the wrong thing).
306 * The rule is that addr 0 and floor 0 refer to the bottom of
307 * the address space, but end 0 and ceiling 0 refer to the top
308 * Comparisons need to use "end - 1" and "ceiling - 1" (though
309 * that end 0 case should be mythical).
310 *
311 * Wherever addr is brought up or ceiling brought down, we
312 * must be careful to reject "the opposite 0" before it
313 * confuses the subsequent tests. But what about where end is
314 * brought down by HUGEPD_SIZE below? no, end can't go down to
315 * 0 there.
316 *
317 * Whereas we round start (addr) and ceiling down, by different
318 * masks at different levels, in order to test whether a table
319 * now has no other vmas using it, so can be freed, we don't
320 * bother to round floor or end up - the tests don't need that.
321 */
322
323 addr &= HUGEPD_MASK;
324 if (addr < floor) {
325 addr += HUGEPD_SIZE;
326 if (!addr)
327 return;
328 }
329 if (ceiling) {
330 ceiling &= HUGEPD_MASK;
331 if (!ceiling)
332 return;
333 }
334 if (end - 1 > ceiling - 1)
335 end -= HUGEPD_SIZE;
336 if (addr > end - 1)
337 return;
338
339 start = addr;
Jan Beulich42b77722008-07-23 21:27:10 -0700340 pgd = pgd_offset(tlb->mm, addr);
David Gibsonf10a04c2006-04-28 15:02:51 +1000341 do {
Jan Beulich42b77722008-07-23 21:27:10 -0700342 BUG_ON(get_slice_psize(tlb->mm, addr) != mmu_huge_psize);
David Gibsonf10a04c2006-04-28 15:02:51 +1000343 next = pgd_addr_end(addr, end);
344 if (pgd_none_or_clear_bad(pgd))
345 continue;
Jan Beulich42b77722008-07-23 21:27:10 -0700346 hugetlb_free_pud_range(tlb, pgd, addr, next, floor, ceiling);
David Gibsonf10a04c2006-04-28 15:02:51 +1000347 } while (pgd++, addr = next, addr != end);
David Gibsone28f7fa2005-08-05 19:39:06 +1000348}
349
David Gibsone28f7fa2005-08-05 19:39:06 +1000350void set_huge_pte_at(struct mm_struct *mm, unsigned long addr,
351 pte_t *ptep, pte_t pte)
352{
David Gibsone28f7fa2005-08-05 19:39:06 +1000353 if (pte_present(*ptep)) {
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100354 /* We open-code pte_clear because we need to pass the right
Benjamin Herrenschmidta741e672007-04-10 17:09:37 +1000355 * argument to hpte_need_flush (huge / !huge). Might not be
356 * necessary anymore if we make hpte_need_flush() get the
357 * page size from the slices
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100358 */
Benjamin Herrenschmidta741e672007-04-10 17:09:37 +1000359 pte_update(mm, addr & HPAGE_MASK, ptep, ~0UL, 1);
David Gibsone28f7fa2005-08-05 19:39:06 +1000360 }
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100361 *ptep = __pte(pte_val(pte) & ~_PAGE_HPTEFLAGS);
David Gibsone28f7fa2005-08-05 19:39:06 +1000362}
363
364pte_t huge_ptep_get_and_clear(struct mm_struct *mm, unsigned long addr,
365 pte_t *ptep)
366{
Benjamin Herrenschmidta741e672007-04-10 17:09:37 +1000367 unsigned long old = pte_update(mm, addr, ptep, ~0UL, 1);
David Gibsone28f7fa2005-08-05 19:39:06 +1000368 return __pte(old);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700369}
370
Linus Torvalds1da177e2005-04-16 15:20:36 -0700371struct page *
372follow_huge_addr(struct mm_struct *mm, unsigned long address, int write)
373{
374 pte_t *ptep;
375 struct page *page;
376
Benjamin Herrenschmidtd0f13e32007-05-08 16:27:27 +1000377 if (get_slice_psize(mm, address) != mmu_huge_psize)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700378 return ERR_PTR(-EINVAL);
379
380 ptep = huge_pte_offset(mm, address);
381 page = pte_page(*ptep);
382 if (page)
383 page += (address % HPAGE_SIZE) / PAGE_SIZE;
384
385 return page;
386}
387
388int pmd_huge(pmd_t pmd)
389{
390 return 0;
391}
392
Andi Kleenceb86872008-07-23 21:27:50 -0700393int pud_huge(pud_t pud)
394{
395 return 0;
396}
397
Linus Torvalds1da177e2005-04-16 15:20:36 -0700398struct page *
399follow_huge_pmd(struct mm_struct *mm, unsigned long address,
400 pmd_t *pmd, int write)
401{
402 BUG();
403 return NULL;
404}
405
Linus Torvalds1da177e2005-04-16 15:20:36 -0700406
407unsigned long hugetlb_get_unmapped_area(struct file *file, unsigned long addr,
408 unsigned long len, unsigned long pgoff,
409 unsigned long flags)
410{
Benjamin Herrenschmidtd0f13e32007-05-08 16:27:27 +1000411 return slice_get_unmapped_area(addr, len, flags,
412 mmu_huge_psize, 1, 0);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700413}
414
David Gibsoncbf52af2005-12-09 14:20:52 +1100415/*
416 * Called by asm hashtable.S for doing lazy icache flush
417 */
418static unsigned int hash_huge_page_do_lazy_icache(unsigned long rflags,
419 pte_t pte, int trap)
420{
421 struct page *page;
422 int i;
423
424 if (!pfn_valid(pte_pfn(pte)))
425 return rflags;
426
427 page = pte_page(pte);
428
429 /* page is dirty */
430 if (!test_bit(PG_arch_1, &page->flags) && !PageReserved(page)) {
431 if (trap == 0x400) {
432 for (i = 0; i < (HPAGE_SIZE / PAGE_SIZE); i++)
433 __flush_dcache_icache(page_address(page+i));
434 set_bit(PG_arch_1, &page->flags);
435 } else {
436 rflags |= HPTE_R_N;
437 }
438 }
439 return rflags;
440}
441
Linus Torvalds1da177e2005-04-16 15:20:36 -0700442int hash_huge_page(struct mm_struct *mm, unsigned long access,
David Gibsoncbf52af2005-12-09 14:20:52 +1100443 unsigned long ea, unsigned long vsid, int local,
444 unsigned long trap)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700445{
446 pte_t *ptep;
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100447 unsigned long old_pte, new_pte;
448 unsigned long va, rflags, pa;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700449 long slot;
450 int err = 1;
Paul Mackerras1189be62007-10-11 20:37:10 +1000451 int ssize = user_segment_size(ea);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700452
Linus Torvalds1da177e2005-04-16 15:20:36 -0700453 ptep = huge_pte_offset(mm, ea);
454
455 /* Search the Linux page table for a match with va */
Paul Mackerras1189be62007-10-11 20:37:10 +1000456 va = hpt_va(ea, vsid, ssize);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700457
458 /*
459 * If no pte found or not present, send the problem up to
460 * do_page_fault
461 */
462 if (unlikely(!ptep || pte_none(*ptep)))
463 goto out;
464
Linus Torvalds1da177e2005-04-16 15:20:36 -0700465 /*
466 * Check the user's access rights to the page. If access should be
467 * prevented then send the problem up to do_page_fault.
468 */
469 if (unlikely(access & ~pte_val(*ptep)))
470 goto out;
471 /*
472 * At this point, we have a pte (old_pte) which can be used to build
473 * or update an HPTE. There are 2 cases:
474 *
475 * 1. There is a valid (present) pte with no associated HPTE (this is
476 * the most common case)
477 * 2. There is a valid (present) pte with an associated HPTE. The
478 * current values of the pp bits in the HPTE prevent access
479 * because we are doing software DIRTY bit management and the
480 * page is currently not DIRTY.
481 */
482
483
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100484 do {
485 old_pte = pte_val(*ptep);
486 if (old_pte & _PAGE_BUSY)
487 goto out;
Benjamin Herrenschmidt41743a42008-06-11 15:37:10 +1000488 new_pte = old_pte | _PAGE_BUSY | _PAGE_ACCESSED;
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100489 } while(old_pte != __cmpxchg_u64((unsigned long *)ptep,
490 old_pte, new_pte));
Linus Torvalds1da177e2005-04-16 15:20:36 -0700491
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100492 rflags = 0x2 | (!(new_pte & _PAGE_RW));
Linus Torvalds1da177e2005-04-16 15:20:36 -0700493 /* _PAGE_EXEC -> HW_NO_EXEC since it's inverted */
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100494 rflags |= ((new_pte & _PAGE_EXEC) ? 0 : HPTE_R_N);
David Gibsoncbf52af2005-12-09 14:20:52 +1100495 if (!cpu_has_feature(CPU_FTR_COHERENT_ICACHE))
496 /* No CPU has hugepages but lacks no execute, so we
497 * don't need to worry about that case */
498 rflags = hash_huge_page_do_lazy_icache(rflags, __pte(old_pte),
499 trap);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700500
501 /* Check if pte already has an hpte (case 2) */
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100502 if (unlikely(old_pte & _PAGE_HASHPTE)) {
Linus Torvalds1da177e2005-04-16 15:20:36 -0700503 /* There MIGHT be an HPTE for this pte */
504 unsigned long hash, slot;
505
Paul Mackerras1189be62007-10-11 20:37:10 +1000506 hash = hpt_hash(va, HPAGE_SHIFT, ssize);
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100507 if (old_pte & _PAGE_F_SECOND)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700508 hash = ~hash;
509 slot = (hash & htab_hash_mask) * HPTES_PER_GROUP;
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100510 slot += (old_pte & _PAGE_F_GIX) >> 12;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700511
Benjamin Herrenschmidt325c82a2005-12-08 16:51:44 +1100512 if (ppc_md.hpte_updatepp(slot, rflags, va, mmu_huge_psize,
Paul Mackerras1189be62007-10-11 20:37:10 +1000513 ssize, local) == -1)
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100514 old_pte &= ~_PAGE_HPTEFLAGS;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700515 }
516
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100517 if (likely(!(old_pte & _PAGE_HASHPTE))) {
Paul Mackerras1189be62007-10-11 20:37:10 +1000518 unsigned long hash = hpt_hash(va, HPAGE_SHIFT, ssize);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700519 unsigned long hpte_group;
520
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100521 pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700522
523repeat:
524 hpte_group = ((hash & htab_hash_mask) *
525 HPTES_PER_GROUP) & ~0x7UL;
526
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100527 /* clear HPTE slot informations in new PTE */
Benjamin Herrenschmidt41743a42008-06-11 15:37:10 +1000528#ifdef CONFIG_PPC_64K_PAGES
529 new_pte = (new_pte & ~_PAGE_HPTEFLAGS) | _PAGE_HPTE_SUB0;
530#else
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100531 new_pte = (new_pte & ~_PAGE_HPTEFLAGS) | _PAGE_HASHPTE;
Benjamin Herrenschmidt41743a42008-06-11 15:37:10 +1000532#endif
Linus Torvalds1da177e2005-04-16 15:20:36 -0700533 /* Add in WIMG bits */
Dave Kleikamp87e9ab12008-06-19 08:32:56 +1000534 rflags |= (new_pte & (_PAGE_WRITETHRU | _PAGE_NO_CACHE |
535 _PAGE_COHERENT | _PAGE_GUARDED));
Linus Torvalds1da177e2005-04-16 15:20:36 -0700536
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100537 /* Insert into the hash table, primary slot */
538 slot = ppc_md.hpte_insert(hpte_group, va, pa, rflags, 0,
Paul Mackerras1189be62007-10-11 20:37:10 +1000539 mmu_huge_psize, ssize);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700540
541 /* Primary is full, try the secondary */
542 if (unlikely(slot == -1)) {
Linus Torvalds1da177e2005-04-16 15:20:36 -0700543 hpte_group = ((~hash & htab_hash_mask) *
544 HPTES_PER_GROUP) & ~0x7UL;
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100545 slot = ppc_md.hpte_insert(hpte_group, va, pa, rflags,
Benjamin Herrenschmidt67b10812005-09-23 13:24:07 -0700546 HPTE_V_SECONDARY,
Paul Mackerras1189be62007-10-11 20:37:10 +1000547 mmu_huge_psize, ssize);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700548 if (slot == -1) {
549 if (mftb() & 0x1)
Benjamin Herrenschmidt67b10812005-09-23 13:24:07 -0700550 hpte_group = ((hash & htab_hash_mask) *
551 HPTES_PER_GROUP)&~0x7UL;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700552
553 ppc_md.hpte_remove(hpte_group);
554 goto repeat;
555 }
556 }
557
558 if (unlikely(slot == -2))
559 panic("hash_huge_page: pte_insert failed\n");
560
Ishizaki Koud649bd72007-01-12 09:54:39 +0900561 new_pte |= (slot << 12) & (_PAGE_F_SECOND | _PAGE_F_GIX);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700562 }
563
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100564 /*
Hugh Dickins01edcd82005-11-23 13:37:39 -0800565 * No need to use ldarx/stdcx here
Benjamin Herrenschmidt3c726f82005-11-07 11:06:55 +1100566 */
567 *ptep = __pte(new_pte & ~_PAGE_BUSY);
568
Linus Torvalds1da177e2005-04-16 15:20:36 -0700569 err = 0;
570
571 out:
Linus Torvalds1da177e2005-04-16 15:20:36 -0700572 return err;
573}
David Gibsonf10a04c2006-04-28 15:02:51 +1000574
Jon Tollefson4ec161c2008-01-04 09:59:50 +1100575void set_huge_psize(int psize)
576{
577 /* Check that it is a page size supported by the hardware and
578 * that it fits within pagetable limits. */
579 if (mmu_psize_defs[psize].shift && mmu_psize_defs[psize].shift < SID_SHIFT &&
580 (mmu_psize_defs[psize].shift > MIN_HUGEPTE_SHIFT ||
581 mmu_psize_defs[psize].shift == HPAGE_SHIFT_64K)) {
582 HPAGE_SHIFT = mmu_psize_defs[psize].shift;
583 mmu_huge_psize = psize;
584#ifdef CONFIG_PPC_64K_PAGES
585 hugepte_shift = (PMD_SHIFT-HPAGE_SHIFT);
586#else
587 if (HPAGE_SHIFT == HPAGE_SHIFT_64K)
588 hugepte_shift = (PMD_SHIFT-HPAGE_SHIFT);
589 else
590 hugepte_shift = (PUD_SHIFT-HPAGE_SHIFT);
591#endif
592
593 } else
594 HPAGE_SHIFT = 0;
595}
596
597static int __init hugepage_setup_sz(char *str)
598{
599 unsigned long long size;
600 int mmu_psize = -1;
601 int shift;
602
603 size = memparse(str, &str);
604
605 shift = __ffs(size);
606 switch (shift) {
607#ifndef CONFIG_PPC_64K_PAGES
608 case HPAGE_SHIFT_64K:
609 mmu_psize = MMU_PAGE_64K;
610 break;
611#endif
612 case HPAGE_SHIFT_16M:
613 mmu_psize = MMU_PAGE_16M;
614 break;
615 }
616
617 if (mmu_psize >=0 && mmu_psize_defs[mmu_psize].shift)
618 set_huge_psize(mmu_psize);
619 else
620 printk(KERN_WARNING "Invalid huge page size specified(%llu)\n", size);
621
622 return 1;
623}
624__setup("hugepagesz=", hugepage_setup_sz);
625
Christoph Lameter4ba9b9d2007-10-16 23:25:51 -0700626static void zero_ctor(struct kmem_cache *cache, void *addr)
David Gibsonf10a04c2006-04-28 15:02:51 +1000627{
628 memset(addr, 0, kmem_cache_size(cache));
629}
630
631static int __init hugetlbpage_init(void)
632{
633 if (!cpu_has_feature(CPU_FTR_16M_PAGE))
634 return -ENODEV;
635
636 huge_pgtable_cache = kmem_cache_create("hugepte_cache",
637 HUGEPTE_TABLE_SIZE,
638 HUGEPTE_TABLE_SIZE,
Christoph Lameterf0f39802007-05-06 14:49:58 -0700639 0,
Paul Mundt20c2df82007-07-20 10:11:58 +0900640 zero_ctor);
David Gibsonf10a04c2006-04-28 15:02:51 +1000641 if (! huge_pgtable_cache)
642 panic("hugetlbpage_init(): could not create hugepte cache\n");
643
644 return 0;
645}
646
647module_init(hugetlbpage_init);