David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 1 | /* arch/sparc64/kernel/ktlb.S: Kernel mapping TLB miss handling. |
| 2 | * |
| 3 | * Copyright (C) 1995, 1997, 2005 David S. Miller <davem@davemloft.net> |
| 4 | * Copyright (C) 1996 Eddie C. Dost (ecd@brainaid.de) |
| 5 | * Copyright (C) 1996 Miguel de Icaza (miguel@nuclecu.unam.mx) |
| 6 | * Copyright (C) 1996,98,99 Jakub Jelinek (jj@sunsite.mff.cuni.cz) |
| 7 | */ |
| 8 | |
| 9 | #include <linux/config.h> |
| 10 | #include <asm/head.h> |
| 11 | #include <asm/asi.h> |
| 12 | #include <asm/page.h> |
| 13 | #include <asm/pgtable.h> |
| 14 | |
| 15 | .text |
| 16 | .align 32 |
| 17 | |
David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 18 | /* |
| 19 | * On a second level vpte miss, check whether the original fault is to the OBP |
| 20 | * range (note that this is only possible for instruction miss, data misses to |
| 21 | * obp range do not use vpte). If so, go back directly to the faulting address. |
| 22 | * This is because we want to read the tpc, otherwise we have no way of knowing |
| 23 | * the 8k aligned faulting address if we are using >8k kernel pagesize. This |
| 24 | * also ensures no vpte range addresses are dropped into tlb while obp is |
| 25 | * executing (see inherit_locked_prom_mappings() rant). |
| 26 | */ |
| 27 | sparc64_vpte_nucleus: |
| 28 | /* Note that kvmap below has verified that the address is |
| 29 | * in the range MODULES_VADDR --> VMALLOC_END already. So |
| 30 | * here we need only check if it is an OBP address or not. |
| 31 | */ |
| 32 | sethi %hi(LOW_OBP_ADDRESS), %g5 |
| 33 | cmp %g4, %g5 |
David S. Miller | 1ac4f5e | 2005-09-21 21:49:32 -0700 | [diff] [blame] | 34 | blu,pn %xcc, kern_vpte |
David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 35 | mov 0x1, %g5 |
| 36 | sllx %g5, 32, %g5 |
| 37 | cmp %g4, %g5 |
David S. Miller | 1ac4f5e | 2005-09-21 21:49:32 -0700 | [diff] [blame] | 38 | blu,pn %xcc, vpte_insn_obp |
David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 39 | nop |
| 40 | |
| 41 | /* These two instructions are patched by paginig_init(). */ |
David S. Miller | 1ac4f5e | 2005-09-21 21:49:32 -0700 | [diff] [blame] | 42 | kern_vpte: |
| 43 | sethi %hi(swapper_pgd_zero), %g5 |
| 44 | lduw [%g5 + %lo(swapper_pgd_zero)], %g5 |
David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 45 | |
| 46 | /* With kernel PGD in %g5, branch back into dtlb_backend. */ |
| 47 | ba,pt %xcc, sparc64_kpte_continue |
| 48 | andn %g1, 0x3, %g1 /* Finish PMD offset adjustment. */ |
| 49 | |
| 50 | vpte_noent: |
| 51 | /* Restore previous TAG_ACCESS, %g5 is zero, and we will |
| 52 | * skip over the trap instruction so that the top level |
| 53 | * TLB miss handler will thing this %g5 value is just an |
| 54 | * invalid PTE, thus branching to full fault processing. |
| 55 | */ |
| 56 | mov TLB_SFSR, %g1 |
| 57 | stxa %g4, [%g1 + %g1] ASI_DMMU |
| 58 | done |
| 59 | |
David S. Miller | 1ac4f5e | 2005-09-21 21:49:32 -0700 | [diff] [blame] | 60 | vpte_insn_obp: |
David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 61 | /* Behave as if we are at TL0. */ |
| 62 | wrpr %g0, 1, %tl |
| 63 | rdpr %tpc, %g4 /* Find original faulting iaddr */ |
| 64 | srlx %g4, 13, %g4 /* Throw out context bits */ |
| 65 | sllx %g4, 13, %g4 /* g4 has vpn + ctx0 now */ |
| 66 | |
| 67 | /* Restore previous TAG_ACCESS. */ |
| 68 | mov TLB_SFSR, %g1 |
| 69 | stxa %g4, [%g1 + %g1] ASI_IMMU |
| 70 | |
David S. Miller | c9c1083 | 2005-10-12 12:22:46 -0700 | [diff] [blame^] | 71 | sethi %hi(prom_trans), %g5 |
| 72 | or %g5, %lo(prom_trans), %g5 |
David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 73 | |
David S. Miller | c9c1083 | 2005-10-12 12:22:46 -0700 | [diff] [blame^] | 74 | 1: ldx [%g5 + 0x00], %g6 ! base |
| 75 | brz,a,pn %g6, longpath ! no more entries, fail |
| 76 | mov TLB_SFSR, %g1 ! and restore %g1 |
| 77 | ldx [%g5 + 0x08], %g1 ! len |
| 78 | add %g6, %g1, %g1 ! end |
| 79 | cmp %g6, %g4 |
| 80 | bgu,pt %xcc, 2f |
| 81 | cmp %g4, %g1 |
| 82 | bgeu,pt %xcc, 2f |
| 83 | ldx [%g5 + 0x10], %g1 ! PTE |
David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 84 | |
David S. Miller | c9c1083 | 2005-10-12 12:22:46 -0700 | [diff] [blame^] | 85 | /* TLB load, restore %g1, and return from trap. */ |
| 86 | sub %g4, %g6, %g6 |
| 87 | add %g1, %g6, %g5 |
| 88 | mov TLB_SFSR, %g1 |
David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 89 | stxa %g5, [%g0] ASI_ITLB_DATA_IN |
| 90 | retry |
| 91 | |
David S. Miller | c9c1083 | 2005-10-12 12:22:46 -0700 | [diff] [blame^] | 92 | 2: ba,pt %xcc, 1b |
| 93 | add %g5, (3 * 8), %g5 ! next entry |
| 94 | |
David S. Miller | 1ac4f5e | 2005-09-21 21:49:32 -0700 | [diff] [blame] | 95 | kvmap_do_obp: |
David S. Miller | c9c1083 | 2005-10-12 12:22:46 -0700 | [diff] [blame^] | 96 | sethi %hi(prom_trans), %g5 |
| 97 | or %g5, %lo(prom_trans), %g5 |
| 98 | srlx %g4, 13, %g4 |
| 99 | sllx %g4, 13, %g4 |
David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 100 | |
David S. Miller | c9c1083 | 2005-10-12 12:22:46 -0700 | [diff] [blame^] | 101 | 1: ldx [%g5 + 0x00], %g6 ! base |
| 102 | brz,a,pn %g6, longpath ! no more entries, fail |
| 103 | mov TLB_SFSR, %g1 ! and restore %g1 |
| 104 | ldx [%g5 + 0x08], %g1 ! len |
| 105 | add %g6, %g1, %g1 ! end |
| 106 | cmp %g6, %g4 |
| 107 | bgu,pt %xcc, 2f |
| 108 | cmp %g4, %g1 |
| 109 | bgeu,pt %xcc, 2f |
| 110 | ldx [%g5 + 0x10], %g1 ! PTE |
David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 111 | |
David S. Miller | c9c1083 | 2005-10-12 12:22:46 -0700 | [diff] [blame^] | 112 | /* TLB load, restore %g1, and return from trap. */ |
| 113 | sub %g4, %g6, %g6 |
| 114 | add %g1, %g6, %g5 |
| 115 | mov TLB_SFSR, %g1 |
David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 116 | stxa %g5, [%g0] ASI_DTLB_DATA_IN |
| 117 | retry |
| 118 | |
David S. Miller | c9c1083 | 2005-10-12 12:22:46 -0700 | [diff] [blame^] | 119 | 2: ba,pt %xcc, 1b |
| 120 | add %g5, (3 * 8), %g5 ! next entry |
| 121 | |
David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 122 | /* |
| 123 | * On a first level data miss, check whether this is to the OBP range (note |
| 124 | * that such accesses can be made by prom, as well as by kernel using |
| 125 | * prom_getproperty on "address"), and if so, do not use vpte access ... |
| 126 | * rather, use information saved during inherit_prom_mappings() using 8k |
| 127 | * pagesize. |
| 128 | */ |
| 129 | .align 32 |
| 130 | kvmap: |
David S. Miller | 5642530 | 2005-09-25 16:46:57 -0700 | [diff] [blame] | 131 | brgez,pn %g4, kvmap_nonlinear |
| 132 | nop |
| 133 | |
| 134 | #ifdef CONFIG_DEBUG_PAGEALLOC |
| 135 | .globl kvmap_linear_patch |
| 136 | kvmap_linear_patch: |
| 137 | #endif |
| 138 | ba,pt %xcc, kvmap_load |
David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 139 | xor %g2, %g4, %g5 |
| 140 | |
David S. Miller | 5642530 | 2005-09-25 16:46:57 -0700 | [diff] [blame] | 141 | #ifdef CONFIG_DEBUG_PAGEALLOC |
| 142 | sethi %hi(swapper_pg_dir), %g5 |
| 143 | or %g5, %lo(swapper_pg_dir), %g5 |
| 144 | sllx %g4, 64 - (PGDIR_SHIFT + PGDIR_BITS), %g6 |
| 145 | srlx %g6, 64 - PAGE_SHIFT, %g6 |
| 146 | andn %g6, 0x3, %g6 |
| 147 | lduw [%g5 + %g6], %g5 |
| 148 | brz,pn %g5, longpath |
| 149 | sllx %g4, 64 - (PMD_SHIFT + PMD_BITS), %g6 |
| 150 | srlx %g6, 64 - PAGE_SHIFT, %g6 |
| 151 | sllx %g5, 11, %g5 |
| 152 | andn %g6, 0x3, %g6 |
| 153 | lduwa [%g5 + %g6] ASI_PHYS_USE_EC, %g5 |
| 154 | brz,pn %g5, longpath |
| 155 | sllx %g4, 64 - PMD_SHIFT, %g6 |
| 156 | srlx %g6, 64 - PAGE_SHIFT, %g6 |
| 157 | sllx %g5, 11, %g5 |
| 158 | andn %g6, 0x7, %g6 |
| 159 | ldxa [%g5 + %g6] ASI_PHYS_USE_EC, %g5 |
| 160 | brz,pn %g5, longpath |
| 161 | nop |
| 162 | ba,a,pt %xcc, kvmap_load |
| 163 | #endif |
| 164 | |
David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 165 | kvmap_nonlinear: |
| 166 | sethi %hi(MODULES_VADDR), %g5 |
| 167 | cmp %g4, %g5 |
| 168 | blu,pn %xcc, longpath |
| 169 | mov (VMALLOC_END >> 24), %g5 |
| 170 | sllx %g5, 24, %g5 |
| 171 | cmp %g4, %g5 |
| 172 | bgeu,pn %xcc, longpath |
| 173 | nop |
| 174 | |
| 175 | kvmap_check_obp: |
| 176 | sethi %hi(LOW_OBP_ADDRESS), %g5 |
| 177 | cmp %g4, %g5 |
| 178 | blu,pn %xcc, kvmap_vmalloc_addr |
| 179 | mov 0x1, %g5 |
| 180 | sllx %g5, 32, %g5 |
| 181 | cmp %g4, %g5 |
David S. Miller | 1ac4f5e | 2005-09-21 21:49:32 -0700 | [diff] [blame] | 182 | blu,pn %xcc, kvmap_do_obp |
David S. Miller | 2a7e299 | 2005-09-21 18:50:51 -0700 | [diff] [blame] | 183 | nop |
| 184 | |
| 185 | kvmap_vmalloc_addr: |
| 186 | /* If we get here, a vmalloc addr was accessed, load kernel VPTE. */ |
| 187 | ldxa [%g3 + %g6] ASI_N, %g5 |
| 188 | brgez,pn %g5, longpath |
| 189 | nop |
| 190 | |
| 191 | kvmap_load: |
| 192 | /* PTE is valid, load into TLB and return from trap. */ |
| 193 | stxa %g5, [%g0] ASI_DTLB_DATA_IN ! Reload TLB |
| 194 | retry |