| Roman Zippel | 2855b97 | 2005-09-03 15:57:09 -0700 | [diff] [blame] | 1 | /* | 
|  | 2 | *  linux/arch/m68k/mm/cache.c | 
|  | 3 | * | 
|  | 4 | *  Instruction cache handling | 
|  | 5 | * | 
|  | 6 | *  Copyright (C) 1995  Hamish Macdonald | 
|  | 7 | */ | 
|  | 8 |  | 
|  | 9 | #include <linux/module.h> | 
|  | 10 | #include <asm/pgalloc.h> | 
|  | 11 | #include <asm/traps.h> | 
|  | 12 |  | 
|  | 13 |  | 
|  | 14 | static unsigned long virt_to_phys_slow(unsigned long vaddr) | 
|  | 15 | { | 
|  | 16 | if (CPU_IS_060) { | 
|  | 17 | unsigned long paddr; | 
|  | 18 |  | 
|  | 19 | /* The PLPAR instruction causes an access error if the translation | 
|  | 20 | * is not possible. To catch this we use the same exception mechanism | 
|  | 21 | * as for user space accesses in <asm/uaccess.h>. */ | 
|  | 22 | asm volatile (".chip 68060\n" | 
|  | 23 | "1: plpar (%0)\n" | 
|  | 24 | ".chip 68k\n" | 
|  | 25 | "2:\n" | 
|  | 26 | ".section .fixup,\"ax\"\n" | 
|  | 27 | "   .even\n" | 
|  | 28 | "3: sub.l %0,%0\n" | 
|  | 29 | "   jra 2b\n" | 
|  | 30 | ".previous\n" | 
|  | 31 | ".section __ex_table,\"a\"\n" | 
|  | 32 | "   .align 4\n" | 
|  | 33 | "   .long 1b,3b\n" | 
|  | 34 | ".previous" | 
|  | 35 | : "=a" (paddr) | 
|  | 36 | : "0" (vaddr)); | 
|  | 37 | return paddr; | 
|  | 38 | } else if (CPU_IS_040) { | 
|  | 39 | unsigned long mmusr; | 
|  | 40 |  | 
|  | 41 | asm volatile (".chip 68040\n\t" | 
|  | 42 | "ptestr (%1)\n\t" | 
|  | 43 | "movec %%mmusr, %0\n\t" | 
|  | 44 | ".chip 68k" | 
|  | 45 | : "=r" (mmusr) | 
|  | 46 | : "a" (vaddr)); | 
|  | 47 |  | 
|  | 48 | if (mmusr & MMU_R_040) | 
|  | 49 | return (mmusr & PAGE_MASK) | (vaddr & ~PAGE_MASK); | 
|  | 50 | } else { | 
|  | 51 | unsigned short mmusr; | 
|  | 52 | unsigned long *descaddr; | 
|  | 53 |  | 
|  | 54 | asm volatile ("ptestr %3,%2@,#7,%0\n\t" | 
|  | 55 | "pmove %%psr,%1@" | 
|  | 56 | : "=a&" (descaddr) | 
|  | 57 | : "a" (&mmusr), "a" (vaddr), "d" (get_fs().seg)); | 
|  | 58 | if (mmusr & (MMU_I|MMU_B|MMU_L)) | 
|  | 59 | return 0; | 
|  | 60 | descaddr = phys_to_virt((unsigned long)descaddr); | 
|  | 61 | switch (mmusr & MMU_NUM) { | 
|  | 62 | case 1: | 
|  | 63 | return (*descaddr & 0xfe000000) | (vaddr & 0x01ffffff); | 
|  | 64 | case 2: | 
|  | 65 | return (*descaddr & 0xfffc0000) | (vaddr & 0x0003ffff); | 
|  | 66 | case 3: | 
|  | 67 | return (*descaddr & PAGE_MASK) | (vaddr & ~PAGE_MASK); | 
|  | 68 | } | 
|  | 69 | } | 
|  | 70 | return 0; | 
|  | 71 | } | 
|  | 72 |  | 
|  | 73 | /* Push n pages at kernel virtual address and clear the icache */ | 
|  | 74 | /* RZ: use cpush %bc instead of cpush %dc, cinv %ic */ | 
|  | 75 | void flush_icache_range(unsigned long address, unsigned long endaddr) | 
|  | 76 | { | 
|  | 77 |  | 
|  | 78 | if (CPU_IS_040_OR_060) { | 
|  | 79 | address &= PAGE_MASK; | 
|  | 80 |  | 
|  | 81 | do { | 
|  | 82 | asm volatile ("nop\n\t" | 
|  | 83 | ".chip 68040\n\t" | 
|  | 84 | "cpushp %%bc,(%0)\n\t" | 
|  | 85 | ".chip 68k" | 
|  | 86 | : : "a" (virt_to_phys_slow(address))); | 
|  | 87 | address += PAGE_SIZE; | 
|  | 88 | } while (address < endaddr); | 
|  | 89 | } else { | 
|  | 90 | unsigned long tmp; | 
|  | 91 | asm volatile ("movec %%cacr,%0\n\t" | 
|  | 92 | "orw %1,%0\n\t" | 
|  | 93 | "movec %0,%%cacr" | 
|  | 94 | : "=&d" (tmp) | 
|  | 95 | : "di" (FLUSH_I)); | 
|  | 96 | } | 
|  | 97 | } | 
|  | 98 | EXPORT_SYMBOL(flush_icache_range); | 
|  | 99 |  | 
|  | 100 | void flush_icache_user_range(struct vm_area_struct *vma, struct page *page, | 
|  | 101 | unsigned long addr, int len) | 
|  | 102 | { | 
|  | 103 | if (CPU_IS_040_OR_060) { | 
|  | 104 | asm volatile ("nop\n\t" | 
|  | 105 | ".chip 68040\n\t" | 
|  | 106 | "cpushp %%bc,(%0)\n\t" | 
|  | 107 | ".chip 68k" | 
|  | 108 | : : "a" (page_to_phys(page))); | 
|  | 109 | } else { | 
|  | 110 | unsigned long tmp; | 
|  | 111 | asm volatile ("movec %%cacr,%0\n\t" | 
|  | 112 | "orw %1,%0\n\t" | 
|  | 113 | "movec %0,%%cacr" | 
|  | 114 | : "=&d" (tmp) | 
|  | 115 | : "di" (FLUSH_I)); | 
|  | 116 | } | 
|  | 117 | } | 
|  | 118 |  |