| Magnus Damm | e7cc9a7 | 2008-02-07 20:18:21 +0900 | [diff] [blame] | 1 | /* | 
 | 2 |  * Trapped io support | 
 | 3 |  * | 
 | 4 |  * Copyright (C) 2008 Magnus Damm | 
 | 5 |  * | 
 | 6 |  * Intercept io operations by trapping. | 
 | 7 |  * | 
 | 8 |  * This file is subject to the terms and conditions of the GNU General Public | 
 | 9 |  * License.  See the file "COPYING" in the main directory of this archive | 
 | 10 |  * for more details. | 
 | 11 |  */ | 
 | 12 | #include <linux/kernel.h> | 
 | 13 | #include <linux/mm.h> | 
 | 14 | #include <linux/bitops.h> | 
 | 15 | #include <linux/vmalloc.h> | 
| Paul Mundt | ecc14e8 | 2008-02-12 16:02:02 +0900 | [diff] [blame] | 16 | #include <linux/module.h> | 
| Magnus Damm | e7cc9a7 | 2008-02-07 20:18:21 +0900 | [diff] [blame] | 17 | #include <asm/system.h> | 
 | 18 | #include <asm/mmu_context.h> | 
 | 19 | #include <asm/uaccess.h> | 
 | 20 | #include <asm/io.h> | 
 | 21 | #include <asm/io_trapped.h> | 
 | 22 |  | 
 | 23 | #define TRAPPED_PAGES_MAX 16 | 
| Magnus Damm | e7cc9a7 | 2008-02-07 20:18:21 +0900 | [diff] [blame] | 24 |  | 
 | 25 | #ifdef CONFIG_HAS_IOPORT | 
 | 26 | LIST_HEAD(trapped_io); | 
| Paul Mundt | ecc14e8 | 2008-02-12 16:02:02 +0900 | [diff] [blame] | 27 | EXPORT_SYMBOL_GPL(trapped_io); | 
| Magnus Damm | e7cc9a7 | 2008-02-07 20:18:21 +0900 | [diff] [blame] | 28 | #endif | 
 | 29 | #ifdef CONFIG_HAS_IOMEM | 
 | 30 | LIST_HEAD(trapped_mem); | 
| Paul Mundt | ecc14e8 | 2008-02-12 16:02:02 +0900 | [diff] [blame] | 31 | EXPORT_SYMBOL_GPL(trapped_mem); | 
| Magnus Damm | e7cc9a7 | 2008-02-07 20:18:21 +0900 | [diff] [blame] | 32 | #endif | 
 | 33 | static DEFINE_SPINLOCK(trapped_lock); | 
 | 34 |  | 
| Paul Mundt | b2839ed | 2008-03-06 12:43:38 +0900 | [diff] [blame] | 35 | int register_trapped_io(struct trapped_io *tiop) | 
| Magnus Damm | e7cc9a7 | 2008-02-07 20:18:21 +0900 | [diff] [blame] | 36 | { | 
 | 37 | 	struct resource *res; | 
 | 38 | 	unsigned long len = 0, flags = 0; | 
 | 39 | 	struct page *pages[TRAPPED_PAGES_MAX]; | 
 | 40 | 	int k, n; | 
 | 41 |  | 
 | 42 | 	/* structure must be page aligned */ | 
 | 43 | 	if ((unsigned long)tiop & (PAGE_SIZE - 1)) | 
 | 44 | 		goto bad; | 
 | 45 |  | 
 | 46 | 	for (k = 0; k < tiop->num_resources; k++) { | 
 | 47 | 		res = tiop->resource + k; | 
 | 48 | 		len += roundup((res->end - res->start) + 1, PAGE_SIZE); | 
 | 49 | 		flags |= res->flags; | 
 | 50 | 	} | 
 | 51 |  | 
 | 52 | 	/* support IORESOURCE_IO _or_ MEM, not both */ | 
 | 53 | 	if (hweight_long(flags) != 1) | 
 | 54 | 		goto bad; | 
 | 55 |  | 
 | 56 | 	n = len >> PAGE_SHIFT; | 
 | 57 |  | 
 | 58 | 	if (n >= TRAPPED_PAGES_MAX) | 
 | 59 | 		goto bad; | 
 | 60 |  | 
 | 61 | 	for (k = 0; k < n; k++) | 
 | 62 | 		pages[k] = virt_to_page(tiop); | 
 | 63 |  | 
 | 64 | 	tiop->virt_base = vmap(pages, n, VM_MAP, PAGE_NONE); | 
 | 65 | 	if (!tiop->virt_base) | 
 | 66 | 		goto bad; | 
 | 67 |  | 
 | 68 | 	len = 0; | 
 | 69 | 	for (k = 0; k < tiop->num_resources; k++) { | 
 | 70 | 		res = tiop->resource + k; | 
 | 71 | 		pr_info("trapped io 0x%08lx overrides %s 0x%08lx\n", | 
 | 72 | 		       (unsigned long)(tiop->virt_base + len), | 
 | 73 | 		       res->flags & IORESOURCE_IO ? "io" : "mmio", | 
 | 74 | 		       (unsigned long)res->start); | 
 | 75 | 		len += roundup((res->end - res->start) + 1, PAGE_SIZE); | 
 | 76 | 	} | 
 | 77 |  | 
 | 78 | 	tiop->magic = IO_TRAPPED_MAGIC; | 
 | 79 | 	INIT_LIST_HEAD(&tiop->list); | 
 | 80 | 	spin_lock_irq(&trapped_lock); | 
 | 81 | 	if (flags & IORESOURCE_IO) | 
 | 82 | 		list_add(&tiop->list, &trapped_io); | 
 | 83 | 	if (flags & IORESOURCE_MEM) | 
 | 84 | 		list_add(&tiop->list, &trapped_mem); | 
 | 85 | 	spin_unlock_irq(&trapped_lock); | 
 | 86 |  | 
 | 87 | 	return 0; | 
 | 88 |  bad: | 
 | 89 | 	pr_warning("unable to install trapped io filter\n"); | 
 | 90 | 	return -1; | 
 | 91 | } | 
| Paul Mundt | ecc14e8 | 2008-02-12 16:02:02 +0900 | [diff] [blame] | 92 | EXPORT_SYMBOL_GPL(register_trapped_io); | 
| Magnus Damm | e7cc9a7 | 2008-02-07 20:18:21 +0900 | [diff] [blame] | 93 |  | 
 | 94 | void __iomem *match_trapped_io_handler(struct list_head *list, | 
 | 95 | 				       unsigned long offset, | 
 | 96 | 				       unsigned long size) | 
 | 97 | { | 
 | 98 | 	unsigned long voffs; | 
 | 99 | 	struct trapped_io *tiop; | 
 | 100 | 	struct resource *res; | 
 | 101 | 	int k, len; | 
 | 102 |  | 
 | 103 | 	spin_lock_irq(&trapped_lock); | 
 | 104 | 	list_for_each_entry(tiop, list, list) { | 
 | 105 | 		voffs = 0; | 
 | 106 | 		for (k = 0; k < tiop->num_resources; k++) { | 
 | 107 | 			res = tiop->resource + k; | 
 | 108 | 			if (res->start == offset) { | 
 | 109 | 				spin_unlock_irq(&trapped_lock); | 
 | 110 | 				return tiop->virt_base + voffs; | 
 | 111 | 			} | 
 | 112 |  | 
 | 113 | 			len = (res->end - res->start) + 1; | 
 | 114 | 			voffs += roundup(len, PAGE_SIZE); | 
 | 115 | 		} | 
 | 116 | 	} | 
 | 117 | 	spin_unlock_irq(&trapped_lock); | 
 | 118 | 	return NULL; | 
 | 119 | } | 
| Paul Mundt | ecc14e8 | 2008-02-12 16:02:02 +0900 | [diff] [blame] | 120 | EXPORT_SYMBOL_GPL(match_trapped_io_handler); | 
| Magnus Damm | e7cc9a7 | 2008-02-07 20:18:21 +0900 | [diff] [blame] | 121 |  | 
 | 122 | static struct trapped_io *lookup_tiop(unsigned long address) | 
 | 123 | { | 
 | 124 | 	pgd_t *pgd_k; | 
 | 125 | 	pud_t *pud_k; | 
 | 126 | 	pmd_t *pmd_k; | 
 | 127 | 	pte_t *pte_k; | 
 | 128 | 	pte_t entry; | 
 | 129 |  | 
 | 130 | 	pgd_k = swapper_pg_dir + pgd_index(address); | 
 | 131 | 	if (!pgd_present(*pgd_k)) | 
 | 132 | 		return NULL; | 
 | 133 |  | 
 | 134 | 	pud_k = pud_offset(pgd_k, address); | 
 | 135 | 	if (!pud_present(*pud_k)) | 
 | 136 | 		return NULL; | 
 | 137 |  | 
 | 138 | 	pmd_k = pmd_offset(pud_k, address); | 
 | 139 | 	if (!pmd_present(*pmd_k)) | 
 | 140 | 		return NULL; | 
 | 141 |  | 
 | 142 | 	pte_k = pte_offset_kernel(pmd_k, address); | 
 | 143 | 	entry = *pte_k; | 
 | 144 |  | 
 | 145 | 	return pfn_to_kaddr(pte_pfn(entry)); | 
 | 146 | } | 
 | 147 |  | 
 | 148 | static unsigned long lookup_address(struct trapped_io *tiop, | 
 | 149 | 				    unsigned long address) | 
 | 150 | { | 
 | 151 | 	struct resource *res; | 
 | 152 | 	unsigned long vaddr = (unsigned long)tiop->virt_base; | 
 | 153 | 	unsigned long len; | 
 | 154 | 	int k; | 
 | 155 |  | 
 | 156 | 	for (k = 0; k < tiop->num_resources; k++) { | 
 | 157 | 		res = tiop->resource + k; | 
 | 158 | 		len = roundup((res->end - res->start) + 1, PAGE_SIZE); | 
 | 159 | 		if (address < (vaddr + len)) | 
 | 160 | 			return res->start + (address - vaddr); | 
 | 161 | 		vaddr += len; | 
 | 162 | 	} | 
 | 163 | 	return 0; | 
 | 164 | } | 
 | 165 |  | 
 | 166 | static unsigned long long copy_word(unsigned long src_addr, int src_len, | 
 | 167 | 				    unsigned long dst_addr, int dst_len) | 
 | 168 | { | 
 | 169 | 	unsigned long long tmp = 0; | 
 | 170 |  | 
 | 171 | 	switch (src_len) { | 
 | 172 | 	case 1: | 
 | 173 | 		tmp = ctrl_inb(src_addr); | 
 | 174 | 		break; | 
 | 175 | 	case 2: | 
 | 176 | 		tmp = ctrl_inw(src_addr); | 
 | 177 | 		break; | 
 | 178 | 	case 4: | 
 | 179 | 		tmp = ctrl_inl(src_addr); | 
 | 180 | 		break; | 
 | 181 | 	case 8: | 
 | 182 | 		tmp = ctrl_inq(src_addr); | 
 | 183 | 		break; | 
 | 184 | 	} | 
 | 185 |  | 
 | 186 | 	switch (dst_len) { | 
 | 187 | 	case 1: | 
 | 188 | 		ctrl_outb(tmp, dst_addr); | 
 | 189 | 		break; | 
 | 190 | 	case 2: | 
 | 191 | 		ctrl_outw(tmp, dst_addr); | 
 | 192 | 		break; | 
 | 193 | 	case 4: | 
 | 194 | 		ctrl_outl(tmp, dst_addr); | 
 | 195 | 		break; | 
 | 196 | 	case 8: | 
 | 197 | 		ctrl_outq(tmp, dst_addr); | 
 | 198 | 		break; | 
 | 199 | 	} | 
 | 200 |  | 
 | 201 | 	return tmp; | 
 | 202 | } | 
 | 203 |  | 
 | 204 | static unsigned long from_device(void *dst, const void *src, unsigned long cnt) | 
 | 205 | { | 
 | 206 | 	struct trapped_io *tiop; | 
 | 207 | 	unsigned long src_addr = (unsigned long)src; | 
 | 208 | 	unsigned long long tmp; | 
 | 209 |  | 
 | 210 | 	pr_debug("trapped io read 0x%08lx (%ld)\n", src_addr, cnt); | 
 | 211 | 	tiop = lookup_tiop(src_addr); | 
 | 212 | 	WARN_ON(!tiop || (tiop->magic != IO_TRAPPED_MAGIC)); | 
 | 213 |  | 
 | 214 | 	src_addr = lookup_address(tiop, src_addr); | 
 | 215 | 	if (!src_addr) | 
 | 216 | 		return cnt; | 
 | 217 |  | 
| Paul Mundt | f1cdd63 | 2008-02-09 19:10:52 +0900 | [diff] [blame] | 218 | 	tmp = copy_word(src_addr, | 
 | 219 | 			max_t(unsigned long, cnt, | 
 | 220 | 			      (tiop->minimum_bus_width / 8)), | 
| Magnus Damm | e7cc9a7 | 2008-02-07 20:18:21 +0900 | [diff] [blame] | 221 | 			(unsigned long)dst, cnt); | 
 | 222 |  | 
 | 223 | 	pr_debug("trapped io read 0x%08lx -> 0x%08llx\n", src_addr, tmp); | 
 | 224 | 	return 0; | 
 | 225 | } | 
 | 226 |  | 
 | 227 | static unsigned long to_device(void *dst, const void *src, unsigned long cnt) | 
 | 228 | { | 
 | 229 | 	struct trapped_io *tiop; | 
 | 230 | 	unsigned long dst_addr = (unsigned long)dst; | 
 | 231 | 	unsigned long long tmp; | 
 | 232 |  | 
 | 233 | 	pr_debug("trapped io write 0x%08lx (%ld)\n", dst_addr, cnt); | 
 | 234 | 	tiop = lookup_tiop(dst_addr); | 
 | 235 | 	WARN_ON(!tiop || (tiop->magic != IO_TRAPPED_MAGIC)); | 
 | 236 |  | 
 | 237 | 	dst_addr = lookup_address(tiop, dst_addr); | 
 | 238 | 	if (!dst_addr) | 
 | 239 | 		return cnt; | 
 | 240 |  | 
 | 241 | 	tmp = copy_word((unsigned long)src, cnt, | 
| Paul Mundt | f1cdd63 | 2008-02-09 19:10:52 +0900 | [diff] [blame] | 242 | 			dst_addr, max_t(unsigned long, cnt, | 
 | 243 | 					(tiop->minimum_bus_width / 8))); | 
| Magnus Damm | e7cc9a7 | 2008-02-07 20:18:21 +0900 | [diff] [blame] | 244 |  | 
 | 245 | 	pr_debug("trapped io write 0x%08lx -> 0x%08llx\n", dst_addr, tmp); | 
 | 246 | 	return 0; | 
 | 247 | } | 
 | 248 |  | 
 | 249 | static struct mem_access trapped_io_access = { | 
 | 250 | 	from_device, | 
 | 251 | 	to_device, | 
 | 252 | }; | 
 | 253 |  | 
 | 254 | int handle_trapped_io(struct pt_regs *regs, unsigned long address) | 
 | 255 | { | 
 | 256 | 	mm_segment_t oldfs; | 
 | 257 | 	opcode_t instruction; | 
 | 258 | 	int tmp; | 
 | 259 |  | 
 | 260 | 	if (!lookup_tiop(address)) | 
 | 261 | 		return 0; | 
 | 262 |  | 
 | 263 | 	WARN_ON(user_mode(regs)); | 
 | 264 |  | 
 | 265 | 	oldfs = get_fs(); | 
 | 266 | 	set_fs(KERNEL_DS); | 
 | 267 | 	if (copy_from_user(&instruction, (void *)(regs->pc), | 
 | 268 | 			   sizeof(instruction))) { | 
 | 269 | 		set_fs(oldfs); | 
 | 270 | 		return 0; | 
 | 271 | 	} | 
 | 272 |  | 
 | 273 | 	tmp = handle_unaligned_access(instruction, regs, &trapped_io_access); | 
 | 274 | 	set_fs(oldfs); | 
 | 275 | 	return tmp == 0; | 
 | 276 | } |