blob: 27eb29ce666b624a59115df89e173eb32d9a4e39 [file] [log] [blame]
Linus Torvalds1da177e2005-04-16 15:20:36 -07001/*
2 * Copyright (C) 2002 Jeff Dike (jdike@karaya.com)
3 * Copyright 2003 PathScale, Inc.
4 * Licensed under the GPL
5 */
6
7#include "linux/stddef.h"
8#include "linux/sched.h"
9#include "linux/mm.h"
10#include "asm/page.h"
11#include "asm/pgtable.h"
12#include "asm/mmu.h"
13#include "user_util.h"
14#include "mem_user.h"
15#include "mem.h"
16#include "skas.h"
17#include "os.h"
18#include "tlb.h"
19
Bodo Stroesser07bf7312005-09-03 15:57:50 -070020static int do_ops(union mm_context *mmu, struct host_vm_op *ops, int last,
21 int finished, void **flush)
Linus Torvalds1da177e2005-04-16 15:20:36 -070022{
23 struct host_vm_op *op;
Bodo Stroesser07bf7312005-09-03 15:57:50 -070024 int i, ret = 0;
Linus Torvalds1da177e2005-04-16 15:20:36 -070025
Bodo Stroesser07bf7312005-09-03 15:57:50 -070026 for(i = 0; i <= last && !ret; i++){
Linus Torvalds1da177e2005-04-16 15:20:36 -070027 op = &ops[i];
28 switch(op->type){
29 case MMAP:
Bodo Stroesser07bf7312005-09-03 15:57:50 -070030 ret = map(&mmu->skas.id, op->u.mmap.addr,
31 op->u.mmap.len, op->u.mmap.r, op->u.mmap.w,
32 op->u.mmap.x, op->u.mmap.fd,
33 op->u.mmap.offset, finished, flush);
Linus Torvalds1da177e2005-04-16 15:20:36 -070034 break;
35 case MUNMAP:
Bodo Stroesser07bf7312005-09-03 15:57:50 -070036 ret = unmap(&mmu->skas.id,
37 (void *) op->u.munmap.addr,
38 op->u.munmap.len, finished, flush);
Linus Torvalds1da177e2005-04-16 15:20:36 -070039 break;
40 case MPROTECT:
Bodo Stroesser07bf7312005-09-03 15:57:50 -070041 ret = protect(&mmu->skas.id, op->u.mprotect.addr,
42 op->u.mprotect.len, op->u.mprotect.r,
43 op->u.mprotect.w, op->u.mprotect.x,
44 finished, flush);
Linus Torvalds1da177e2005-04-16 15:20:36 -070045 break;
46 default:
47 printk("Unknown op type %d in do_ops\n", op->type);
48 break;
49 }
50 }
Jeff Dikec5600492005-09-03 15:57:36 -070051
Bodo Stroesser07bf7312005-09-03 15:57:50 -070052 return ret;
Linus Torvalds1da177e2005-04-16 15:20:36 -070053}
54
Jeff Diked67b5692005-07-07 17:56:49 -070055extern int proc_mm;
56
Linus Torvalds1da177e2005-04-16 15:20:36 -070057static void fix_range(struct mm_struct *mm, unsigned long start_addr,
58 unsigned long end_addr, int force)
59{
Jeff Diked67b5692005-07-07 17:56:49 -070060 if(!proc_mm && (end_addr > CONFIG_STUB_START))
61 end_addr = CONFIG_STUB_START;
Linus Torvalds1da177e2005-04-16 15:20:36 -070062
Jeff Diked67b5692005-07-07 17:56:49 -070063 fix_range_common(mm, start_addr, end_addr, force, do_ops);
Linus Torvalds1da177e2005-04-16 15:20:36 -070064}
65
66void __flush_tlb_one_skas(unsigned long addr)
67{
68 flush_tlb_kernel_range_common(addr, addr + PAGE_SIZE);
69}
70
71void flush_tlb_range_skas(struct vm_area_struct *vma, unsigned long start,
72 unsigned long end)
73{
74 if(vma->vm_mm == NULL)
75 flush_tlb_kernel_range_common(start, end);
76 else fix_range(vma->vm_mm, start, end, 0);
77}
78
79void flush_tlb_mm_skas(struct mm_struct *mm)
80{
Jeff Diked67b5692005-07-07 17:56:49 -070081 unsigned long end;
82
Linus Torvalds1da177e2005-04-16 15:20:36 -070083 /* Don't bother flushing if this address space is about to be
84 * destroyed.
85 */
86 if(atomic_read(&mm->mm_users) == 0)
87 return;
88
Jeff Diked67b5692005-07-07 17:56:49 -070089 end = proc_mm ? task_size : CONFIG_STUB_START;
90 fix_range(mm, 0, end, 0);
Linus Torvalds1da177e2005-04-16 15:20:36 -070091}
92
93void force_flush_all_skas(void)
94{
Jeff Diked67b5692005-07-07 17:56:49 -070095 unsigned long end = proc_mm ? task_size : CONFIG_STUB_START;
96 fix_range(current->mm, 0, end, 1);
Linus Torvalds1da177e2005-04-16 15:20:36 -070097}