|  | /* | 
|  | * Copyright (C) 2002 Jeff Dike (jdike@karaya.com) | 
|  | * Licensed under the GPL | 
|  | */ | 
|  |  | 
|  | #include "linux/sched.h" | 
|  | #include "linux/slab.h" | 
|  | #include "linux/ptrace.h" | 
|  | #include "linux/proc_fs.h" | 
|  | #include "linux/file.h" | 
|  | #include "linux/errno.h" | 
|  | #include "linux/init.h" | 
|  | #include "asm/uaccess.h" | 
|  | #include "asm/atomic.h" | 
|  | #include "kern_util.h" | 
|  | #include "as-layout.h" | 
|  | #include "skas.h" | 
|  | #include "os.h" | 
|  | #include "tlb.h" | 
|  | #include "kern.h" | 
|  | #include "mode.h" | 
|  | #include "registers.h" | 
|  |  | 
|  | void switch_to_skas(void *prev, void *next) | 
|  | { | 
|  | struct task_struct *from, *to; | 
|  |  | 
|  | from = prev; | 
|  | to = next; | 
|  |  | 
|  | /* XXX need to check runqueues[cpu].idle */ | 
|  | if(current->pid == 0) | 
|  | switch_timers(0); | 
|  |  | 
|  | switch_threads(&from->thread.mode.skas.switch_buf, | 
|  | &to->thread.mode.skas.switch_buf); | 
|  |  | 
|  | arch_switch_to_skas(current->thread.prev_sched, current); | 
|  |  | 
|  | if(current->pid == 0) | 
|  | switch_timers(1); | 
|  | } | 
|  |  | 
|  | extern void schedule_tail(struct task_struct *prev); | 
|  |  | 
|  | /* This is called magically, by its address being stuffed in a jmp_buf | 
|  | * and being longjmp-d to. | 
|  | */ | 
|  | void new_thread_handler(void) | 
|  | { | 
|  | int (*fn)(void *), n; | 
|  | void *arg; | 
|  |  | 
|  | if(current->thread.prev_sched != NULL) | 
|  | schedule_tail(current->thread.prev_sched); | 
|  | current->thread.prev_sched = NULL; | 
|  |  | 
|  | fn = current->thread.request.u.thread.proc; | 
|  | arg = current->thread.request.u.thread.arg; | 
|  |  | 
|  | /* The return value is 1 if the kernel thread execs a process, | 
|  | * 0 if it just exits | 
|  | */ | 
|  | n = run_kernel_thread(fn, arg, ¤t->thread.exec_buf); | 
|  | if(n == 1){ | 
|  | /* Handle any immediate reschedules or signals */ | 
|  | interrupt_end(); | 
|  | userspace(¤t->thread.regs.regs); | 
|  | } | 
|  | else do_exit(0); | 
|  | } | 
|  |  | 
|  | void release_thread_skas(struct task_struct *task) | 
|  | { | 
|  | } | 
|  |  | 
|  | /* Called magically, see new_thread_handler above */ | 
|  | void fork_handler(void) | 
|  | { | 
|  | force_flush_all(); | 
|  | if(current->thread.prev_sched == NULL) | 
|  | panic("blech"); | 
|  |  | 
|  | schedule_tail(current->thread.prev_sched); | 
|  |  | 
|  | /* XXX: if interrupt_end() calls schedule, this call to | 
|  | * arch_switch_to_skas isn't needed. We could want to apply this to | 
|  | * improve performance. -bb */ | 
|  | arch_switch_to_skas(current->thread.prev_sched, current); | 
|  |  | 
|  | current->thread.prev_sched = NULL; | 
|  |  | 
|  | /* Handle any immediate reschedules or signals */ | 
|  | interrupt_end(); | 
|  |  | 
|  | userspace(¤t->thread.regs.regs); | 
|  | } | 
|  |  | 
|  | int copy_thread_skas(int nr, unsigned long clone_flags, unsigned long sp, | 
|  | unsigned long stack_top, struct task_struct * p, | 
|  | struct pt_regs *regs) | 
|  | { | 
|  | void (*handler)(void); | 
|  |  | 
|  | if(current->thread.forking){ | 
|  | memcpy(&p->thread.regs.regs.skas, ®s->regs.skas, | 
|  | sizeof(p->thread.regs.regs.skas)); | 
|  | REGS_SET_SYSCALL_RETURN(p->thread.regs.regs.skas.regs, 0); | 
|  | if(sp != 0) REGS_SP(p->thread.regs.regs.skas.regs) = sp; | 
|  |  | 
|  | handler = fork_handler; | 
|  |  | 
|  | arch_copy_thread(¤t->thread.arch, &p->thread.arch); | 
|  | } | 
|  | else { | 
|  | init_thread_registers(&p->thread.regs.regs); | 
|  | p->thread.request.u.thread = current->thread.request.u.thread; | 
|  | handler = new_thread_handler; | 
|  | } | 
|  |  | 
|  | new_thread(task_stack_page(p), &p->thread.mode.skas.switch_buf, | 
|  | handler); | 
|  | return(0); | 
|  | } | 
|  |  | 
|  | int new_mm(unsigned long stack) | 
|  | { | 
|  | int fd; | 
|  |  | 
|  | fd = os_open_file("/proc/mm", of_cloexec(of_write(OPENFLAGS())), 0); | 
|  | if(fd < 0) | 
|  | return(fd); | 
|  |  | 
|  | if(skas_needs_stub) | 
|  | map_stub_pages(fd, CONFIG_STUB_CODE, CONFIG_STUB_DATA, stack); | 
|  |  | 
|  | return(fd); | 
|  | } | 
|  |  | 
|  | void init_idle_skas(void) | 
|  | { | 
|  | cpu_tasks[current_thread->cpu].pid = os_getpid(); | 
|  | default_idle(); | 
|  | } | 
|  |  | 
|  | extern void start_kernel(void); | 
|  |  | 
|  | static int start_kernel_proc(void *unused) | 
|  | { | 
|  | int pid; | 
|  |  | 
|  | block_signals(); | 
|  | pid = os_getpid(); | 
|  |  | 
|  | cpu_tasks[0].pid = pid; | 
|  | cpu_tasks[0].task = current; | 
|  | #ifdef CONFIG_SMP | 
|  | cpu_online_map = cpumask_of_cpu(0); | 
|  | #endif | 
|  | start_kernel(); | 
|  | return(0); | 
|  | } | 
|  |  | 
|  | extern int userspace_pid[]; | 
|  |  | 
|  | extern char cpu0_irqstack[]; | 
|  |  | 
|  | int start_uml_skas(void) | 
|  | { | 
|  | stack_protections((unsigned long) &cpu0_irqstack); | 
|  | set_sigstack(cpu0_irqstack, THREAD_SIZE); | 
|  | if(proc_mm) | 
|  | userspace_pid[0] = start_userspace(0); | 
|  |  | 
|  | init_new_thread_signals(); | 
|  |  | 
|  | init_task.thread.request.u.thread.proc = start_kernel_proc; | 
|  | init_task.thread.request.u.thread.arg = NULL; | 
|  | return(start_idle_thread(task_stack_page(&init_task), | 
|  | &init_task.thread.mode.skas.switch_buf)); | 
|  | } | 
|  |  | 
|  | int external_pid_skas(struct task_struct *task) | 
|  | { | 
|  | /* FIXME: Need to look up userspace_pid by cpu */ | 
|  | return(userspace_pid[0]); | 
|  | } | 
|  |  | 
|  | int thread_pid_skas(struct task_struct *task) | 
|  | { | 
|  | /* FIXME: Need to look up userspace_pid by cpu */ | 
|  | return(userspace_pid[0]); | 
|  | } | 
|  |  | 
|  | void kill_off_processes_skas(void) | 
|  | { | 
|  | if(proc_mm) | 
|  | /* | 
|  | * FIXME: need to loop over userspace_pids in | 
|  | * kill_off_processes_skas | 
|  | */ | 
|  | os_kill_ptraced_process(userspace_pid[0], 1); | 
|  | else { | 
|  | struct task_struct *p; | 
|  | int pid, me; | 
|  |  | 
|  | me = os_getpid(); | 
|  | for_each_process(p){ | 
|  | if(p->mm == NULL) | 
|  | continue; | 
|  |  | 
|  | pid = p->mm->context.skas.id.u.pid; | 
|  | os_kill_ptraced_process(pid, 1); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | unsigned long current_stub_stack(void) | 
|  | { | 
|  | if(current->mm == NULL) | 
|  | return(0); | 
|  |  | 
|  | return(current->mm->context.skas.id.stack); | 
|  | } |