|  | /* | 
|  | *  arch/ppc/kernel/signal.c | 
|  | * | 
|  | *  PowerPC version | 
|  | *    Copyright (C) 1995-1996 Gary Thomas (gdt@linuxppc.org) | 
|  | * | 
|  | *  Derived from "arch/i386/kernel/signal.c" | 
|  | *    Copyright (C) 1991, 1992 Linus Torvalds | 
|  | *    1997-11-28  Modified for POSIX.1b signals by Richard Henderson | 
|  | * | 
|  | *  This program is free software; you can redistribute it and/or | 
|  | *  modify it under the terms of the GNU General Public License | 
|  | *  as published by the Free Software Foundation; either version | 
|  | *  2 of the License, or (at your option) any later version. | 
|  | */ | 
|  |  | 
|  | #include <linux/sched.h> | 
|  | #include <linux/mm.h> | 
|  | #include <linux/smp.h> | 
|  | #include <linux/smp_lock.h> | 
|  | #include <linux/kernel.h> | 
|  | #include <linux/signal.h> | 
|  | #include <linux/errno.h> | 
|  | #include <linux/wait.h> | 
|  | #include <linux/ptrace.h> | 
|  | #include <linux/unistd.h> | 
|  | #include <linux/stddef.h> | 
|  | #include <linux/elf.h> | 
|  | #include <linux/tty.h> | 
|  | #include <linux/binfmts.h> | 
|  | #include <linux/suspend.h> | 
|  | #include <asm/ucontext.h> | 
|  | #include <asm/uaccess.h> | 
|  | #include <asm/pgtable.h> | 
|  | #include <asm/cacheflush.h> | 
|  |  | 
|  | #undef DEBUG_SIG | 
|  |  | 
|  | #define _BLOCKABLE (~(sigmask(SIGKILL) | sigmask(SIGSTOP))) | 
|  |  | 
|  | extern void sigreturn_exit(struct pt_regs *); | 
|  |  | 
|  | #define GP_REGS_SIZE	min(sizeof(elf_gregset_t), sizeof(struct pt_regs)) | 
|  |  | 
|  | int do_signal(sigset_t *oldset, struct pt_regs *regs); | 
|  |  | 
|  | /* | 
|  | * Atomically swap in the new signal mask, and wait for a signal. | 
|  | */ | 
|  | int | 
|  | sys_sigsuspend(old_sigset_t mask, int p2, int p3, int p4, int p6, int p7, | 
|  | struct pt_regs *regs) | 
|  | { | 
|  | sigset_t saveset; | 
|  |  | 
|  | mask &= _BLOCKABLE; | 
|  | spin_lock_irq(¤t->sighand->siglock); | 
|  | saveset = current->blocked; | 
|  | siginitset(¤t->blocked, mask); | 
|  | recalc_sigpending(); | 
|  | spin_unlock_irq(¤t->sighand->siglock); | 
|  |  | 
|  | regs->result = -EINTR; | 
|  | regs->gpr[3] = EINTR; | 
|  | regs->ccr |= 0x10000000; | 
|  | while (1) { | 
|  | current->state = TASK_INTERRUPTIBLE; | 
|  | schedule(); | 
|  | if (do_signal(&saveset, regs)) | 
|  | sigreturn_exit(regs); | 
|  | } | 
|  | } | 
|  |  | 
|  | int | 
|  | sys_rt_sigsuspend(sigset_t __user *unewset, size_t sigsetsize, int p3, int p4, | 
|  | int p6, int p7, struct pt_regs *regs) | 
|  | { | 
|  | sigset_t saveset, newset; | 
|  |  | 
|  | /* XXX: Don't preclude handling different sized sigset_t's.  */ | 
|  | if (sigsetsize != sizeof(sigset_t)) | 
|  | return -EINVAL; | 
|  |  | 
|  | if (copy_from_user(&newset, unewset, sizeof(newset))) | 
|  | return -EFAULT; | 
|  | sigdelsetmask(&newset, ~_BLOCKABLE); | 
|  |  | 
|  | spin_lock_irq(¤t->sighand->siglock); | 
|  | saveset = current->blocked; | 
|  | current->blocked = newset; | 
|  | recalc_sigpending(); | 
|  | spin_unlock_irq(¤t->sighand->siglock); | 
|  |  | 
|  | regs->result = -EINTR; | 
|  | regs->gpr[3] = EINTR; | 
|  | regs->ccr |= 0x10000000; | 
|  | while (1) { | 
|  | current->state = TASK_INTERRUPTIBLE; | 
|  | schedule(); | 
|  | if (do_signal(&saveset, regs)) | 
|  | sigreturn_exit(regs); | 
|  | } | 
|  | } | 
|  |  | 
|  |  | 
|  | int | 
|  | sys_sigaltstack(const stack_t __user *uss, stack_t __user *uoss, int r5, | 
|  | int r6, int r7, int r8, struct pt_regs *regs) | 
|  | { | 
|  | return do_sigaltstack(uss, uoss, regs->gpr[1]); | 
|  | } | 
|  |  | 
|  | int | 
|  | sys_sigaction(int sig, const struct old_sigaction __user *act, | 
|  | struct old_sigaction __user *oact) | 
|  | { | 
|  | struct k_sigaction new_ka, old_ka; | 
|  | int ret; | 
|  |  | 
|  | if (act) { | 
|  | old_sigset_t mask; | 
|  | if (!access_ok(VERIFY_READ, act, sizeof(*act)) || | 
|  | __get_user(new_ka.sa.sa_handler, &act->sa_handler) || | 
|  | __get_user(new_ka.sa.sa_restorer, &act->sa_restorer)) | 
|  | return -EFAULT; | 
|  | __get_user(new_ka.sa.sa_flags, &act->sa_flags); | 
|  | __get_user(mask, &act->sa_mask); | 
|  | siginitset(&new_ka.sa.sa_mask, mask); | 
|  | } | 
|  |  | 
|  | ret = do_sigaction(sig, (act? &new_ka: NULL), (oact? &old_ka: NULL)); | 
|  |  | 
|  | if (!ret && oact) { | 
|  | if (!access_ok(VERIFY_WRITE, oact, sizeof(*oact)) || | 
|  | __put_user(old_ka.sa.sa_handler, &oact->sa_handler) || | 
|  | __put_user(old_ka.sa.sa_restorer, &oact->sa_restorer)) | 
|  | return -EFAULT; | 
|  | __put_user(old_ka.sa.sa_flags, &oact->sa_flags); | 
|  | __put_user(old_ka.sa.sa_mask.sig[0], &oact->sa_mask); | 
|  | } | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * When we have signals to deliver, we set up on the | 
|  | * user stack, going down from the original stack pointer: | 
|  | *	a sigregs struct | 
|  | *	a sigcontext struct | 
|  | *	a gap of __SIGNAL_FRAMESIZE bytes | 
|  | * | 
|  | * Each of these things must be a multiple of 16 bytes in size. | 
|  | * | 
|  | */ | 
|  | struct sigregs { | 
|  | struct mcontext	mctx;		/* all the register values */ | 
|  | /* Programs using the rs6000/xcoff abi can save up to 19 gp regs | 
|  | and 18 fp regs below sp before decrementing it. */ | 
|  | int		abigap[56]; | 
|  | }; | 
|  |  | 
|  | /* We use the mc_pad field for the signal return trampoline. */ | 
|  | #define tramp	mc_pad | 
|  |  | 
|  | /* | 
|  | *  When we have rt signals to deliver, we set up on the | 
|  | *  user stack, going down from the original stack pointer: | 
|  | *	one rt_sigframe struct (siginfo + ucontext + ABI gap) | 
|  | *	a gap of __SIGNAL_FRAMESIZE+16 bytes | 
|  | *  (the +16 is to get the siginfo and ucontext in the same | 
|  | *  positions as in older kernels). | 
|  | * | 
|  | *  Each of these things must be a multiple of 16 bytes in size. | 
|  | * | 
|  | */ | 
|  | struct rt_sigframe | 
|  | { | 
|  | struct siginfo info; | 
|  | struct ucontext uc; | 
|  | /* Programs using the rs6000/xcoff abi can save up to 19 gp regs | 
|  | and 18 fp regs below sp before decrementing it. */ | 
|  | int		abigap[56]; | 
|  | }; | 
|  |  | 
|  | /* | 
|  | * Save the current user registers on the user stack. | 
|  | * We only save the altivec/spe registers if the process has used | 
|  | * altivec/spe instructions at some point. | 
|  | */ | 
|  | static int | 
|  | save_user_regs(struct pt_regs *regs, struct mcontext __user *frame, int sigret) | 
|  | { | 
|  | /* save general and floating-point registers */ | 
|  | CHECK_FULL_REGS(regs); | 
|  | preempt_disable(); | 
|  | if (regs->msr & MSR_FP) | 
|  | giveup_fpu(current); | 
|  | #ifdef CONFIG_ALTIVEC | 
|  | if (current->thread.used_vr && (regs->msr & MSR_VEC)) | 
|  | giveup_altivec(current); | 
|  | #endif /* CONFIG_ALTIVEC */ | 
|  | #ifdef CONFIG_SPE | 
|  | if (current->thread.used_spe && (regs->msr & MSR_SPE)) | 
|  | giveup_spe(current); | 
|  | #endif /* CONFIG_ALTIVEC */ | 
|  | preempt_enable(); | 
|  |  | 
|  | if (__copy_to_user(&frame->mc_gregs, regs, GP_REGS_SIZE) | 
|  | || __copy_to_user(&frame->mc_fregs, current->thread.fpr, | 
|  | ELF_NFPREG * sizeof(double))) | 
|  | return 1; | 
|  |  | 
|  | current->thread.fpscr = 0;	/* turn off all fp exceptions */ | 
|  |  | 
|  | #ifdef CONFIG_ALTIVEC | 
|  | /* save altivec registers */ | 
|  | if (current->thread.used_vr) { | 
|  | if (__copy_to_user(&frame->mc_vregs, current->thread.vr, | 
|  | ELF_NVRREG * sizeof(vector128))) | 
|  | return 1; | 
|  | /* set MSR_VEC in the saved MSR value to indicate that | 
|  | frame->mc_vregs contains valid data */ | 
|  | if (__put_user(regs->msr | MSR_VEC, &frame->mc_gregs[PT_MSR])) | 
|  | return 1; | 
|  | } | 
|  | /* else assert((regs->msr & MSR_VEC) == 0) */ | 
|  |  | 
|  | /* We always copy to/from vrsave, it's 0 if we don't have or don't | 
|  | * use altivec. Since VSCR only contains 32 bits saved in the least | 
|  | * significant bits of a vector, we "cheat" and stuff VRSAVE in the | 
|  | * most significant bits of that same vector. --BenH | 
|  | */ | 
|  | if (__put_user(current->thread.vrsave, (u32 __user *)&frame->mc_vregs[32])) | 
|  | return 1; | 
|  | #endif /* CONFIG_ALTIVEC */ | 
|  |  | 
|  | #ifdef CONFIG_SPE | 
|  | /* save spe registers */ | 
|  | if (current->thread.used_spe) { | 
|  | if (__copy_to_user(&frame->mc_vregs, current->thread.evr, | 
|  | ELF_NEVRREG * sizeof(u32))) | 
|  | return 1; | 
|  | /* set MSR_SPE in the saved MSR value to indicate that | 
|  | frame->mc_vregs contains valid data */ | 
|  | if (__put_user(regs->msr | MSR_SPE, &frame->mc_gregs[PT_MSR])) | 
|  | return 1; | 
|  | } | 
|  | /* else assert((regs->msr & MSR_SPE) == 0) */ | 
|  |  | 
|  | /* We always copy to/from spefscr */ | 
|  | if (__put_user(current->thread.spefscr, (u32 *)&frame->mc_vregs + ELF_NEVRREG)) | 
|  | return 1; | 
|  | #endif /* CONFIG_SPE */ | 
|  |  | 
|  | if (sigret) { | 
|  | /* Set up the sigreturn trampoline: li r0,sigret; sc */ | 
|  | if (__put_user(0x38000000UL + sigret, &frame->tramp[0]) | 
|  | || __put_user(0x44000002UL, &frame->tramp[1])) | 
|  | return 1; | 
|  | flush_icache_range((unsigned long) &frame->tramp[0], | 
|  | (unsigned long) &frame->tramp[2]); | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Restore the current user register values from the user stack, | 
|  | * (except for MSR). | 
|  | */ | 
|  | static int | 
|  | restore_user_regs(struct pt_regs *regs, struct mcontext __user *sr, int sig) | 
|  | { | 
|  | unsigned long save_r2 = 0; | 
|  | #if defined(CONFIG_ALTIVEC) || defined(CONFIG_SPE) | 
|  | unsigned long msr; | 
|  | #endif | 
|  |  | 
|  | /* backup/restore the TLS as we don't want it to be modified */ | 
|  | if (!sig) | 
|  | save_r2 = regs->gpr[2]; | 
|  | /* copy up to but not including MSR */ | 
|  | if (__copy_from_user(regs, &sr->mc_gregs, PT_MSR * sizeof(elf_greg_t))) | 
|  | return 1; | 
|  | /* copy from orig_r3 (the word after the MSR) up to the end */ | 
|  | if (__copy_from_user(®s->orig_gpr3, &sr->mc_gregs[PT_ORIG_R3], | 
|  | GP_REGS_SIZE - PT_ORIG_R3 * sizeof(elf_greg_t))) | 
|  | return 1; | 
|  | if (!sig) | 
|  | regs->gpr[2] = save_r2; | 
|  |  | 
|  | /* force the process to reload the FP registers from | 
|  | current->thread when it next does FP instructions */ | 
|  | regs->msr &= ~(MSR_FP | MSR_FE0 | MSR_FE1); | 
|  | if (__copy_from_user(current->thread.fpr, &sr->mc_fregs, | 
|  | sizeof(sr->mc_fregs))) | 
|  | return 1; | 
|  |  | 
|  | #ifdef CONFIG_ALTIVEC | 
|  | /* force the process to reload the altivec registers from | 
|  | current->thread when it next does altivec instructions */ | 
|  | regs->msr &= ~MSR_VEC; | 
|  | if (!__get_user(msr, &sr->mc_gregs[PT_MSR]) && (msr & MSR_VEC) != 0) { | 
|  | /* restore altivec registers from the stack */ | 
|  | if (__copy_from_user(current->thread.vr, &sr->mc_vregs, | 
|  | sizeof(sr->mc_vregs))) | 
|  | return 1; | 
|  | } else if (current->thread.used_vr) | 
|  | memset(¤t->thread.vr, 0, ELF_NVRREG * sizeof(vector128)); | 
|  |  | 
|  | /* Always get VRSAVE back */ | 
|  | if (__get_user(current->thread.vrsave, (u32 __user *)&sr->mc_vregs[32])) | 
|  | return 1; | 
|  | #endif /* CONFIG_ALTIVEC */ | 
|  |  | 
|  | #ifdef CONFIG_SPE | 
|  | /* force the process to reload the spe registers from | 
|  | current->thread when it next does spe instructions */ | 
|  | regs->msr &= ~MSR_SPE; | 
|  | if (!__get_user(msr, &sr->mc_gregs[PT_MSR]) && (msr & MSR_SPE) != 0) { | 
|  | /* restore spe registers from the stack */ | 
|  | if (__copy_from_user(current->thread.evr, &sr->mc_vregs, | 
|  | ELF_NEVRREG * sizeof(u32))) | 
|  | return 1; | 
|  | } else if (current->thread.used_spe) | 
|  | memset(¤t->thread.evr, 0, ELF_NEVRREG * sizeof(u32)); | 
|  |  | 
|  | /* Always get SPEFSCR back */ | 
|  | if (__get_user(current->thread.spefscr, (u32 *)&sr->mc_vregs + ELF_NEVRREG)) | 
|  | return 1; | 
|  | #endif /* CONFIG_SPE */ | 
|  |  | 
|  | #ifndef CONFIG_SMP | 
|  | preempt_disable(); | 
|  | if (last_task_used_math == current) | 
|  | last_task_used_math = NULL; | 
|  | if (last_task_used_altivec == current) | 
|  | last_task_used_altivec = NULL; | 
|  | if (last_task_used_spe == current) | 
|  | last_task_used_spe = NULL; | 
|  | preempt_enable(); | 
|  | #endif | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Restore the user process's signal mask | 
|  | */ | 
|  | static void | 
|  | restore_sigmask(sigset_t *set) | 
|  | { | 
|  | sigdelsetmask(set, ~_BLOCKABLE); | 
|  | spin_lock_irq(¤t->sighand->siglock); | 
|  | current->blocked = *set; | 
|  | recalc_sigpending(); | 
|  | spin_unlock_irq(¤t->sighand->siglock); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Set up a signal frame for a "real-time" signal handler | 
|  | * (one which gets siginfo). | 
|  | */ | 
|  | static void | 
|  | handle_rt_signal(unsigned long sig, struct k_sigaction *ka, | 
|  | siginfo_t *info, sigset_t *oldset, struct pt_regs * regs, | 
|  | unsigned long newsp) | 
|  | { | 
|  | struct rt_sigframe __user *rt_sf; | 
|  | struct mcontext __user *frame; | 
|  | unsigned long origsp = newsp; | 
|  |  | 
|  | /* Set up Signal Frame */ | 
|  | /* Put a Real Time Context onto stack */ | 
|  | newsp -= sizeof(*rt_sf); | 
|  | rt_sf = (struct rt_sigframe __user *) newsp; | 
|  |  | 
|  | /* create a stack frame for the caller of the handler */ | 
|  | newsp -= __SIGNAL_FRAMESIZE + 16; | 
|  |  | 
|  | if (!access_ok(VERIFY_WRITE, (void __user *) newsp, origsp - newsp)) | 
|  | goto badframe; | 
|  |  | 
|  | /* Put the siginfo & fill in most of the ucontext */ | 
|  | if (copy_siginfo_to_user(&rt_sf->info, info) | 
|  | || __put_user(0, &rt_sf->uc.uc_flags) | 
|  | || __put_user(0, &rt_sf->uc.uc_link) | 
|  | || __put_user(current->sas_ss_sp, &rt_sf->uc.uc_stack.ss_sp) | 
|  | || __put_user(sas_ss_flags(regs->gpr[1]), | 
|  | &rt_sf->uc.uc_stack.ss_flags) | 
|  | || __put_user(current->sas_ss_size, &rt_sf->uc.uc_stack.ss_size) | 
|  | || __put_user(&rt_sf->uc.uc_mcontext, &rt_sf->uc.uc_regs) | 
|  | || __copy_to_user(&rt_sf->uc.uc_sigmask, oldset, sizeof(*oldset))) | 
|  | goto badframe; | 
|  |  | 
|  | /* Save user registers on the stack */ | 
|  | frame = &rt_sf->uc.uc_mcontext; | 
|  | if (save_user_regs(regs, frame, __NR_rt_sigreturn)) | 
|  | goto badframe; | 
|  |  | 
|  | if (put_user(regs->gpr[1], (unsigned long __user *)newsp)) | 
|  | goto badframe; | 
|  | regs->gpr[1] = newsp; | 
|  | regs->gpr[3] = sig; | 
|  | regs->gpr[4] = (unsigned long) &rt_sf->info; | 
|  | regs->gpr[5] = (unsigned long) &rt_sf->uc; | 
|  | regs->gpr[6] = (unsigned long) rt_sf; | 
|  | regs->nip = (unsigned long) ka->sa.sa_handler; | 
|  | regs->link = (unsigned long) frame->tramp; | 
|  | regs->trap = 0; | 
|  |  | 
|  | return; | 
|  |  | 
|  | badframe: | 
|  | #ifdef DEBUG_SIG | 
|  | printk("badframe in handle_rt_signal, regs=%p frame=%p newsp=%lx\n", | 
|  | regs, frame, newsp); | 
|  | #endif | 
|  | force_sigsegv(sig, current); | 
|  | } | 
|  |  | 
|  | static int do_setcontext(struct ucontext __user *ucp, struct pt_regs *regs, int sig) | 
|  | { | 
|  | sigset_t set; | 
|  | struct mcontext __user *mcp; | 
|  |  | 
|  | if (__copy_from_user(&set, &ucp->uc_sigmask, sizeof(set)) | 
|  | || __get_user(mcp, &ucp->uc_regs)) | 
|  | return -EFAULT; | 
|  | restore_sigmask(&set); | 
|  | if (restore_user_regs(regs, mcp, sig)) | 
|  | return -EFAULT; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | int sys_swapcontext(struct ucontext __user *old_ctx, | 
|  | struct ucontext __user *new_ctx, | 
|  | int ctx_size, int r6, int r7, int r8, struct pt_regs *regs) | 
|  | { | 
|  | unsigned char tmp; | 
|  |  | 
|  | /* Context size is for future use. Right now, we only make sure | 
|  | * we are passed something we understand | 
|  | */ | 
|  | if (ctx_size < sizeof(struct ucontext)) | 
|  | return -EINVAL; | 
|  |  | 
|  | if (old_ctx != NULL) { | 
|  | if (!access_ok(VERIFY_WRITE, old_ctx, sizeof(*old_ctx)) | 
|  | || save_user_regs(regs, &old_ctx->uc_mcontext, 0) | 
|  | || __copy_to_user(&old_ctx->uc_sigmask, | 
|  | ¤t->blocked, sizeof(sigset_t)) | 
|  | || __put_user(&old_ctx->uc_mcontext, &old_ctx->uc_regs)) | 
|  | return -EFAULT; | 
|  | } | 
|  | if (new_ctx == NULL) | 
|  | return 0; | 
|  | if (!access_ok(VERIFY_READ, new_ctx, sizeof(*new_ctx)) | 
|  | || __get_user(tmp, (u8 __user *) new_ctx) | 
|  | || __get_user(tmp, (u8 __user *) (new_ctx + 1) - 1)) | 
|  | return -EFAULT; | 
|  |  | 
|  | /* | 
|  | * If we get a fault copying the context into the kernel's | 
|  | * image of the user's registers, we can't just return -EFAULT | 
|  | * because the user's registers will be corrupted.  For instance | 
|  | * the NIP value may have been updated but not some of the | 
|  | * other registers.  Given that we have done the access_ok | 
|  | * and successfully read the first and last bytes of the region | 
|  | * above, this should only happen in an out-of-memory situation | 
|  | * or if another thread unmaps the region containing the context. | 
|  | * We kill the task with a SIGSEGV in this situation. | 
|  | */ | 
|  | if (do_setcontext(new_ctx, regs, 0)) | 
|  | do_exit(SIGSEGV); | 
|  | sigreturn_exit(regs); | 
|  | /* doesn't actually return back to here */ | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | int sys_rt_sigreturn(int r3, int r4, int r5, int r6, int r7, int r8, | 
|  | struct pt_regs *regs) | 
|  | { | 
|  | struct rt_sigframe __user *rt_sf; | 
|  |  | 
|  | /* Always make any pending restarted system calls return -EINTR */ | 
|  | current_thread_info()->restart_block.fn = do_no_restart_syscall; | 
|  |  | 
|  | rt_sf = (struct rt_sigframe __user *) | 
|  | (regs->gpr[1] + __SIGNAL_FRAMESIZE + 16); | 
|  | if (!access_ok(VERIFY_READ, rt_sf, sizeof(struct rt_sigframe))) | 
|  | goto bad; | 
|  | if (do_setcontext(&rt_sf->uc, regs, 1)) | 
|  | goto bad; | 
|  |  | 
|  | /* | 
|  | * It's not clear whether or why it is desirable to save the | 
|  | * sigaltstack setting on signal delivery and restore it on | 
|  | * signal return.  But other architectures do this and we have | 
|  | * always done it up until now so it is probably better not to | 
|  | * change it.  -- paulus | 
|  | */ | 
|  | do_sigaltstack(&rt_sf->uc.uc_stack, NULL, regs->gpr[1]); | 
|  |  | 
|  | sigreturn_exit(regs);		/* doesn't return here */ | 
|  | return 0; | 
|  |  | 
|  | bad: | 
|  | force_sig(SIGSEGV, current); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | int sys_debug_setcontext(struct ucontext __user *ctx, | 
|  | int ndbg, struct sig_dbg_op __user *dbg, | 
|  | int r6, int r7, int r8, | 
|  | struct pt_regs *regs) | 
|  | { | 
|  | struct sig_dbg_op op; | 
|  | int i; | 
|  | unsigned long new_msr = regs->msr; | 
|  | #if defined(CONFIG_4xx) || defined(CONFIG_BOOKE) | 
|  | unsigned long new_dbcr0 = current->thread.dbcr0; | 
|  | #endif | 
|  |  | 
|  | for (i=0; i<ndbg; i++) { | 
|  | if (__copy_from_user(&op, dbg, sizeof(op))) | 
|  | return -EFAULT; | 
|  | switch (op.dbg_type) { | 
|  | case SIG_DBG_SINGLE_STEPPING: | 
|  | #if defined(CONFIG_4xx) || defined(CONFIG_BOOKE) | 
|  | if (op.dbg_value) { | 
|  | new_msr |= MSR_DE; | 
|  | new_dbcr0 |= (DBCR0_IDM | DBCR0_IC); | 
|  | } else { | 
|  | new_msr &= ~MSR_DE; | 
|  | new_dbcr0 &= ~(DBCR0_IDM | DBCR0_IC); | 
|  | } | 
|  | #else | 
|  | if (op.dbg_value) | 
|  | new_msr |= MSR_SE; | 
|  | else | 
|  | new_msr &= ~MSR_SE; | 
|  | #endif | 
|  | break; | 
|  | case SIG_DBG_BRANCH_TRACING: | 
|  | #if defined(CONFIG_4xx) || defined(CONFIG_BOOKE) | 
|  | return -EINVAL; | 
|  | #else | 
|  | if (op.dbg_value) | 
|  | new_msr |= MSR_BE; | 
|  | else | 
|  | new_msr &= ~MSR_BE; | 
|  | #endif | 
|  | break; | 
|  |  | 
|  | default: | 
|  | return -EINVAL; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* We wait until here to actually install the values in the | 
|  | registers so if we fail in the above loop, it will not | 
|  | affect the contents of these registers.  After this point, | 
|  | failure is a problem, anyway, and it's very unlikely unless | 
|  | the user is really doing something wrong. */ | 
|  | regs->msr = new_msr; | 
|  | #if defined(CONFIG_4xx) || defined(CONFIG_BOOKE) | 
|  | current->thread.dbcr0 = new_dbcr0; | 
|  | #endif | 
|  |  | 
|  | /* | 
|  | * If we get a fault copying the context into the kernel's | 
|  | * image of the user's registers, we can't just return -EFAULT | 
|  | * because the user's registers will be corrupted.  For instance | 
|  | * the NIP value may have been updated but not some of the | 
|  | * other registers.  Given that we have done the access_ok | 
|  | * and successfully read the first and last bytes of the region | 
|  | * above, this should only happen in an out-of-memory situation | 
|  | * or if another thread unmaps the region containing the context. | 
|  | * We kill the task with a SIGSEGV in this situation. | 
|  | */ | 
|  | if (do_setcontext(ctx, regs, 1)) { | 
|  | force_sig(SIGSEGV, current); | 
|  | goto out; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * It's not clear whether or why it is desirable to save the | 
|  | * sigaltstack setting on signal delivery and restore it on | 
|  | * signal return.  But other architectures do this and we have | 
|  | * always done it up until now so it is probably better not to | 
|  | * change it.  -- paulus | 
|  | */ | 
|  | do_sigaltstack(&ctx->uc_stack, NULL, regs->gpr[1]); | 
|  |  | 
|  | sigreturn_exit(regs); | 
|  | /* doesn't actually return back to here */ | 
|  |  | 
|  | out: | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * OK, we're invoking a handler | 
|  | */ | 
|  | static void | 
|  | handle_signal(unsigned long sig, struct k_sigaction *ka, | 
|  | siginfo_t *info, sigset_t *oldset, struct pt_regs * regs, | 
|  | unsigned long newsp) | 
|  | { | 
|  | struct sigcontext __user *sc; | 
|  | struct sigregs __user *frame; | 
|  | unsigned long origsp = newsp; | 
|  |  | 
|  | /* Set up Signal Frame */ | 
|  | newsp -= sizeof(struct sigregs); | 
|  | frame = (struct sigregs __user *) newsp; | 
|  |  | 
|  | /* Put a sigcontext on the stack */ | 
|  | newsp -= sizeof(*sc); | 
|  | sc = (struct sigcontext __user *) newsp; | 
|  |  | 
|  | /* create a stack frame for the caller of the handler */ | 
|  | newsp -= __SIGNAL_FRAMESIZE; | 
|  |  | 
|  | if (!access_ok(VERIFY_WRITE, (void __user *) newsp, origsp - newsp)) | 
|  | goto badframe; | 
|  |  | 
|  | #if _NSIG != 64 | 
|  | #error "Please adjust handle_signal()" | 
|  | #endif | 
|  | if (__put_user((unsigned long) ka->sa.sa_handler, &sc->handler) | 
|  | || __put_user(oldset->sig[0], &sc->oldmask) | 
|  | || __put_user(oldset->sig[1], &sc->_unused[3]) | 
|  | || __put_user((struct pt_regs __user *)frame, &sc->regs) | 
|  | || __put_user(sig, &sc->signal)) | 
|  | goto badframe; | 
|  |  | 
|  | if (save_user_regs(regs, &frame->mctx, __NR_sigreturn)) | 
|  | goto badframe; | 
|  |  | 
|  | if (put_user(regs->gpr[1], (unsigned long __user *)newsp)) | 
|  | goto badframe; | 
|  | regs->gpr[1] = newsp; | 
|  | regs->gpr[3] = sig; | 
|  | regs->gpr[4] = (unsigned long) sc; | 
|  | regs->nip = (unsigned long) ka->sa.sa_handler; | 
|  | regs->link = (unsigned long) frame->mctx.tramp; | 
|  | regs->trap = 0; | 
|  |  | 
|  | return; | 
|  |  | 
|  | badframe: | 
|  | #ifdef DEBUG_SIG | 
|  | printk("badframe in handle_signal, regs=%p frame=%p newsp=%lx\n", | 
|  | regs, frame, newsp); | 
|  | #endif | 
|  | force_sigsegv(sig, current); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Do a signal return; undo the signal stack. | 
|  | */ | 
|  | int sys_sigreturn(int r3, int r4, int r5, int r6, int r7, int r8, | 
|  | struct pt_regs *regs) | 
|  | { | 
|  | struct sigcontext __user *sc; | 
|  | struct sigcontext sigctx; | 
|  | struct mcontext __user *sr; | 
|  | sigset_t set; | 
|  |  | 
|  | /* Always make any pending restarted system calls return -EINTR */ | 
|  | current_thread_info()->restart_block.fn = do_no_restart_syscall; | 
|  |  | 
|  | sc = (struct sigcontext __user *)(regs->gpr[1] + __SIGNAL_FRAMESIZE); | 
|  | if (copy_from_user(&sigctx, sc, sizeof(sigctx))) | 
|  | goto badframe; | 
|  |  | 
|  | set.sig[0] = sigctx.oldmask; | 
|  | set.sig[1] = sigctx._unused[3]; | 
|  | restore_sigmask(&set); | 
|  |  | 
|  | sr = (struct mcontext __user *) sigctx.regs; | 
|  | if (!access_ok(VERIFY_READ, sr, sizeof(*sr)) | 
|  | || restore_user_regs(regs, sr, 1)) | 
|  | goto badframe; | 
|  |  | 
|  | sigreturn_exit(regs);		/* doesn't return */ | 
|  | return 0; | 
|  |  | 
|  | badframe: | 
|  | force_sig(SIGSEGV, current); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Note that 'init' is a special process: it doesn't get signals it doesn't | 
|  | * want to handle. Thus you cannot kill init even with a SIGKILL even by | 
|  | * mistake. | 
|  | */ | 
|  | int do_signal(sigset_t *oldset, struct pt_regs *regs) | 
|  | { | 
|  | siginfo_t info; | 
|  | struct k_sigaction ka; | 
|  | unsigned long frame, newsp; | 
|  | int signr, ret; | 
|  |  | 
|  | if (try_to_freeze()) { | 
|  | signr = 0; | 
|  | if (!signal_pending(current)) | 
|  | goto no_signal; | 
|  | } | 
|  |  | 
|  | if (!oldset) | 
|  | oldset = ¤t->blocked; | 
|  |  | 
|  | newsp = frame = 0; | 
|  |  | 
|  | signr = get_signal_to_deliver(&info, &ka, regs, NULL); | 
|  | no_signal: | 
|  | if (TRAP(regs) == 0x0C00		/* System Call! */ | 
|  | && regs->ccr & 0x10000000		/* error signalled */ | 
|  | && ((ret = regs->gpr[3]) == ERESTARTSYS | 
|  | || ret == ERESTARTNOHAND || ret == ERESTARTNOINTR | 
|  | || ret == ERESTART_RESTARTBLOCK)) { | 
|  |  | 
|  | if (signr > 0 | 
|  | && (ret == ERESTARTNOHAND || ret == ERESTART_RESTARTBLOCK | 
|  | || (ret == ERESTARTSYS | 
|  | && !(ka.sa.sa_flags & SA_RESTART)))) { | 
|  | /* make the system call return an EINTR error */ | 
|  | regs->result = -EINTR; | 
|  | regs->gpr[3] = EINTR; | 
|  | /* note that the cr0.SO bit is already set */ | 
|  | } else { | 
|  | regs->nip -= 4;	/* Back up & retry system call */ | 
|  | regs->result = 0; | 
|  | regs->trap = 0; | 
|  | if (ret == ERESTART_RESTARTBLOCK) | 
|  | regs->gpr[0] = __NR_restart_syscall; | 
|  | else | 
|  | regs->gpr[3] = regs->orig_gpr3; | 
|  | } | 
|  | } | 
|  |  | 
|  | if (signr == 0) | 
|  | return 0;		/* no signals delivered */ | 
|  |  | 
|  | if ((ka.sa.sa_flags & SA_ONSTACK) && current->sas_ss_size | 
|  | && !on_sig_stack(regs->gpr[1])) | 
|  | newsp = current->sas_ss_sp + current->sas_ss_size; | 
|  | else | 
|  | newsp = regs->gpr[1]; | 
|  | newsp &= ~0xfUL; | 
|  |  | 
|  | /* Whee!  Actually deliver the signal.  */ | 
|  | if (ka.sa.sa_flags & SA_SIGINFO) | 
|  | handle_rt_signal(signr, &ka, &info, oldset, regs, newsp); | 
|  | else | 
|  | handle_signal(signr, &ka, &info, oldset, regs, newsp); | 
|  |  | 
|  | spin_lock_irq(¤t->sighand->siglock); | 
|  | sigorsets(¤t->blocked,¤t->blocked,&ka.sa.sa_mask); | 
|  | if (!(ka.sa.sa_flags & SA_NODEFER)) | 
|  | sigaddset(¤t->blocked, signr); | 
|  | recalc_sigpending(); | 
|  | spin_unlock_irq(¤t->sighand->siglock); | 
|  |  | 
|  | return 1; | 
|  | } | 
|  |  |