|  | /* $Id: etrap.S,v 1.31 2000/01/08 16:38:18 anton Exp $ | 
|  | * etrap.S: Sparc trap window preparation for entry into the | 
|  | *          Linux kernel. | 
|  | * | 
|  | * Copyright (C) 1995 David S. Miller (davem@caip.rutgers.edu) | 
|  | */ | 
|  |  | 
|  | #include <asm/head.h> | 
|  | #include <asm/asi.h> | 
|  | #include <asm/contregs.h> | 
|  | #include <asm/page.h> | 
|  | #include <asm/psr.h> | 
|  | #include <asm/ptrace.h> | 
|  | #include <asm/winmacro.h> | 
|  | #include <asm/asmmacro.h> | 
|  | #include <asm/thread_info.h> | 
|  |  | 
|  | /* Registers to not touch at all. */ | 
|  | #define t_psr        l0 /* Set by caller */ | 
|  | #define t_pc         l1 /* Set by caller */ | 
|  | #define t_npc        l2 /* Set by caller */ | 
|  | #define t_wim        l3 /* Set by caller */ | 
|  | #define t_twinmask   l4 /* Set at beginning of this entry routine. */ | 
|  | #define t_kstack     l5 /* Set right before pt_regs frame is built */ | 
|  | #define t_retpc      l6 /* If you change this, change winmacro.h header file */ | 
|  | #define t_systable   l7 /* Never touch this, could be the syscall table ptr. */ | 
|  | #define curptr       g6 /* Set after pt_regs frame is built */ | 
|  |  | 
|  | .text | 
|  | .align 4 | 
|  |  | 
|  | /* SEVEN WINDOW PATCH INSTRUCTIONS */ | 
|  | .globl	tsetup_7win_patch1, tsetup_7win_patch2 | 
|  | .globl	tsetup_7win_patch3, tsetup_7win_patch4 | 
|  | .globl	tsetup_7win_patch5, tsetup_7win_patch6 | 
|  | tsetup_7win_patch1:	sll	%t_wim, 0x6, %t_wim | 
|  | tsetup_7win_patch2:	and	%g2, 0x7f, %g2 | 
|  | tsetup_7win_patch3:	and	%g2, 0x7f, %g2 | 
|  | tsetup_7win_patch4:	and	%g1, 0x7f, %g1 | 
|  | tsetup_7win_patch5:	sll	%t_wim, 0x6, %t_wim | 
|  | tsetup_7win_patch6:	and	%g2, 0x7f, %g2 | 
|  | /* END OF PATCH INSTRUCTIONS */ | 
|  |  | 
|  | /* At trap time, interrupts and all generic traps do the | 
|  | * following: | 
|  | * | 
|  | * rd	%psr, %l0 | 
|  | * b	some_handler | 
|  | * rd	%wim, %l3 | 
|  | * nop | 
|  | * | 
|  | * Then 'some_handler' if it needs a trap frame (ie. it has | 
|  | * to call c-code and the trap cannot be handled in-window) | 
|  | * then it does the SAVE_ALL macro in entry.S which does | 
|  | * | 
|  | * sethi	%hi(trap_setup), %l4 | 
|  | * jmpl		%l4 + %lo(trap_setup), %l6 | 
|  | * nop | 
|  | */ | 
|  |  | 
|  | /* 2 3 4  window number | 
|  | * ----- | 
|  | * O T S  mnemonic | 
|  | * | 
|  | * O == Current window before trap | 
|  | * T == Window entered when trap occurred | 
|  | * S == Window we will need to save if (1<<T) == %wim | 
|  | * | 
|  | * Before execution gets here, it must be guaranteed that | 
|  | * %l0 contains trap time %psr, %l1 and %l2 contain the | 
|  | * trap pc and npc, and %l3 contains the trap time %wim. | 
|  | */ | 
|  |  | 
|  | .globl	trap_setup, tsetup_patch1, tsetup_patch2 | 
|  | .globl	tsetup_patch3, tsetup_patch4 | 
|  | .globl	tsetup_patch5, tsetup_patch6 | 
|  | trap_setup: | 
|  | /* Calculate mask of trap window.  See if from user | 
|  | * or kernel and branch conditionally. | 
|  | */ | 
|  | mov	1, %t_twinmask | 
|  | andcc	%t_psr, PSR_PS, %g0		 ! fromsupv_p = (psr & PSR_PS) | 
|  | be	trap_setup_from_user		 ! nope, from user mode | 
|  | sll	%t_twinmask, %t_psr, %t_twinmask ! t_twinmask = (1 << psr) | 
|  |  | 
|  | /* From kernel, allocate more kernel stack and | 
|  | * build a pt_regs trap frame. | 
|  | */ | 
|  | sub	%fp, (STACKFRAME_SZ + TRACEREG_SZ), %t_kstack | 
|  | STORE_PT_ALL(t_kstack, t_psr, t_pc, t_npc, g2) | 
|  |  | 
|  | /* See if we are in the trap window. */ | 
|  | andcc	%t_twinmask, %t_wim, %g0 | 
|  | bne	trap_setup_kernel_spill		! in trap window, clean up | 
|  | nop | 
|  |  | 
|  | /* Trap from kernel with a window available. | 
|  | * Just do it... | 
|  | */ | 
|  | jmpl	%t_retpc + 0x8, %g0	! return to caller | 
|  | mov	%t_kstack, %sp		! jump onto new stack | 
|  |  | 
|  | trap_setup_kernel_spill: | 
|  | ld	[%curptr + TI_UWINMASK], %g1 | 
|  | orcc	%g0, %g1, %g0 | 
|  | bne	trap_setup_user_spill	! there are some user windows, yuck | 
|  | /* Spill from kernel, but only kernel windows, adjust | 
|  | * %wim and go. | 
|  | */ | 
|  | srl	%t_wim, 0x1, %g2	! begin computation of new %wim | 
|  | tsetup_patch1: | 
|  | sll	%t_wim, 0x7, %t_wim	! patched on 7 window Sparcs | 
|  | or	%t_wim, %g2, %g2 | 
|  | tsetup_patch2: | 
|  | and	%g2, 0xff, %g2		! patched on 7 window Sparcs | 
|  |  | 
|  | save	%g0, %g0, %g0 | 
|  |  | 
|  | /* Set new %wim value */ | 
|  | wr	%g2, 0x0, %wim | 
|  |  | 
|  | /* Save the kernel window onto the corresponding stack. */ | 
|  | STORE_WINDOW(sp) | 
|  |  | 
|  | restore	%g0, %g0, %g0 | 
|  |  | 
|  | jmpl	%t_retpc + 0x8, %g0	! return to caller | 
|  | mov	%t_kstack, %sp		! and onto new kernel stack | 
|  |  | 
|  | #define STACK_OFFSET (THREAD_SIZE - TRACEREG_SZ - STACKFRAME_SZ) | 
|  |  | 
|  | trap_setup_from_user: | 
|  | /* We can't use %curptr yet. */ | 
|  | LOAD_CURRENT(t_kstack, t_twinmask) | 
|  |  | 
|  | sethi	%hi(STACK_OFFSET), %t_twinmask | 
|  | or	%t_twinmask, %lo(STACK_OFFSET), %t_twinmask | 
|  | add	%t_kstack, %t_twinmask, %t_kstack | 
|  |  | 
|  | mov	1, %t_twinmask | 
|  | sll	%t_twinmask, %t_psr, %t_twinmask ! t_twinmask = (1 << psr) | 
|  |  | 
|  | /* Build pt_regs frame. */ | 
|  | STORE_PT_ALL(t_kstack, t_psr, t_pc, t_npc, g2) | 
|  |  | 
|  | #if 0 | 
|  | /* If we're sure every task_struct is THREAD_SIZE aligned, | 
|  | we can speed this up. */ | 
|  | sethi	%hi(STACK_OFFSET), %curptr | 
|  | or	%curptr, %lo(STACK_OFFSET), %curptr | 
|  | sub	%t_kstack, %curptr, %curptr | 
|  | #else | 
|  | sethi	%hi(~(THREAD_SIZE - 1)), %curptr | 
|  | and	%t_kstack, %curptr, %curptr | 
|  | #endif | 
|  |  | 
|  | /* Clear current_thread_info->w_saved */ | 
|  | st	%g0, [%curptr + TI_W_SAVED] | 
|  |  | 
|  | /* See if we are in the trap window. */ | 
|  | andcc	%t_twinmask, %t_wim, %g0 | 
|  | bne	trap_setup_user_spill		! yep we are | 
|  | orn	%g0, %t_twinmask, %g1		! negate trap win mask into %g1 | 
|  |  | 
|  | /* Trap from user, but not into the invalid window. | 
|  | * Calculate new umask.  The way this works is, | 
|  | * any window from the %wim at trap time until | 
|  | * the window right before the one we are in now, | 
|  | * is a user window.  A diagram: | 
|  | * | 
|  | *      7 6 5 4 3 2 1 0    window number | 
|  | *      --------------- | 
|  | *        I     L T        mnemonic | 
|  | * | 
|  | * Window 'I' is the invalid window in our example, | 
|  | * window 'L' is the window the user was in when | 
|  | * the trap occurred, window T is the trap window | 
|  | * we are in now.  So therefore, windows 5, 4 and | 
|  | * 3 are user windows.  The following sequence | 
|  | * computes the user winmask to represent this. | 
|  | */ | 
|  | subcc	%t_wim, %t_twinmask, %g2 | 
|  | bneg,a	1f | 
|  | sub	%g2, 0x1, %g2 | 
|  | 1: | 
|  | andn	%g2, %t_twinmask, %g2 | 
|  | tsetup_patch3: | 
|  | and	%g2, 0xff, %g2			! patched on 7win Sparcs | 
|  | st	%g2, [%curptr + TI_UWINMASK]	! store new umask | 
|  |  | 
|  | jmpl	%t_retpc + 0x8, %g0		! return to caller | 
|  | mov	%t_kstack, %sp			! and onto kernel stack | 
|  |  | 
|  | trap_setup_user_spill: | 
|  | /* A spill occurred from either kernel or user mode | 
|  | * and there exist some user windows to deal with. | 
|  | * A mask of the currently valid user windows | 
|  | * is in %g1 upon entry to here. | 
|  | */ | 
|  |  | 
|  | tsetup_patch4: | 
|  | and	%g1, 0xff, %g1		! patched on 7win Sparcs, mask | 
|  | srl	%t_wim, 0x1, %g2	! compute new %wim | 
|  | tsetup_patch5: | 
|  | sll	%t_wim, 0x7, %t_wim	! patched on 7win Sparcs | 
|  | or	%t_wim, %g2, %g2	! %g2 is new %wim | 
|  | tsetup_patch6: | 
|  | and	%g2, 0xff, %g2		! patched on 7win Sparcs | 
|  | andn	%g1, %g2, %g1		! clear this bit in %g1 | 
|  | st	%g1, [%curptr + TI_UWINMASK] | 
|  |  | 
|  | save	%g0, %g0, %g0 | 
|  |  | 
|  | wr	%g2, 0x0, %wim | 
|  |  | 
|  | /* Call MMU-architecture dependent stack checking | 
|  | * routine. | 
|  | */ | 
|  | .globl	tsetup_mmu_patchme | 
|  | tsetup_mmu_patchme: | 
|  | b	tsetup_sun4c_stackchk | 
|  | andcc	%sp, 0x7, %g0 | 
|  |  | 
|  | /* Architecture specific stack checking routines.  When either | 
|  | * of these routines are called, the globals are free to use | 
|  | * as they have been safely stashed on the new kernel stack | 
|  | * pointer.  Thus the definition below for simplicity. | 
|  | */ | 
|  | #define glob_tmp     g1 | 
|  |  | 
|  | .globl	tsetup_sun4c_stackchk | 
|  | tsetup_sun4c_stackchk: | 
|  | /* Done by caller: andcc %sp, 0x7, %g0 */ | 
|  | bne	trap_setup_user_stack_is_bolixed | 
|  | sra	%sp, 29, %glob_tmp | 
|  |  | 
|  | add	%glob_tmp, 0x1, %glob_tmp | 
|  | andncc	%glob_tmp, 0x1, %g0 | 
|  | bne	trap_setup_user_stack_is_bolixed | 
|  | and	%sp, 0xfff, %glob_tmp		! delay slot | 
|  |  | 
|  | /* See if our dump area will be on more than one | 
|  | * page. | 
|  | */ | 
|  | add	%glob_tmp, 0x38, %glob_tmp | 
|  | andncc	%glob_tmp, 0xff8, %g0 | 
|  | be	tsetup_sun4c_onepage		! only one page to check | 
|  | lda	[%sp] ASI_PTE, %glob_tmp	! have to check first page anyways | 
|  |  | 
|  | tsetup_sun4c_twopages: | 
|  | /* Is first page ok permission wise? */ | 
|  | srl	%glob_tmp, 29, %glob_tmp | 
|  | cmp	%glob_tmp, 0x6 | 
|  | bne	trap_setup_user_stack_is_bolixed | 
|  | add	%sp, 0x38, %glob_tmp		/* Is second page in vma hole? */ | 
|  |  | 
|  | sra	%glob_tmp, 29, %glob_tmp | 
|  | add	%glob_tmp, 0x1, %glob_tmp | 
|  | andncc	%glob_tmp, 0x1, %g0 | 
|  | bne	trap_setup_user_stack_is_bolixed | 
|  | add	%sp, 0x38, %glob_tmp | 
|  |  | 
|  | lda	[%glob_tmp] ASI_PTE, %glob_tmp | 
|  |  | 
|  | tsetup_sun4c_onepage: | 
|  | srl	%glob_tmp, 29, %glob_tmp | 
|  | cmp	%glob_tmp, 0x6				! can user write to it? | 
|  | bne	trap_setup_user_stack_is_bolixed	! failure | 
|  | nop | 
|  |  | 
|  | STORE_WINDOW(sp) | 
|  |  | 
|  | restore %g0, %g0, %g0 | 
|  |  | 
|  | jmpl	%t_retpc + 0x8, %g0 | 
|  | mov	%t_kstack, %sp | 
|  |  | 
|  | .globl	tsetup_srmmu_stackchk | 
|  | tsetup_srmmu_stackchk: | 
|  | /* Check results of callers andcc %sp, 0x7, %g0 */ | 
|  | bne	trap_setup_user_stack_is_bolixed | 
|  | sethi   %hi(PAGE_OFFSET), %glob_tmp | 
|  |  | 
|  | cmp	%glob_tmp, %sp | 
|  | bleu,a	1f | 
|  | lda	[%g0] ASI_M_MMUREGS, %glob_tmp		! read MMU control | 
|  |  | 
|  | trap_setup_user_stack_is_bolixed: | 
|  | /* From user/kernel into invalid window w/bad user | 
|  | * stack. Save bad user stack, and return to caller. | 
|  | */ | 
|  | SAVE_BOLIXED_USER_STACK(curptr, g3) | 
|  | restore	%g0, %g0, %g0 | 
|  |  | 
|  | jmpl	%t_retpc + 0x8, %g0 | 
|  | mov	%t_kstack, %sp | 
|  |  | 
|  | 1: | 
|  | /* Clear the fault status and turn on the no_fault bit. */ | 
|  | or	%glob_tmp, 0x2, %glob_tmp		! or in no_fault bit | 
|  | sta	%glob_tmp, [%g0] ASI_M_MMUREGS		! set it | 
|  |  | 
|  | /* Dump the registers and cross fingers. */ | 
|  | STORE_WINDOW(sp) | 
|  |  | 
|  | /* Clear the no_fault bit and check the status. */ | 
|  | andn	%glob_tmp, 0x2, %glob_tmp | 
|  | sta	%glob_tmp, [%g0] ASI_M_MMUREGS | 
|  | mov	AC_M_SFAR, %glob_tmp | 
|  | lda	[%glob_tmp] ASI_M_MMUREGS, %g0 | 
|  | mov	AC_M_SFSR, %glob_tmp | 
|  | lda	[%glob_tmp] ASI_M_MMUREGS, %glob_tmp	! save away status of winstore | 
|  | andcc	%glob_tmp, 0x2, %g0			! did we fault? | 
|  | bne	trap_setup_user_stack_is_bolixed	! failure | 
|  | nop | 
|  |  | 
|  | restore %g0, %g0, %g0 | 
|  |  | 
|  | jmpl	%t_retpc + 0x8, %g0 | 
|  | mov	%t_kstack, %sp | 
|  |  |