| /*  | 
 |  * Linux/PA-RISC Project (http://www.parisc-linux.org/) | 
 |  *  | 
 |  * System call entry code Copyright (c) Matthew Wilcox 1999 <willy@bofh.ai> | 
 |  * Licensed under the GNU GPL. | 
 |  * thanks to Philipp Rumpf, Mike Shaver and various others | 
 |  * sorry about the wall, puffin.. | 
 |  */ | 
 |  | 
 | #include <asm/asm-offsets.h> | 
 | #include <asm/unistd.h> | 
 | #include <asm/errno.h> | 
 | #include <asm/page.h> | 
 | #include <asm/psw.h> | 
 | #include <asm/thread_info.h> | 
 | #include <asm/assembly.h> | 
 | #include <asm/processor.h> | 
 |  | 
 | #include <linux/linkage.h> | 
 |  | 
 | 	/* We fill the empty parts of the gateway page with | 
 |  	 * something that will kill the kernel or a | 
 |  	 * userspace application. | 
 | 	 */ | 
 | #define KILL_INSN	break	0,0 | 
 |  | 
 | 	.level          LEVEL | 
 |  | 
 | 	.text | 
 |  | 
 | 	.import syscall_exit,code | 
 | 	.import syscall_exit_rfi,code | 
 |  | 
 | 	/* Linux gateway page is aliased to virtual page 0 in the kernel | 
 | 	 * address space. Since it is a gateway page it cannot be | 
 | 	 * dereferenced, so null pointers will still fault. We start | 
 | 	 * the actual entry point at 0x100. We put break instructions | 
 | 	 * at the beginning of the page to trap null indirect function | 
 | 	 * pointers. | 
 | 	 */ | 
 |  | 
 | 	.align PAGE_SIZE | 
 | ENTRY(linux_gateway_page) | 
 |  | 
 |         /* ADDRESS 0x00 to 0xb0 = 176 bytes / 4 bytes per insn = 44 insns */ | 
 | 	.rept 44 | 
 | 	KILL_INSN | 
 | 	.endr | 
 |  | 
 | 	/* ADDRESS 0xb0 to 0xb4, lws uses 1 insns for entry */ | 
 | 	/* Light-weight-syscall entry must always be located at 0xb0 */ | 
 | 	/* WARNING: Keep this number updated with table size changes */ | 
 | #define __NR_lws_entries (2) | 
 |  | 
 | lws_entry: | 
 | 	/* Unconditional branch to lws_start, located on the  | 
 | 	   same gateway page */ | 
 | 	b,n	lws_start | 
 |  | 
 | 	/* Fill from 0xb4 to 0xe0 */ | 
 | 	.rept 11 | 
 | 	KILL_INSN | 
 | 	.endr | 
 |  | 
 | 	/* This function MUST be located at 0xe0 for glibc's threading  | 
 | 	mechanism to work. DO NOT MOVE THIS CODE EVER! */ | 
 | set_thread_pointer: | 
 | 	gate	.+8, %r0		/* increase privilege */ | 
 | 	depi	3, 31, 2, %r31		/* Ensure we return into user mode. */ | 
 | 	be	0(%sr7,%r31)		/* return to user space */ | 
 | 	mtctl	%r26, %cr27		/* move arg0 to the control register */ | 
 |  | 
 | 	/* Increase the chance of trapping if random jumps occur to this | 
 | 	address, fill from 0xf0 to 0x100 */ | 
 | 	.rept 4 | 
 | 	KILL_INSN | 
 | 	.endr | 
 |  | 
 | /* This address must remain fixed at 0x100 for glibc's syscalls to work */ | 
 | 	.align 256 | 
 | linux_gateway_entry: | 
 | 	gate	.+8, %r0			/* become privileged */ | 
 | 	mtsp	%r0,%sr4			/* get kernel space into sr4 */ | 
 | 	mtsp	%r0,%sr5			/* get kernel space into sr5 */ | 
 | 	mtsp	%r0,%sr6			/* get kernel space into sr6 */ | 
 | 	mfsp    %sr7,%r1                        /* save user sr7 */ | 
 | 	mtsp    %r1,%sr3                        /* and store it in sr3 */ | 
 |  | 
 | #ifdef CONFIG_64BIT | 
 | 	/* for now we can *always* set the W bit on entry to the syscall | 
 | 	 * since we don't support wide userland processes.  We could | 
 | 	 * also save the current SM other than in r0 and restore it on | 
 | 	 * exit from the syscall, and also use that value to know | 
 | 	 * whether to do narrow or wide syscalls. -PB | 
 | 	 */ | 
 | 	ssm	PSW_SM_W, %r1 | 
 | 	extrd,u	%r1,PSW_W_BIT,1,%r1 | 
 | 	/* sp must be aligned on 4, so deposit the W bit setting into | 
 | 	 * the bottom of sp temporarily */ | 
 | 	or,ev	%r1,%r30,%r30 | 
 | 	b,n	1f | 
 | 	/* The top halves of argument registers must be cleared on syscall | 
 | 	 * entry from narrow executable. | 
 | 	 */ | 
 | 	depdi	0, 31, 32, %r26 | 
 | 	depdi	0, 31, 32, %r25 | 
 | 	depdi	0, 31, 32, %r24 | 
 | 	depdi	0, 31, 32, %r23 | 
 | 	depdi	0, 31, 32, %r22 | 
 | 	depdi	0, 31, 32, %r21 | 
 | 1:	 | 
 | #endif | 
 | 	mfctl   %cr30,%r1 | 
 | 	xor     %r1,%r30,%r30                   /* ye olde xor trick */ | 
 | 	xor     %r1,%r30,%r1 | 
 | 	xor     %r1,%r30,%r30 | 
 | 	 | 
 | 	ldo     THREAD_SZ_ALGN+FRAME_SIZE(%r30),%r30  /* set up kernel stack */ | 
 |  | 
 | 	/* N.B.: It is critical that we don't set sr7 to 0 until r30 | 
 | 	 *       contains a valid kernel stack pointer. It is also | 
 | 	 *       critical that we don't start using the kernel stack | 
 | 	 *       until after sr7 has been set to 0. | 
 | 	 */ | 
 |  | 
 | 	mtsp	%r0,%sr7			/* get kernel space into sr7 */ | 
 | 	STREGM	%r1,FRAME_SIZE(%r30)		/* save r1 (usp) here for now */ | 
 | 	mfctl	%cr30,%r1			/* get task ptr in %r1 */ | 
 | 	LDREG	TI_TASK(%r1),%r1 | 
 |  | 
 | 	/* Save some registers for sigcontext and potential task | 
 | 	   switch (see entry.S for the details of which ones are | 
 | 	   saved/restored).  TASK_PT_PSW is zeroed so we can see whether | 
 | 	   a process is on a syscall or not.  For an interrupt the real | 
 | 	   PSW value is stored.  This is needed for gdb and sys_ptrace. */ | 
 | 	STREG	%r0,  TASK_PT_PSW(%r1) | 
 | 	STREG	%r2,  TASK_PT_GR2(%r1)		/* preserve rp */ | 
 | 	STREG	%r19, TASK_PT_GR19(%r1) | 
 |  | 
 | 	LDREGM	-FRAME_SIZE(%r30), %r2		/* get users sp back */ | 
 | #ifdef CONFIG_64BIT | 
 | 	extrd,u	%r2,63,1,%r19			/* W hidden in bottom bit */ | 
 | #if 0 | 
 | 	xor	%r19,%r2,%r2			/* clear bottom bit */ | 
 | 	depd,z	%r19,1,1,%r19 | 
 | 	std	%r19,TASK_PT_PSW(%r1) | 
 | #endif | 
 | #endif | 
 | 	STREG	%r2,  TASK_PT_GR30(%r1)		/* ... and save it */ | 
 | 	 | 
 | 	STREG	%r20, TASK_PT_GR20(%r1)		/* Syscall number */ | 
 | 	STREG	%r21, TASK_PT_GR21(%r1) | 
 | 	STREG	%r22, TASK_PT_GR22(%r1) | 
 | 	STREG	%r23, TASK_PT_GR23(%r1)		/* 4th argument */ | 
 | 	STREG	%r24, TASK_PT_GR24(%r1)		/* 3rd argument */ | 
 | 	STREG	%r25, TASK_PT_GR25(%r1)		/* 2nd argument */ | 
 | 	STREG	%r26, TASK_PT_GR26(%r1)	 	/* 1st argument */ | 
 | 	STREG	%r27, TASK_PT_GR27(%r1)		/* user dp */ | 
 | 	STREG   %r28, TASK_PT_GR28(%r1)         /* return value 0 */ | 
 | 	STREG   %r28, TASK_PT_ORIG_R28(%r1)     /* return value 0 (saved for signals) */ | 
 | 	STREG	%r29, TASK_PT_GR29(%r1)		/* return value 1 */ | 
 | 	STREG	%r31, TASK_PT_GR31(%r1)		/* preserve syscall return ptr */ | 
 | 	 | 
 | 	ldo	TASK_PT_FR0(%r1), %r27		/* save fpregs from the kernel */ | 
 | 	save_fp	%r27				/* or potential task switch  */ | 
 |  | 
 | 	mfctl	%cr11, %r27			/* i.e. SAR */ | 
 | 	STREG	%r27, TASK_PT_SAR(%r1) | 
 |  | 
 | 	loadgp | 
 |  | 
 | #ifdef CONFIG_64BIT | 
 | 	ldo	-16(%r30),%r29			/* Reference param save area */ | 
 | 	copy	%r19,%r2			/* W bit back to r2 */ | 
 | #else | 
 | 	/* no need to save these on stack in wide mode because the first 8 | 
 | 	 * args are passed in registers */ | 
 | 	stw     %r22, -52(%r30)                 /* 5th argument */ | 
 | 	stw     %r21, -56(%r30)                 /* 6th argument */ | 
 | #endif | 
 |  | 
 | 	/* Are we being ptraced? */ | 
 | 	mfctl	%cr30, %r1 | 
 | 	LDREG	TI_TASK(%r1),%r1 | 
 | 	ldw	TASK_PTRACE(%r1), %r1 | 
 | 	bb,<,n	%r1,31,.Ltracesys | 
 | 	 | 
 | 	/* Note!  We cannot use the syscall table that is mapped | 
 | 	nearby since the gateway page is mapped execute-only. */ | 
 |  | 
 | #ifdef CONFIG_64BIT | 
 | 	ldil	L%sys_call_table, %r1 | 
 | 	or,=	%r2,%r2,%r2 | 
 | 	addil	L%(sys_call_table64-sys_call_table), %r1 | 
 | 	ldo	R%sys_call_table(%r1), %r19 | 
 | 	or,=	%r2,%r2,%r2 | 
 | 	ldo	R%sys_call_table64(%r1), %r19 | 
 | #else | 
 | 	ldil	L%sys_call_table, %r1 | 
 | 	ldo     R%sys_call_table(%r1), %r19 | 
 | #endif	 | 
 | 	comiclr,>>	__NR_Linux_syscalls, %r20, %r0 | 
 | 	b,n	.Lsyscall_nosys | 
 | 	 | 
 | 	LDREGX  %r20(%r19), %r19 | 
 |  | 
 | 	/* If this is a sys_rt_sigreturn call, and the signal was received | 
 | 	 * when not in_syscall, then we want to return via syscall_exit_rfi, | 
 | 	 * not syscall_exit.  Signal no. in r20, in_syscall in r25 (see | 
 | 	 * trampoline code in signal.c). | 
 | 	 */ | 
 | 	ldi	__NR_rt_sigreturn,%r2 | 
 | 	comb,=	%r2,%r20,.Lrt_sigreturn | 
 | .Lin_syscall: | 
 | 	ldil	L%syscall_exit,%r2 | 
 | 	be      0(%sr7,%r19) | 
 | 	ldo	R%syscall_exit(%r2),%r2 | 
 | .Lrt_sigreturn: | 
 | 	comib,<> 0,%r25,.Lin_syscall | 
 | 	ldil	L%syscall_exit_rfi,%r2 | 
 | 	be      0(%sr7,%r19) | 
 | 	ldo	R%syscall_exit_rfi(%r2),%r2 | 
 |  | 
 | 	/* Note!  Because we are not running where we were linked, any | 
 | 	calls to functions external to this file must be indirect.  To | 
 | 	be safe, we apply the opposite rule to functions within this | 
 | 	file, with local labels given to them to ensure correctness. */ | 
 | 	 | 
 | .Lsyscall_nosys: | 
 | syscall_nosys: | 
 | 	ldil	L%syscall_exit,%r1 | 
 | 	be	R%syscall_exit(%sr7,%r1) | 
 | 	ldo	-ENOSYS(%r0),%r28		   /* set errno */ | 
 |  | 
 |  | 
 | /* Warning! This trace code is a virtual duplicate of the code above so be | 
 |  * sure to maintain both! */ | 
 | .Ltracesys: | 
 | tracesys: | 
 | 	/* Need to save more registers so the debugger can see where we | 
 | 	 * are.  This saves only the lower 8 bits of PSW, so that the C | 
 | 	 * bit is still clear on syscalls, and the D bit is set if this | 
 | 	 * full register save path has been executed.  We check the D | 
 | 	 * bit on syscall_return_rfi to determine which registers to | 
 | 	 * restore.  An interrupt results in a full PSW saved with the | 
 | 	 * C bit set, a non-straced syscall entry results in C and D clear | 
 | 	 * in the saved PSW. | 
 | 	 */ | 
 | 	ldo     -THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1      /* get task ptr */ | 
 | 	LDREG	TI_TASK(%r1), %r1 | 
 | 	ssm	0,%r2 | 
 | 	STREG	%r2,TASK_PT_PSW(%r1)		/* Lower 8 bits only!! */ | 
 | 	mfsp	%sr0,%r2 | 
 | 	STREG	%r2,TASK_PT_SR0(%r1) | 
 | 	mfsp	%sr1,%r2 | 
 | 	STREG	%r2,TASK_PT_SR1(%r1) | 
 | 	mfsp	%sr2,%r2 | 
 | 	STREG	%r2,TASK_PT_SR2(%r1) | 
 | 	mfsp	%sr3,%r2 | 
 | 	STREG	%r2,TASK_PT_SR3(%r1) | 
 | 	STREG	%r2,TASK_PT_SR4(%r1) | 
 | 	STREG	%r2,TASK_PT_SR5(%r1) | 
 | 	STREG	%r2,TASK_PT_SR6(%r1) | 
 | 	STREG	%r2,TASK_PT_SR7(%r1) | 
 | 	STREG	%r2,TASK_PT_IASQ0(%r1) | 
 | 	STREG	%r2,TASK_PT_IASQ1(%r1) | 
 | 	LDREG	TASK_PT_GR31(%r1),%r2 | 
 | 	STREG	%r2,TASK_PT_IAOQ0(%r1) | 
 | 	ldo	4(%r2),%r2 | 
 | 	STREG	%r2,TASK_PT_IAOQ1(%r1) | 
 | 	ldo	TASK_REGS(%r1),%r2 | 
 | 	/* reg_save %r2 */ | 
 | 	STREG	%r3,PT_GR3(%r2) | 
 | 	STREG	%r4,PT_GR4(%r2) | 
 | 	STREG	%r5,PT_GR5(%r2) | 
 | 	STREG	%r6,PT_GR6(%r2) | 
 | 	STREG	%r7,PT_GR7(%r2) | 
 | 	STREG	%r8,PT_GR8(%r2) | 
 | 	STREG	%r9,PT_GR9(%r2) | 
 | 	STREG	%r10,PT_GR10(%r2) | 
 | 	STREG	%r11,PT_GR11(%r2) | 
 | 	STREG	%r12,PT_GR12(%r2) | 
 | 	STREG	%r13,PT_GR13(%r2) | 
 | 	STREG	%r14,PT_GR14(%r2) | 
 | 	STREG	%r15,PT_GR15(%r2) | 
 | 	STREG	%r16,PT_GR16(%r2) | 
 | 	STREG	%r17,PT_GR17(%r2) | 
 | 	STREG	%r18,PT_GR18(%r2) | 
 | 	/* Finished saving things for the debugger */ | 
 |  | 
 | 	copy	%r2,%r26 | 
 | 	ldil	L%do_syscall_trace_enter,%r1 | 
 | 	ldil	L%tracesys_next,%r2 | 
 | 	be	R%do_syscall_trace_enter(%sr7,%r1) | 
 | 	ldo	R%tracesys_next(%r2),%r2 | 
 | 	 | 
 | tracesys_next: | 
 | 	/* do_syscall_trace_enter either returned the syscallno, or -1L, | 
 | 	 *  so we skip restoring the PT_GR20 below, since we pulled it from | 
 | 	 *  task->thread.regs.gr[20] above. | 
 | 	 */ | 
 | 	copy	%ret0,%r20 | 
 | 	ldil	L%sys_call_table,%r1 | 
 | 	ldo     R%sys_call_table(%r1), %r19 | 
 |  | 
 | 	ldo     -THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1      /* get task ptr */ | 
 | 	LDREG	TI_TASK(%r1), %r1 | 
 | 	LDREG   TASK_PT_GR26(%r1), %r26		/* Restore the users args */ | 
 | 	LDREG   TASK_PT_GR25(%r1), %r25 | 
 | 	LDREG   TASK_PT_GR24(%r1), %r24 | 
 | 	LDREG   TASK_PT_GR23(%r1), %r23 | 
 | #ifdef CONFIG_64BIT | 
 | 	LDREG   TASK_PT_GR22(%r1), %r22 | 
 | 	LDREG   TASK_PT_GR21(%r1), %r21 | 
 | 	ldo	-16(%r30),%r29			/* Reference param save area */ | 
 | #endif | 
 |  | 
 | 	comiclr,>>=	__NR_Linux_syscalls, %r20, %r0 | 
 | 	b,n	.Lsyscall_nosys | 
 |  | 
 | 	LDREGX  %r20(%r19), %r19 | 
 |  | 
 | 	/* If this is a sys_rt_sigreturn call, and the signal was received | 
 | 	 * when not in_syscall, then we want to return via syscall_exit_rfi, | 
 | 	 * not syscall_exit.  Signal no. in r20, in_syscall in r25 (see | 
 | 	 * trampoline code in signal.c). | 
 | 	 */ | 
 | 	ldi	__NR_rt_sigreturn,%r2 | 
 | 	comb,=	%r2,%r20,.Ltrace_rt_sigreturn | 
 | .Ltrace_in_syscall: | 
 | 	ldil	L%tracesys_exit,%r2 | 
 | 	be      0(%sr7,%r19) | 
 | 	ldo	R%tracesys_exit(%r2),%r2 | 
 |  | 
 | 	/* Do *not* call this function on the gateway page, because it | 
 | 	makes a direct call to syscall_trace. */ | 
 | 	 | 
 | tracesys_exit: | 
 | 	ldo     -THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1      /* get task ptr */ | 
 | 	LDREG	TI_TASK(%r1), %r1 | 
 | #ifdef CONFIG_64BIT | 
 | 	ldo	-16(%r30),%r29			/* Reference param save area */ | 
 | #endif | 
 | 	ldo	TASK_REGS(%r1),%r26 | 
 | 	bl	do_syscall_trace_exit,%r2 | 
 | 	STREG   %r28,TASK_PT_GR28(%r1)          /* save return value now */ | 
 | 	ldo     -THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1      /* get task ptr */ | 
 | 	LDREG	TI_TASK(%r1), %r1 | 
 | 	LDREG   TASK_PT_GR28(%r1), %r28		/* Restore return val. */ | 
 |  | 
 | 	ldil	L%syscall_exit,%r1 | 
 | 	be,n	R%syscall_exit(%sr7,%r1) | 
 |  | 
 | .Ltrace_rt_sigreturn: | 
 | 	comib,<> 0,%r25,.Ltrace_in_syscall | 
 | 	ldil	L%tracesys_sigexit,%r2 | 
 | 	be      0(%sr7,%r19) | 
 | 	ldo	R%tracesys_sigexit(%r2),%r2 | 
 |  | 
 | tracesys_sigexit: | 
 | 	ldo     -THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1      /* get task ptr */ | 
 | 	LDREG	TI_TASK(%r1), %r1 | 
 | #ifdef CONFIG_64BIT | 
 | 	ldo	-16(%r30),%r29			/* Reference param save area */ | 
 | #endif | 
 | 	bl	do_syscall_trace_exit,%r2 | 
 | 	ldo	TASK_REGS(%r1),%r26 | 
 |  | 
 | 	ldil	L%syscall_exit_rfi,%r1 | 
 | 	be,n	R%syscall_exit_rfi(%sr7,%r1) | 
 |  | 
 |  | 
 | 	/********************************************************* | 
 | 		32/64-bit Light-Weight-Syscall ABI | 
 |  | 
 | 		* - Indicates a hint for userspace inline asm | 
 | 		implementations. | 
 |  | 
 | 		Syscall number (caller-saves) | 
 | 	        - %r20 | 
 | 	        * In asm clobber. | 
 |  | 
 | 		Argument registers (caller-saves) | 
 | 	        - %r26, %r25, %r24, %r23, %r22 | 
 | 	        * In asm input. | 
 |  | 
 | 		Return registers (caller-saves) | 
 | 	        - %r28 (return), %r21 (errno) | 
 | 	        * In asm output. | 
 |  | 
 | 		Caller-saves registers | 
 | 	        - %r1, %r27, %r29 | 
 | 	        - %r2 (return pointer) | 
 | 	        - %r31 (ble link register) | 
 | 	        * In asm clobber. | 
 |  | 
 | 		Callee-saves registers | 
 | 	        - %r3-%r18 | 
 | 	        - %r30 (stack pointer) | 
 | 	        * Not in asm clobber. | 
 |  | 
 | 		If userspace is 32-bit: | 
 | 		Callee-saves registers | 
 | 	        - %r19 (32-bit PIC register) | 
 |  | 
 | 		Differences from 32-bit calling convention: | 
 | 		- Syscall number in %r20 | 
 | 		- Additional argument register %r22 (arg4) | 
 | 		- Callee-saves %r19. | 
 |  | 
 | 		If userspace is 64-bit: | 
 | 		Callee-saves registers | 
 | 		- %r27 (64-bit PIC register) | 
 |  | 
 | 		Differences from 64-bit calling convention: | 
 | 		- Syscall number in %r20 | 
 | 		- Additional argument register %r22 (arg4) | 
 | 		- Callee-saves %r27. | 
 |  | 
 | 		Error codes returned by entry path: | 
 |  | 
 | 		ENOSYS - r20 was an invalid LWS number. | 
 |  | 
 | 	*********************************************************/ | 
 | lws_start: | 
 | 	/* Gate and ensure we return to userspace */ | 
 | 	gate	.+8, %r0 | 
 | 	depi	3, 31, 2, %r31	/* Ensure we return to userspace */ | 
 |  | 
 | #ifdef CONFIG_64BIT | 
 | 	/* FIXME: If we are a 64-bit kernel just | 
 | 	 *        turn this on unconditionally. | 
 | 	 */ | 
 | 	ssm	PSW_SM_W, %r1 | 
 | 	extrd,u	%r1,PSW_W_BIT,1,%r1 | 
 | 	/* sp must be aligned on 4, so deposit the W bit setting into | 
 | 	 * the bottom of sp temporarily */ | 
 | 	or,ev	%r1,%r30,%r30 | 
 |  | 
 | 	/* Clip LWS number to a 32-bit value always */ | 
 | 	depdi	0, 31, 32, %r20 | 
 | #endif	 | 
 |  | 
 |         /* Is the lws entry number valid? */ | 
 | 	comiclr,>>=	__NR_lws_entries, %r20, %r0 | 
 | 	b,n	lws_exit_nosys | 
 |  | 
 | 	/* WARNING: Trashing sr2 and sr3 */ | 
 | 	mfsp	%sr7,%r1			/* get userspace into sr3 */ | 
 | 	mtsp	%r1,%sr3 | 
 | 	mtsp	%r0,%sr2			/* get kernel space into sr2 */ | 
 |  | 
 | 	/* Load table start */ | 
 | 	ldil	L%lws_table, %r1 | 
 | 	ldo	R%lws_table(%r1), %r28	/* Scratch use of r28 */ | 
 | 	LDREGX	%r20(%sr2,r28), %r21	/* Scratch use of r21 */ | 
 |  | 
 | 	/* Jump to lws, lws table pointers already relocated */ | 
 | 	be,n	0(%sr2,%r21) | 
 |  | 
 | lws_exit_nosys: | 
 | 	ldo	-ENOSYS(%r0),%r21		   /* set errno */ | 
 | 	/* Fall through: Return to userspace */ | 
 |  | 
 | lws_exit: | 
 | #ifdef CONFIG_64BIT | 
 | 	/* decide whether to reset the wide mode bit | 
 | 	 * | 
 | 	 * For a syscall, the W bit is stored in the lowest bit | 
 | 	 * of sp.  Extract it and reset W if it is zero */ | 
 | 	extrd,u,*<>	%r30,63,1,%r1 | 
 | 	rsm	PSW_SM_W, %r0 | 
 | 	/* now reset the lowest bit of sp if it was set */ | 
 | 	xor	%r30,%r1,%r30 | 
 | #endif | 
 | 	be,n	0(%sr3, %r31) | 
 |  | 
 |  | 
 | 	 | 
 | 	/*************************************************** | 
 | 		Implementing CAS as an atomic operation: | 
 |  | 
 | 		%r26 - Address to examine | 
 | 		%r25 - Old value to check (old) | 
 | 		%r24 - New value to set (new) | 
 | 		%r28 - Return prev through this register. | 
 | 		%r21 - Kernel error code | 
 |  | 
 | 		If debugging is DISabled: | 
 |  | 
 | 		%r21 has the following meanings: | 
 |  | 
 | 		EAGAIN - CAS is busy, ldcw failed, try again. | 
 | 		EFAULT - Read or write failed.		 | 
 |  | 
 | 		If debugging is enabled: | 
 |  | 
 | 		EDEADLOCK - CAS called recursively. | 
 | 		EAGAIN && r28 == 1 - CAS is busy. Lock contended. | 
 | 		EAGAIN && r28 == 2 - CAS is busy. ldcw failed. | 
 | 		EFAULT - Read or write failed. | 
 |  | 
 | 		Scratch: r20, r28, r1 | 
 |  | 
 | 	****************************************************/ | 
 |  | 
 | 	/* Do not enable LWS debugging */ | 
 | #define ENABLE_LWS_DEBUG 0  | 
 |  | 
 | 	/* ELF64 Process entry path */ | 
 | lws_compare_and_swap64: | 
 | #ifdef CONFIG_64BIT | 
 | 	b,n	lws_compare_and_swap | 
 | #else | 
 | 	/* If we are not a 64-bit kernel, then we don't | 
 | 	 * have 64-bit input registers, and calling | 
 | 	 * the 64-bit LWS CAS returns ENOSYS. | 
 | 	 */ | 
 | 	b,n	lws_exit_nosys | 
 | #endif | 
 |  | 
 | 	/* ELF32 Process entry path */ | 
 | lws_compare_and_swap32: | 
 | #ifdef CONFIG_64BIT | 
 | 	/* Clip all the input registers */ | 
 | 	depdi	0, 31, 32, %r26 | 
 | 	depdi	0, 31, 32, %r25 | 
 | 	depdi	0, 31, 32, %r24 | 
 | #endif | 
 |  | 
 | lws_compare_and_swap: | 
 | #ifdef CONFIG_SMP | 
 | 	/* Load start of lock table */ | 
 | 	ldil	L%lws_lock_start, %r20 | 
 | 	ldo	R%lws_lock_start(%r20), %r28 | 
 |  | 
 | 	/* Extract four bits from r26 and hash lock (Bits 4-7) */ | 
 | 	extru  %r26, 27, 4, %r20 | 
 |  | 
 | 	/* Find lock to use, the hash is either one of 0 to | 
 | 	   15, multiplied by 16 (keep it 16-byte aligned) | 
 | 	   and add to the lock table offset. */ | 
 | 	shlw	%r20, 4, %r20 | 
 | 	add	%r20, %r28, %r20 | 
 |  | 
 | # if ENABLE_LWS_DEBUG | 
 | 	/*	 | 
 | 		DEBUG, check for deadlock!  | 
 | 		If the thread register values are the same | 
 | 		then we were the one that locked it last and | 
 | 		this is a recurisve call that will deadlock. | 
 | 		We *must* giveup this call and fail. | 
 | 	*/ | 
 | 	ldw	4(%sr2,%r20), %r28			/* Load thread register */ | 
 | 	/* WARNING: If cr27 cycles to the same value we have problems */ | 
 | 	mfctl	%cr27, %r21				/* Get current thread register */ | 
 | 	cmpb,<>,n	%r21, %r28, cas_lock		/* Called recursive? */ | 
 | 	b	lws_exit				/* Return error! */ | 
 | 	ldo	-EDEADLOCK(%r0), %r21 | 
 | cas_lock: | 
 | 	cmpb,=,n	%r0, %r28, cas_nocontend	/* Is nobody using it? */ | 
 | 	ldo	1(%r0), %r28				/* 1st case */ | 
 | 	b	lws_exit				/* Contended... */ | 
 | 	ldo	-EAGAIN(%r0), %r21			/* Spin in userspace */ | 
 | cas_nocontend: | 
 | # endif | 
 | /* ENABLE_LWS_DEBUG */ | 
 |  | 
 | 	LDCW	0(%sr2,%r20), %r28			/* Try to acquire the lock */ | 
 | 	cmpb,<>,n	%r0, %r28, cas_action		/* Did we get it? */ | 
 | cas_wouldblock: | 
 | 	ldo	2(%r0), %r28				/* 2nd case */ | 
 | 	b	lws_exit				/* Contended... */ | 
 | 	ldo	-EAGAIN(%r0), %r21			/* Spin in userspace */ | 
 | #endif | 
 | /* CONFIG_SMP */ | 
 |  | 
 | 	/* | 
 | 		prev = *addr; | 
 | 		if ( prev == old ) | 
 | 		  *addr = new; | 
 | 		return prev; | 
 | 	*/ | 
 |  | 
 | 	/* NOTES: | 
 | 		This all works becuse intr_do_signal | 
 | 		and schedule both check the return iasq | 
 | 		and see that we are on the kernel page | 
 | 		so this process is never scheduled off | 
 | 		or is ever sent any signal of any sort, | 
 | 		thus it is wholly atomic from usrspaces | 
 | 		perspective | 
 | 	*/ | 
 | cas_action: | 
 | #if defined CONFIG_SMP && ENABLE_LWS_DEBUG | 
 | 	/* DEBUG */ | 
 | 	mfctl	%cr27, %r1 | 
 | 	stw	%r1, 4(%sr2,%r20) | 
 | #endif | 
 | 	/* The load and store could fail */ | 
 | 1:	ldw	0(%sr3,%r26), %r28 | 
 | 	sub,<>	%r28, %r25, %r0 | 
 | 2:	stw	%r24, 0(%sr3,%r26) | 
 | #ifdef CONFIG_SMP | 
 | 	/* Free lock */ | 
 | 	stw	%r20, 0(%sr2,%r20) | 
 | # if ENABLE_LWS_DEBUG | 
 | 	/* Clear thread register indicator */ | 
 | 	stw	%r0, 4(%sr2,%r20) | 
 | # endif | 
 | #endif | 
 | 	/* Return to userspace, set no error */ | 
 | 	b	lws_exit | 
 | 	copy	%r0, %r21 | 
 |  | 
 | 3:		 | 
 | 	/* Error occured on load or store */ | 
 | #ifdef CONFIG_SMP | 
 | 	/* Free lock */ | 
 | 	stw	%r20, 0(%sr2,%r20) | 
 | # if ENABLE_LWS_DEBUG | 
 | 	stw	%r0, 4(%sr2,%r20) | 
 | # endif | 
 | #endif | 
 | 	b	lws_exit | 
 | 	ldo	-EFAULT(%r0),%r21	/* set errno */ | 
 | 	nop | 
 | 	nop | 
 | 	nop | 
 | 	nop | 
 |  | 
 | 	/* Two exception table entries, one for the load, | 
 | 	   the other for the store. Either return -EFAULT. | 
 | 	   Each of the entries must be relocated. */ | 
 | 	.section __ex_table,"aw" | 
 | 	ASM_ULONG_INSN (1b - linux_gateway_page), (3b - linux_gateway_page) | 
 | 	ASM_ULONG_INSN (2b - linux_gateway_page), (3b - linux_gateway_page) | 
 | 	.previous | 
 |  | 
 |  | 
 | 	/* Make sure nothing else is placed on this page */ | 
 | 	.align PAGE_SIZE | 
 | END(linux_gateway_page) | 
 | ENTRY(end_linux_gateway_page) | 
 |  | 
 | 	/* Relocate symbols assuming linux_gateway_page is mapped | 
 | 	   to virtual address 0x0 */ | 
 |  | 
 | #define LWS_ENTRY(_name_) ASM_ULONG_INSN (lws_##_name_ - linux_gateway_page) | 
 |  | 
 | 	.section .rodata,"a" | 
 |  | 
 | 	.align PAGE_SIZE | 
 | 	/* Light-weight-syscall table */ | 
 | 	/* Start of lws table. */ | 
 | ENTRY(lws_table) | 
 | 	LWS_ENTRY(compare_and_swap32)	/* 0 - ELF32 Atomic compare and swap */ | 
 | 	LWS_ENTRY(compare_and_swap64)	/* 1 - ELF64 Atomic compare and swap */ | 
 | END(lws_table) | 
 | 	/* End of lws table */ | 
 |  | 
 | 	.align PAGE_SIZE | 
 | ENTRY(sys_call_table) | 
 | #include "syscall_table.S" | 
 | END(sys_call_table) | 
 |  | 
 | #ifdef CONFIG_64BIT | 
 | 	.align PAGE_SIZE | 
 | ENTRY(sys_call_table64) | 
 | #define SYSCALL_TABLE_64BIT | 
 | #include "syscall_table.S" | 
 | END(sys_call_table64) | 
 | #endif | 
 |  | 
 | #ifdef CONFIG_SMP | 
 | 	/* | 
 | 		All light-weight-syscall atomic operations  | 
 | 		will use this set of locks  | 
 |  | 
 | 		NOTE: The lws_lock_start symbol must be | 
 | 		at least 16-byte aligned for safe use | 
 | 		with ldcw. | 
 | 	*/ | 
 | 	.section .data | 
 | 	.align	PAGE_SIZE | 
 | ENTRY(lws_lock_start) | 
 | 	/* lws locks */ | 
 | 	.rept 16 | 
 | 	/* Keep locks aligned at 16-bytes */ | 
 | 	.word 1 | 
 | 	.word 0  | 
 | 	.word 0 | 
 | 	.word 0 | 
 | 	.endr | 
 | END(lws_lock_start) | 
 | 	.previous | 
 | #endif | 
 | /* CONFIG_SMP for lws_lock_start */ | 
 |  | 
 | .end | 
 |  | 
 |  |