|  | /* linux/arch/sparc/lib/memset.S: Sparc optimized memset, bzero and clear_user code | 
|  | * Copyright (C) 1991,1996 Free Software Foundation | 
|  | * Copyright (C) 1996,1997 Jakub Jelinek (jj@sunsite.mff.cuni.cz) | 
|  | * Copyright (C) 1996 David S. Miller (davem@caip.rutgers.edu) | 
|  | * | 
|  | * Returns 0, if ok, and number of bytes not yet set if exception | 
|  | * occurs and we were called as clear_user. | 
|  | */ | 
|  |  | 
|  | #include <asm/ptrace.h> | 
|  |  | 
|  | /* Work around cpp -rob */ | 
|  | #define ALLOC #alloc | 
|  | #define EXECINSTR #execinstr | 
|  | #define EX(x,y,a,b) 				\ | 
|  | 98: 	x,y;					\ | 
|  | .section .fixup,ALLOC,EXECINSTR;	\ | 
|  | .align	4;				\ | 
|  | 99:	ba 30f;					\ | 
|  | a, b, %o0;				\ | 
|  | .section __ex_table,ALLOC;		\ | 
|  | .align	4;				\ | 
|  | .word	98b, 99b;			\ | 
|  | .text;					\ | 
|  | .align	4 | 
|  |  | 
|  | #define EXT(start,end,handler) 			\ | 
|  | .section __ex_table,ALLOC;		\ | 
|  | .align	4;				\ | 
|  | .word	start, 0, end, handler;		\ | 
|  | .text;					\ | 
|  | .align	4 | 
|  |  | 
|  | /* Please don't change these macros, unless you change the logic | 
|  | * in the .fixup section below as well. | 
|  | * Store 64 bytes at (BASE + OFFSET) using value SOURCE. */ | 
|  | #define ZERO_BIG_BLOCK(base, offset, source)    \ | 
|  | std	source, [base + offset + 0x00]; \ | 
|  | std	source, [base + offset + 0x08]; \ | 
|  | std	source, [base + offset + 0x10]; \ | 
|  | std	source, [base + offset + 0x18]; \ | 
|  | std	source, [base + offset + 0x20]; \ | 
|  | std	source, [base + offset + 0x28]; \ | 
|  | std	source, [base + offset + 0x30]; \ | 
|  | std	source, [base + offset + 0x38]; | 
|  |  | 
|  | #define ZERO_LAST_BLOCKS(base, offset, source)	\ | 
|  | std	source, [base - offset - 0x38]; \ | 
|  | std	source, [base - offset - 0x30]; \ | 
|  | std	source, [base - offset - 0x28]; \ | 
|  | std	source, [base - offset - 0x20]; \ | 
|  | std	source, [base - offset - 0x18]; \ | 
|  | std	source, [base - offset - 0x10]; \ | 
|  | std	source, [base - offset - 0x08]; \ | 
|  | std	source, [base - offset - 0x00]; | 
|  |  | 
|  | .text | 
|  | .align 4 | 
|  |  | 
|  | .globl  __bzero_begin | 
|  | __bzero_begin: | 
|  |  | 
|  | .globl	__bzero, __memset, | 
|  | .globl	memset | 
|  | .globl	__memset_start, __memset_end | 
|  | __memset_start: | 
|  | __memset: | 
|  | memset: | 
|  | and	%o1, 0xff, %g3 | 
|  | sll	%g3, 8, %g2 | 
|  | or	%g3, %g2, %g3 | 
|  | sll	%g3, 16, %g2 | 
|  | or	%g3, %g2, %g3 | 
|  | b	1f | 
|  | mov	%o2, %o1 | 
|  | 3: | 
|  | cmp	%o2, 3 | 
|  | be	2f | 
|  | EX(stb	%g3, [%o0], sub %o1, 0) | 
|  |  | 
|  | cmp	%o2, 2 | 
|  | be	2f | 
|  | EX(stb	%g3, [%o0 + 0x01], sub %o1, 1) | 
|  |  | 
|  | EX(stb	%g3, [%o0 + 0x02], sub %o1, 2) | 
|  | 2: | 
|  | sub	%o2, 4, %o2 | 
|  | add	%o1, %o2, %o1 | 
|  | b	4f | 
|  | sub	%o0, %o2, %o0 | 
|  |  | 
|  | __bzero: | 
|  | mov	%g0, %g3 | 
|  | 1: | 
|  | cmp	%o1, 7 | 
|  | bleu	7f | 
|  | andcc	%o0, 3, %o2 | 
|  |  | 
|  | bne	3b | 
|  | 4: | 
|  | andcc	%o0, 4, %g0 | 
|  |  | 
|  | be	2f | 
|  | mov	%g3, %g2 | 
|  |  | 
|  | EX(st	%g3, [%o0], sub %o1, 0) | 
|  | sub	%o1, 4, %o1 | 
|  | add	%o0, 4, %o0 | 
|  | 2: | 
|  | andcc	%o1, 0xffffff80, %o3	! Now everything is 8 aligned and o1 is len to run | 
|  | be	9f | 
|  | andcc	%o1, 0x78, %o2 | 
|  | 10: | 
|  | ZERO_BIG_BLOCK(%o0, 0x00, %g2) | 
|  | subcc	%o3, 128, %o3 | 
|  | ZERO_BIG_BLOCK(%o0, 0x40, %g2) | 
|  | 11: | 
|  | EXT(10b, 11b, 20f) | 
|  | bne	10b | 
|  | add	%o0, 128, %o0 | 
|  |  | 
|  | orcc	%o2, %g0, %g0 | 
|  | 9: | 
|  | be	13f | 
|  | andcc	%o1, 7, %o1 | 
|  |  | 
|  | srl	%o2, 1, %o3 | 
|  | set	13f, %o4 | 
|  | sub	%o4, %o3, %o4 | 
|  | jmp	%o4 | 
|  | add	%o0, %o2, %o0 | 
|  |  | 
|  | 12: | 
|  | ZERO_LAST_BLOCKS(%o0, 0x48, %g2) | 
|  | ZERO_LAST_BLOCKS(%o0, 0x08, %g2) | 
|  | 13: | 
|  | be	8f | 
|  | andcc	%o1, 4, %g0 | 
|  |  | 
|  | be	1f | 
|  | andcc	%o1, 2, %g0 | 
|  |  | 
|  | EX(st	%g3, [%o0], and %o1, 7) | 
|  | add	%o0, 4, %o0 | 
|  | 1: | 
|  | be	1f | 
|  | andcc	%o1, 1, %g0 | 
|  |  | 
|  | EX(sth	%g3, [%o0], and %o1, 3) | 
|  | add	%o0, 2, %o0 | 
|  | 1: | 
|  | bne,a	8f | 
|  | EX(stb	%g3, [%o0], and %o1, 1) | 
|  | 8: | 
|  | retl | 
|  | clr	%o0 | 
|  | 7: | 
|  | be	13b | 
|  | orcc	%o1, 0, %g0 | 
|  |  | 
|  | be	0f | 
|  | 8: | 
|  | add	%o0, 1, %o0 | 
|  | subcc	%o1, 1, %o1 | 
|  | bne,a	8b | 
|  | EX(stb	%g3, [%o0 - 1], add %o1, 1) | 
|  | 0: | 
|  | retl | 
|  | clr	%o0 | 
|  | __memset_end: | 
|  |  | 
|  | .section .fixup,#alloc,#execinstr | 
|  | .align	4 | 
|  | 20: | 
|  | cmp	%g2, 8 | 
|  | bleu	1f | 
|  | and	%o1, 0x7f, %o1 | 
|  | sub	%g2, 9, %g2 | 
|  | add	%o3, 64, %o3 | 
|  | 1: | 
|  | sll	%g2, 3, %g2 | 
|  | add	%o3, %o1, %o0 | 
|  | b 30f | 
|  | sub	%o0, %g2, %o0 | 
|  | 21: | 
|  | mov	8, %o0 | 
|  | and	%o1, 7, %o1 | 
|  | sub	%o0, %g2, %o0 | 
|  | sll	%o0, 3, %o0 | 
|  | b 30f | 
|  | add	%o0, %o1, %o0 | 
|  | 30: | 
|  | /* %o4 is faulting address, %o5 is %pc where fault occurred */ | 
|  | save	%sp, -104, %sp | 
|  | mov	%i5, %o0 | 
|  | mov	%i7, %o1 | 
|  | call	lookup_fault | 
|  | mov	%i4, %o2 | 
|  | ret | 
|  | restore | 
|  |  | 
|  | .globl __bzero_end | 
|  | __bzero_end: |