| David S. Miller | 6c70b6f | 2007-08-08 17:11:39 -0700 | [diff] [blame] | 1 | /* GENmemcpy.S: Generic sparc64 memcpy. | 
 | 2 |  * | 
 | 3 |  * Copyright (C) 2007 David S. Miller (davem@davemloft.net) | 
 | 4 |  */ | 
 | 5 |  | 
 | 6 | #ifdef __KERNEL__ | 
 | 7 | #define GLOBAL_SPARE	%g7 | 
 | 8 | #else | 
 | 9 | #define GLOBAL_SPARE	%g5 | 
 | 10 | #endif | 
 | 11 |  | 
 | 12 | #ifndef EX_LD | 
 | 13 | #define EX_LD(x)	x | 
 | 14 | #endif | 
 | 15 |  | 
 | 16 | #ifndef EX_ST | 
 | 17 | #define EX_ST(x)	x | 
 | 18 | #endif | 
 | 19 |  | 
 | 20 | #ifndef EX_RETVAL | 
 | 21 | #define EX_RETVAL(x)	x | 
 | 22 | #endif | 
 | 23 |  | 
 | 24 | #ifndef LOAD | 
 | 25 | #define LOAD(type,addr,dest)	type [addr], dest | 
 | 26 | #endif | 
 | 27 |  | 
 | 28 | #ifndef STORE | 
 | 29 | #define STORE(type,src,addr)	type src, [addr] | 
 | 30 | #endif | 
 | 31 |  | 
 | 32 | #ifndef FUNC_NAME | 
 | 33 | #define FUNC_NAME	GENmemcpy | 
 | 34 | #endif | 
 | 35 |  | 
 | 36 | #ifndef PREAMBLE | 
 | 37 | #define PREAMBLE | 
 | 38 | #endif | 
 | 39 |  | 
 | 40 | #ifndef XCC | 
 | 41 | #define XCC xcc | 
 | 42 | #endif | 
 | 43 |  | 
 | 44 | 	.register	%g2,#scratch | 
 | 45 | 	.register	%g3,#scratch | 
 | 46 |  | 
 | 47 | 	.text | 
 | 48 | 	.align		64 | 
 | 49 |  | 
 | 50 | 	.globl	FUNC_NAME | 
 | 51 | 	.type	FUNC_NAME,#function | 
 | 52 | FUNC_NAME:	/* %o0=dst, %o1=src, %o2=len */ | 
 | 53 | 	srlx		%o2, 31, %g2 | 
 | 54 | 	cmp		%g2, 0 | 
 | 55 | 	tne		%XCC, 5 | 
 | 56 | 	PREAMBLE | 
 | 57 | 	mov		%o0, GLOBAL_SPARE | 
 | 58 |  | 
 | 59 | 	cmp		%o2, 0 | 
 | 60 | 	be,pn		%XCC, 85f | 
 | 61 | 	 or		%o0, %o1, %o3 | 
 | 62 | 	cmp		%o2, 16 | 
 | 63 | 	blu,a,pn	%XCC, 80f | 
 | 64 | 	 or		%o3, %o2, %o3 | 
 | 65 |  | 
 | 66 | 	xor		%o0, %o1, %o4 | 
 | 67 | 	andcc		%o4, 0x7, %g0 | 
 | 68 | 	bne,a,pn	%XCC, 90f | 
 | 69 | 	 sub		%o0, %o1, %o3 | 
 | 70 |  | 
 | 71 | 	and		%o0, 0x7, %o4 | 
 | 72 | 	sub		%o4, 0x8, %o4 | 
 | 73 | 	sub		%g0, %o4, %o4 | 
 | 74 | 	sub		%o2, %o4, %o2 | 
 | 75 | 1:	subcc		%o4, 1, %o4 | 
 | 76 | 	EX_LD(LOAD(ldub, %o1, %g1)) | 
 | 77 | 	EX_ST(STORE(stb, %g1, %o0)) | 
 | 78 | 	add		%o1, 1, %o1 | 
 | 79 | 	bne,pt		%XCC, 1b | 
 | 80 | 	add		%o0, 1, %o0 | 
 | 81 |  | 
 | 82 | 	andn		%o2, 0x7, %g1 | 
 | 83 | 	sub		%o2, %g1, %o2 | 
 | 84 | 1:	subcc		%g1, 0x8, %g1 | 
 | 85 | 	EX_LD(LOAD(ldx, %o1, %g2)) | 
 | 86 | 	EX_ST(STORE(stx, %g2, %o0)) | 
 | 87 | 	add		%o1, 0x8, %o1 | 
 | 88 | 	bne,pt		%XCC, 1b | 
 | 89 | 	 add		%o0, 0x8, %o0 | 
 | 90 |  | 
 | 91 | 	brz,pt		%o2, 85f | 
 | 92 | 	 sub		%o0, %o1, %o3 | 
 | 93 | 	ba,a,pt		%XCC, 90f | 
 | 94 |  | 
 | 95 | 	.align		64 | 
 | 96 | 80: /* 0 < len <= 16 */ | 
 | 97 | 	andcc		%o3, 0x3, %g0 | 
 | 98 | 	bne,pn		%XCC, 90f | 
 | 99 | 	 sub		%o0, %o1, %o3 | 
 | 100 |  | 
 | 101 | 1: | 
 | 102 | 	subcc		%o2, 4, %o2 | 
 | 103 | 	EX_LD(LOAD(lduw, %o1, %g1)) | 
 | 104 | 	EX_ST(STORE(stw, %g1, %o1 + %o3)) | 
 | 105 | 	bgu,pt		%XCC, 1b | 
 | 106 | 	 add		%o1, 4, %o1 | 
 | 107 |  | 
 | 108 | 85:	retl | 
 | 109 | 	 mov		EX_RETVAL(GLOBAL_SPARE), %o0 | 
 | 110 |  | 
 | 111 | 	.align		32 | 
 | 112 | 90: | 
 | 113 | 	subcc		%o2, 1, %o2 | 
 | 114 | 	EX_LD(LOAD(ldub, %o1, %g1)) | 
 | 115 | 	EX_ST(STORE(stb, %g1, %o1 + %o3)) | 
 | 116 | 	bgu,pt		%XCC, 90b | 
 | 117 | 	 add		%o1, 1, %o1 | 
 | 118 | 	retl | 
 | 119 | 	 mov		EX_RETVAL(GLOBAL_SPARE), %o0 | 
 | 120 |  | 
 | 121 | 	.size		FUNC_NAME, .-FUNC_NAME |