| /* | 
 |  * String handling functions for PowerPC. | 
 |  * | 
 |  * Copyright (C) 1996 Paul Mackerras. | 
 |  * | 
 |  * This program is free software; you can redistribute it and/or | 
 |  * modify it under the terms of the GNU General Public License | 
 |  * as published by the Free Software Foundation; either version | 
 |  * 2 of the License, or (at your option) any later version. | 
 |  */ | 
 | #include <asm/processor.h> | 
 | #include <asm/errno.h> | 
 | #include <asm/ppc_asm.h> | 
 |  | 
 | _GLOBAL(memset) | 
 | 	neg	r0,r3 | 
 | 	rlwimi	r4,r4,8,16,23 | 
 | 	andi.	r0,r0,7			/* # bytes to be 8-byte aligned */ | 
 | 	rlwimi	r4,r4,16,0,15 | 
 | 	cmplw	cr1,r5,r0		/* do we get that far? */ | 
 | 	rldimi	r4,r4,32,0 | 
 | 	PPC_MTOCRF	1,r0 | 
 | 	mr	r6,r3 | 
 | 	blt	cr1,8f | 
 | 	beq+	3f			/* if already 8-byte aligned */ | 
 | 	subf	r5,r0,r5 | 
 | 	bf	31,1f | 
 | 	stb	r4,0(r6) | 
 | 	addi	r6,r6,1 | 
 | 1:	bf	30,2f | 
 | 	sth	r4,0(r6) | 
 | 	addi	r6,r6,2 | 
 | 2:	bf	29,3f | 
 | 	stw	r4,0(r6) | 
 | 	addi	r6,r6,4 | 
 | 3:	srdi.	r0,r5,6 | 
 | 	clrldi	r5,r5,58 | 
 | 	mtctr	r0 | 
 | 	beq	5f | 
 | 4:	std	r4,0(r6) | 
 | 	std	r4,8(r6) | 
 | 	std	r4,16(r6) | 
 | 	std	r4,24(r6) | 
 | 	std	r4,32(r6) | 
 | 	std	r4,40(r6) | 
 | 	std	r4,48(r6) | 
 | 	std	r4,56(r6) | 
 | 	addi	r6,r6,64 | 
 | 	bdnz	4b | 
 | 5:	srwi.	r0,r5,3 | 
 | 	clrlwi	r5,r5,29 | 
 | 	PPC_MTOCRF	1,r0 | 
 | 	beq	8f | 
 | 	bf	29,6f | 
 | 	std	r4,0(r6) | 
 | 	std	r4,8(r6) | 
 | 	std	r4,16(r6) | 
 | 	std	r4,24(r6) | 
 | 	addi	r6,r6,32 | 
 | 6:	bf	30,7f | 
 | 	std	r4,0(r6) | 
 | 	std	r4,8(r6) | 
 | 	addi	r6,r6,16 | 
 | 7:	bf	31,8f | 
 | 	std	r4,0(r6) | 
 | 	addi	r6,r6,8 | 
 | 8:	cmpwi	r5,0 | 
 | 	PPC_MTOCRF	1,r5 | 
 | 	beqlr+ | 
 | 	bf	29,9f | 
 | 	stw	r4,0(r6) | 
 | 	addi	r6,r6,4 | 
 | 9:	bf	30,10f | 
 | 	sth	r4,0(r6) | 
 | 	addi	r6,r6,2 | 
 | 10:	bflr	31 | 
 | 	stb	r4,0(r6) | 
 | 	blr | 
 |  | 
 | _GLOBAL(memmove) | 
 | 	cmplw	0,r3,r4 | 
 | 	bgt	.backwards_memcpy | 
 | 	b	.memcpy | 
 |  | 
 | _GLOBAL(backwards_memcpy) | 
 | 	rlwinm.	r7,r5,32-3,3,31		/* r0 = r5 >> 3 */ | 
 | 	add	r6,r3,r5 | 
 | 	add	r4,r4,r5 | 
 | 	beq	2f | 
 | 	andi.	r0,r6,3 | 
 | 	mtctr	r7 | 
 | 	bne	5f | 
 | 1:	lwz	r7,-4(r4) | 
 | 	lwzu	r8,-8(r4) | 
 | 	stw	r7,-4(r6) | 
 | 	stwu	r8,-8(r6) | 
 | 	bdnz	1b | 
 | 	andi.	r5,r5,7 | 
 | 2:	cmplwi	0,r5,4 | 
 | 	blt	3f | 
 | 	lwzu	r0,-4(r4) | 
 | 	subi	r5,r5,4 | 
 | 	stwu	r0,-4(r6) | 
 | 3:	cmpwi	0,r5,0 | 
 | 	beqlr | 
 | 	mtctr	r5 | 
 | 4:	lbzu	r0,-1(r4) | 
 | 	stbu	r0,-1(r6) | 
 | 	bdnz	4b | 
 | 	blr | 
 | 5:	mtctr	r0 | 
 | 6:	lbzu	r7,-1(r4) | 
 | 	stbu	r7,-1(r6) | 
 | 	bdnz	6b | 
 | 	subf	r5,r0,r5 | 
 | 	rlwinm.	r7,r5,32-3,3,31 | 
 | 	beq	2b | 
 | 	mtctr	r7 | 
 | 	b	1b |