| /* align.c - handle alignment exceptions for the Power PC. | 
 |  * | 
 |  * Copyright (c) 1996 Paul Mackerras <paulus@cs.anu.edu.au> | 
 |  * Copyright (c) 1998-1999 TiVo, Inc. | 
 |  *   PowerPC 403GCX modifications. | 
 |  * Copyright (c) 1999 Grant Erickson <grant@lcse.umn.edu> | 
 |  *   PowerPC 403GCX/405GP modifications. | 
 |  * Copyright (c) 2001-2002 PPC64 team, IBM Corp | 
 |  *   64-bit and Power4 support | 
 |  * Copyright (c) 2005 Benjamin Herrenschmidt, IBM Corp | 
 |  *                    <benh@kernel.crashing.org> | 
 |  *   Merge ppc32 and ppc64 implementations | 
 |  * | 
 |  * This program is free software; you can redistribute it and/or | 
 |  * modify it under the terms of the GNU General Public License | 
 |  * as published by the Free Software Foundation; either version | 
 |  * 2 of the License, or (at your option) any later version. | 
 |  */ | 
 |  | 
 | #include <linux/kernel.h> | 
 | #include <linux/mm.h> | 
 | #include <asm/processor.h> | 
 | #include <asm/uaccess.h> | 
 | #include <asm/system.h> | 
 | #include <asm/cache.h> | 
 | #include <asm/cputable.h> | 
 |  | 
 | struct aligninfo { | 
 | 	unsigned char len; | 
 | 	unsigned char flags; | 
 | }; | 
 |  | 
 | #define IS_XFORM(inst)	(((inst) >> 26) == 31) | 
 | #define IS_DSFORM(inst)	(((inst) >> 26) >= 56) | 
 |  | 
 | #define INVALID	{ 0, 0 } | 
 |  | 
 | /* Bits in the flags field */ | 
 | #define LD	0	/* load */ | 
 | #define ST	1	/* store */ | 
 | #define	SE	2	/* sign-extend value */ | 
 | #define F	4	/* to/from fp regs */ | 
 | #define U	8	/* update index register */ | 
 | #define M	0x10	/* multiple load/store */ | 
 | #define SW	0x20	/* byte swap */ | 
 | #define S	0x40	/* single-precision fp or... */ | 
 | #define SX	0x40	/* ... byte count in XER */ | 
 | #define HARD	0x80	/* string, stwcx. */ | 
 |  | 
 | /* DSISR bits reported for a DCBZ instruction: */ | 
 | #define DCBZ	0x5f	/* 8xx/82xx dcbz faults when cache not enabled */ | 
 |  | 
 | #define SWAP(a, b)	(t = (a), (a) = (b), (b) = t) | 
 |  | 
 | /* | 
 |  * The PowerPC stores certain bits of the instruction that caused the | 
 |  * alignment exception in the DSISR register.  This array maps those | 
 |  * bits to information about the operand length and what the | 
 |  * instruction would do. | 
 |  */ | 
 | static struct aligninfo aligninfo[128] = { | 
 | 	{ 4, LD },		/* 00 0 0000: lwz / lwarx */ | 
 | 	INVALID,		/* 00 0 0001 */ | 
 | 	{ 4, ST },		/* 00 0 0010: stw */ | 
 | 	INVALID,		/* 00 0 0011 */ | 
 | 	{ 2, LD },		/* 00 0 0100: lhz */ | 
 | 	{ 2, LD+SE },		/* 00 0 0101: lha */ | 
 | 	{ 2, ST },		/* 00 0 0110: sth */ | 
 | 	{ 4, LD+M },		/* 00 0 0111: lmw */ | 
 | 	{ 4, LD+F+S },		/* 00 0 1000: lfs */ | 
 | 	{ 8, LD+F },		/* 00 0 1001: lfd */ | 
 | 	{ 4, ST+F+S },		/* 00 0 1010: stfs */ | 
 | 	{ 8, ST+F },		/* 00 0 1011: stfd */ | 
 | 	INVALID,		/* 00 0 1100 */ | 
 | 	{ 8, LD },		/* 00 0 1101: ld/ldu/lwa */ | 
 | 	INVALID,		/* 00 0 1110 */ | 
 | 	{ 8, ST },		/* 00 0 1111: std/stdu */ | 
 | 	{ 4, LD+U },		/* 00 1 0000: lwzu */ | 
 | 	INVALID,		/* 00 1 0001 */ | 
 | 	{ 4, ST+U },		/* 00 1 0010: stwu */ | 
 | 	INVALID,		/* 00 1 0011 */ | 
 | 	{ 2, LD+U },		/* 00 1 0100: lhzu */ | 
 | 	{ 2, LD+SE+U },		/* 00 1 0101: lhau */ | 
 | 	{ 2, ST+U },		/* 00 1 0110: sthu */ | 
 | 	{ 4, ST+M },		/* 00 1 0111: stmw */ | 
 | 	{ 4, LD+F+S+U },	/* 00 1 1000: lfsu */ | 
 | 	{ 8, LD+F+U },		/* 00 1 1001: lfdu */ | 
 | 	{ 4, ST+F+S+U },	/* 00 1 1010: stfsu */ | 
 | 	{ 8, ST+F+U },		/* 00 1 1011: stfdu */ | 
 | 	INVALID,		/* 00 1 1100 */ | 
 | 	INVALID,		/* 00 1 1101 */ | 
 | 	INVALID,		/* 00 1 1110 */ | 
 | 	INVALID,		/* 00 1 1111 */ | 
 | 	{ 8, LD },		/* 01 0 0000: ldx */ | 
 | 	INVALID,		/* 01 0 0001 */ | 
 | 	{ 8, ST },		/* 01 0 0010: stdx */ | 
 | 	INVALID,		/* 01 0 0011 */ | 
 | 	INVALID,		/* 01 0 0100 */ | 
 | 	{ 4, LD+SE },		/* 01 0 0101: lwax */ | 
 | 	INVALID,		/* 01 0 0110 */ | 
 | 	INVALID,		/* 01 0 0111 */ | 
 | 	{ 4, LD+M+HARD+SX },	/* 01 0 1000: lswx */ | 
 | 	{ 4, LD+M+HARD },	/* 01 0 1001: lswi */ | 
 | 	{ 4, ST+M+HARD+SX },	/* 01 0 1010: stswx */ | 
 | 	{ 4, ST+M+HARD },	/* 01 0 1011: stswi */ | 
 | 	INVALID,		/* 01 0 1100 */ | 
 | 	{ 8, LD+U },		/* 01 0 1101: ldu */ | 
 | 	INVALID,		/* 01 0 1110 */ | 
 | 	{ 8, ST+U },		/* 01 0 1111: stdu */ | 
 | 	{ 8, LD+U },		/* 01 1 0000: ldux */ | 
 | 	INVALID,		/* 01 1 0001 */ | 
 | 	{ 8, ST+U },		/* 01 1 0010: stdux */ | 
 | 	INVALID,		/* 01 1 0011 */ | 
 | 	INVALID,		/* 01 1 0100 */ | 
 | 	{ 4, LD+SE+U },		/* 01 1 0101: lwaux */ | 
 | 	INVALID,		/* 01 1 0110 */ | 
 | 	INVALID,		/* 01 1 0111 */ | 
 | 	INVALID,		/* 01 1 1000 */ | 
 | 	INVALID,		/* 01 1 1001 */ | 
 | 	INVALID,		/* 01 1 1010 */ | 
 | 	INVALID,		/* 01 1 1011 */ | 
 | 	INVALID,		/* 01 1 1100 */ | 
 | 	INVALID,		/* 01 1 1101 */ | 
 | 	INVALID,		/* 01 1 1110 */ | 
 | 	INVALID,		/* 01 1 1111 */ | 
 | 	INVALID,		/* 10 0 0000 */ | 
 | 	INVALID,		/* 10 0 0001 */ | 
 | 	INVALID,		/* 10 0 0010: stwcx. */ | 
 | 	INVALID,		/* 10 0 0011 */ | 
 | 	INVALID,		/* 10 0 0100 */ | 
 | 	INVALID,		/* 10 0 0101 */ | 
 | 	INVALID,		/* 10 0 0110 */ | 
 | 	INVALID,		/* 10 0 0111 */ | 
 | 	{ 4, LD+SW },		/* 10 0 1000: lwbrx */ | 
 | 	INVALID,		/* 10 0 1001 */ | 
 | 	{ 4, ST+SW },		/* 10 0 1010: stwbrx */ | 
 | 	INVALID,		/* 10 0 1011 */ | 
 | 	{ 2, LD+SW },		/* 10 0 1100: lhbrx */ | 
 | 	{ 4, LD+SE },		/* 10 0 1101  lwa */ | 
 | 	{ 2, ST+SW },		/* 10 0 1110: sthbrx */ | 
 | 	INVALID,		/* 10 0 1111 */ | 
 | 	INVALID,		/* 10 1 0000 */ | 
 | 	INVALID,		/* 10 1 0001 */ | 
 | 	INVALID,		/* 10 1 0010 */ | 
 | 	INVALID,		/* 10 1 0011 */ | 
 | 	INVALID,		/* 10 1 0100 */ | 
 | 	INVALID,		/* 10 1 0101 */ | 
 | 	INVALID,		/* 10 1 0110 */ | 
 | 	INVALID,		/* 10 1 0111 */ | 
 | 	INVALID,		/* 10 1 1000 */ | 
 | 	INVALID,		/* 10 1 1001 */ | 
 | 	INVALID,		/* 10 1 1010 */ | 
 | 	INVALID,		/* 10 1 1011 */ | 
 | 	INVALID,		/* 10 1 1100 */ | 
 | 	INVALID,		/* 10 1 1101 */ | 
 | 	INVALID,		/* 10 1 1110 */ | 
 | 	{ 0, ST+HARD },		/* 10 1 1111: dcbz */ | 
 | 	{ 4, LD },		/* 11 0 0000: lwzx */ | 
 | 	INVALID,		/* 11 0 0001 */ | 
 | 	{ 4, ST },		/* 11 0 0010: stwx */ | 
 | 	INVALID,		/* 11 0 0011 */ | 
 | 	{ 2, LD },		/* 11 0 0100: lhzx */ | 
 | 	{ 2, LD+SE },		/* 11 0 0101: lhax */ | 
 | 	{ 2, ST },		/* 11 0 0110: sthx */ | 
 | 	INVALID,		/* 11 0 0111 */ | 
 | 	{ 4, LD+F+S },		/* 11 0 1000: lfsx */ | 
 | 	{ 8, LD+F },		/* 11 0 1001: lfdx */ | 
 | 	{ 4, ST+F+S },		/* 11 0 1010: stfsx */ | 
 | 	{ 8, ST+F },		/* 11 0 1011: stfdx */ | 
 | 	INVALID,		/* 11 0 1100 */ | 
 | 	{ 8, LD+M },		/* 11 0 1101: lmd */ | 
 | 	INVALID,		/* 11 0 1110 */ | 
 | 	{ 8, ST+M },		/* 11 0 1111: stmd */ | 
 | 	{ 4, LD+U },		/* 11 1 0000: lwzux */ | 
 | 	INVALID,		/* 11 1 0001 */ | 
 | 	{ 4, ST+U },		/* 11 1 0010: stwux */ | 
 | 	INVALID,		/* 11 1 0011 */ | 
 | 	{ 2, LD+U },		/* 11 1 0100: lhzux */ | 
 | 	{ 2, LD+SE+U },		/* 11 1 0101: lhaux */ | 
 | 	{ 2, ST+U },		/* 11 1 0110: sthux */ | 
 | 	INVALID,		/* 11 1 0111 */ | 
 | 	{ 4, LD+F+S+U },	/* 11 1 1000: lfsux */ | 
 | 	{ 8, LD+F+U },		/* 11 1 1001: lfdux */ | 
 | 	{ 4, ST+F+S+U },	/* 11 1 1010: stfsux */ | 
 | 	{ 8, ST+F+U },		/* 11 1 1011: stfdux */ | 
 | 	INVALID,		/* 11 1 1100 */ | 
 | 	INVALID,		/* 11 1 1101 */ | 
 | 	INVALID,		/* 11 1 1110 */ | 
 | 	INVALID,		/* 11 1 1111 */ | 
 | }; | 
 |  | 
 | /* | 
 |  * Create a DSISR value from the instruction | 
 |  */ | 
 | static inline unsigned make_dsisr(unsigned instr) | 
 | { | 
 | 	unsigned dsisr; | 
 |  | 
 |  | 
 | 	/* bits  6:15 --> 22:31 */ | 
 | 	dsisr = (instr & 0x03ff0000) >> 16; | 
 |  | 
 | 	if (IS_XFORM(instr)) { | 
 | 		/* bits 29:30 --> 15:16 */ | 
 | 		dsisr |= (instr & 0x00000006) << 14; | 
 | 		/* bit     25 -->    17 */ | 
 | 		dsisr |= (instr & 0x00000040) << 8; | 
 | 		/* bits 21:24 --> 18:21 */ | 
 | 		dsisr |= (instr & 0x00000780) << 3; | 
 | 	} else { | 
 | 		/* bit      5 -->    17 */ | 
 | 		dsisr |= (instr & 0x04000000) >> 12; | 
 | 		/* bits  1: 4 --> 18:21 */ | 
 | 		dsisr |= (instr & 0x78000000) >> 17; | 
 | 		/* bits 30:31 --> 12:13 */ | 
 | 		if (IS_DSFORM(instr)) | 
 | 			dsisr |= (instr & 0x00000003) << 18; | 
 | 	} | 
 |  | 
 | 	return dsisr; | 
 | } | 
 |  | 
 | /* | 
 |  * The dcbz (data cache block zero) instruction | 
 |  * gives an alignment fault if used on non-cacheable | 
 |  * memory.  We handle the fault mainly for the | 
 |  * case when we are running with the cache disabled | 
 |  * for debugging. | 
 |  */ | 
 | static int emulate_dcbz(struct pt_regs *regs, unsigned char __user *addr) | 
 | { | 
 | 	long __user *p; | 
 | 	int i, size; | 
 |  | 
 | #ifdef __powerpc64__ | 
 | 	size = ppc64_caches.dline_size; | 
 | #else | 
 | 	size = L1_CACHE_BYTES; | 
 | #endif | 
 | 	p = (long __user *) (regs->dar & -size); | 
 | 	if (user_mode(regs) && !access_ok(VERIFY_WRITE, p, size)) | 
 | 		return -EFAULT; | 
 | 	for (i = 0; i < size / sizeof(long); ++i) | 
 | 		if (__put_user(0, p+i)) | 
 | 			return -EFAULT; | 
 | 	return 1; | 
 | } | 
 |  | 
 | /* | 
 |  * Emulate load & store multiple instructions | 
 |  * On 64-bit machines, these instructions only affect/use the | 
 |  * bottom 4 bytes of each register, and the loads clear the | 
 |  * top 4 bytes of the affected register. | 
 |  */ | 
 | #ifdef CONFIG_PPC64 | 
 | #define REG_BYTE(rp, i)		*((u8 *)((rp) + ((i) >> 2)) + ((i) & 3) + 4) | 
 | #else | 
 | #define REG_BYTE(rp, i)		*((u8 *)(rp) + (i)) | 
 | #endif | 
 |  | 
 | #define SWIZ_PTR(p)		((unsigned char __user *)((p) ^ swiz)) | 
 |  | 
 | static int emulate_multiple(struct pt_regs *regs, unsigned char __user *addr, | 
 | 			    unsigned int reg, unsigned int nb, | 
 | 			    unsigned int flags, unsigned int instr, | 
 | 			    unsigned long swiz) | 
 | { | 
 | 	unsigned long *rptr; | 
 | 	unsigned int nb0, i, bswiz; | 
 | 	unsigned long p; | 
 |  | 
 | 	/* | 
 | 	 * We do not try to emulate 8 bytes multiple as they aren't really | 
 | 	 * available in our operating environments and we don't try to | 
 | 	 * emulate multiples operations in kernel land as they should never | 
 | 	 * be used/generated there at least not on unaligned boundaries | 
 | 	 */ | 
 | 	if (unlikely((nb > 4) || !user_mode(regs))) | 
 | 		return 0; | 
 |  | 
 | 	/* lmw, stmw, lswi/x, stswi/x */ | 
 | 	nb0 = 0; | 
 | 	if (flags & HARD) { | 
 | 		if (flags & SX) { | 
 | 			nb = regs->xer & 127; | 
 | 			if (nb == 0) | 
 | 				return 1; | 
 | 		} else { | 
 | 			unsigned long pc = regs->nip ^ (swiz & 4); | 
 |  | 
 | 			if (__get_user(instr, (unsigned int __user *)pc)) | 
 | 				return -EFAULT; | 
 | 			if (swiz == 0 && (flags & SW)) | 
 | 				instr = cpu_to_le32(instr); | 
 | 			nb = (instr >> 11) & 0x1f; | 
 | 			if (nb == 0) | 
 | 				nb = 32; | 
 | 		} | 
 | 		if (nb + reg * 4 > 128) { | 
 | 			nb0 = nb + reg * 4 - 128; | 
 | 			nb = 128 - reg * 4; | 
 | 		} | 
 | 	} else { | 
 | 		/* lwm, stmw */ | 
 | 		nb = (32 - reg) * 4; | 
 | 	} | 
 |  | 
 | 	if (!access_ok((flags & ST ? VERIFY_WRITE: VERIFY_READ), addr, nb+nb0)) | 
 | 		return -EFAULT;	/* bad address */ | 
 |  | 
 | 	rptr = ®s->gpr[reg]; | 
 | 	p = (unsigned long) addr; | 
 | 	bswiz = (flags & SW)? 3: 0; | 
 |  | 
 | 	if (!(flags & ST)) { | 
 | 		/* | 
 | 		 * This zeroes the top 4 bytes of the affected registers | 
 | 		 * in 64-bit mode, and also zeroes out any remaining | 
 | 		 * bytes of the last register for lsw*. | 
 | 		 */ | 
 | 		memset(rptr, 0, ((nb + 3) / 4) * sizeof(unsigned long)); | 
 | 		if (nb0 > 0) | 
 | 			memset(®s->gpr[0], 0, | 
 | 			       ((nb0 + 3) / 4) * sizeof(unsigned long)); | 
 |  | 
 | 		for (i = 0; i < nb; ++i, ++p) | 
 | 			if (__get_user(REG_BYTE(rptr, i ^ bswiz), SWIZ_PTR(p))) | 
 | 				return -EFAULT; | 
 | 		if (nb0 > 0) { | 
 | 			rptr = ®s->gpr[0]; | 
 | 			addr += nb; | 
 | 			for (i = 0; i < nb0; ++i, ++p) | 
 | 				if (__get_user(REG_BYTE(rptr, i ^ bswiz), | 
 | 					       SWIZ_PTR(p))) | 
 | 					return -EFAULT; | 
 | 		} | 
 |  | 
 | 	} else { | 
 | 		for (i = 0; i < nb; ++i, ++p) | 
 | 			if (__put_user(REG_BYTE(rptr, i ^ bswiz), SWIZ_PTR(p))) | 
 | 				return -EFAULT; | 
 | 		if (nb0 > 0) { | 
 | 			rptr = ®s->gpr[0]; | 
 | 			addr += nb; | 
 | 			for (i = 0; i < nb0; ++i, ++p) | 
 | 				if (__put_user(REG_BYTE(rptr, i ^ bswiz), | 
 | 					       SWIZ_PTR(p))) | 
 | 					return -EFAULT; | 
 | 		} | 
 | 	} | 
 | 	return 1; | 
 | } | 
 |  | 
 |  | 
 | /* | 
 |  * Called on alignment exception. Attempts to fixup | 
 |  * | 
 |  * Return 1 on success | 
 |  * Return 0 if unable to handle the interrupt | 
 |  * Return -EFAULT if data address is bad | 
 |  */ | 
 |  | 
 | int fix_alignment(struct pt_regs *regs) | 
 | { | 
 | 	unsigned int instr, nb, flags; | 
 | 	unsigned int reg, areg; | 
 | 	unsigned int dsisr; | 
 | 	unsigned char __user *addr; | 
 | 	unsigned long p, swiz; | 
 | 	int ret, t; | 
 | 	union { | 
 | 		u64 ll; | 
 | 		double dd; | 
 | 		unsigned char v[8]; | 
 | 		struct { | 
 | 			unsigned hi32; | 
 | 			int	 low32; | 
 | 		} x32; | 
 | 		struct { | 
 | 			unsigned char hi48[6]; | 
 | 			short	      low16; | 
 | 		} x16; | 
 | 	} data; | 
 |  | 
 | 	/* | 
 | 	 * We require a complete register set, if not, then our assembly | 
 | 	 * is broken | 
 | 	 */ | 
 | 	CHECK_FULL_REGS(regs); | 
 |  | 
 | 	dsisr = regs->dsisr; | 
 |  | 
 | 	/* Some processors don't provide us with a DSISR we can use here, | 
 | 	 * let's make one up from the instruction | 
 | 	 */ | 
 | 	if (cpu_has_feature(CPU_FTR_NODSISRALIGN)) { | 
 | 		unsigned long pc = regs->nip; | 
 |  | 
 | 		if (cpu_has_feature(CPU_FTR_PPC_LE) && (regs->msr & MSR_LE)) | 
 | 			pc ^= 4; | 
 | 		if (unlikely(__get_user(instr, (unsigned int __user *)pc))) | 
 | 			return -EFAULT; | 
 | 		if (cpu_has_feature(CPU_FTR_REAL_LE) && (regs->msr & MSR_LE)) | 
 | 			instr = cpu_to_le32(instr); | 
 | 		dsisr = make_dsisr(instr); | 
 | 	} | 
 |  | 
 | 	/* extract the operation and registers from the dsisr */ | 
 | 	reg = (dsisr >> 5) & 0x1f;	/* source/dest register */ | 
 | 	areg = dsisr & 0x1f;		/* register to update */ | 
 | 	instr = (dsisr >> 10) & 0x7f; | 
 | 	instr |= (dsisr >> 13) & 0x60; | 
 |  | 
 | 	/* Lookup the operation in our table */ | 
 | 	nb = aligninfo[instr].len; | 
 | 	flags = aligninfo[instr].flags; | 
 |  | 
 | 	/* Byteswap little endian loads and stores */ | 
 | 	swiz = 0; | 
 | 	if (regs->msr & MSR_LE) { | 
 | 		flags ^= SW; | 
 | 		/* | 
 | 		 * So-called "PowerPC little endian" mode works by | 
 | 		 * swizzling addresses rather than by actually doing | 
 | 		 * any byte-swapping.  To emulate this, we XOR each | 
 | 		 * byte address with 7.  We also byte-swap, because | 
 | 		 * the processor's address swizzling depends on the | 
 | 		 * operand size (it xors the address with 7 for bytes, | 
 | 		 * 6 for halfwords, 4 for words, 0 for doublewords) but | 
 | 		 * we will xor with 7 and load/store each byte separately. | 
 | 		 */ | 
 | 		if (cpu_has_feature(CPU_FTR_PPC_LE)) | 
 | 			swiz = 7; | 
 | 	} | 
 |  | 
 | 	/* DAR has the operand effective address */ | 
 | 	addr = (unsigned char __user *)regs->dar; | 
 |  | 
 | 	/* A size of 0 indicates an instruction we don't support, with | 
 | 	 * the exception of DCBZ which is handled as a special case here | 
 | 	 */ | 
 | 	if (instr == DCBZ) | 
 | 		return emulate_dcbz(regs, addr); | 
 | 	if (unlikely(nb == 0)) | 
 | 		return 0; | 
 |  | 
 | 	/* Load/Store Multiple instructions are handled in their own | 
 | 	 * function | 
 | 	 */ | 
 | 	if (flags & M) | 
 | 		return emulate_multiple(regs, addr, reg, nb, | 
 | 					flags, instr, swiz); | 
 |  | 
 | 	/* Verify the address of the operand */ | 
 | 	if (unlikely(user_mode(regs) && | 
 | 		     !access_ok((flags & ST ? VERIFY_WRITE : VERIFY_READ), | 
 | 				addr, nb))) | 
 | 		return -EFAULT; | 
 |  | 
 | 	/* Force the fprs into the save area so we can reference them */ | 
 | 	if (flags & F) { | 
 | 		/* userland only */ | 
 | 		if (unlikely(!user_mode(regs))) | 
 | 			return 0; | 
 | 		flush_fp_to_thread(current); | 
 | 	} | 
 |  | 
 | 	/* If we are loading, get the data from user space, else | 
 | 	 * get it from register values | 
 | 	 */ | 
 | 	if (!(flags & ST)) { | 
 | 		data.ll = 0; | 
 | 		ret = 0; | 
 | 		p = (unsigned long) addr; | 
 | 		switch (nb) { | 
 | 		case 8: | 
 | 			ret |= __get_user(data.v[0], SWIZ_PTR(p++)); | 
 | 			ret |= __get_user(data.v[1], SWIZ_PTR(p++)); | 
 | 			ret |= __get_user(data.v[2], SWIZ_PTR(p++)); | 
 | 			ret |= __get_user(data.v[3], SWIZ_PTR(p++)); | 
 | 		case 4: | 
 | 			ret |= __get_user(data.v[4], SWIZ_PTR(p++)); | 
 | 			ret |= __get_user(data.v[5], SWIZ_PTR(p++)); | 
 | 		case 2: | 
 | 			ret |= __get_user(data.v[6], SWIZ_PTR(p++)); | 
 | 			ret |= __get_user(data.v[7], SWIZ_PTR(p++)); | 
 | 			if (unlikely(ret)) | 
 | 				return -EFAULT; | 
 | 		} | 
 | 	} else if (flags & F) { | 
 | 		data.dd = current->thread.fpr[reg]; | 
 | 		if (flags & S) { | 
 | 			/* Single-precision FP store requires conversion... */ | 
 | #ifdef CONFIG_PPC_FPU | 
 | 			preempt_disable(); | 
 | 			enable_kernel_fp(); | 
 | 			cvt_df(&data.dd, (float *)&data.v[4], ¤t->thread); | 
 | 			preempt_enable(); | 
 | #else | 
 | 			return 0; | 
 | #endif | 
 | 		} | 
 | 	} else | 
 | 		data.ll = regs->gpr[reg]; | 
 |  | 
 | 	if (flags & SW) { | 
 | 		switch (nb) { | 
 | 		case 8: | 
 | 			SWAP(data.v[0], data.v[7]); | 
 | 			SWAP(data.v[1], data.v[6]); | 
 | 			SWAP(data.v[2], data.v[5]); | 
 | 			SWAP(data.v[3], data.v[4]); | 
 | 			break; | 
 | 		case 4: | 
 | 			SWAP(data.v[4], data.v[7]); | 
 | 			SWAP(data.v[5], data.v[6]); | 
 | 			break; | 
 | 		case 2: | 
 | 			SWAP(data.v[6], data.v[7]); | 
 | 			break; | 
 | 		} | 
 | 	} | 
 |  | 
 | 	/* Perform other misc operations like sign extension | 
 | 	 * or floating point single precision conversion | 
 | 	 */ | 
 | 	switch (flags & ~(U|SW)) { | 
 | 	case LD+SE:	/* sign extend */ | 
 | 		if ( nb == 2 ) | 
 | 			data.ll = data.x16.low16; | 
 | 		else	/* nb must be 4 */ | 
 | 			data.ll = data.x32.low32; | 
 | 		break; | 
 |  | 
 | 	/* Single-precision FP load requires conversion... */ | 
 | 	case LD+F+S: | 
 | #ifdef CONFIG_PPC_FPU | 
 | 		preempt_disable(); | 
 | 		enable_kernel_fp(); | 
 | 		cvt_fd((float *)&data.v[4], &data.dd, ¤t->thread); | 
 | 		preempt_enable(); | 
 | #else | 
 | 		return 0; | 
 | #endif | 
 | 		break; | 
 | 	} | 
 |  | 
 | 	/* Store result to memory or update registers */ | 
 | 	if (flags & ST) { | 
 | 		ret = 0; | 
 | 		p = (unsigned long) addr; | 
 | 		switch (nb) { | 
 | 		case 8: | 
 | 			ret |= __put_user(data.v[0], SWIZ_PTR(p++)); | 
 | 			ret |= __put_user(data.v[1], SWIZ_PTR(p++)); | 
 | 			ret |= __put_user(data.v[2], SWIZ_PTR(p++)); | 
 | 			ret |= __put_user(data.v[3], SWIZ_PTR(p++)); | 
 | 		case 4: | 
 | 			ret |= __put_user(data.v[4], SWIZ_PTR(p++)); | 
 | 			ret |= __put_user(data.v[5], SWIZ_PTR(p++)); | 
 | 		case 2: | 
 | 			ret |= __put_user(data.v[6], SWIZ_PTR(p++)); | 
 | 			ret |= __put_user(data.v[7], SWIZ_PTR(p++)); | 
 | 		} | 
 | 		if (unlikely(ret)) | 
 | 			return -EFAULT; | 
 | 	} else if (flags & F) | 
 | 		current->thread.fpr[reg] = data.dd; | 
 | 	else | 
 | 		regs->gpr[reg] = data.ll; | 
 |  | 
 | 	/* Update RA as needed */ | 
 | 	if (flags & U) | 
 | 		regs->gpr[areg] = regs->dar; | 
 |  | 
 | 	return 1; | 
 | } |