|  | #ifndef _ASM_FUTEX_H | 
|  | #define _ASM_FUTEX_H | 
|  |  | 
|  | #include <linux/futex.h> | 
|  | #include <linux/uaccess.h> | 
|  | #include <asm/errno.h> | 
|  |  | 
|  | #define __futex_atomic_op1(insn, ret, oldval, uaddr, oparg) \ | 
|  | do {									\ | 
|  | register unsigned long r8 __asm ("r8") = 0;			\ | 
|  | __asm__ __volatile__(						\ | 
|  | "	mf;;					\n"	\ | 
|  | "[1:] "	insn ";;				\n"	\ | 
|  | "	.xdata4 \"__ex_table\", 1b-., 2f-.	\n"	\ | 
|  | "[2:]"							\ | 
|  | : "+r" (r8), "=r" (oldval)				\ | 
|  | : "r" (uaddr), "r" (oparg)				\ | 
|  | : "memory");						\ | 
|  | ret = r8;							\ | 
|  | } while (0) | 
|  |  | 
|  | #define __futex_atomic_op2(insn, ret, oldval, uaddr, oparg) \ | 
|  | do {									\ | 
|  | register unsigned long r8 __asm ("r8") = 0;			\ | 
|  | int val, newval;						\ | 
|  | do {								\ | 
|  | __asm__ __volatile__(					\ | 
|  | "	mf;;				  \n"	\ | 
|  | "[1:]	ld4 %3=[%4];;			  \n"	\ | 
|  | "	mov %2=%3			  \n"	\ | 
|  | insn	";;			  \n"	\ | 
|  | "	mov ar.ccv=%2;;			  \n"	\ | 
|  | "[2:]	cmpxchg4.acq %1=[%4],%3,ar.ccv;;  \n"	\ | 
|  | "	.xdata4 \"__ex_table\", 1b-., 3f-.\n"	\ | 
|  | "	.xdata4 \"__ex_table\", 2b-., 3f-.\n"	\ | 
|  | "[3:]"						\ | 
|  | : "+r" (r8), "=r" (val), "=&r" (oldval),	\ | 
|  | "=&r" (newval)				\ | 
|  | : "r" (uaddr), "r" (oparg)			\ | 
|  | : "memory");					\ | 
|  | if (unlikely (r8))					\ | 
|  | break;						\ | 
|  | } while (unlikely (val != oldval));				\ | 
|  | ret = r8;							\ | 
|  | } while (0) | 
|  |  | 
|  | static inline int | 
|  | futex_atomic_op_inuser (int encoded_op, u32 __user *uaddr) | 
|  | { | 
|  | int op = (encoded_op >> 28) & 7; | 
|  | int cmp = (encoded_op >> 24) & 15; | 
|  | int oparg = (encoded_op << 8) >> 20; | 
|  | int cmparg = (encoded_op << 20) >> 20; | 
|  | int oldval = 0, ret; | 
|  | if (encoded_op & (FUTEX_OP_OPARG_SHIFT << 28)) | 
|  | oparg = 1 << oparg; | 
|  |  | 
|  | if (! access_ok (VERIFY_WRITE, uaddr, sizeof(u32))) | 
|  | return -EFAULT; | 
|  |  | 
|  | pagefault_disable(); | 
|  |  | 
|  | switch (op) { | 
|  | case FUTEX_OP_SET: | 
|  | __futex_atomic_op1("xchg4 %1=[%2],%3", ret, oldval, uaddr, | 
|  | oparg); | 
|  | break; | 
|  | case FUTEX_OP_ADD: | 
|  | __futex_atomic_op2("add %3=%3,%5", ret, oldval, uaddr, oparg); | 
|  | break; | 
|  | case FUTEX_OP_OR: | 
|  | __futex_atomic_op2("or %3=%3,%5", ret, oldval, uaddr, oparg); | 
|  | break; | 
|  | case FUTEX_OP_ANDN: | 
|  | __futex_atomic_op2("and %3=%3,%5", ret, oldval, uaddr, | 
|  | ~oparg); | 
|  | break; | 
|  | case FUTEX_OP_XOR: | 
|  | __futex_atomic_op2("xor %3=%3,%5", ret, oldval, uaddr, oparg); | 
|  | break; | 
|  | default: | 
|  | ret = -ENOSYS; | 
|  | } | 
|  |  | 
|  | pagefault_enable(); | 
|  |  | 
|  | if (!ret) { | 
|  | switch (cmp) { | 
|  | case FUTEX_OP_CMP_EQ: ret = (oldval == cmparg); break; | 
|  | case FUTEX_OP_CMP_NE: ret = (oldval != cmparg); break; | 
|  | case FUTEX_OP_CMP_LT: ret = (oldval < cmparg); break; | 
|  | case FUTEX_OP_CMP_GE: ret = (oldval >= cmparg); break; | 
|  | case FUTEX_OP_CMP_LE: ret = (oldval <= cmparg); break; | 
|  | case FUTEX_OP_CMP_GT: ret = (oldval > cmparg); break; | 
|  | default: ret = -ENOSYS; | 
|  | } | 
|  | } | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static inline int | 
|  | futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, | 
|  | u32 oldval, u32 newval) | 
|  | { | 
|  | if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32))) | 
|  | return -EFAULT; | 
|  |  | 
|  | { | 
|  | register unsigned long r8 __asm ("r8"); | 
|  | unsigned long prev; | 
|  | __asm__ __volatile__( | 
|  | "	mf;;					\n" | 
|  | "	mov %0=r0				\n" | 
|  | "	mov ar.ccv=%4;;				\n" | 
|  | "[1:]	cmpxchg4.acq %1=[%2],%3,ar.ccv		\n" | 
|  | "	.xdata4 \"__ex_table\", 1b-., 2f-.	\n" | 
|  | "[2:]" | 
|  | : "=r" (r8), "=r" (prev) | 
|  | : "r" (uaddr), "r" (newval), | 
|  | "rO" ((long) (unsigned) oldval) | 
|  | : "memory"); | 
|  | *uval = prev; | 
|  | return r8; | 
|  | } | 
|  | } | 
|  |  | 
|  | #endif /* _ASM_FUTEX_H */ |