| Paul Mundt | ec723fb | 2006-12-07 20:33:38 +0900 | [diff] [blame] | 1 | #ifndef __ASM_SH_ATOMIC_LLSC_H | 
|  | 2 | #define __ASM_SH_ATOMIC_LLSC_H | 
|  | 3 |  | 
|  | 4 | /* | 
|  | 5 | * To get proper branch prediction for the main line, we must branch | 
|  | 6 | * forward to code at the end of this object's .text section, then | 
|  | 7 | * branch back to restart the operation. | 
|  | 8 | */ | 
|  | 9 | static inline void atomic_add(int i, atomic_t *v) | 
|  | 10 | { | 
|  | 11 | unsigned long tmp; | 
|  | 12 |  | 
|  | 13 | __asm__ __volatile__ ( | 
|  | 14 | "1:	movli.l @%2, %0		! atomic_add	\n" | 
|  | 15 | "	add	%1, %0				\n" | 
|  | 16 | "	movco.l	%0, @%2				\n" | 
|  | 17 | "	bf	1b				\n" | 
|  | 18 | : "=&z" (tmp) | 
|  | 19 | : "r" (i), "r" (&v->counter) | 
|  | 20 | : "t"); | 
|  | 21 | } | 
|  | 22 |  | 
|  | 23 | static inline void atomic_sub(int i, atomic_t *v) | 
|  | 24 | { | 
|  | 25 | unsigned long tmp; | 
|  | 26 |  | 
|  | 27 | __asm__ __volatile__ ( | 
|  | 28 | "1:	movli.l @%2, %0		! atomic_sub	\n" | 
|  | 29 | "	sub	%1, %0				\n" | 
|  | 30 | "	movco.l	%0, @%2				\n" | 
|  | 31 | "	bf	1b				\n" | 
|  | 32 | : "=&z" (tmp) | 
|  | 33 | : "r" (i), "r" (&v->counter) | 
|  | 34 | : "t"); | 
|  | 35 | } | 
|  | 36 |  | 
|  | 37 | /* | 
|  | 38 | * SH-4A note: | 
|  | 39 | * | 
|  | 40 | * We basically get atomic_xxx_return() for free compared with | 
|  | 41 | * atomic_xxx(). movli.l/movco.l require r0 due to the instruction | 
|  | 42 | * encoding, so the retval is automatically set without having to | 
|  | 43 | * do any special work. | 
|  | 44 | */ | 
|  | 45 | static inline int atomic_add_return(int i, atomic_t *v) | 
|  | 46 | { | 
|  | 47 | unsigned long temp; | 
|  | 48 |  | 
|  | 49 | __asm__ __volatile__ ( | 
|  | 50 | "1:	movli.l @%2, %0		! atomic_add_return	\n" | 
|  | 51 | "	add	%1, %0					\n" | 
|  | 52 | "	movco.l	%0, @%2					\n" | 
|  | 53 | "	bf	1b					\n" | 
|  | 54 | "	synco						\n" | 
|  | 55 | : "=&z" (temp) | 
|  | 56 | : "r" (i), "r" (&v->counter) | 
|  | 57 | : "t"); | 
|  | 58 |  | 
|  | 59 | return temp; | 
|  | 60 | } | 
|  | 61 |  | 
|  | 62 | static inline int atomic_sub_return(int i, atomic_t *v) | 
|  | 63 | { | 
|  | 64 | unsigned long temp; | 
|  | 65 |  | 
|  | 66 | __asm__ __volatile__ ( | 
|  | 67 | "1:	movli.l @%2, %0		! atomic_sub_return	\n" | 
|  | 68 | "	sub	%1, %0					\n" | 
|  | 69 | "	movco.l	%0, @%2					\n" | 
|  | 70 | "	bf	1b					\n" | 
|  | 71 | "	synco						\n" | 
|  | 72 | : "=&z" (temp) | 
|  | 73 | : "r" (i), "r" (&v->counter) | 
|  | 74 | : "t"); | 
|  | 75 |  | 
|  | 76 | return temp; | 
|  | 77 | } | 
|  | 78 |  | 
|  | 79 | static inline void atomic_clear_mask(unsigned int mask, atomic_t *v) | 
|  | 80 | { | 
|  | 81 | unsigned long tmp; | 
|  | 82 |  | 
|  | 83 | __asm__ __volatile__ ( | 
|  | 84 | "1:	movli.l @%2, %0		! atomic_clear_mask	\n" | 
|  | 85 | "	and	%1, %0					\n" | 
|  | 86 | "	movco.l	%0, @%2					\n" | 
|  | 87 | "	bf	1b					\n" | 
|  | 88 | : "=&z" (tmp) | 
|  | 89 | : "r" (~mask), "r" (&v->counter) | 
|  | 90 | : "t"); | 
|  | 91 | } | 
|  | 92 |  | 
|  | 93 | static inline void atomic_set_mask(unsigned int mask, atomic_t *v) | 
|  | 94 | { | 
|  | 95 | unsigned long tmp; | 
|  | 96 |  | 
|  | 97 | __asm__ __volatile__ ( | 
|  | 98 | "1:	movli.l @%2, %0		! atomic_set_mask	\n" | 
|  | 99 | "	or	%1, %0					\n" | 
|  | 100 | "	movco.l	%0, @%2					\n" | 
|  | 101 | "	bf	1b					\n" | 
|  | 102 | : "=&z" (tmp) | 
|  | 103 | : "r" (mask), "r" (&v->counter) | 
|  | 104 | : "t"); | 
|  | 105 | } | 
|  | 106 |  | 
| Paul Mundt | ec723fb | 2006-12-07 20:33:38 +0900 | [diff] [blame] | 107 | #endif /* __ASM_SH_ATOMIC_LLSC_H */ |