| #ifndef __ASM_SH_ATOMIC_LLSC_H | 
 | #define __ASM_SH_ATOMIC_LLSC_H | 
 |  | 
 | /* | 
 |  * To get proper branch prediction for the main line, we must branch | 
 |  * forward to code at the end of this object's .text section, then | 
 |  * branch back to restart the operation. | 
 |  */ | 
 | static inline void atomic_add(int i, atomic_t *v) | 
 | { | 
 | 	unsigned long tmp; | 
 |  | 
 | 	__asm__ __volatile__ ( | 
 | "1:	movli.l @%2, %0		! atomic_add	\n" | 
 | "	add	%1, %0				\n" | 
 | "	movco.l	%0, @%2				\n" | 
 | "	bf	1b				\n" | 
 | 	: "=&z" (tmp) | 
 | 	: "r" (i), "r" (&v->counter) | 
 | 	: "t"); | 
 | } | 
 |  | 
 | static inline void atomic_sub(int i, atomic_t *v) | 
 | { | 
 | 	unsigned long tmp; | 
 |  | 
 | 	__asm__ __volatile__ ( | 
 | "1:	movli.l @%2, %0		! atomic_sub	\n" | 
 | "	sub	%1, %0				\n" | 
 | "	movco.l	%0, @%2				\n" | 
 | "	bf	1b				\n" | 
 | 	: "=&z" (tmp) | 
 | 	: "r" (i), "r" (&v->counter) | 
 | 	: "t"); | 
 | } | 
 |  | 
 | /* | 
 |  * SH-4A note: | 
 |  * | 
 |  * We basically get atomic_xxx_return() for free compared with | 
 |  * atomic_xxx(). movli.l/movco.l require r0 due to the instruction | 
 |  * encoding, so the retval is automatically set without having to | 
 |  * do any special work. | 
 |  */ | 
 | static inline int atomic_add_return(int i, atomic_t *v) | 
 | { | 
 | 	unsigned long temp; | 
 |  | 
 | 	__asm__ __volatile__ ( | 
 | "1:	movli.l @%2, %0		! atomic_add_return	\n" | 
 | "	add	%1, %0					\n" | 
 | "	movco.l	%0, @%2					\n" | 
 | "	bf	1b					\n" | 
 | "	synco						\n" | 
 | 	: "=&z" (temp) | 
 | 	: "r" (i), "r" (&v->counter) | 
 | 	: "t"); | 
 |  | 
 | 	return temp; | 
 | } | 
 |  | 
 | static inline int atomic_sub_return(int i, atomic_t *v) | 
 | { | 
 | 	unsigned long temp; | 
 |  | 
 | 	__asm__ __volatile__ ( | 
 | "1:	movli.l @%2, %0		! atomic_sub_return	\n" | 
 | "	sub	%1, %0					\n" | 
 | "	movco.l	%0, @%2					\n" | 
 | "	bf	1b					\n" | 
 | "	synco						\n" | 
 | 	: "=&z" (temp) | 
 | 	: "r" (i), "r" (&v->counter) | 
 | 	: "t"); | 
 |  | 
 | 	return temp; | 
 | } | 
 |  | 
 | static inline void atomic_clear_mask(unsigned int mask, atomic_t *v) | 
 | { | 
 | 	unsigned long tmp; | 
 |  | 
 | 	__asm__ __volatile__ ( | 
 | "1:	movli.l @%2, %0		! atomic_clear_mask	\n" | 
 | "	and	%1, %0					\n" | 
 | "	movco.l	%0, @%2					\n" | 
 | "	bf	1b					\n" | 
 | 	: "=&z" (tmp) | 
 | 	: "r" (~mask), "r" (&v->counter) | 
 | 	: "t"); | 
 | } | 
 |  | 
 | static inline void atomic_set_mask(unsigned int mask, atomic_t *v) | 
 | { | 
 | 	unsigned long tmp; | 
 |  | 
 | 	__asm__ __volatile__ ( | 
 | "1:	movli.l @%2, %0		! atomic_set_mask	\n" | 
 | "	or	%1, %0					\n" | 
 | "	movco.l	%0, @%2					\n" | 
 | "	bf	1b					\n" | 
 | 	: "=&z" (tmp) | 
 | 	: "r" (mask), "r" (&v->counter) | 
 | 	: "t"); | 
 | } | 
 |  | 
 | #endif /* __ASM_SH_ATOMIC_LLSC_H */ |