| /* | 
 |  * Copyright IBM Corp. 1999, 2009 | 
 |  * | 
 |  * Author(s): Martin Schwidefsky <schwidefsky@de.ibm.com> | 
 |  */ | 
 |  | 
 | #ifndef __ASM_BARRIER_H | 
 | #define __ASM_BARRIER_H | 
 |  | 
 | /* | 
 |  * Force strict CPU ordering. | 
 |  * And yes, this is required on UP too when we're talking | 
 |  * to devices. | 
 |  */ | 
 |  | 
 | #ifdef CONFIG_HAVE_MARCH_Z196_FEATURES | 
 | /* Fast-BCR without checkpoint synchronization */ | 
 | #define __ASM_BARRIER "bcr 14,0\n" | 
 | #else | 
 | #define __ASM_BARRIER "bcr 15,0\n" | 
 | #endif | 
 |  | 
 | #define mb() do {  asm volatile(__ASM_BARRIER : : : "memory"); } while (0) | 
 |  | 
 | #define rmb()				mb() | 
 | #define wmb()				mb() | 
 | #define dma_rmb()			rmb() | 
 | #define dma_wmb()			wmb() | 
 | #define smp_mb()			mb() | 
 | #define smp_rmb()			rmb() | 
 | #define smp_wmb()			wmb() | 
 |  | 
 | #define read_barrier_depends()		do { } while (0) | 
 | #define smp_read_barrier_depends()	do { } while (0) | 
 |  | 
 | #define smp_mb__before_atomic()		smp_mb() | 
 | #define smp_mb__after_atomic()		smp_mb() | 
 |  | 
 | #define smp_store_mb(var, value)		do { WRITE_ONCE(var, value); mb(); } while (0) | 
 |  | 
 | #define smp_store_release(p, v)						\ | 
 | do {									\ | 
 | 	compiletime_assert_atomic_type(*p);				\ | 
 | 	barrier();							\ | 
 | 	ACCESS_ONCE(*p) = (v);						\ | 
 | } while (0) | 
 |  | 
 | #define smp_load_acquire(p)						\ | 
 | ({									\ | 
 | 	typeof(*p) ___p1 = ACCESS_ONCE(*p);				\ | 
 | 	compiletime_assert_atomic_type(*p);				\ | 
 | 	barrier();							\ | 
 | 	___p1;								\ | 
 | }) | 
 |  | 
 | #endif /* __ASM_BARRIER_H */ |