X-Git-Url: http://demsky.eecs.uci.edu/git/?a=blobdiff_plain;f=include%2Fimpatomic.h;h=2fde095ee67e6ce5498c4a8dd61d62c0779f2fec;hb=c451049894af068c8ba22011e3094e66a45d20c3;hp=472e6638a603a70a09db0d06b28a9b40b3fb1001;hpb=5a8ba5648b3b56f963bd6b165791631c4c5ff6be;p=model-checker.git diff --git a/include/impatomic.h b/include/impatomic.h index 472e663..2fde095 100644 --- a/include/impatomic.h +++ b/include/impatomic.h @@ -1,22 +1,12 @@ +#include "memoryorder.h" +#include "cmodelint.h" #ifdef __cplusplus -#include namespace std { -#else -#include -#include #endif - #define CPP0X( feature ) - -typedef enum memory_order { - memory_order_relaxed, memory_order_acquire, memory_order_release, - memory_order_acq_rel, memory_order_seq_cst -} memory_order; - - typedef struct atomic_flag { #ifdef __cplusplus @@ -82,43 +72,47 @@ inline void atomic_flag::fence( memory_order __x__ ) const volatile __x__=memory-ordering, and __y__=memory-ordering. */ -#define _ATOMIC_LOAD_( __a__, __x__ ) \ - ({ volatile __typeof__((__a__)->__f__)* __p__ = ((__a__)->__f__); \ - model->switch_to_master(new ModelAction(ATOMIC_READ, __x__, __p__)); \ - ((__typeof__((__a__)->__f__)) (thread_current()->get_return_value())); \ - }) - - -#define _ATOMIC_STORE_( __a__, __m__, __x__ ) \ - ({ volatile __typeof__((__a__)->__f__)* __p__ = ((__a__)->__f__); \ - __typeof__(__m__) __v__ = (__m__); \ - model->switch_to_master(new ModelAction(ATOMIC_WRITE, __x__, __p__, __v__)); \ - __v__; }) - -#define _ATOMIC_MODIFY_( __a__, __o__, __m__, __x__ ) \ - ({ volatile __typeof__((__a__)->__f__)* __p__ = ((__a__)->__f__); \ - model->switch_to_master(new ModelAction(ATOMIC_READ, __x__, __p__)); \ - __typeof__((__a__)->__f__) __old__=(__typeof__((__a__)->__f__)) thread_current()->get_return_value(); \ - __typeof__(__m__) __v__ = (__m__); \ - __typeof__((__a__)->__f__) __copy__= __old__; \ - __copy__ __o__ __v__; \ - model->switch_to_master(new ModelAction(ATOMIC_RMW, __x__, __p__, __copy__)); \ - __old__; }) - -#define _ATOMIC_CMPSWP_( __a__, __e__, __m__, __x__ ) \ - ({ volatile __typeof__((__a__)->__f__)* __p__ = ((__a__)->__f__); \ - __typeof__(__e__) __q__ = (__e__); \ - __typeof__(__m__) __v__ = (__m__); \ - bool __r__; \ - model->switch_to_master(new ModelAction(ATOMIC_READ, __x__, __p__)); \ - __typeof__((__a__)->__f__) __t__=(__typeof__((__a__)->__f__)) thread_current()->get_return_value(); \ - if (__t__ == * __q__ ) { \ - model->switch_to_master(new ModelAction(ATOMIC_RMW, __x__, __p__, __v__)); __r__ = true; } \ - else { *__q__ = __t__; __r__ = false;} \ - __r__; }) - -#define _ATOMIC_FENCE_( __a__, __x__ ) \ - ({ ASSERT(0);}) +#define _ATOMIC_LOAD_( __a__, __x__ ) \ + ({ volatile __typeof__((__a__)->__f__)* __p__ = & ((__a__)->__f__); \ + __typeof__((__a__)->__f__) __r__ = (__typeof__((__a__)->__f__))model_read_action((void *)__p__, __x__); \ + __r__; }) + +#define _ATOMIC_STORE_( __a__, __m__, __x__ ) \ + ({ volatile __typeof__((__a__)->__f__)* __p__ = & ((__a__)->__f__); \ + __typeof__(__m__) __v__ = (__m__); \ + model_write_action((void *) __p__, __x__, (uint64_t) __v__); \ + __v__; }) + + +#define _ATOMIC_INIT_( __a__, __m__ ) \ + ({ volatile __typeof__((__a__)->__f__)* __p__ = & ((__a__)->__f__); \ + __typeof__(__m__) __v__ = (__m__); \ + model_init_action((void *) __p__, (uint64_t) __v__); \ + __v__; }) + +#define _ATOMIC_MODIFY_( __a__, __o__, __m__, __x__ ) \ + ({ volatile __typeof__((__a__)->__f__)* __p__ = & ((__a__)->__f__); \ + __typeof__((__a__)->__f__) __old__=(__typeof__((__a__)->__f__)) model_rmwr_action((void *)__p__, __x__); \ + __typeof__(__m__) __v__ = (__m__); \ + __typeof__((__a__)->__f__) __copy__= __old__; \ + __copy__ __o__ __v__; \ + model_rmw_action((void *)__p__, __x__, (uint64_t) __copy__); \ + __old__; }) + +#define _ATOMIC_CMPSWP_( __a__, __e__, __m__, __x__ ) \ + ({ volatile __typeof__((__a__)->__f__)* __p__ = & ((__a__)->__f__); \ + __typeof__(__e__) __q__ = (__e__); \ + __typeof__(__m__) __v__ = (__m__); \ + bool __r__; \ + __typeof__((__a__)->__f__) __t__=(__typeof__((__a__)->__f__)) model_rmwr_action((void *)__p__, __x__); \ + if (__t__ == * __q__ ) { \ + model_rmw_action((void *)__p__, __x__, (uint64_t) __v__); __r__ = true; } \ + else { model_rmwc_action((void *)__p__, __x__); *__q__ = __t__; __r__ = false;} \ + __r__; }) + +#define _ATOMIC_FENCE_( __a__, __x__ ) \ + ({ model_fence_action(__x__);}) + #define ATOMIC_CHAR_LOCK_FREE 1 #define ATOMIC_CHAR16_T_LOCK_FREE 1 @@ -1547,15 +1541,16 @@ template<> struct atomic< wchar_t > : atomic_wchar_t #ifdef __cplusplus -inline bool atomic_is_lock_free( const volatile atomic_bool* __a__ ) +inline bool atomic_is_lock_free +( const volatile atomic_bool* __a__ ) { return false; } inline bool atomic_load_explicit ( volatile atomic_bool* __a__, memory_order __x__ ) { return _ATOMIC_LOAD_( __a__, __x__ ); } -inline bool atomic_load( volatile atomic_bool* __a__ ) -{ return atomic_load_explicit( __a__, memory_order_seq_cst ); } +inline bool atomic_load +( volatile atomic_bool* __a__ ) { return atomic_load_explicit( __a__, memory_order_seq_cst ); } inline void atomic_store_explicit ( volatile atomic_bool* __a__, bool __m__, memory_order __x__ ) @@ -1608,7 +1603,7 @@ inline void atomic_store inline void* atomic_swap_explicit ( volatile atomic_address* __a__, void* __m__, memory_order __x__ ) -{ return _ATOMIC_MODIFY_( __a__, =, __m__, __x__ ); } +{ return _ATOMIC_MODIFY_( __a__, =, __m__, __x__ ); } inline void* atomic_swap ( volatile atomic_address* __a__, void* __m__ ) @@ -2123,10 +2118,10 @@ inline void atomic_fence inline void* atomic_fetch_add_explicit ( volatile atomic_address* __a__, ptrdiff_t __m__, memory_order __x__ ) -{ void* volatile* __p__ = &((__a__)->__f__); - model->switch_to_master(new ModelAction(ATOMIC_READ, __x__, __p__)); - void* __r__ = (void *) thread_current()->get_return_value(); - model->switch_to_master(new ModelAction(ATOMIC_RMW, __x__, __p__, (void*)((char*)(*__p__) + __m__))); +{ + void* volatile* __p__ = &((__a__)->__f__); + void* __r__ = (void *) model_rmwr_action((void *)__p__, __x__); + model_rmw_action((void *)__p__, __x__, (uint64_t) ((char*)(*__p__) + __m__)); return __r__; } inline void* atomic_fetch_add @@ -2136,24 +2131,23 @@ inline void* atomic_fetch_add inline void* atomic_fetch_sub_explicit ( volatile atomic_address* __a__, ptrdiff_t __m__, memory_order __x__ ) -{ void* volatile* __p__ = &((__a__)->__f__); - model->switch_to_master(new ModelAction(ATOMIC_READ, __x__, __p__)); - void* __r__ = (void *) thread_current()->get_return_value(); - model->switch_to_master(new ModelAction(ATOMIC_RMW, __x__, __p__, (void*)((char*)(*__p__) - __m__))); +{ + void* volatile* __p__ = &((__a__)->__f__); + void* __r__ = (void *) model_rmwr_action((void *)__p__, __x__); + model_rmw_action((void *)__p__, __x__, (uint64_t)((char*)(*__p__) - __m__)); return __r__; } inline void* atomic_fetch_sub ( volatile atomic_address* __a__, ptrdiff_t __m__ ) { return atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } - inline char atomic_fetch_add_explicit ( volatile atomic_char* __a__, char __m__, memory_order __x__ ) { return _ATOMIC_MODIFY_( __a__, +=, __m__, __x__ ); } inline char atomic_fetch_add ( volatile atomic_char* __a__, char __m__ ) -{ atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } inline char atomic_fetch_sub_explicit @@ -2162,7 +2156,7 @@ inline char atomic_fetch_sub_explicit inline char atomic_fetch_sub ( volatile atomic_char* __a__, char __m__ ) -{ atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } inline char atomic_fetch_and_explicit @@ -2171,7 +2165,7 @@ inline char atomic_fetch_and_explicit inline char atomic_fetch_and ( volatile atomic_char* __a__, char __m__ ) -{ atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } inline char atomic_fetch_or_explicit @@ -2180,7 +2174,7 @@ inline char atomic_fetch_or_explicit inline char atomic_fetch_or ( volatile atomic_char* __a__, char __m__ ) -{ atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } inline char atomic_fetch_xor_explicit @@ -2189,7 +2183,7 @@ inline char atomic_fetch_xor_explicit inline char atomic_fetch_xor ( volatile atomic_char* __a__, char __m__ ) -{ atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } inline signed char atomic_fetch_add_explicit @@ -2198,7 +2192,7 @@ inline signed char atomic_fetch_add_explicit inline signed char atomic_fetch_add ( volatile atomic_schar* __a__, signed char __m__ ) -{ atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } inline signed char atomic_fetch_sub_explicit @@ -2207,7 +2201,7 @@ inline signed char atomic_fetch_sub_explicit inline signed char atomic_fetch_sub ( volatile atomic_schar* __a__, signed char __m__ ) -{ atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } inline signed char atomic_fetch_and_explicit @@ -2216,7 +2210,7 @@ inline signed char atomic_fetch_and_explicit inline signed char atomic_fetch_and ( volatile atomic_schar* __a__, signed char __m__ ) -{ atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } inline signed char atomic_fetch_or_explicit @@ -2225,7 +2219,7 @@ inline signed char atomic_fetch_or_explicit inline signed char atomic_fetch_or ( volatile atomic_schar* __a__, signed char __m__ ) -{ atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } inline signed char atomic_fetch_xor_explicit @@ -2234,7 +2228,7 @@ inline signed char atomic_fetch_xor_explicit inline signed char atomic_fetch_xor ( volatile atomic_schar* __a__, signed char __m__ ) -{ atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned char atomic_fetch_add_explicit @@ -2243,7 +2237,7 @@ inline unsigned char atomic_fetch_add_explicit inline unsigned char atomic_fetch_add ( volatile atomic_uchar* __a__, unsigned char __m__ ) -{ atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned char atomic_fetch_sub_explicit @@ -2252,7 +2246,7 @@ inline unsigned char atomic_fetch_sub_explicit inline unsigned char atomic_fetch_sub ( volatile atomic_uchar* __a__, unsigned char __m__ ) -{ atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned char atomic_fetch_and_explicit @@ -2261,7 +2255,7 @@ inline unsigned char atomic_fetch_and_explicit inline unsigned char atomic_fetch_and ( volatile atomic_uchar* __a__, unsigned char __m__ ) -{ atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned char atomic_fetch_or_explicit @@ -2270,7 +2264,7 @@ inline unsigned char atomic_fetch_or_explicit inline unsigned char atomic_fetch_or ( volatile atomic_uchar* __a__, unsigned char __m__ ) -{ atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned char atomic_fetch_xor_explicit @@ -2279,7 +2273,7 @@ inline unsigned char atomic_fetch_xor_explicit inline unsigned char atomic_fetch_xor ( volatile atomic_uchar* __a__, unsigned char __m__ ) -{ atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } inline short atomic_fetch_add_explicit @@ -2288,7 +2282,7 @@ inline short atomic_fetch_add_explicit inline short atomic_fetch_add ( volatile atomic_short* __a__, short __m__ ) -{ atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } inline short atomic_fetch_sub_explicit @@ -2297,7 +2291,7 @@ inline short atomic_fetch_sub_explicit inline short atomic_fetch_sub ( volatile atomic_short* __a__, short __m__ ) -{ atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } inline short atomic_fetch_and_explicit @@ -2306,7 +2300,7 @@ inline short atomic_fetch_and_explicit inline short atomic_fetch_and ( volatile atomic_short* __a__, short __m__ ) -{ atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } inline short atomic_fetch_or_explicit @@ -2315,7 +2309,7 @@ inline short atomic_fetch_or_explicit inline short atomic_fetch_or ( volatile atomic_short* __a__, short __m__ ) -{ atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } inline short atomic_fetch_xor_explicit @@ -2324,7 +2318,7 @@ inline short atomic_fetch_xor_explicit inline short atomic_fetch_xor ( volatile atomic_short* __a__, short __m__ ) -{ atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned short atomic_fetch_add_explicit @@ -2333,7 +2327,7 @@ inline unsigned short atomic_fetch_add_explicit inline unsigned short atomic_fetch_add ( volatile atomic_ushort* __a__, unsigned short __m__ ) -{ atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned short atomic_fetch_sub_explicit @@ -2342,7 +2336,7 @@ inline unsigned short atomic_fetch_sub_explicit inline unsigned short atomic_fetch_sub ( volatile atomic_ushort* __a__, unsigned short __m__ ) -{ atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned short atomic_fetch_and_explicit @@ -2351,7 +2345,7 @@ inline unsigned short atomic_fetch_and_explicit inline unsigned short atomic_fetch_and ( volatile atomic_ushort* __a__, unsigned short __m__ ) -{ atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned short atomic_fetch_or_explicit @@ -2360,7 +2354,7 @@ inline unsigned short atomic_fetch_or_explicit inline unsigned short atomic_fetch_or ( volatile atomic_ushort* __a__, unsigned short __m__ ) -{ atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned short atomic_fetch_xor_explicit @@ -2369,7 +2363,7 @@ inline unsigned short atomic_fetch_xor_explicit inline unsigned short atomic_fetch_xor ( volatile atomic_ushort* __a__, unsigned short __m__ ) -{ atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } inline int atomic_fetch_add_explicit @@ -2378,7 +2372,7 @@ inline int atomic_fetch_add_explicit inline int atomic_fetch_add ( volatile atomic_int* __a__, int __m__ ) -{ atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } inline int atomic_fetch_sub_explicit @@ -2387,7 +2381,7 @@ inline int atomic_fetch_sub_explicit inline int atomic_fetch_sub ( volatile atomic_int* __a__, int __m__ ) -{ atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } inline int atomic_fetch_and_explicit @@ -2396,7 +2390,7 @@ inline int atomic_fetch_and_explicit inline int atomic_fetch_and ( volatile atomic_int* __a__, int __m__ ) -{ atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } inline int atomic_fetch_or_explicit @@ -2405,7 +2399,7 @@ inline int atomic_fetch_or_explicit inline int atomic_fetch_or ( volatile atomic_int* __a__, int __m__ ) -{ atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } inline int atomic_fetch_xor_explicit @@ -2414,7 +2408,7 @@ inline int atomic_fetch_xor_explicit inline int atomic_fetch_xor ( volatile atomic_int* __a__, int __m__ ) -{ atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned int atomic_fetch_add_explicit @@ -2423,7 +2417,7 @@ inline unsigned int atomic_fetch_add_explicit inline unsigned int atomic_fetch_add ( volatile atomic_uint* __a__, unsigned int __m__ ) -{ atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned int atomic_fetch_sub_explicit @@ -2432,7 +2426,7 @@ inline unsigned int atomic_fetch_sub_explicit inline unsigned int atomic_fetch_sub ( volatile atomic_uint* __a__, unsigned int __m__ ) -{ atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned int atomic_fetch_and_explicit @@ -2441,7 +2435,7 @@ inline unsigned int atomic_fetch_and_explicit inline unsigned int atomic_fetch_and ( volatile atomic_uint* __a__, unsigned int __m__ ) -{ atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned int atomic_fetch_or_explicit @@ -2450,7 +2444,7 @@ inline unsigned int atomic_fetch_or_explicit inline unsigned int atomic_fetch_or ( volatile atomic_uint* __a__, unsigned int __m__ ) -{ atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned int atomic_fetch_xor_explicit @@ -2459,7 +2453,7 @@ inline unsigned int atomic_fetch_xor_explicit inline unsigned int atomic_fetch_xor ( volatile atomic_uint* __a__, unsigned int __m__ ) -{ atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } inline long atomic_fetch_add_explicit @@ -2468,7 +2462,7 @@ inline long atomic_fetch_add_explicit inline long atomic_fetch_add ( volatile atomic_long* __a__, long __m__ ) -{ atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } inline long atomic_fetch_sub_explicit @@ -2477,7 +2471,7 @@ inline long atomic_fetch_sub_explicit inline long atomic_fetch_sub ( volatile atomic_long* __a__, long __m__ ) -{ atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } inline long atomic_fetch_and_explicit @@ -2486,7 +2480,7 @@ inline long atomic_fetch_and_explicit inline long atomic_fetch_and ( volatile atomic_long* __a__, long __m__ ) -{ atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } inline long atomic_fetch_or_explicit @@ -2495,7 +2489,7 @@ inline long atomic_fetch_or_explicit inline long atomic_fetch_or ( volatile atomic_long* __a__, long __m__ ) -{ atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } inline long atomic_fetch_xor_explicit @@ -2504,7 +2498,7 @@ inline long atomic_fetch_xor_explicit inline long atomic_fetch_xor ( volatile atomic_long* __a__, long __m__ ) -{ atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned long atomic_fetch_add_explicit @@ -2513,7 +2507,7 @@ inline unsigned long atomic_fetch_add_explicit inline unsigned long atomic_fetch_add ( volatile atomic_ulong* __a__, unsigned long __m__ ) -{ atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned long atomic_fetch_sub_explicit @@ -2522,7 +2516,7 @@ inline unsigned long atomic_fetch_sub_explicit inline unsigned long atomic_fetch_sub ( volatile atomic_ulong* __a__, unsigned long __m__ ) -{ atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned long atomic_fetch_and_explicit @@ -2531,7 +2525,7 @@ inline unsigned long atomic_fetch_and_explicit inline unsigned long atomic_fetch_and ( volatile atomic_ulong* __a__, unsigned long __m__ ) -{ atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned long atomic_fetch_or_explicit @@ -2540,7 +2534,7 @@ inline unsigned long atomic_fetch_or_explicit inline unsigned long atomic_fetch_or ( volatile atomic_ulong* __a__, unsigned long __m__ ) -{ atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned long atomic_fetch_xor_explicit @@ -2549,7 +2543,7 @@ inline unsigned long atomic_fetch_xor_explicit inline unsigned long atomic_fetch_xor ( volatile atomic_ulong* __a__, unsigned long __m__ ) -{ atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } inline long long atomic_fetch_add_explicit @@ -2558,7 +2552,7 @@ inline long long atomic_fetch_add_explicit inline long long atomic_fetch_add ( volatile atomic_llong* __a__, long long __m__ ) -{ atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } inline long long atomic_fetch_sub_explicit @@ -2567,7 +2561,7 @@ inline long long atomic_fetch_sub_explicit inline long long atomic_fetch_sub ( volatile atomic_llong* __a__, long long __m__ ) -{ atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } inline long long atomic_fetch_and_explicit @@ -2576,7 +2570,7 @@ inline long long atomic_fetch_and_explicit inline long long atomic_fetch_and ( volatile atomic_llong* __a__, long long __m__ ) -{ atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } inline long long atomic_fetch_or_explicit @@ -2585,7 +2579,7 @@ inline long long atomic_fetch_or_explicit inline long long atomic_fetch_or ( volatile atomic_llong* __a__, long long __m__ ) -{ atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } inline long long atomic_fetch_xor_explicit @@ -2594,7 +2588,7 @@ inline long long atomic_fetch_xor_explicit inline long long atomic_fetch_xor ( volatile atomic_llong* __a__, long long __m__ ) -{ atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned long long atomic_fetch_add_explicit @@ -2603,7 +2597,7 @@ inline unsigned long long atomic_fetch_add_explicit inline unsigned long long atomic_fetch_add ( volatile atomic_ullong* __a__, unsigned long long __m__ ) -{ atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned long long atomic_fetch_sub_explicit @@ -2612,7 +2606,7 @@ inline unsigned long long atomic_fetch_sub_explicit inline unsigned long long atomic_fetch_sub ( volatile atomic_ullong* __a__, unsigned long long __m__ ) -{ atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned long long atomic_fetch_and_explicit @@ -2621,7 +2615,7 @@ inline unsigned long long atomic_fetch_and_explicit inline unsigned long long atomic_fetch_and ( volatile atomic_ullong* __a__, unsigned long long __m__ ) -{ atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned long long atomic_fetch_or_explicit @@ -2630,7 +2624,7 @@ inline unsigned long long atomic_fetch_or_explicit inline unsigned long long atomic_fetch_or ( volatile atomic_ullong* __a__, unsigned long long __m__ ) -{ atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } inline unsigned long long atomic_fetch_xor_explicit @@ -2639,7 +2633,7 @@ inline unsigned long long atomic_fetch_xor_explicit inline unsigned long long atomic_fetch_xor ( volatile atomic_ullong* __a__, unsigned long long __m__ ) -{ atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } inline wchar_t atomic_fetch_add_explicit @@ -2648,7 +2642,7 @@ inline wchar_t atomic_fetch_add_explicit inline wchar_t atomic_fetch_add ( volatile atomic_wchar_t* __a__, wchar_t __m__ ) -{ atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_add_explicit( __a__, __m__, memory_order_seq_cst ); } inline wchar_t atomic_fetch_sub_explicit @@ -2657,7 +2651,7 @@ inline wchar_t atomic_fetch_sub_explicit inline wchar_t atomic_fetch_sub ( volatile atomic_wchar_t* __a__, wchar_t __m__ ) -{ atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_sub_explicit( __a__, __m__, memory_order_seq_cst ); } inline wchar_t atomic_fetch_and_explicit @@ -2666,7 +2660,7 @@ inline wchar_t atomic_fetch_and_explicit inline wchar_t atomic_fetch_and ( volatile atomic_wchar_t* __a__, wchar_t __m__ ) -{ atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_and_explicit( __a__, __m__, memory_order_seq_cst ); } inline wchar_t atomic_fetch_or_explicit @@ -2675,7 +2669,7 @@ inline wchar_t atomic_fetch_or_explicit inline wchar_t atomic_fetch_or ( volatile atomic_wchar_t* __a__, wchar_t __m__ ) -{ atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_or_explicit( __a__, __m__, memory_order_seq_cst ); } inline wchar_t atomic_fetch_xor_explicit @@ -2684,7 +2678,7 @@ inline wchar_t atomic_fetch_xor_explicit inline wchar_t atomic_fetch_xor ( volatile atomic_wchar_t* __a__, wchar_t __m__ ) -{ atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } +{ return atomic_fetch_xor_explicit( __a__, __m__, memory_order_seq_cst ); } #else @@ -2699,6 +2693,9 @@ _ATOMIC_LOAD_( __a__, memory_order_seq_cst ) #define atomic_load_explicit( __a__, __x__ ) \ _ATOMIC_LOAD_( __a__, __x__ ) +#define atomic_init( __a__, __m__ ) \ +_ATOMIC_INIT_( __a__, __m__ ) + #define atomic_store( __a__, __m__ ) \ _ATOMIC_STORE_( __a__, __m__, memory_order_seq_cst )