9 #include "snapshot-interface.h"
10 #include "threads-model.h"
13 memory_order orders[6] = {
14 memory_order_relaxed, memory_order_consume, memory_order_acquire,
15 memory_order_release, memory_order_acq_rel, memory_order_seq_cst,
18 /* --- helper functions --- */
19 uint64_t model_rmwrcas_action_helper(void *obj, int atomic_index, uint64_t oldval, int size, const char *position) {
20 createModelIfNotExist();
21 return model->switch_thread(new ModelAction(ATOMIC_RMWRCAS, position, orders[atomic_index], obj, oldval, size));
24 uint64_t model_rmwr_action_helper(void *obj, int atomic_index, const char *position) {
25 createModelIfNotExist();
26 return model->switch_thread(new ModelAction(ATOMIC_RMWR, position, orders[atomic_index], obj));
29 void model_rmw_action_helper(void *obj, uint64_t val, int atomic_index, const char * position) {
30 createModelIfNotExist();
31 model->switch_thread(new ModelAction(ATOMIC_RMW, position, orders[atomic_index], obj, val));
34 void model_rmwc_action_helper(void *obj, int atomic_index, const char *position) {
35 createModelIfNotExist();
36 model->switch_thread(new ModelAction(ATOMIC_RMWC, position, orders[atomic_index], obj));
40 #define VOLATILELOAD(size) \
41 uint ## size ## _t cds_volatile_load ## size(void * obj, const char * position) { \
42 createModelIfNotExist(); \
43 return (uint ## size ## _t)model->switch_thread(new ModelAction(ATOMIC_READ, position, memory_order_volatile_load, obj)); \
51 // cds volatile stores
52 #define VOLATILESTORE(size) \
53 void cds_volatile_store ## size (void * obj, uint ## size ## _t val, const char * position) { \
54 createModelIfNotExist(); \
55 model->switch_thread(new ModelAction(ATOMIC_WRITE, position, memory_order_volatile_store, obj, (uint64_t) val)); \
56 *((volatile uint ## size ## _t *)obj) = val; \
57 thread_id_t tid = thread_current_id(); \
58 for(int i=0;i < size / 8;i++) { \
59 atomraceCheckWrite(tid, (void *)(((char *)obj)+i)); \
69 #define CDSATOMICINT(size) \
70 void cds_atomic_init ## size (void * obj, uint ## size ## _t val, const char * position) { \
71 createModelIfNotExist(); \
72 model->switch_thread(new ModelAction(ATOMIC_INIT, position, memory_order_relaxed, obj, (uint64_t) val)); \
73 *((volatile uint ## size ## _t *)obj) = val; \
74 thread_id_t tid = thread_current_id(); \
75 for(int i=0;i < size / 8;i++) { \
76 atomraceCheckWrite(tid, (void *)(((char *)obj)+i)); \
86 #define CDSATOMICLOAD(size) \
87 uint ## size ## _t cds_atomic_load ## size(void * obj, int atomic_index, const char * position) { \
88 createModelIfNotExist(); \
89 uint ## size ## _t val = (uint ## size ## _t)model->switch_thread( \
90 new ModelAction(ATOMIC_READ, position, orders[atomic_index], obj)); \
91 thread_id_t tid = thread_current_id(); \
92 for(int i=0;i < size / 8;i++) { \
93 atomraceCheckRead(tid, (void *)(((char *)obj)+i)); \
104 #define CDSATOMICSTORE(size) \
105 void cds_atomic_store ## size(void * obj, uint ## size ## _t val, int atomic_index, const char * position) { \
106 createModelIfNotExist(); \
107 model->switch_thread(new ModelAction(ATOMIC_WRITE, position, orders[atomic_index], obj, (uint64_t) val)); \
108 *((volatile uint ## size ## _t *)obj) = val; \
109 thread_id_t tid = thread_current_id(); \
110 for(int i=0;i < size / 8;i++) { \
111 atomraceCheckWrite(tid, (void *)(((char *)obj)+i)); \
121 #define _ATOMIC_RMW_(__op__, size, addr, val, atomic_index, position) \
123 uint ## size ## _t _old = model_rmwr_action_helper(addr, atomic_index, position); \
124 uint ## size ## _t _copy = _old; \
125 uint ## size ## _t _val = val; \
127 model_rmw_action_helper(addr, (uint64_t) _copy, atomic_index, position); \
128 *((volatile uint ## size ## _t *)addr) = _copy; \
129 thread_id_t tid = thread_current_id(); \
130 for(int i=0;i < size / 8;i++) { \
131 atomraceCheckRead(tid, (void *)(((char *)addr)+i)); \
132 recordWrite(tid, (void *)(((char *)addr)+i)); \
137 // cds atomic exchange
138 #define CDSATOMICEXCHANGE(size) \
139 uint ## size ## _t cds_atomic_exchange ## size(void* addr, uint ## size ## _t val, int atomic_index, const char * position) { \
140 _ATOMIC_RMW_( =, size, addr, val, atomic_index, position); \
144 CDSATOMICEXCHANGE(16)
145 CDSATOMICEXCHANGE(32)
146 CDSATOMICEXCHANGE(64)
148 // cds atomic fetch add
149 #define CDSATOMICADD(size) \
150 uint ## size ## _t cds_atomic_fetch_add ## size(void* addr, uint ## size ## _t val, int atomic_index, const char * position) { \
151 _ATOMIC_RMW_( +=, size, addr, val, atomic_index, position); \
159 // cds atomic fetch sub
160 #define CDSATOMICSUB(size) \
161 uint ## size ## _t cds_atomic_fetch_sub ## size(void* addr, uint ## size ## _t val, int atomic_index, const char * position) { \
162 _ATOMIC_RMW_( -=, size, addr, val, atomic_index, position); \
170 // cds atomic fetch and
171 #define CDSATOMICAND(size) \
172 uint ## size ## _t cds_atomic_fetch_and ## size(void* addr, uint ## size ## _t val, int atomic_index, const char * position) { \
173 _ATOMIC_RMW_( &=, size, addr, val, atomic_index, position); \
181 // cds atomic fetch or
182 #define CDSATOMICOR(size) \
183 uint ## size ## _t cds_atomic_fetch_or ## size(void* addr, uint ## size ## _t val, int atomic_index, const char * position) { \
184 _ATOMIC_RMW_( |=, size, addr, val, atomic_index, position); \
192 // cds atomic fetch xor
193 #define CDSATOMICXOR(size) \
194 uint ## size ## _t cds_atomic_fetch_xor ## size(void* addr, uint ## size ## _t val, int atomic_index, const char * position) { \
195 _ATOMIC_RMW_( ^=, size, addr, val, atomic_index, position); \
203 // cds atomic compare and exchange
204 // In order to accomodate the LLVM PASS, the return values are not true or false.
206 #define _ATOMIC_CMPSWP_WEAK_ _ATOMIC_CMPSWP_
207 #define _ATOMIC_CMPSWP_(size, addr, expected, desired, atomic_index, position) \
209 uint ## size ## _t _desired = desired; \
210 uint ## size ## _t _expected = expected; \
211 uint ## size ## _t _old = model_rmwrcas_action_helper(addr, atomic_index, _expected, sizeof(_expected), position); \
212 if (_old == _expected) { \
213 model_rmw_action_helper(addr, (uint64_t) _desired, atomic_index, position); \
214 *((volatile uint ## size ## _t *)addr) = desired; \
215 thread_id_t tid = thread_current_id(); \
216 for(int i=0;i < size / 8;i++) { \
217 recordWrite(tid, (void *)(((char *)addr)+i)); \
219 return _expected; } \
221 model_rmwc_action_helper(addr, atomic_index, position); _expected = _old; return _old; } \
224 // atomic_compare_exchange version 1: the CmpOperand (corresponds to expected)
225 // extracted from LLVM IR is an integer type.
226 #define CDSATOMICCASV1(size) \
227 uint ## size ## _t cds_atomic_compare_exchange ## size ## _v1(void* addr, uint ## size ## _t expected, uint ## size ## _t desired, int atomic_index_succ, int atomic_index_fail, const char *position) { \
228 _ATOMIC_CMPSWP_(size, addr, expected, desired, atomic_index_succ, position); \
236 // atomic_compare_exchange version 2
237 #define CDSATOMICCASV2(size) \
238 bool cds_atomic_compare_exchange ## size ## _v2(void* addr, uint ## size ## _t* expected, uint ## size ## _t desired, int atomic_index_succ, int atomic_index_fail, const char *position) { \
239 uint ## size ## _t ret = cds_atomic_compare_exchange ## size ## _v1(addr, *expected, desired, atomic_index_succ, atomic_index_fail, position); \
240 if (ret == *expected) {return true;} else {return false;} \
248 // cds atomic thread fence
250 void cds_atomic_thread_fence(int atomic_index, const char * position) {
251 model->switch_thread(
252 new ModelAction(ATOMIC_FENCE, position, orders[atomic_index], FENCE_LOCATION)
257 #define _ATOMIC_CMPSWP_( __a__, __e__, __m__, __x__ ) \
258 ({ volatile __typeof__((__a__)->__f__)* __p__ = & ((__a__)->__f__); \
259 __typeof__(__e__) __q__ = (__e__); \
260 __typeof__(__m__) __v__ = (__m__); \
262 __typeof__((__a__)->__f__) __t__=(__typeof__((__a__)->__f__)) model_rmwr_action((void *)__p__, __x__); \
263 if (__t__ == * __q__ ) { \
264 model_rmw_action((void *)__p__, __x__, (uint64_t) __v__); __r__ = true; } \
265 else { model_rmwc_action((void *)__p__, __x__); *__q__ = __t__; __r__ = false;} \
268 #define _ATOMIC_FENCE_( __x__ ) \
269 ({ model_fence_action(__x__);})
274 #define _ATOMIC_MODIFY_( __a__, __o__, __m__, __x__ ) \
275 ({ volatile __typeof__((__a__)->__f__)* __p__ = & ((__a__)->__f__); \
276 __typeof__((__a__)->__f__) __old__=(__typeof__((__a__)->__f__)) model_rmwr_action((void *)__p__, __x__); \
277 __typeof__(__m__) __v__ = (__m__); \
278 __typeof__((__a__)->__f__) __copy__= __old__; \
279 __copy__ __o__ __v__; \
280 model_rmw_action((void *)__p__, __x__, (uint64_t) __copy__); \
281 __old__ = __old__; Silence clang (-Wunused-value) \
285 void cds_func_entry(const char * funcName) {
287 createModelIfNotExist();
288 thread_id_t tid = thread_current_id();
291 ModelHistory *history = model->get_history();
292 if ( !history->getFuncMap()->contains(funcName) ) {
293 // add func id to func map
294 func_id = history->get_func_counter();
295 history->incr_func_counter();
296 history->getFuncMap()->put(funcName, func_id);
298 // add func id to reverse func map
299 ModelVector<const char *> * func_map_rev = history->getFuncMapRev();
300 if ( func_map_rev->size() <= func_id )
301 func_map_rev->resize( func_id + 1 );
303 func_map_rev->at(func_id) = funcName;
305 func_id = history->getFuncMap()->get(funcName);
308 history->enter_function(func_id, tid);
312 void cds_func_exit(const char * funcName) {
314 createModelIfNotExist();
315 thread_id_t tid = thread_current_id();
318 ModelHistory *history = model->get_history();
319 func_id = history->getFuncMap()->get(funcName);
321 * func_id not found; this could happen in the case where a function calls cds_func_entry
322 * when the model has been defined yet, but then an atomic inside the function initializes
323 * the model. And then cds_func_exit is called upon the function exiting.
329 history->exit_function(func_id, tid);