… | |
… | |
79 | #else |
79 | #else |
80 | #define ECB_GCC_VERSION(major,minor) (__GNUC__ > (major) || (__GNUC__ == (major) && __GNUC_MINOR__ >= (minor))) |
80 | #define ECB_GCC_VERSION(major,minor) (__GNUC__ > (major) || (__GNUC__ == (major) && __GNUC_MINOR__ >= (minor))) |
81 | #endif |
81 | #endif |
82 | #endif |
82 | #endif |
83 | |
83 | |
|
|
84 | #define ECB_C (__STDC__+0) /* this assumes that __STDC__ is either empty or a number */ |
|
|
85 | #define ECB_C99 (__STDC_VERSION__ >= 199901L) |
|
|
86 | #define ECB_C11 (__STDC_VERSION__ >= 201112L) |
|
|
87 | #define ECB_CPP (__cplusplus+0) |
|
|
88 | #define ECB_CPP11 (__cplusplus >= 201103L) |
|
|
89 | |
84 | /*****************************************************************************/ |
90 | /*****************************************************************************/ |
85 | |
91 | |
86 | /* ECB_NO_THREADS - ecb is not used by multiple threads, ever */ |
92 | /* ECB_NO_THREADS - ecb is not used by multiple threads, ever */ |
87 | /* ECB_NO_SMP - ecb might be used in multiple threads, but only on a single cpu */ |
93 | /* ECB_NO_SMP - ecb might be used in multiple threads, but only on a single cpu */ |
88 | |
94 | |
89 | #if ECB_NO_THREADS |
95 | #if ECB_NO_THREADS |
90 | # define ECB_NO_SMP 1 |
96 | #define ECB_NO_SMP 1 |
91 | #endif |
97 | #endif |
92 | |
98 | |
93 | #if ECB_NO_THREADS || ECB_NO_SMP |
99 | #if ECB_NO_SMP |
94 | #define ECB_MEMORY_FENCE do { } while (0) |
100 | #define ECB_MEMORY_FENCE do { } while (0) |
95 | #endif |
101 | #endif |
96 | |
102 | |
97 | #ifndef ECB_MEMORY_FENCE |
103 | #ifndef ECB_MEMORY_FENCE |
98 | #if ECB_GCC_VERSION(2,5) || defined __INTEL_COMPILER || (__llvm__ && __GNUC__) || __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
104 | #if ECB_GCC_VERSION(2,5) || defined __INTEL_COMPILER || (__llvm__ && __GNUC__) || __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
99 | #if __i386 || __i386__ |
105 | #if __i386 || __i386__ |
100 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
106 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
101 | #define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE /* non-lock xchg might be enough */ |
107 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory") |
102 | #define ECB_MEMORY_FENCE_RELEASE do { } while (0) /* unlikely to change in future cpus */ |
108 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
103 | #elif __amd64 || __amd64__ || __x86_64 || __x86_64__ |
109 | #elif __amd64 || __amd64__ || __x86_64 || __x86_64__ |
104 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
110 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
105 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("lfence" : : : "memory") |
111 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory") |
106 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("sfence") /* play safe - not needed in any current cpu */ |
112 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
107 | #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ |
113 | #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ |
108 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
114 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
109 | #elif defined __ARM_ARCH_6__ || defined __ARM_ARCH_6J__ \ |
115 | #elif defined __ARM_ARCH_6__ || defined __ARM_ARCH_6J__ \ |
110 | || defined __ARM_ARCH_6K__ || defined __ARM_ARCH_6ZK__ |
116 | || defined __ARM_ARCH_6K__ || defined __ARM_ARCH_6ZK__ |
111 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,5" : : "r" (0) : "memory") |
117 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,5" : : "r" (0) : "memory") |
112 | #elif defined __ARM_ARCH_7__ || defined __ARM_ARCH_7A__ \ |
118 | #elif defined __ARM_ARCH_7__ || defined __ARM_ARCH_7A__ \ |
113 | || defined __ARM_ARCH_7M__ || defined __ARM_ARCH_7R__ |
119 | || defined __ARM_ARCH_7M__ || defined __ARM_ARCH_7R__ |
114 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") |
120 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") |
115 | #elif __sparc || __sparc__ |
121 | #elif __sparc || __sparc__ |
116 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad | " : : : "memory") |
122 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad" : : : "memory") |
117 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory") |
123 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory") |
118 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore") |
124 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore") |
119 | #elif defined __s390__ || defined __s390x__ |
125 | #elif defined __s390__ || defined __s390x__ |
120 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory") |
126 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory") |
121 | #elif defined __mips__ |
127 | #elif defined __mips__ |
122 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
128 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
123 | #elif defined __alpha__ |
129 | #elif defined __alpha__ |
124 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mb" : : : "memory") |
130 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mb" : : : "memory") |
|
|
131 | #elif defined __hppa__ |
|
|
132 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("" : : : "memory") |
|
|
133 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
|
|
134 | #elif defined __ia64__ |
|
|
135 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mf" : : : "memory") |
125 | #endif |
136 | #endif |
126 | #endif |
137 | #endif |
127 | #endif |
138 | #endif |
128 | |
139 | |
129 | #ifndef ECB_MEMORY_FENCE |
140 | #ifndef ECB_MEMORY_FENCE |
|
|
141 | #if ECB_GCC_VERSION(4,7) |
|
|
142 | /* see comment below (stdatomic.h) about the C11 memory model. */ |
|
|
143 | #define ECB_MEMORY_FENCE __atomic_thread_fence (__ATOMIC_SEQ_CST) |
|
|
144 | #elif defined __clang && __has_feature (cxx_atomic) |
|
|
145 | /* see comment below (stdatomic.h) about the C11 memory model. */ |
|
|
146 | #define ECB_MEMORY_FENCE __c11_atomic_thread_fence (__ATOMIC_SEQ_CST) |
130 | #if ECB_GCC_VERSION(4,4) || defined __INTEL_COMPILER || defined __clang__ |
147 | #elif ECB_GCC_VERSION(4,4) || defined __INTEL_COMPILER || defined __clang__ |
131 | #define ECB_MEMORY_FENCE __sync_synchronize () |
148 | #define ECB_MEMORY_FENCE __sync_synchronize () |
132 | /*#define ECB_MEMORY_FENCE_ACQUIRE ({ char dummy = 0; __sync_lock_test_and_set (&dummy, 1); }) */ |
|
|
133 | /*#define ECB_MEMORY_FENCE_RELEASE ({ char dummy = 1; __sync_lock_release (&dummy ); }) */ |
|
|
134 | #elif _MSC_VER >= 1400 /* VC++ 2005 */ |
149 | #elif _MSC_VER >= 1400 /* VC++ 2005 */ |
135 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
150 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
136 | #define ECB_MEMORY_FENCE _ReadWriteBarrier () |
151 | #define ECB_MEMORY_FENCE _ReadWriteBarrier () |
137 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ |
152 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ |
138 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () |
153 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () |
… | |
… | |
148 | #define ECB_MEMORY_FENCE __sync () |
163 | #define ECB_MEMORY_FENCE __sync () |
149 | #endif |
164 | #endif |
150 | #endif |
165 | #endif |
151 | |
166 | |
152 | #ifndef ECB_MEMORY_FENCE |
167 | #ifndef ECB_MEMORY_FENCE |
|
|
168 | #if ECB_C11 && !defined __STDC_NO_ATOMICS__ |
|
|
169 | /* we assume that these memory fences work on all variables/all memory accesses, */ |
|
|
170 | /* not just C11 atomics and atomic accesses */ |
|
|
171 | #include <stdatomic.h> |
|
|
172 | /* Unfortunately, neither gcc 4.7 nor clang 3.1 generate any instructions for */ |
|
|
173 | /* any fence other than seq_cst, which isn't very efficient for us. */ |
|
|
174 | /* Why that is, we don't know - either the C11 memory model is quite useless */ |
|
|
175 | /* for most usages, or gcc and clang have a bug */ |
|
|
176 | /* I *currently* lean towards the latter, and inefficiently implement */ |
|
|
177 | /* all three of ecb's fences as a seq_cst fence */ |
|
|
178 | #define ECB_MEMORY_FENCE atomic_thread_fence (memory_order_seq_cst) |
|
|
179 | #endif |
|
|
180 | #endif |
|
|
181 | |
|
|
182 | #ifndef ECB_MEMORY_FENCE |
153 | #if !ECB_AVOID_PTHREADS |
183 | #if !ECB_AVOID_PTHREADS |
154 | /* |
184 | /* |
155 | * if you get undefined symbol references to pthread_mutex_lock, |
185 | * if you get undefined symbol references to pthread_mutex_lock, |
156 | * or failure to find pthread.h, then you should implement |
186 | * or failure to find pthread.h, then you should implement |
157 | * the ECB_MEMORY_FENCE operations for your cpu/compiler |
187 | * the ECB_MEMORY_FENCE operations for your cpu/compiler |
… | |
… | |
175 | #define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE |
205 | #define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE |
176 | #endif |
206 | #endif |
177 | |
207 | |
178 | /*****************************************************************************/ |
208 | /*****************************************************************************/ |
179 | |
209 | |
180 | #define ECB_C99 (__STDC_VERSION__ >= 199901L) |
|
|
181 | |
|
|
182 | #if __cplusplus |
210 | #if __cplusplus |
183 | #define ecb_inline static inline |
211 | #define ecb_inline static inline |
184 | #elif ECB_GCC_VERSION(2,5) |
212 | #elif ECB_GCC_VERSION(2,5) |
185 | #define ecb_inline static __inline__ |
213 | #define ecb_inline static __inline__ |
186 | #elif ECB_C99 |
214 | #elif ECB_C99 |
… | |
… | |
224 | #elif ECB_GCC_VERSION(3,0) |
252 | #elif ECB_GCC_VERSION(3,0) |
225 | #define ecb_decltype(x) __typeof(x) |
253 | #define ecb_decltype(x) __typeof(x) |
226 | #endif |
254 | #endif |
227 | |
255 | |
228 | #define ecb_noinline ecb_attribute ((__noinline__)) |
256 | #define ecb_noinline ecb_attribute ((__noinline__)) |
229 | #define ecb_noreturn ecb_attribute ((__noreturn__)) |
|
|
230 | #define ecb_unused ecb_attribute ((__unused__)) |
257 | #define ecb_unused ecb_attribute ((__unused__)) |
231 | #define ecb_const ecb_attribute ((__const__)) |
258 | #define ecb_const ecb_attribute ((__const__)) |
232 | #define ecb_pure ecb_attribute ((__pure__)) |
259 | #define ecb_pure ecb_attribute ((__pure__)) |
|
|
260 | |
|
|
261 | #if ECB_C11 |
|
|
262 | #define ecb_noreturn _Noreturn |
|
|
263 | #else |
|
|
264 | #define ecb_noreturn ecb_attribute ((__noreturn__)) |
|
|
265 | #endif |
233 | |
266 | |
234 | #if ECB_GCC_VERSION(4,3) |
267 | #if ECB_GCC_VERSION(4,3) |
235 | #define ecb_artificial ecb_attribute ((__artificial__)) |
268 | #define ecb_artificial ecb_attribute ((__artificial__)) |
236 | #define ecb_hot ecb_attribute ((__hot__)) |
269 | #define ecb_hot ecb_attribute ((__hot__)) |
237 | #define ecb_cold ecb_attribute ((__cold__)) |
270 | #define ecb_cold ecb_attribute ((__cold__)) |