… | |
… | |
74 | #else |
74 | #else |
75 | #define ECB_PTRSIZE 4 |
75 | #define ECB_PTRSIZE 4 |
76 | #endif |
76 | #endif |
77 | #endif |
77 | #endif |
78 | |
78 | |
|
|
79 | #define ECB_GCC_AMD64 (__amd64 || __amd64__ || __x86_64 || __x86_64__) |
|
|
80 | #define ECB_MSVC_AMD64 (_M_AMD64 || _M_X64) |
|
|
81 | |
79 | /* work around x32 idiocy by defining proper macros */ |
82 | /* work around x32 idiocy by defining proper macros */ |
80 | #if __amd64 || __x86_64 || _M_AMD64 || _M_X64 |
83 | #if ECB_GCC_AMD64 || ECB_MSVC_AMD64 |
81 | #if _ILP32 |
84 | #if _ILP32 |
82 | #define ECB_AMD64_X32 1 |
85 | #define ECB_AMD64_X32 1 |
83 | #else |
86 | #else |
84 | #define ECB_AMD64 1 |
87 | #define ECB_AMD64 1 |
85 | #endif |
88 | #endif |
… | |
… | |
147 | |
150 | |
148 | #if ECB_NO_SMP |
151 | #if ECB_NO_SMP |
149 | #define ECB_MEMORY_FENCE do { } while (0) |
152 | #define ECB_MEMORY_FENCE do { } while (0) |
150 | #endif |
153 | #endif |
151 | |
154 | |
|
|
155 | /* http://www-01.ibm.com/support/knowledgecenter/SSGH3R_13.1.0/com.ibm.xlcpp131.aix.doc/compiler_ref/compiler_builtins.html */ |
|
|
156 | #if __xlC__ && ECB_CPP |
|
|
157 | #include <builtins.h> |
|
|
158 | #endif |
|
|
159 | |
152 | #ifndef ECB_MEMORY_FENCE |
160 | #ifndef ECB_MEMORY_FENCE |
153 | #if ECB_GCC_VERSION(2,5) || defined __INTEL_COMPILER || (__llvm__ && __GNUC__) || __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
161 | #if ECB_GCC_VERSION(2,5) || defined __INTEL_COMPILER || (__llvm__ && __GNUC__) || __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
154 | #if __i386 || __i386__ |
162 | #if __i386 || __i386__ |
155 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
163 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
156 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory") |
164 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory") |
157 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
165 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
158 | #elif __amd64 || __amd64__ || __x86_64 || __x86_64__ |
166 | #elif ECB_GCC_AMD64 |
159 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
167 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
160 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory") |
168 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory") |
161 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
169 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
162 | #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ |
170 | #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ |
163 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
171 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
… | |
… | |
167 | #elif defined __ARM_ARCH_7__ || defined __ARM_ARCH_7A__ \ |
175 | #elif defined __ARM_ARCH_7__ || defined __ARM_ARCH_7A__ \ |
168 | || defined __ARM_ARCH_7M__ || defined __ARM_ARCH_7R__ |
176 | || defined __ARM_ARCH_7M__ || defined __ARM_ARCH_7R__ |
169 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") |
177 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") |
170 | #elif __aarch64__ |
178 | #elif __aarch64__ |
171 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb ish" : : : "memory") |
179 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb ish" : : : "memory") |
172 | #elif (__sparc || __sparc__) && !__sparcv8 |
180 | #elif (__sparc || __sparc__) && !(__sparc_v8__ || defined __sparcv8) |
173 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad" : : : "memory") |
181 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad" : : : "memory") |
174 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory") |
182 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory") |
175 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore") |
183 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore") |
176 | #elif defined __s390__ || defined __s390x__ |
184 | #elif defined __s390__ || defined __s390x__ |
177 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory") |
185 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory") |
… | |
… | |
349 | #define ecb_deprecated __declspec (deprecated) |
357 | #define ecb_deprecated __declspec (deprecated) |
350 | #else |
358 | #else |
351 | #define ecb_deprecated ecb_attribute ((__deprecated__)) |
359 | #define ecb_deprecated ecb_attribute ((__deprecated__)) |
352 | #endif |
360 | #endif |
353 | |
361 | |
354 | #if __MSC_VER >= 1500 |
362 | #if _MSC_VER >= 1500 |
355 | #define ecb_deprecated_message(msg) __declspec (deprecated (msg)) |
363 | #define ecb_deprecated_message(msg) __declspec (deprecated (msg)) |
356 | #elif ECB_GCC_VERSION(4,5) |
364 | #elif ECB_GCC_VERSION(4,5) |
357 | #define ecb_deprecated_message(msg) ecb_attribute ((__deprecated__ (msg)) |
365 | #define ecb_deprecated_message(msg) ecb_attribute ((__deprecated__ (msg)) |
358 | #else |
366 | #else |
359 | #define ecb_deprecated_message(msg) ecb_deprecated |
367 | #define ecb_deprecated_message(msg) ecb_deprecated |
… | |
… | |
367 | |
375 | |
368 | #define ecb_unused ecb_attribute ((__unused__)) |
376 | #define ecb_unused ecb_attribute ((__unused__)) |
369 | #define ecb_const ecb_attribute ((__const__)) |
377 | #define ecb_const ecb_attribute ((__const__)) |
370 | #define ecb_pure ecb_attribute ((__pure__)) |
378 | #define ecb_pure ecb_attribute ((__pure__)) |
371 | |
379 | |
372 | /* TODO http://msdn.microsoft.com/en-us/library/k6ktzx3s.aspx __declspec(noreturn) */ |
|
|
373 | #if ECB_C11 || __IBMC_NORETURN |
380 | #if ECB_C11 || __IBMC_NORETURN |
374 | /* http://pic.dhe.ibm.com/infocenter/compbg/v121v141/topic/com.ibm.xlcpp121.bg.doc/language_ref/noreturn.html */ |
381 | /* http://www-01.ibm.com/support/knowledgecenter/SSGH3R_13.1.0/com.ibm.xlcpp131.aix.doc/language_ref/noreturn.html */ |
375 | #define ecb_noreturn _Noreturn |
382 | #define ecb_noreturn _Noreturn |
376 | #elif ECB_CPP11 |
383 | #elif ECB_CPP11 |
377 | #define ecb_noreturn [[noreturn]] |
384 | #define ecb_noreturn [[noreturn]] |
378 | #elif _MSC_VER >= 1200 |
385 | #elif _MSC_VER >= 1200 |
|
|
386 | /* http://msdn.microsoft.com/en-us/library/k6ktzx3s.aspx */ |
379 | #define ecb_noreturn __declspec (noreturn) |
387 | #define ecb_noreturn __declspec (noreturn) |
380 | #else |
388 | #else |
381 | #define ecb_noreturn ecb_attribute ((__noreturn__)) |
389 | #define ecb_noreturn ecb_attribute ((__noreturn__)) |
382 | #endif |
390 | #endif |
383 | |
391 | |
… | |
… | |
544 | ecb_inline ecb_const uint32_t ecb_rotr32 (uint32_t x, unsigned int count) { return (x << (32 - count)) | (x >> count); } |
552 | ecb_inline ecb_const uint32_t ecb_rotr32 (uint32_t x, unsigned int count) { return (x << (32 - count)) | (x >> count); } |
545 | ecb_inline ecb_const uint64_t ecb_rotl64 (uint64_t x, unsigned int count) { return (x >> (64 - count)) | (x << count); } |
553 | ecb_inline ecb_const uint64_t ecb_rotl64 (uint64_t x, unsigned int count) { return (x >> (64 - count)) | (x << count); } |
546 | ecb_inline ecb_const uint64_t ecb_rotr64 (uint64_t x, unsigned int count) { return (x << (64 - count)) | (x >> count); } |
554 | ecb_inline ecb_const uint64_t ecb_rotr64 (uint64_t x, unsigned int count) { return (x << (64 - count)) | (x >> count); } |
547 | |
555 | |
548 | #if ECB_GCC_VERSION(4,3) || (ECB_CLANG_BUILTIN(__builtin_bswap32) && ECB_CLANG_BUILTIN(__builtin_bswap64)) |
556 | #if ECB_GCC_VERSION(4,3) || (ECB_CLANG_BUILTIN(__builtin_bswap32) && ECB_CLANG_BUILTIN(__builtin_bswap64)) |
|
|
557 | #if ECB_GCC_VERSION(4,8) || ECB_CLANG_BUILTIN(__builtin_bswap16) |
|
|
558 | #define ecb_bswap16(x) __builtin_bswap16 (x) |
|
|
559 | #else |
549 | #define ecb_bswap16(x) (__builtin_bswap32 (x) >> 16) |
560 | #define ecb_bswap16(x) (__builtin_bswap32 (x) >> 16) |
|
|
561 | #endif |
550 | #define ecb_bswap32(x) __builtin_bswap32 (x) |
562 | #define ecb_bswap32(x) __builtin_bswap32 (x) |
551 | #define ecb_bswap64(x) __builtin_bswap64 (x) |
563 | #define ecb_bswap64(x) __builtin_bswap64 (x) |
|
|
564 | #elif _MSC_VER |
|
|
565 | #include <stdlib.h> |
|
|
566 | #define ecb_bswap16(x) ((uint16_t)_byteswap_ushort ((uint16_t)(x))) |
|
|
567 | #define ecb_bswap32(x) ((uint32_t)_byteswap_ulong ((uint32_t)(x))) |
|
|
568 | #define ecb_bswap64(x) ((uint64_t)_byteswap_uint64 ((uint64_t)(x))) |
552 | #else |
569 | #else |
553 | ecb_function_ ecb_const uint16_t ecb_bswap16 (uint16_t x); |
570 | ecb_function_ ecb_const uint16_t ecb_bswap16 (uint16_t x); |
554 | ecb_function_ ecb_const uint16_t |
571 | ecb_function_ ecb_const uint16_t |
555 | ecb_bswap16 (uint16_t x) |
572 | ecb_bswap16 (uint16_t x) |
556 | { |
573 | { |
… | |
… | |
591 | /* but less than using pointers, and always seems to */ |
608 | /* but less than using pointers, and always seems to */ |
592 | /* successfully return a constant. */ |
609 | /* successfully return a constant. */ |
593 | /* the reason why we have this horrible preprocessor mess */ |
610 | /* the reason why we have this horrible preprocessor mess */ |
594 | /* is to avoid it in all cases, at least on common architectures */ |
611 | /* is to avoid it in all cases, at least on common architectures */ |
595 | /* or when using a recent enough gcc version (>= 4.6) */ |
612 | /* or when using a recent enough gcc version (>= 4.6) */ |
596 | #if __i386 || __i386__ || _M_X86 || __amd64 || __amd64__ || _M_X64 |
613 | #if ((__i386 || __i386__) && !__VOS__) || _M_IX86 || ECB_GCC_AMD64 || ECB_MSVC_AMD64 |
597 | return 0x44; |
614 | return 0x44; |
598 | #elif __BYTE_ORDER__ && __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ |
615 | #elif __BYTE_ORDER__ && __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ |
599 | return 0x44; |
616 | return 0x44; |
600 | #elif __BYTE_ORDER__ && __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ |
617 | #elif __BYTE_ORDER__ && __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ |
601 | return 0x11; |
618 | return 0x11; |
… | |
… | |
652 | |
669 | |
653 | /* basically, everything uses "ieee pure-endian" floating point numbers */ |
670 | /* basically, everything uses "ieee pure-endian" floating point numbers */ |
654 | /* the only noteworthy exception is ancient armle, which uses order 43218765 */ |
671 | /* the only noteworthy exception is ancient armle, which uses order 43218765 */ |
655 | #if 0 \ |
672 | #if 0 \ |
656 | || __i386 || __i386__ \ |
673 | || __i386 || __i386__ \ |
657 | || __amd64 || __amd64__ || __x86_64 || __x86_64__ \ |
674 | || ECB_GCC_AMD64 \ |
658 | || __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ \ |
675 | || __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ \ |
659 | || defined __s390__ || defined __s390x__ \ |
676 | || defined __s390__ || defined __s390x__ \ |
660 | || defined __mips__ \ |
677 | || defined __mips__ \ |
661 | || defined __alpha__ \ |
678 | || defined __alpha__ \ |
662 | || defined __hppa__ \ |
679 | || defined __hppa__ \ |
663 | || defined __ia64__ \ |
680 | || defined __ia64__ \ |
664 | || defined __m68k__ \ |
681 | || defined __m68k__ \ |
665 | || defined __m88k__ \ |
682 | || defined __m88k__ \ |
666 | || defined __sh__ \ |
683 | || defined __sh__ \ |
667 | || defined _M_IX86 || defined _M_AMD64 || defined _M_IA64 \ |
684 | || defined _M_IX86 || defined ECB_MSVC_AMD64 || defined _M_IA64 \ |
668 | || (defined __arm__ && (defined __ARM_EABI__ || defined __EABI__ || defined __VFP_FP__ || defined _WIN32_WCE || defined __ANDROID__)) \ |
685 | || (defined __arm__ && (defined __ARM_EABI__ || defined __EABI__ || defined __VFP_FP__ || defined _WIN32_WCE || defined __ANDROID__)) \ |
669 | || defined __aarch64__ |
686 | || defined __aarch64__ |
670 | #define ECB_STDFP 1 |
687 | #define ECB_STDFP 1 |
671 | #include <string.h> /* for memcpy */ |
688 | #include <string.h> /* for memcpy */ |
672 | #else |
689 | #else |
… | |
… | |
690 | #define ECB_NAN ECB_INFINITY |
707 | #define ECB_NAN ECB_INFINITY |
691 | #endif |
708 | #endif |
692 | |
709 | |
693 | #if ECB_C99 || _XOPEN_VERSION >= 600 || _POSIX_VERSION >= 200112L |
710 | #if ECB_C99 || _XOPEN_VERSION >= 600 || _POSIX_VERSION >= 200112L |
694 | #define ecb_ldexpf(x,e) ldexpf ((x), (e)) |
711 | #define ecb_ldexpf(x,e) ldexpf ((x), (e)) |
|
|
712 | #define ecb_frexpf(x,e) frexpf ((x), (e)) |
695 | #else |
713 | #else |
696 | #define ecb_ldexpf(x,e) (float) ldexp ((x), (e)) |
714 | #define ecb_ldexpf(x,e) (float) ldexp ((double) (x), (e)) |
|
|
715 | #define ecb_frexpf(x,e) (float) frexp ((double) (x), (e)) |
697 | #endif |
716 | #endif |
698 | |
717 | |
699 | /* converts an ieee half/binary16 to a float */ |
718 | /* converts an ieee half/binary16 to a float */ |
700 | ecb_function_ ecb_const float ecb_binary16_to_float (uint16_t x); |
719 | ecb_function_ ecb_const float ecb_binary16_to_float (uint16_t x); |
701 | ecb_function_ ecb_const float |
720 | ecb_function_ ecb_const float |
… | |
… | |
730 | if (x == 0e0f ) return 0x00000000U; |
749 | if (x == 0e0f ) return 0x00000000U; |
731 | if (x > +3.40282346638528860e+38f) return 0x7f800000U; |
750 | if (x > +3.40282346638528860e+38f) return 0x7f800000U; |
732 | if (x < -3.40282346638528860e+38f) return 0xff800000U; |
751 | if (x < -3.40282346638528860e+38f) return 0xff800000U; |
733 | if (x != x ) return 0x7fbfffffU; |
752 | if (x != x ) return 0x7fbfffffU; |
734 | |
753 | |
735 | m = frexpf (x, &e) * 0x1000000U; |
754 | m = ecb_frexpf (x, &e) * 0x1000000U; |
736 | |
755 | |
737 | r = m & 0x80000000U; |
756 | r = m & 0x80000000U; |
738 | |
757 | |
739 | if (r) |
758 | if (r) |
740 | m = -m; |
759 | m = -m; |