1 | /* |
1 | /* |
2 | * libecb - http://software.schmorp.de/pkg/libecb |
2 | * libecb - http://software.schmorp.de/pkg/libecb |
3 | * |
3 | * |
4 | * Copyright (©) 2009-2012 Marc Alexander Lehmann <libecb@schmorp.de> |
4 | * Copyright (©) 2009-2013 Marc Alexander Lehmann <libecb@schmorp.de> |
5 | * Copyright (©) 2011 Emanuele Giaquinta |
5 | * Copyright (©) 2011 Emanuele Giaquinta |
6 | * All rights reserved. |
6 | * All rights reserved. |
7 | * |
7 | * |
8 | * Redistribution and use in source and binary forms, with or without modifica- |
8 | * Redistribution and use in source and binary forms, with or without modifica- |
9 | * tion, are permitted provided that the following conditions are met: |
9 | * tion, are permitted provided that the following conditions are met: |
… | |
… | |
64 | #define ECB_PTRSIZE 4 |
64 | #define ECB_PTRSIZE 4 |
65 | #endif |
65 | #endif |
66 | #endif |
66 | #endif |
67 | |
67 | |
68 | /* work around x32 idiocy by defining proper macros */ |
68 | /* work around x32 idiocy by defining proper macros */ |
69 | #if __x86_64 || _M_AMD64 |
69 | #if __amd64 || __x86_64 || _M_AMD64 || _M_X64 |
70 | #if _ILP32 |
70 | #if _ILP32 |
71 | #define ECB_AMD64_X32 1 |
71 | #define ECB_AMD64_X32 1 |
72 | #else |
72 | #else |
73 | #define ECB_AMD64 1 |
73 | #define ECB_AMD64 1 |
74 | #endif |
74 | #endif |
… | |
… | |
134 | || defined __ARM_ARCH_6K__ || defined __ARM_ARCH_6ZK__ |
134 | || defined __ARM_ARCH_6K__ || defined __ARM_ARCH_6ZK__ |
135 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,5" : : "r" (0) : "memory") |
135 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,5" : : "r" (0) : "memory") |
136 | #elif defined __ARM_ARCH_7__ || defined __ARM_ARCH_7A__ \ |
136 | #elif defined __ARM_ARCH_7__ || defined __ARM_ARCH_7A__ \ |
137 | || defined __ARM_ARCH_7M__ || defined __ARM_ARCH_7R__ |
137 | || defined __ARM_ARCH_7M__ || defined __ARM_ARCH_7R__ |
138 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") |
138 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") |
139 | #elif __sparc || __sparc__ |
139 | #elif (__sparc || __sparc__) && !__sparcv8 |
140 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad" : : : "memory") |
140 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad" : : : "memory") |
141 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory") |
141 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory") |
142 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore") |
142 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore") |
143 | #elif defined __s390__ || defined __s390x__ |
143 | #elif defined __s390__ || defined __s390x__ |
144 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory") |
144 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory") |
… | |
… | |
177 | * #define ECB_MEMORY_FENCE __c11_atomic_thread_fence (__ATOMIC_SEQ_CST) |
177 | * #define ECB_MEMORY_FENCE __c11_atomic_thread_fence (__ATOMIC_SEQ_CST) |
178 | */ |
178 | */ |
179 | |
179 | |
180 | #elif ECB_GCC_VERSION(4,4) || defined __INTEL_COMPILER || defined __clang__ |
180 | #elif ECB_GCC_VERSION(4,4) || defined __INTEL_COMPILER || defined __clang__ |
181 | #define ECB_MEMORY_FENCE __sync_synchronize () |
181 | #define ECB_MEMORY_FENCE __sync_synchronize () |
|
|
182 | #elif _MSC_VER >= 1500 /* VC++ 2008 */ |
|
|
183 | /* apparently, microsoft broke all the memory barrier stuff in Visual Studio 2008... */ |
|
|
184 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
|
|
185 | #define ECB_MEMORY_FENCE _ReadWriteBarrier (); MemoryBarrier() |
|
|
186 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier (); MemoryBarrier() /* according to msdn, _ReadBarrier is not a load fence */ |
|
|
187 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier (); MemoryBarrier() |
182 | #elif _MSC_VER >= 1400 /* VC++ 2005 */ |
188 | #elif _MSC_VER >= 1400 /* VC++ 2005 */ |
183 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
189 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
184 | #define ECB_MEMORY_FENCE _ReadWriteBarrier () |
190 | #define ECB_MEMORY_FENCE _ReadWriteBarrier () |
185 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ |
191 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ |
186 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () |
192 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () |
… | |
… | |
612 | float r; |
618 | float r; |
613 | |
619 | |
614 | if (!e ) r = ldexpf (m , -24); |
620 | if (!e ) r = ldexpf (m , -24); |
615 | else if (e != 31) r = ldexpf (m + 0x400, e - 25); |
621 | else if (e != 31) r = ldexpf (m + 0x400, e - 25); |
616 | else if (m ) r = ECB_NAN; |
622 | else if (m ) r = ECB_NAN; |
617 | else r = INFINITY; |
623 | else r = ECB_INFINITY; |
618 | |
624 | |
619 | return x & 0x8000 ? -r : r; |
625 | return x & 0x8000 ? -r : r; |
620 | } |
626 | } |
621 | |
627 | |
622 | /* convert a float to ieee single/binary32 */ |
628 | /* convert a float to ieee single/binary32 */ |