1 | /* |
1 | /* |
2 | * libev event processing core, watcher management |
2 | * libev event processing core, watcher management |
3 | * |
3 | * |
4 | * Copyright (c) 2007,2008,2009,2010,2011,2012 Marc Alexander Lehmann <libev@schmorp.de> |
4 | * Copyright (c) 2007,2008,2009,2010,2011,2012,2013 Marc Alexander Lehmann <libev@schmorp.de> |
5 | * All rights reserved. |
5 | * All rights reserved. |
6 | * |
6 | * |
7 | * Redistribution and use in source and binary forms, with or without modifica- |
7 | * Redistribution and use in source and binary forms, with or without modifica- |
8 | * tion, are permitted provided that the following conditions are met: |
8 | * tion, are permitted provided that the following conditions are met: |
9 | * |
9 | * |
… | |
… | |
43 | # include EV_CONFIG_H |
43 | # include EV_CONFIG_H |
44 | # else |
44 | # else |
45 | # include "config.h" |
45 | # include "config.h" |
46 | # endif |
46 | # endif |
47 | |
47 | |
48 | #if HAVE_FLOOR |
48 | # if HAVE_FLOOR |
49 | # ifndef EV_USE_FLOOR |
49 | # ifndef EV_USE_FLOOR |
50 | # define EV_USE_FLOOR 1 |
50 | # define EV_USE_FLOOR 1 |
|
|
51 | # endif |
51 | # endif |
52 | # endif |
52 | #endif |
|
|
53 | |
53 | |
54 | # if HAVE_CLOCK_SYSCALL |
54 | # if HAVE_CLOCK_SYSCALL |
55 | # ifndef EV_USE_CLOCK_SYSCALL |
55 | # ifndef EV_USE_CLOCK_SYSCALL |
56 | # define EV_USE_CLOCK_SYSCALL 1 |
56 | # define EV_USE_CLOCK_SYSCALL 1 |
57 | # ifndef EV_USE_REALTIME |
57 | # ifndef EV_USE_REALTIME |
… | |
… | |
241 | #elif defined SIGARRAYSIZE |
241 | #elif defined SIGARRAYSIZE |
242 | # define EV_NSIG (SIGARRAYSIZE) /* Assume ary[SIGARRAYSIZE] */ |
242 | # define EV_NSIG (SIGARRAYSIZE) /* Assume ary[SIGARRAYSIZE] */ |
243 | #elif defined _sys_nsig |
243 | #elif defined _sys_nsig |
244 | # define EV_NSIG (_sys_nsig) /* Solaris 2.5 */ |
244 | # define EV_NSIG (_sys_nsig) /* Solaris 2.5 */ |
245 | #else |
245 | #else |
246 | # error "unable to find value for NSIG, please report" |
246 | # define EV_NSIG (8 * sizeof (sigset_t) + 1) |
247 | /* to make it compile regardless, just remove the above line, */ |
|
|
248 | /* but consider reporting it, too! :) */ |
|
|
249 | # define EV_NSIG 65 |
|
|
250 | #endif |
247 | #endif |
251 | |
248 | |
252 | #ifndef EV_USE_FLOOR |
249 | #ifndef EV_USE_FLOOR |
253 | # define EV_USE_FLOOR 0 |
250 | # define EV_USE_FLOOR 0 |
254 | #endif |
251 | #endif |
255 | |
252 | |
256 | #ifndef EV_USE_CLOCK_SYSCALL |
253 | #ifndef EV_USE_CLOCK_SYSCALL |
257 | # if __linux && __GLIBC__ >= 2 |
254 | # if __linux && __GLIBC__ == 2 && __GLIBC_MINOR__ < 17 |
258 | # define EV_USE_CLOCK_SYSCALL EV_FEATURE_OS |
255 | # define EV_USE_CLOCK_SYSCALL EV_FEATURE_OS |
259 | # else |
256 | # else |
260 | # define EV_USE_CLOCK_SYSCALL 0 |
257 | # define EV_USE_CLOCK_SYSCALL 0 |
|
|
258 | # endif |
|
|
259 | #endif |
|
|
260 | |
|
|
261 | #if !(_POSIX_TIMERS > 0) |
|
|
262 | # ifndef EV_USE_MONOTONIC |
|
|
263 | # define EV_USE_MONOTONIC 0 |
|
|
264 | # endif |
|
|
265 | # ifndef EV_USE_REALTIME |
|
|
266 | # define EV_USE_REALTIME 0 |
261 | # endif |
267 | # endif |
262 | #endif |
268 | #endif |
263 | |
269 | |
264 | #ifndef EV_USE_MONOTONIC |
270 | #ifndef EV_USE_MONOTONIC |
265 | # if defined _POSIX_MONOTONIC_CLOCK && _POSIX_MONOTONIC_CLOCK >= 0 |
271 | # if defined _POSIX_MONOTONIC_CLOCK && _POSIX_MONOTONIC_CLOCK >= 0 |
… | |
… | |
355 | # define EV_USE_4HEAP EV_FEATURE_DATA |
361 | # define EV_USE_4HEAP EV_FEATURE_DATA |
356 | #endif |
362 | #endif |
357 | |
363 | |
358 | #ifndef EV_HEAP_CACHE_AT |
364 | #ifndef EV_HEAP_CACHE_AT |
359 | # define EV_HEAP_CACHE_AT EV_FEATURE_DATA |
365 | # define EV_HEAP_CACHE_AT EV_FEATURE_DATA |
|
|
366 | #endif |
|
|
367 | |
|
|
368 | #ifdef ANDROID |
|
|
369 | /* supposedly, android doesn't typedef fd_mask */ |
|
|
370 | # undef EV_USE_SELECT |
|
|
371 | # define EV_USE_SELECT 0 |
|
|
372 | /* supposedly, we need to include syscall.h, not sys/syscall.h, so just disable */ |
|
|
373 | # undef EV_USE_CLOCK_SYSCALL |
|
|
374 | # define EV_USE_CLOCK_SYSCALL 0 |
|
|
375 | #endif |
|
|
376 | |
|
|
377 | /* aix's poll.h seems to cause lots of trouble */ |
|
|
378 | #ifdef _AIX |
|
|
379 | /* AIX has a completely broken poll.h header */ |
|
|
380 | # undef EV_USE_POLL |
|
|
381 | # define EV_USE_POLL 0 |
360 | #endif |
382 | #endif |
361 | |
383 | |
362 | /* on linux, we can use a (slow) syscall to avoid a dependency on pthread, */ |
384 | /* on linux, we can use a (slow) syscall to avoid a dependency on pthread, */ |
363 | /* which makes programs even slower. might work on other unices, too. */ |
385 | /* which makes programs even slower. might work on other unices, too. */ |
364 | #if EV_USE_CLOCK_SYSCALL |
386 | #if EV_USE_CLOCK_SYSCALL |
… | |
… | |
372 | # define EV_USE_CLOCK_SYSCALL 0 |
394 | # define EV_USE_CLOCK_SYSCALL 0 |
373 | # endif |
395 | # endif |
374 | #endif |
396 | #endif |
375 | |
397 | |
376 | /* this block fixes any misconfiguration where we know we run into trouble otherwise */ |
398 | /* this block fixes any misconfiguration where we know we run into trouble otherwise */ |
377 | |
|
|
378 | #ifdef _AIX |
|
|
379 | /* AIX has a completely broken poll.h header */ |
|
|
380 | # undef EV_USE_POLL |
|
|
381 | # define EV_USE_POLL 0 |
|
|
382 | #endif |
|
|
383 | |
399 | |
384 | #ifndef CLOCK_MONOTONIC |
400 | #ifndef CLOCK_MONOTONIC |
385 | # undef EV_USE_MONOTONIC |
401 | # undef EV_USE_MONOTONIC |
386 | # define EV_USE_MONOTONIC 0 |
402 | # define EV_USE_MONOTONIC 0 |
387 | #endif |
403 | #endif |
… | |
… | |
475 | /* the following is ecb.h embedded into libev - use update_ev_c to update from an external copy */ |
491 | /* the following is ecb.h embedded into libev - use update_ev_c to update from an external copy */ |
476 | /* ECB.H BEGIN */ |
492 | /* ECB.H BEGIN */ |
477 | /* |
493 | /* |
478 | * libecb - http://software.schmorp.de/pkg/libecb |
494 | * libecb - http://software.schmorp.de/pkg/libecb |
479 | * |
495 | * |
480 | * Copyright (©) 2009-2012 Marc Alexander Lehmann <libecb@schmorp.de> |
496 | * Copyright (©) 2009-2015 Marc Alexander Lehmann <libecb@schmorp.de> |
481 | * Copyright (©) 2011 Emanuele Giaquinta |
497 | * Copyright (©) 2011 Emanuele Giaquinta |
482 | * All rights reserved. |
498 | * All rights reserved. |
483 | * |
499 | * |
484 | * Redistribution and use in source and binary forms, with or without modifica- |
500 | * Redistribution and use in source and binary forms, with or without modifica- |
485 | * tion, are permitted provided that the following conditions are met: |
501 | * tion, are permitted provided that the following conditions are met: |
… | |
… | |
499 | * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; |
515 | * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; |
500 | * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, |
516 | * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, |
501 | * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTH- |
517 | * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTH- |
502 | * ERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED |
518 | * ERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED |
503 | * OF THE POSSIBILITY OF SUCH DAMAGE. |
519 | * OF THE POSSIBILITY OF SUCH DAMAGE. |
|
|
520 | * |
|
|
521 | * Alternatively, the contents of this file may be used under the terms of |
|
|
522 | * the GNU General Public License ("GPL") version 2 or any later version, |
|
|
523 | * in which case the provisions of the GPL are applicable instead of |
|
|
524 | * the above. If you wish to allow the use of your version of this file |
|
|
525 | * only under the terms of the GPL and not to allow others to use your |
|
|
526 | * version of this file under the BSD license, indicate your decision |
|
|
527 | * by deleting the provisions above and replace them with the notice |
|
|
528 | * and other provisions required by the GPL. If you do not delete the |
|
|
529 | * provisions above, a recipient may use your version of this file under |
|
|
530 | * either the BSD or the GPL. |
504 | */ |
531 | */ |
505 | |
532 | |
506 | #ifndef ECB_H |
533 | #ifndef ECB_H |
507 | #define ECB_H |
534 | #define ECB_H |
508 | |
535 | |
509 | /* 16 bits major, 16 bits minor */ |
536 | /* 16 bits major, 16 bits minor */ |
510 | #define ECB_VERSION 0x00010001 |
537 | #define ECB_VERSION 0x00010005 |
511 | |
538 | |
512 | #ifdef _WIN32 |
539 | #ifdef _WIN32 |
513 | typedef signed char int8_t; |
540 | typedef signed char int8_t; |
514 | typedef unsigned char uint8_t; |
541 | typedef unsigned char uint8_t; |
515 | typedef signed short int16_t; |
542 | typedef signed short int16_t; |
… | |
… | |
530 | #else |
557 | #else |
531 | #define ECB_PTRSIZE 4 |
558 | #define ECB_PTRSIZE 4 |
532 | typedef uint32_t uintptr_t; |
559 | typedef uint32_t uintptr_t; |
533 | typedef int32_t intptr_t; |
560 | typedef int32_t intptr_t; |
534 | #endif |
561 | #endif |
535 | typedef intptr_t ptrdiff_t; |
|
|
536 | #else |
562 | #else |
537 | #include <inttypes.h> |
563 | #include <inttypes.h> |
538 | #if UINTMAX_MAX > 0xffffffffU |
564 | #if (defined INTPTR_MAX ? INTPTR_MAX : ULONG_MAX) > 0xffffffffU |
539 | #define ECB_PTRSIZE 8 |
565 | #define ECB_PTRSIZE 8 |
540 | #else |
566 | #else |
541 | #define ECB_PTRSIZE 4 |
567 | #define ECB_PTRSIZE 4 |
|
|
568 | #endif |
|
|
569 | #endif |
|
|
570 | |
|
|
571 | #define ECB_GCC_AMD64 (__amd64 || __amd64__ || __x86_64 || __x86_64__) |
|
|
572 | #define ECB_MSVC_AMD64 (_M_AMD64 || _M_X64) |
|
|
573 | |
|
|
574 | /* work around x32 idiocy by defining proper macros */ |
|
|
575 | #if ECB_GCC_AMD64 || ECB_MSVC_AMD64 |
|
|
576 | #if _ILP32 |
|
|
577 | #define ECB_AMD64_X32 1 |
|
|
578 | #else |
|
|
579 | #define ECB_AMD64 1 |
542 | #endif |
580 | #endif |
543 | #endif |
581 | #endif |
544 | |
582 | |
545 | /* many compilers define _GNUC_ to some versions but then only implement |
583 | /* many compilers define _GNUC_ to some versions but then only implement |
546 | * what their idiot authors think are the "more important" extensions, |
584 | * what their idiot authors think are the "more important" extensions, |
547 | * causing enormous grief in return for some better fake benchmark numbers. |
585 | * causing enormous grief in return for some better fake benchmark numbers. |
548 | * or so. |
586 | * or so. |
549 | * we try to detect these and simply assume they are not gcc - if they have |
587 | * we try to detect these and simply assume they are not gcc - if they have |
550 | * an issue with that they should have done it right in the first place. |
588 | * an issue with that they should have done it right in the first place. |
551 | */ |
589 | */ |
552 | #ifndef ECB_GCC_VERSION |
|
|
553 | #if !defined __GNUC_MINOR__ || defined __INTEL_COMPILER || defined __SUNPRO_C || defined __SUNPRO_CC || defined __llvm__ || defined __clang__ |
590 | #if !defined __GNUC_MINOR__ || defined __INTEL_COMPILER || defined __SUNPRO_C || defined __SUNPRO_CC || defined __llvm__ || defined __clang__ |
554 | #define ECB_GCC_VERSION(major,minor) 0 |
591 | #define ECB_GCC_VERSION(major,minor) 0 |
555 | #else |
592 | #else |
556 | #define ECB_GCC_VERSION(major,minor) (__GNUC__ > (major) || (__GNUC__ == (major) && __GNUC_MINOR__ >= (minor))) |
593 | #define ECB_GCC_VERSION(major,minor) (__GNUC__ > (major) || (__GNUC__ == (major) && __GNUC_MINOR__ >= (minor))) |
557 | #endif |
594 | #endif |
558 | #endif |
|
|
559 | |
595 | |
560 | #define ECB_C (__STDC__+0) /* this assumes that __STDC__ is either empty or a number */ |
596 | #define ECB_CLANG_VERSION(major,minor) (__clang_major__ > (major) || (__clang_major__ == (major) && __clang_minor__ >= (minor))) |
561 | #define ECB_C99 (__STDC_VERSION__ >= 199901L) |
597 | |
562 | #define ECB_C11 (__STDC_VERSION__ >= 201112L) |
598 | #if __clang__ && defined __has_builtin |
|
|
599 | #define ECB_CLANG_BUILTIN(x) __has_builtin (x) |
|
|
600 | #else |
|
|
601 | #define ECB_CLANG_BUILTIN(x) 0 |
|
|
602 | #endif |
|
|
603 | |
|
|
604 | #if __clang__ && defined __has_extension |
|
|
605 | #define ECB_CLANG_EXTENSION(x) __has_extension (x) |
|
|
606 | #else |
|
|
607 | #define ECB_CLANG_EXTENSION(x) 0 |
|
|
608 | #endif |
|
|
609 | |
563 | #define ECB_CPP (__cplusplus+0) |
610 | #define ECB_CPP (__cplusplus+0) |
564 | #define ECB_CPP98 (__cplusplus >= 199711L) |
|
|
565 | #define ECB_CPP11 (__cplusplus >= 201103L) |
611 | #define ECB_CPP11 (__cplusplus >= 201103L) |
|
|
612 | |
|
|
613 | #if ECB_CPP |
|
|
614 | #define ECB_C 0 |
|
|
615 | #define ECB_STDC_VERSION 0 |
|
|
616 | #else |
|
|
617 | #define ECB_C 1 |
|
|
618 | #define ECB_STDC_VERSION __STDC_VERSION__ |
|
|
619 | #endif |
|
|
620 | |
|
|
621 | #define ECB_C99 (ECB_STDC_VERSION >= 199901L) |
|
|
622 | #define ECB_C11 (ECB_STDC_VERSION >= 201112L) |
|
|
623 | |
|
|
624 | #if ECB_CPP |
|
|
625 | #define ECB_EXTERN_C extern "C" |
|
|
626 | #define ECB_EXTERN_C_BEG ECB_EXTERN_C { |
|
|
627 | #define ECB_EXTERN_C_END } |
|
|
628 | #else |
|
|
629 | #define ECB_EXTERN_C extern |
|
|
630 | #define ECB_EXTERN_C_BEG |
|
|
631 | #define ECB_EXTERN_C_END |
|
|
632 | #endif |
566 | |
633 | |
567 | /*****************************************************************************/ |
634 | /*****************************************************************************/ |
568 | |
635 | |
569 | /* ECB_NO_THREADS - ecb is not used by multiple threads, ever */ |
636 | /* ECB_NO_THREADS - ecb is not used by multiple threads, ever */ |
570 | /* ECB_NO_SMP - ecb might be used in multiple threads, but only on a single cpu */ |
637 | /* ECB_NO_SMP - ecb might be used in multiple threads, but only on a single cpu */ |
571 | |
638 | |
572 | #if ECB_NO_THREADS |
639 | #if ECB_NO_THREADS |
573 | # define ECB_NO_SMP 1 |
640 | #define ECB_NO_SMP 1 |
574 | #endif |
641 | #endif |
575 | |
642 | |
576 | #if ECB_NO_SMP |
643 | #if ECB_NO_SMP |
577 | #define ECB_MEMORY_FENCE do { } while (0) |
644 | #define ECB_MEMORY_FENCE do { } while (0) |
|
|
645 | #endif |
|
|
646 | |
|
|
647 | /* http://www-01.ibm.com/support/knowledgecenter/SSGH3R_13.1.0/com.ibm.xlcpp131.aix.doc/compiler_ref/compiler_builtins.html */ |
|
|
648 | #if __xlC__ && ECB_CPP |
|
|
649 | #include <builtins.h> |
|
|
650 | #endif |
|
|
651 | |
|
|
652 | #if 1400 <= _MSC_VER |
|
|
653 | #include <intrin.h> /* fence functions _ReadBarrier, also bit search functions _BitScanReverse */ |
578 | #endif |
654 | #endif |
579 | |
655 | |
580 | #ifndef ECB_MEMORY_FENCE |
656 | #ifndef ECB_MEMORY_FENCE |
581 | #if ECB_GCC_VERSION(2,5) || defined __INTEL_COMPILER || (__llvm__ && __GNUC__) || __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
657 | #if ECB_GCC_VERSION(2,5) || defined __INTEL_COMPILER || (__llvm__ && __GNUC__) || __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
582 | #if __i386 || __i386__ |
658 | #if __i386 || __i386__ |
583 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
659 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
584 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory") |
660 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory") |
585 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
661 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
586 | #elif __amd64 || __amd64__ || __x86_64 || __x86_64__ |
662 | #elif ECB_GCC_AMD64 |
587 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
663 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
588 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory") |
664 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory") |
589 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
665 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
590 | #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ |
666 | #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ |
591 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
667 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
|
|
668 | #elif defined __ARM_ARCH_2__ \ |
|
|
669 | || defined __ARM_ARCH_3__ || defined __ARM_ARCH_3M__ \ |
|
|
670 | || defined __ARM_ARCH_4__ || defined __ARM_ARCH_4T__ \ |
|
|
671 | || defined __ARM_ARCH_5__ || defined __ARM_ARCH_5E__ \ |
|
|
672 | || defined __ARM_ARCH_5T__ || defined __ARM_ARCH_5TE__ \ |
|
|
673 | || defined __ARM_ARCH_5TEJ__ |
|
|
674 | /* should not need any, unless running old code on newer cpu - arm doesn't support that */ |
592 | #elif defined __ARM_ARCH_6__ || defined __ARM_ARCH_6J__ \ |
675 | #elif defined __ARM_ARCH_6__ || defined __ARM_ARCH_6J__ \ |
593 | || defined __ARM_ARCH_6K__ || defined __ARM_ARCH_6ZK__ |
676 | || defined __ARM_ARCH_6K__ || defined __ARM_ARCH_6ZK__ \ |
|
|
677 | || defined __ARM_ARCH_6T2__ |
594 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,5" : : "r" (0) : "memory") |
678 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,5" : : "r" (0) : "memory") |
595 | #elif defined __ARM_ARCH_7__ || defined __ARM_ARCH_7A__ \ |
679 | #elif defined __ARM_ARCH_7__ || defined __ARM_ARCH_7A__ \ |
596 | || defined __ARM_ARCH_7M__ || defined __ARM_ARCH_7R__ |
680 | || defined __ARM_ARCH_7R__ || defined __ARM_ARCH_7M__ |
597 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") |
681 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") |
598 | #elif __sparc || __sparc__ |
682 | #elif __aarch64__ |
|
|
683 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb ish" : : : "memory") |
|
|
684 | #elif (__sparc || __sparc__) && !(__sparc_v8__ || defined __sparcv8) |
599 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad" : : : "memory") |
685 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad" : : : "memory") |
600 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory") |
686 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory") |
601 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore") |
687 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore") |
602 | #elif defined __s390__ || defined __s390x__ |
688 | #elif defined __s390__ || defined __s390x__ |
603 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory") |
689 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory") |
604 | #elif defined __mips__ |
690 | #elif defined __mips__ |
|
|
691 | /* GNU/Linux emulates sync on mips1 architectures, so we force its use */ |
|
|
692 | /* anybody else who still uses mips1 is supposed to send in their version, with detection code. */ |
605 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
693 | #define ECB_MEMORY_FENCE __asm__ __volatile__ (".set mips2; sync; .set mips0" : : : "memory") |
606 | #elif defined __alpha__ |
694 | #elif defined __alpha__ |
607 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mb" : : : "memory") |
695 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mb" : : : "memory") |
608 | #elif defined __hppa__ |
696 | #elif defined __hppa__ |
609 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("" : : : "memory") |
697 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("" : : : "memory") |
610 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
698 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
611 | #elif defined __ia64__ |
699 | #elif defined __ia64__ |
612 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mf" : : : "memory") |
700 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mf" : : : "memory") |
|
|
701 | #elif defined __m68k__ |
|
|
702 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("" : : : "memory") |
|
|
703 | #elif defined __m88k__ |
|
|
704 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("tb1 0,%%r0,128" : : : "memory") |
|
|
705 | #elif defined __sh__ |
|
|
706 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("" : : : "memory") |
613 | #endif |
707 | #endif |
614 | #endif |
708 | #endif |
615 | #endif |
709 | #endif |
616 | |
710 | |
617 | #ifndef ECB_MEMORY_FENCE |
711 | #ifndef ECB_MEMORY_FENCE |
618 | #if ECB_GCC_VERSION(4,7) |
712 | #if ECB_GCC_VERSION(4,7) |
619 | /* see comment below about the C11 memory model. in short - avoid */ |
713 | /* see comment below (stdatomic.h) about the C11 memory model. */ |
620 | #define ECB_MEMORY_FENCE __atomic_thread_fence (__ATOMIC_SEQ_CST) |
714 | #define ECB_MEMORY_FENCE __atomic_thread_fence (__ATOMIC_SEQ_CST) |
621 | #elif defined __clang && __has_feature (cxx_atomic) |
715 | #define ECB_MEMORY_FENCE_ACQUIRE __atomic_thread_fence (__ATOMIC_ACQUIRE) |
622 | /* see above */ |
716 | #define ECB_MEMORY_FENCE_RELEASE __atomic_thread_fence (__ATOMIC_RELEASE) |
|
|
717 | |
|
|
718 | #elif ECB_CLANG_EXTENSION(c_atomic) |
|
|
719 | /* see comment below (stdatomic.h) about the C11 memory model. */ |
623 | #define ECB_MEMORY_FENCE __c11_atomic_thread_fence (__ATOMIC_SEQ_CST) |
720 | #define ECB_MEMORY_FENCE __c11_atomic_thread_fence (__ATOMIC_SEQ_CST) |
|
|
721 | #define ECB_MEMORY_FENCE_ACQUIRE __c11_atomic_thread_fence (__ATOMIC_ACQUIRE) |
|
|
722 | #define ECB_MEMORY_FENCE_RELEASE __c11_atomic_thread_fence (__ATOMIC_RELEASE) |
|
|
723 | |
624 | #elif ECB_GCC_VERSION(4,4) || defined __INTEL_COMPILER || defined __clang__ |
724 | #elif ECB_GCC_VERSION(4,4) || defined __INTEL_COMPILER || defined __clang__ |
625 | #define ECB_MEMORY_FENCE __sync_synchronize () |
725 | #define ECB_MEMORY_FENCE __sync_synchronize () |
626 | /*#define ECB_MEMORY_FENCE_ACQUIRE ({ char dummy = 0; __sync_lock_test_and_set (&dummy, 1); }) */ |
726 | #elif _MSC_VER >= 1500 /* VC++ 2008 */ |
627 | /*#define ECB_MEMORY_FENCE_RELEASE ({ char dummy = 1; __sync_lock_release (&dummy ); }) */ |
727 | /* apparently, microsoft broke all the memory barrier stuff in Visual Studio 2008... */ |
|
|
728 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
|
|
729 | #define ECB_MEMORY_FENCE _ReadWriteBarrier (); MemoryBarrier() |
|
|
730 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier (); MemoryBarrier() /* according to msdn, _ReadBarrier is not a load fence */ |
|
|
731 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier (); MemoryBarrier() |
628 | #elif _MSC_VER >= 1400 /* VC++ 2005 */ |
732 | #elif _MSC_VER >= 1400 /* VC++ 2005 */ |
629 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
733 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
630 | #define ECB_MEMORY_FENCE _ReadWriteBarrier () |
734 | #define ECB_MEMORY_FENCE _ReadWriteBarrier () |
631 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ |
735 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ |
632 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () |
736 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () |
… | |
… | |
646 | #ifndef ECB_MEMORY_FENCE |
750 | #ifndef ECB_MEMORY_FENCE |
647 | #if ECB_C11 && !defined __STDC_NO_ATOMICS__ |
751 | #if ECB_C11 && !defined __STDC_NO_ATOMICS__ |
648 | /* we assume that these memory fences work on all variables/all memory accesses, */ |
752 | /* we assume that these memory fences work on all variables/all memory accesses, */ |
649 | /* not just C11 atomics and atomic accesses */ |
753 | /* not just C11 atomics and atomic accesses */ |
650 | #include <stdatomic.h> |
754 | #include <stdatomic.h> |
651 | /* unfortunately, the C11 memory model seems to be very limited, and unable to express */ |
755 | /* Unfortunately, neither gcc 4.7 nor clang 3.1 generate any instructions for */ |
652 | /* simple barrier semantics. That means we need to take out thor's hammer. */ |
756 | /* any fence other than seq_cst, which isn't very efficient for us. */ |
|
|
757 | /* Why that is, we don't know - either the C11 memory model is quite useless */ |
|
|
758 | /* for most usages, or gcc and clang have a bug */ |
|
|
759 | /* I *currently* lean towards the latter, and inefficiently implement */ |
|
|
760 | /* all three of ecb's fences as a seq_cst fence */ |
|
|
761 | /* Update, gcc-4.8 generates mfence for all c++ fences, but nothing */ |
|
|
762 | /* for all __atomic_thread_fence's except seq_cst */ |
653 | #define ECB_MEMORY_FENCE atomic_thread_fence (memory_order_seq_cst) |
763 | #define ECB_MEMORY_FENCE atomic_thread_fence (memory_order_seq_cst) |
654 | #endif |
|
|
655 | #endif |
764 | #endif |
656 | #endif |
765 | #endif |
657 | |
766 | |
658 | #ifndef ECB_MEMORY_FENCE |
767 | #ifndef ECB_MEMORY_FENCE |
659 | #if !ECB_AVOID_PTHREADS |
768 | #if !ECB_AVOID_PTHREADS |
… | |
… | |
681 | #define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE |
790 | #define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE |
682 | #endif |
791 | #endif |
683 | |
792 | |
684 | /*****************************************************************************/ |
793 | /*****************************************************************************/ |
685 | |
794 | |
686 | #if __cplusplus |
795 | #if ECB_CPP |
687 | #define ecb_inline static inline |
796 | #define ecb_inline static inline |
688 | #elif ECB_GCC_VERSION(2,5) |
797 | #elif ECB_GCC_VERSION(2,5) |
689 | #define ecb_inline static __inline__ |
798 | #define ecb_inline static __inline__ |
690 | #elif ECB_C99 |
799 | #elif ECB_C99 |
691 | #define ecb_inline static inline |
800 | #define ecb_inline static inline |
… | |
… | |
705 | |
814 | |
706 | #define ECB_CONCAT_(a, b) a ## b |
815 | #define ECB_CONCAT_(a, b) a ## b |
707 | #define ECB_CONCAT(a, b) ECB_CONCAT_(a, b) |
816 | #define ECB_CONCAT(a, b) ECB_CONCAT_(a, b) |
708 | #define ECB_STRINGIFY_(a) # a |
817 | #define ECB_STRINGIFY_(a) # a |
709 | #define ECB_STRINGIFY(a) ECB_STRINGIFY_(a) |
818 | #define ECB_STRINGIFY(a) ECB_STRINGIFY_(a) |
|
|
819 | #define ECB_STRINGIFY_EXPR(expr) ((expr), ECB_STRINGIFY_ (expr)) |
710 | |
820 | |
711 | #define ecb_function_ ecb_inline |
821 | #define ecb_function_ ecb_inline |
712 | |
822 | |
713 | #if ECB_GCC_VERSION(3,1) |
823 | #if ECB_GCC_VERSION(3,1) || ECB_CLANG_VERSION(2,8) |
714 | #define ecb_attribute(attrlist) __attribute__(attrlist) |
824 | #define ecb_attribute(attrlist) __attribute__ (attrlist) |
|
|
825 | #else |
|
|
826 | #define ecb_attribute(attrlist) |
|
|
827 | #endif |
|
|
828 | |
|
|
829 | #if ECB_GCC_VERSION(3,1) || ECB_CLANG_BUILTIN(__builtin_constant_p) |
715 | #define ecb_is_constant(expr) __builtin_constant_p (expr) |
830 | #define ecb_is_constant(expr) __builtin_constant_p (expr) |
|
|
831 | #else |
|
|
832 | /* possible C11 impl for integral types |
|
|
833 | typedef struct ecb_is_constant_struct ecb_is_constant_struct; |
|
|
834 | #define ecb_is_constant(expr) _Generic ((1 ? (struct ecb_is_constant_struct *)0 : (void *)((expr) - (expr)), ecb_is_constant_struct *: 0, default: 1)) */ |
|
|
835 | |
|
|
836 | #define ecb_is_constant(expr) 0 |
|
|
837 | #endif |
|
|
838 | |
|
|
839 | #if ECB_GCC_VERSION(3,1) || ECB_CLANG_BUILTIN(__builtin_expect) |
716 | #define ecb_expect(expr,value) __builtin_expect ((expr),(value)) |
840 | #define ecb_expect(expr,value) __builtin_expect ((expr),(value)) |
|
|
841 | #else |
|
|
842 | #define ecb_expect(expr,value) (expr) |
|
|
843 | #endif |
|
|
844 | |
|
|
845 | #if ECB_GCC_VERSION(3,1) || ECB_CLANG_BUILTIN(__builtin_prefetch) |
717 | #define ecb_prefetch(addr,rw,locality) __builtin_prefetch (addr, rw, locality) |
846 | #define ecb_prefetch(addr,rw,locality) __builtin_prefetch (addr, rw, locality) |
718 | #else |
847 | #else |
719 | #define ecb_attribute(attrlist) |
|
|
720 | #define ecb_is_constant(expr) 0 |
|
|
721 | #define ecb_expect(expr,value) (expr) |
|
|
722 | #define ecb_prefetch(addr,rw,locality) |
848 | #define ecb_prefetch(addr,rw,locality) |
723 | #endif |
849 | #endif |
724 | |
850 | |
725 | /* no emulation for ecb_decltype */ |
851 | /* no emulation for ecb_decltype */ |
726 | #if ECB_GCC_VERSION(4,5) |
852 | #if ECB_CPP11 |
|
|
853 | // older implementations might have problems with decltype(x)::type, work around it |
|
|
854 | template<class T> struct ecb_decltype_t { typedef T type; }; |
727 | #define ecb_decltype(x) __decltype(x) |
855 | #define ecb_decltype(x) ecb_decltype_t<decltype (x)>::type |
728 | #elif ECB_GCC_VERSION(3,0) |
856 | #elif ECB_GCC_VERSION(3,0) || ECB_CLANG_VERSION(2,8) |
729 | #define ecb_decltype(x) __typeof(x) |
857 | #define ecb_decltype(x) __typeof__ (x) |
730 | #endif |
858 | #endif |
731 | |
859 | |
|
|
860 | #if _MSC_VER >= 1300 |
|
|
861 | #define ecb_deprecated __declspec (deprecated) |
|
|
862 | #else |
|
|
863 | #define ecb_deprecated ecb_attribute ((__deprecated__)) |
|
|
864 | #endif |
|
|
865 | |
|
|
866 | #if _MSC_VER >= 1500 |
|
|
867 | #define ecb_deprecated_message(msg) __declspec (deprecated (msg)) |
|
|
868 | #elif ECB_GCC_VERSION(4,5) |
|
|
869 | #define ecb_deprecated_message(msg) ecb_attribute ((__deprecated__ (msg)) |
|
|
870 | #else |
|
|
871 | #define ecb_deprecated_message(msg) ecb_deprecated |
|
|
872 | #endif |
|
|
873 | |
|
|
874 | #if _MSC_VER >= 1400 |
|
|
875 | #define ecb_noinline __declspec (noinline) |
|
|
876 | #else |
732 | #define ecb_noinline ecb_attribute ((__noinline__)) |
877 | #define ecb_noinline ecb_attribute ((__noinline__)) |
|
|
878 | #endif |
|
|
879 | |
733 | #define ecb_unused ecb_attribute ((__unused__)) |
880 | #define ecb_unused ecb_attribute ((__unused__)) |
734 | #define ecb_const ecb_attribute ((__const__)) |
881 | #define ecb_const ecb_attribute ((__const__)) |
735 | #define ecb_pure ecb_attribute ((__pure__)) |
882 | #define ecb_pure ecb_attribute ((__pure__)) |
736 | |
883 | |
737 | #if ECB_C11 |
884 | #if ECB_C11 || __IBMC_NORETURN |
|
|
885 | /* http://www-01.ibm.com/support/knowledgecenter/SSGH3R_13.1.0/com.ibm.xlcpp131.aix.doc/language_ref/noreturn.html */ |
738 | #define ecb_noreturn _Noreturn |
886 | #define ecb_noreturn _Noreturn |
|
|
887 | #elif ECB_CPP11 |
|
|
888 | #define ecb_noreturn [[noreturn]] |
|
|
889 | #elif _MSC_VER >= 1200 |
|
|
890 | /* http://msdn.microsoft.com/en-us/library/k6ktzx3s.aspx */ |
|
|
891 | #define ecb_noreturn __declspec (noreturn) |
739 | #else |
892 | #else |
740 | #define ecb_noreturn ecb_attribute ((__noreturn__)) |
893 | #define ecb_noreturn ecb_attribute ((__noreturn__)) |
741 | #endif |
894 | #endif |
742 | |
895 | |
743 | #if ECB_GCC_VERSION(4,3) |
896 | #if ECB_GCC_VERSION(4,3) |
… | |
… | |
758 | /* for compatibility to the rest of the world */ |
911 | /* for compatibility to the rest of the world */ |
759 | #define ecb_likely(expr) ecb_expect_true (expr) |
912 | #define ecb_likely(expr) ecb_expect_true (expr) |
760 | #define ecb_unlikely(expr) ecb_expect_false (expr) |
913 | #define ecb_unlikely(expr) ecb_expect_false (expr) |
761 | |
914 | |
762 | /* count trailing zero bits and count # of one bits */ |
915 | /* count trailing zero bits and count # of one bits */ |
763 | #if ECB_GCC_VERSION(3,4) |
916 | #if ECB_GCC_VERSION(3,4) \ |
|
|
917 | || (ECB_CLANG_BUILTIN(__builtin_clz) && ECB_CLANG_BUILTIN(__builtin_clzll) \ |
|
|
918 | && ECB_CLANG_BUILTIN(__builtin_ctz) && ECB_CLANG_BUILTIN(__builtin_ctzll) \ |
|
|
919 | && ECB_CLANG_BUILTIN(__builtin_popcount)) |
764 | /* we assume int == 32 bit, long == 32 or 64 bit and long long == 64 bit */ |
920 | /* we assume int == 32 bit, long == 32 or 64 bit and long long == 64 bit */ |
765 | #define ecb_ld32(x) (__builtin_clz (x) ^ 31) |
921 | #define ecb_ld32(x) (__builtin_clz (x) ^ 31) |
766 | #define ecb_ld64(x) (__builtin_clzll (x) ^ 63) |
922 | #define ecb_ld64(x) (__builtin_clzll (x) ^ 63) |
767 | #define ecb_ctz32(x) __builtin_ctz (x) |
923 | #define ecb_ctz32(x) __builtin_ctz (x) |
768 | #define ecb_ctz64(x) __builtin_ctzll (x) |
924 | #define ecb_ctz64(x) __builtin_ctzll (x) |
769 | #define ecb_popcount32(x) __builtin_popcount (x) |
925 | #define ecb_popcount32(x) __builtin_popcount (x) |
770 | /* no popcountll */ |
926 | /* no popcountll */ |
771 | #else |
927 | #else |
772 | ecb_function_ int ecb_ctz32 (uint32_t x) ecb_const; |
928 | ecb_function_ ecb_const int ecb_ctz32 (uint32_t x); |
773 | ecb_function_ int |
929 | ecb_function_ ecb_const int |
774 | ecb_ctz32 (uint32_t x) |
930 | ecb_ctz32 (uint32_t x) |
775 | { |
931 | { |
|
|
932 | #if 1400 <= _MSC_VER && (_M_IX86 || _M_X64 || _M_IA64 || _M_ARM) |
|
|
933 | unsigned long r; |
|
|
934 | _BitScanForward (&r, x); |
|
|
935 | return (int)r; |
|
|
936 | #else |
776 | int r = 0; |
937 | int r = 0; |
777 | |
938 | |
778 | x &= ~x + 1; /* this isolates the lowest bit */ |
939 | x &= ~x + 1; /* this isolates the lowest bit */ |
779 | |
940 | |
780 | #if ECB_branchless_on_i386 |
941 | #if ECB_branchless_on_i386 |
… | |
… | |
790 | if (x & 0xff00ff00) r += 8; |
951 | if (x & 0xff00ff00) r += 8; |
791 | if (x & 0xffff0000) r += 16; |
952 | if (x & 0xffff0000) r += 16; |
792 | #endif |
953 | #endif |
793 | |
954 | |
794 | return r; |
955 | return r; |
|
|
956 | #endif |
795 | } |
957 | } |
796 | |
958 | |
797 | ecb_function_ int ecb_ctz64 (uint64_t x) ecb_const; |
959 | ecb_function_ ecb_const int ecb_ctz64 (uint64_t x); |
798 | ecb_function_ int |
960 | ecb_function_ ecb_const int |
799 | ecb_ctz64 (uint64_t x) |
961 | ecb_ctz64 (uint64_t x) |
800 | { |
962 | { |
|
|
963 | #if 1400 <= _MSC_VER && (_M_X64 || _M_IA64 || _M_ARM) |
|
|
964 | unsigned long r; |
|
|
965 | _BitScanForward64 (&r, x); |
|
|
966 | return (int)r; |
|
|
967 | #else |
801 | int shift = x & 0xffffffffU ? 0 : 32; |
968 | int shift = x & 0xffffffff ? 0 : 32; |
802 | return ecb_ctz32 (x >> shift) + shift; |
969 | return ecb_ctz32 (x >> shift) + shift; |
|
|
970 | #endif |
803 | } |
971 | } |
804 | |
972 | |
805 | ecb_function_ int ecb_popcount32 (uint32_t x) ecb_const; |
973 | ecb_function_ ecb_const int ecb_popcount32 (uint32_t x); |
806 | ecb_function_ int |
974 | ecb_function_ ecb_const int |
807 | ecb_popcount32 (uint32_t x) |
975 | ecb_popcount32 (uint32_t x) |
808 | { |
976 | { |
809 | x -= (x >> 1) & 0x55555555; |
977 | x -= (x >> 1) & 0x55555555; |
810 | x = ((x >> 2) & 0x33333333) + (x & 0x33333333); |
978 | x = ((x >> 2) & 0x33333333) + (x & 0x33333333); |
811 | x = ((x >> 4) + x) & 0x0f0f0f0f; |
979 | x = ((x >> 4) + x) & 0x0f0f0f0f; |
812 | x *= 0x01010101; |
980 | x *= 0x01010101; |
813 | |
981 | |
814 | return x >> 24; |
982 | return x >> 24; |
815 | } |
983 | } |
816 | |
984 | |
817 | ecb_function_ int ecb_ld32 (uint32_t x) ecb_const; |
985 | ecb_function_ ecb_const int ecb_ld32 (uint32_t x); |
818 | ecb_function_ int ecb_ld32 (uint32_t x) |
986 | ecb_function_ ecb_const int ecb_ld32 (uint32_t x) |
819 | { |
987 | { |
|
|
988 | #if 1400 <= _MSC_VER && (_M_IX86 || _M_X64 || _M_IA64 || _M_ARM) |
|
|
989 | unsigned long r; |
|
|
990 | _BitScanReverse (&r, x); |
|
|
991 | return (int)r; |
|
|
992 | #else |
820 | int r = 0; |
993 | int r = 0; |
821 | |
994 | |
822 | if (x >> 16) { x >>= 16; r += 16; } |
995 | if (x >> 16) { x >>= 16; r += 16; } |
823 | if (x >> 8) { x >>= 8; r += 8; } |
996 | if (x >> 8) { x >>= 8; r += 8; } |
824 | if (x >> 4) { x >>= 4; r += 4; } |
997 | if (x >> 4) { x >>= 4; r += 4; } |
825 | if (x >> 2) { x >>= 2; r += 2; } |
998 | if (x >> 2) { x >>= 2; r += 2; } |
826 | if (x >> 1) { r += 1; } |
999 | if (x >> 1) { r += 1; } |
827 | |
1000 | |
828 | return r; |
1001 | return r; |
|
|
1002 | #endif |
829 | } |
1003 | } |
830 | |
1004 | |
831 | ecb_function_ int ecb_ld64 (uint64_t x) ecb_const; |
1005 | ecb_function_ ecb_const int ecb_ld64 (uint64_t x); |
832 | ecb_function_ int ecb_ld64 (uint64_t x) |
1006 | ecb_function_ ecb_const int ecb_ld64 (uint64_t x) |
833 | { |
1007 | { |
|
|
1008 | #if 1400 <= _MSC_VER && (_M_X64 || _M_IA64 || _M_ARM) |
|
|
1009 | unsigned long r; |
|
|
1010 | _BitScanReverse64 (&r, x); |
|
|
1011 | return (int)r; |
|
|
1012 | #else |
834 | int r = 0; |
1013 | int r = 0; |
835 | |
1014 | |
836 | if (x >> 32) { x >>= 32; r += 32; } |
1015 | if (x >> 32) { x >>= 32; r += 32; } |
837 | |
1016 | |
838 | return r + ecb_ld32 (x); |
1017 | return r + ecb_ld32 (x); |
|
|
1018 | #endif |
839 | } |
1019 | } |
840 | #endif |
1020 | #endif |
841 | |
1021 | |
842 | ecb_function_ ecb_bool ecb_is_pot32 (uint32_t x) ecb_const; |
1022 | ecb_function_ ecb_const ecb_bool ecb_is_pot32 (uint32_t x); |
843 | ecb_function_ ecb_bool ecb_is_pot32 (uint32_t x) { return !(x & (x - 1)); } |
1023 | ecb_function_ ecb_const ecb_bool ecb_is_pot32 (uint32_t x) { return !(x & (x - 1)); } |
844 | ecb_function_ ecb_bool ecb_is_pot64 (uint64_t x) ecb_const; |
1024 | ecb_function_ ecb_const ecb_bool ecb_is_pot64 (uint64_t x); |
845 | ecb_function_ ecb_bool ecb_is_pot64 (uint64_t x) { return !(x & (x - 1)); } |
1025 | ecb_function_ ecb_const ecb_bool ecb_is_pot64 (uint64_t x) { return !(x & (x - 1)); } |
846 | |
1026 | |
847 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) ecb_const; |
1027 | ecb_function_ ecb_const uint8_t ecb_bitrev8 (uint8_t x); |
848 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) |
1028 | ecb_function_ ecb_const uint8_t ecb_bitrev8 (uint8_t x) |
849 | { |
1029 | { |
850 | return ( (x * 0x0802U & 0x22110U) |
1030 | return ( (x * 0x0802U & 0x22110U) |
851 | | (x * 0x8020U & 0x88440U)) * 0x10101U >> 16; |
1031 | | (x * 0x8020U & 0x88440U)) * 0x10101U >> 16; |
852 | } |
1032 | } |
853 | |
1033 | |
854 | ecb_function_ uint16_t ecb_bitrev16 (uint16_t x) ecb_const; |
1034 | ecb_function_ ecb_const uint16_t ecb_bitrev16 (uint16_t x); |
855 | ecb_function_ uint16_t ecb_bitrev16 (uint16_t x) |
1035 | ecb_function_ ecb_const uint16_t ecb_bitrev16 (uint16_t x) |
856 | { |
1036 | { |
857 | x = ((x >> 1) & 0x5555) | ((x & 0x5555) << 1); |
1037 | x = ((x >> 1) & 0x5555) | ((x & 0x5555) << 1); |
858 | x = ((x >> 2) & 0x3333) | ((x & 0x3333) << 2); |
1038 | x = ((x >> 2) & 0x3333) | ((x & 0x3333) << 2); |
859 | x = ((x >> 4) & 0x0f0f) | ((x & 0x0f0f) << 4); |
1039 | x = ((x >> 4) & 0x0f0f) | ((x & 0x0f0f) << 4); |
860 | x = ( x >> 8 ) | ( x << 8); |
1040 | x = ( x >> 8 ) | ( x << 8); |
861 | |
1041 | |
862 | return x; |
1042 | return x; |
863 | } |
1043 | } |
864 | |
1044 | |
865 | ecb_function_ uint32_t ecb_bitrev32 (uint32_t x) ecb_const; |
1045 | ecb_function_ ecb_const uint32_t ecb_bitrev32 (uint32_t x); |
866 | ecb_function_ uint32_t ecb_bitrev32 (uint32_t x) |
1046 | ecb_function_ ecb_const uint32_t ecb_bitrev32 (uint32_t x) |
867 | { |
1047 | { |
868 | x = ((x >> 1) & 0x55555555) | ((x & 0x55555555) << 1); |
1048 | x = ((x >> 1) & 0x55555555) | ((x & 0x55555555) << 1); |
869 | x = ((x >> 2) & 0x33333333) | ((x & 0x33333333) << 2); |
1049 | x = ((x >> 2) & 0x33333333) | ((x & 0x33333333) << 2); |
870 | x = ((x >> 4) & 0x0f0f0f0f) | ((x & 0x0f0f0f0f) << 4); |
1050 | x = ((x >> 4) & 0x0f0f0f0f) | ((x & 0x0f0f0f0f) << 4); |
871 | x = ((x >> 8) & 0x00ff00ff) | ((x & 0x00ff00ff) << 8); |
1051 | x = ((x >> 8) & 0x00ff00ff) | ((x & 0x00ff00ff) << 8); |
… | |
… | |
874 | return x; |
1054 | return x; |
875 | } |
1055 | } |
876 | |
1056 | |
877 | /* popcount64 is only available on 64 bit cpus as gcc builtin */ |
1057 | /* popcount64 is only available on 64 bit cpus as gcc builtin */ |
878 | /* so for this version we are lazy */ |
1058 | /* so for this version we are lazy */ |
879 | ecb_function_ int ecb_popcount64 (uint64_t x) ecb_const; |
1059 | ecb_function_ ecb_const int ecb_popcount64 (uint64_t x); |
880 | ecb_function_ int |
1060 | ecb_function_ ecb_const int |
881 | ecb_popcount64 (uint64_t x) |
1061 | ecb_popcount64 (uint64_t x) |
882 | { |
1062 | { |
883 | return ecb_popcount32 (x) + ecb_popcount32 (x >> 32); |
1063 | return ecb_popcount32 (x) + ecb_popcount32 (x >> 32); |
884 | } |
1064 | } |
885 | |
1065 | |
886 | ecb_inline uint8_t ecb_rotl8 (uint8_t x, unsigned int count) ecb_const; |
1066 | ecb_inline ecb_const uint8_t ecb_rotl8 (uint8_t x, unsigned int count); |
887 | ecb_inline uint8_t ecb_rotr8 (uint8_t x, unsigned int count) ecb_const; |
1067 | ecb_inline ecb_const uint8_t ecb_rotr8 (uint8_t x, unsigned int count); |
888 | ecb_inline uint16_t ecb_rotl16 (uint16_t x, unsigned int count) ecb_const; |
1068 | ecb_inline ecb_const uint16_t ecb_rotl16 (uint16_t x, unsigned int count); |
889 | ecb_inline uint16_t ecb_rotr16 (uint16_t x, unsigned int count) ecb_const; |
1069 | ecb_inline ecb_const uint16_t ecb_rotr16 (uint16_t x, unsigned int count); |
890 | ecb_inline uint32_t ecb_rotl32 (uint32_t x, unsigned int count) ecb_const; |
1070 | ecb_inline ecb_const uint32_t ecb_rotl32 (uint32_t x, unsigned int count); |
891 | ecb_inline uint32_t ecb_rotr32 (uint32_t x, unsigned int count) ecb_const; |
1071 | ecb_inline ecb_const uint32_t ecb_rotr32 (uint32_t x, unsigned int count); |
892 | ecb_inline uint64_t ecb_rotl64 (uint64_t x, unsigned int count) ecb_const; |
1072 | ecb_inline ecb_const uint64_t ecb_rotl64 (uint64_t x, unsigned int count); |
893 | ecb_inline uint64_t ecb_rotr64 (uint64_t x, unsigned int count) ecb_const; |
1073 | ecb_inline ecb_const uint64_t ecb_rotr64 (uint64_t x, unsigned int count); |
894 | |
1074 | |
895 | ecb_inline uint8_t ecb_rotl8 (uint8_t x, unsigned int count) { return (x >> ( 8 - count)) | (x << count); } |
1075 | ecb_inline ecb_const uint8_t ecb_rotl8 (uint8_t x, unsigned int count) { return (x >> ( 8 - count)) | (x << count); } |
896 | ecb_inline uint8_t ecb_rotr8 (uint8_t x, unsigned int count) { return (x << ( 8 - count)) | (x >> count); } |
1076 | ecb_inline ecb_const uint8_t ecb_rotr8 (uint8_t x, unsigned int count) { return (x << ( 8 - count)) | (x >> count); } |
897 | ecb_inline uint16_t ecb_rotl16 (uint16_t x, unsigned int count) { return (x >> (16 - count)) | (x << count); } |
1077 | ecb_inline ecb_const uint16_t ecb_rotl16 (uint16_t x, unsigned int count) { return (x >> (16 - count)) | (x << count); } |
898 | ecb_inline uint16_t ecb_rotr16 (uint16_t x, unsigned int count) { return (x << (16 - count)) | (x >> count); } |
1078 | ecb_inline ecb_const uint16_t ecb_rotr16 (uint16_t x, unsigned int count) { return (x << (16 - count)) | (x >> count); } |
899 | ecb_inline uint32_t ecb_rotl32 (uint32_t x, unsigned int count) { return (x >> (32 - count)) | (x << count); } |
1079 | ecb_inline ecb_const uint32_t ecb_rotl32 (uint32_t x, unsigned int count) { return (x >> (32 - count)) | (x << count); } |
900 | ecb_inline uint32_t ecb_rotr32 (uint32_t x, unsigned int count) { return (x << (32 - count)) | (x >> count); } |
1080 | ecb_inline ecb_const uint32_t ecb_rotr32 (uint32_t x, unsigned int count) { return (x << (32 - count)) | (x >> count); } |
901 | ecb_inline uint64_t ecb_rotl64 (uint64_t x, unsigned int count) { return (x >> (64 - count)) | (x << count); } |
1081 | ecb_inline ecb_const uint64_t ecb_rotl64 (uint64_t x, unsigned int count) { return (x >> (64 - count)) | (x << count); } |
902 | ecb_inline uint64_t ecb_rotr64 (uint64_t x, unsigned int count) { return (x << (64 - count)) | (x >> count); } |
1082 | ecb_inline ecb_const uint64_t ecb_rotr64 (uint64_t x, unsigned int count) { return (x << (64 - count)) | (x >> count); } |
903 | |
1083 | |
904 | #if ECB_GCC_VERSION(4,3) |
1084 | #if ECB_GCC_VERSION(4,3) || (ECB_CLANG_BUILTIN(__builtin_bswap32) && ECB_CLANG_BUILTIN(__builtin_bswap64)) |
|
|
1085 | #if ECB_GCC_VERSION(4,8) || ECB_CLANG_BUILTIN(__builtin_bswap16) |
|
|
1086 | #define ecb_bswap16(x) __builtin_bswap16 (x) |
|
|
1087 | #else |
905 | #define ecb_bswap16(x) (__builtin_bswap32 (x) >> 16) |
1088 | #define ecb_bswap16(x) (__builtin_bswap32 (x) >> 16) |
|
|
1089 | #endif |
906 | #define ecb_bswap32(x) __builtin_bswap32 (x) |
1090 | #define ecb_bswap32(x) __builtin_bswap32 (x) |
907 | #define ecb_bswap64(x) __builtin_bswap64 (x) |
1091 | #define ecb_bswap64(x) __builtin_bswap64 (x) |
|
|
1092 | #elif _MSC_VER |
|
|
1093 | #include <stdlib.h> |
|
|
1094 | #define ecb_bswap16(x) ((uint16_t)_byteswap_ushort ((uint16_t)(x))) |
|
|
1095 | #define ecb_bswap32(x) ((uint32_t)_byteswap_ulong ((uint32_t)(x))) |
|
|
1096 | #define ecb_bswap64(x) ((uint64_t)_byteswap_uint64 ((uint64_t)(x))) |
908 | #else |
1097 | #else |
909 | ecb_function_ uint16_t ecb_bswap16 (uint16_t x) ecb_const; |
1098 | ecb_function_ ecb_const uint16_t ecb_bswap16 (uint16_t x); |
910 | ecb_function_ uint16_t |
1099 | ecb_function_ ecb_const uint16_t |
911 | ecb_bswap16 (uint16_t x) |
1100 | ecb_bswap16 (uint16_t x) |
912 | { |
1101 | { |
913 | return ecb_rotl16 (x, 8); |
1102 | return ecb_rotl16 (x, 8); |
914 | } |
1103 | } |
915 | |
1104 | |
916 | ecb_function_ uint32_t ecb_bswap32 (uint32_t x) ecb_const; |
1105 | ecb_function_ ecb_const uint32_t ecb_bswap32 (uint32_t x); |
917 | ecb_function_ uint32_t |
1106 | ecb_function_ ecb_const uint32_t |
918 | ecb_bswap32 (uint32_t x) |
1107 | ecb_bswap32 (uint32_t x) |
919 | { |
1108 | { |
920 | return (((uint32_t)ecb_bswap16 (x)) << 16) | ecb_bswap16 (x >> 16); |
1109 | return (((uint32_t)ecb_bswap16 (x)) << 16) | ecb_bswap16 (x >> 16); |
921 | } |
1110 | } |
922 | |
1111 | |
923 | ecb_function_ uint64_t ecb_bswap64 (uint64_t x) ecb_const; |
1112 | ecb_function_ ecb_const uint64_t ecb_bswap64 (uint64_t x); |
924 | ecb_function_ uint64_t |
1113 | ecb_function_ ecb_const uint64_t |
925 | ecb_bswap64 (uint64_t x) |
1114 | ecb_bswap64 (uint64_t x) |
926 | { |
1115 | { |
927 | return (((uint64_t)ecb_bswap32 (x)) << 32) | ecb_bswap32 (x >> 32); |
1116 | return (((uint64_t)ecb_bswap32 (x)) << 32) | ecb_bswap32 (x >> 32); |
928 | } |
1117 | } |
929 | #endif |
1118 | #endif |
930 | |
1119 | |
931 | #if ECB_GCC_VERSION(4,5) |
1120 | #if ECB_GCC_VERSION(4,5) || ECB_CLANG_BUILTIN(__builtin_unreachable) |
932 | #define ecb_unreachable() __builtin_unreachable () |
1121 | #define ecb_unreachable() __builtin_unreachable () |
933 | #else |
1122 | #else |
934 | /* this seems to work fine, but gcc always emits a warning for it :/ */ |
1123 | /* this seems to work fine, but gcc always emits a warning for it :/ */ |
935 | ecb_inline void ecb_unreachable (void) ecb_noreturn; |
1124 | ecb_inline ecb_noreturn void ecb_unreachable (void); |
936 | ecb_inline void ecb_unreachable (void) { } |
1125 | ecb_inline ecb_noreturn void ecb_unreachable (void) { } |
937 | #endif |
1126 | #endif |
938 | |
1127 | |
939 | /* try to tell the compiler that some condition is definitely true */ |
1128 | /* try to tell the compiler that some condition is definitely true */ |
940 | #define ecb_assume(cond) do { if (!(cond)) ecb_unreachable (); } while (0) |
1129 | #define ecb_assume(cond) if (!(cond)) ecb_unreachable (); else 0 |
941 | |
1130 | |
942 | ecb_inline unsigned char ecb_byteorder_helper (void) ecb_const; |
1131 | ecb_inline ecb_const uint32_t ecb_byteorder_helper (void); |
943 | ecb_inline unsigned char |
1132 | ecb_inline ecb_const uint32_t |
944 | ecb_byteorder_helper (void) |
1133 | ecb_byteorder_helper (void) |
945 | { |
1134 | { |
946 | const uint32_t u = 0x11223344; |
1135 | /* the union code still generates code under pressure in gcc, */ |
947 | return *(unsigned char *)&u; |
1136 | /* but less than using pointers, and always seems to */ |
|
|
1137 | /* successfully return a constant. */ |
|
|
1138 | /* the reason why we have this horrible preprocessor mess */ |
|
|
1139 | /* is to avoid it in all cases, at least on common architectures */ |
|
|
1140 | /* or when using a recent enough gcc version (>= 4.6) */ |
|
|
1141 | #if (defined __BYTE_ORDER__ && __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__) \ |
|
|
1142 | || ((__i386 || __i386__ || _M_IX86 || ECB_GCC_AMD64 || ECB_MSVC_AMD64) && !__VOS__) |
|
|
1143 | #define ECB_LITTLE_ENDIAN 1 |
|
|
1144 | return 0x44332211; |
|
|
1145 | #elif (defined __BYTE_ORDER__ && __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__) \ |
|
|
1146 | || ((__AARCH64EB__ || __MIPSEB__ || __ARMEB__) && !__VOS__) |
|
|
1147 | #define ECB_BIG_ENDIAN 1 |
|
|
1148 | return 0x11223344; |
|
|
1149 | #else |
|
|
1150 | union |
|
|
1151 | { |
|
|
1152 | uint8_t c[4]; |
|
|
1153 | uint32_t u; |
|
|
1154 | } u = { 0x11, 0x22, 0x33, 0x44 }; |
|
|
1155 | return u.u; |
|
|
1156 | #endif |
948 | } |
1157 | } |
949 | |
1158 | |
950 | ecb_inline ecb_bool ecb_big_endian (void) ecb_const; |
1159 | ecb_inline ecb_const ecb_bool ecb_big_endian (void); |
951 | ecb_inline ecb_bool ecb_big_endian (void) { return ecb_byteorder_helper () == 0x11; } |
1160 | ecb_inline ecb_const ecb_bool ecb_big_endian (void) { return ecb_byteorder_helper () == 0x11223344; } |
952 | ecb_inline ecb_bool ecb_little_endian (void) ecb_const; |
1161 | ecb_inline ecb_const ecb_bool ecb_little_endian (void); |
953 | ecb_inline ecb_bool ecb_little_endian (void) { return ecb_byteorder_helper () == 0x44; } |
1162 | ecb_inline ecb_const ecb_bool ecb_little_endian (void) { return ecb_byteorder_helper () == 0x44332211; } |
954 | |
1163 | |
955 | #if ECB_GCC_VERSION(3,0) || ECB_C99 |
1164 | #if ECB_GCC_VERSION(3,0) || ECB_C99 |
956 | #define ecb_mod(m,n) ((m) % (n) + ((m) % (n) < 0 ? (n) : 0)) |
1165 | #define ecb_mod(m,n) ((m) % (n) + ((m) % (n) < 0 ? (n) : 0)) |
957 | #else |
1166 | #else |
958 | #define ecb_mod(m,n) ((m) < 0 ? ((n) - 1 - ((-1 - (m)) % (n))) : ((m) % (n))) |
1167 | #define ecb_mod(m,n) ((m) < 0 ? ((n) - 1 - ((-1 - (m)) % (n))) : ((m) % (n))) |
959 | #endif |
1168 | #endif |
960 | |
1169 | |
961 | #if __cplusplus |
1170 | #if ECB_CPP |
962 | template<typename T> |
1171 | template<typename T> |
963 | static inline T ecb_div_rd (T val, T div) |
1172 | static inline T ecb_div_rd (T val, T div) |
964 | { |
1173 | { |
965 | return val < 0 ? - ((-val + div - 1) / div) : (val ) / div; |
1174 | return val < 0 ? - ((-val + div - 1) / div) : (val ) / div; |
966 | } |
1175 | } |
… | |
… | |
983 | } |
1192 | } |
984 | #else |
1193 | #else |
985 | #define ecb_array_length(name) (sizeof (name) / sizeof (name [0])) |
1194 | #define ecb_array_length(name) (sizeof (name) / sizeof (name [0])) |
986 | #endif |
1195 | #endif |
987 | |
1196 | |
|
|
1197 | ecb_function_ ecb_const uint32_t ecb_binary16_to_binary32 (uint32_t x); |
|
|
1198 | ecb_function_ ecb_const uint32_t |
|
|
1199 | ecb_binary16_to_binary32 (uint32_t x) |
|
|
1200 | { |
|
|
1201 | unsigned int s = (x & 0x8000) << (31 - 15); |
|
|
1202 | int e = (x >> 10) & 0x001f; |
|
|
1203 | unsigned int m = x & 0x03ff; |
|
|
1204 | |
|
|
1205 | if (ecb_expect_false (e == 31)) |
|
|
1206 | /* infinity or NaN */ |
|
|
1207 | e = 255 - (127 - 15); |
|
|
1208 | else if (ecb_expect_false (!e)) |
|
|
1209 | { |
|
|
1210 | if (ecb_expect_true (!m)) |
|
|
1211 | /* zero, handled by code below by forcing e to 0 */ |
|
|
1212 | e = 0 - (127 - 15); |
|
|
1213 | else |
|
|
1214 | { |
|
|
1215 | /* subnormal, renormalise */ |
|
|
1216 | unsigned int s = 10 - ecb_ld32 (m); |
|
|
1217 | |
|
|
1218 | m = (m << s) & 0x3ff; /* mask implicit bit */ |
|
|
1219 | e -= s - 1; |
|
|
1220 | } |
|
|
1221 | } |
|
|
1222 | |
|
|
1223 | /* e and m now are normalised, or zero, (or inf or nan) */ |
|
|
1224 | e += 127 - 15; |
|
|
1225 | |
|
|
1226 | return s | (e << 23) | (m << (23 - 10)); |
|
|
1227 | } |
|
|
1228 | |
|
|
1229 | ecb_function_ ecb_const uint16_t ecb_binary32_to_binary16 (uint32_t x); |
|
|
1230 | ecb_function_ ecb_const uint16_t |
|
|
1231 | ecb_binary32_to_binary16 (uint32_t x) |
|
|
1232 | { |
|
|
1233 | unsigned int s = (x >> 16) & 0x00008000; /* sign bit, the easy part */ |
|
|
1234 | unsigned int e = ((x >> 23) & 0x000000ff) - (127 - 15); /* the desired exponent */ |
|
|
1235 | unsigned int m = x & 0x007fffff; |
|
|
1236 | |
|
|
1237 | x &= 0x7fffffff; |
|
|
1238 | |
|
|
1239 | /* if it's within range of binary16 normals, use fast path */ |
|
|
1240 | if (ecb_expect_true (0x38800000 <= x && x <= 0x477fefff)) |
|
|
1241 | { |
|
|
1242 | /* mantissa round-to-even */ |
|
|
1243 | m += 0x00000fff + ((m >> (23 - 10)) & 1); |
|
|
1244 | |
|
|
1245 | /* handle overflow */ |
|
|
1246 | if (ecb_expect_false (m >= 0x00800000)) |
|
|
1247 | { |
|
|
1248 | m >>= 1; |
|
|
1249 | e += 1; |
|
|
1250 | } |
|
|
1251 | |
|
|
1252 | return s | (e << 10) | (m >> (23 - 10)); |
|
|
1253 | } |
|
|
1254 | |
|
|
1255 | /* handle large numbers and infinity */ |
|
|
1256 | if (ecb_expect_true (0x477fefff < x && x <= 0x7f800000)) |
|
|
1257 | return s | 0x7c00; |
|
|
1258 | |
|
|
1259 | /* handle zero, subnormals and small numbers */ |
|
|
1260 | if (ecb_expect_true (x < 0x38800000)) |
|
|
1261 | { |
|
|
1262 | /* zero */ |
|
|
1263 | if (ecb_expect_true (!x)) |
|
|
1264 | return s; |
|
|
1265 | |
|
|
1266 | /* handle subnormals */ |
|
|
1267 | |
|
|
1268 | /* too small, will be zero */ |
|
|
1269 | if (e < (14 - 24)) /* might not be sharp, but is good enough */ |
|
|
1270 | return s; |
|
|
1271 | |
|
|
1272 | m |= 0x00800000; /* make implicit bit explicit */ |
|
|
1273 | |
|
|
1274 | /* very tricky - we need to round to the nearest e (+10) bit value */ |
|
|
1275 | { |
|
|
1276 | unsigned int bits = 14 - e; |
|
|
1277 | unsigned int half = (1 << (bits - 1)) - 1; |
|
|
1278 | unsigned int even = (m >> bits) & 1; |
|
|
1279 | |
|
|
1280 | /* if this overflows, we will end up with a normalised number */ |
|
|
1281 | m = (m + half + even) >> bits; |
|
|
1282 | } |
|
|
1283 | |
|
|
1284 | return s | m; |
|
|
1285 | } |
|
|
1286 | |
|
|
1287 | /* handle NaNs, preserve leftmost nan bits, but make sure we don't turn them into infinities */ |
|
|
1288 | m >>= 13; |
|
|
1289 | |
|
|
1290 | return s | 0x7c00 | m | !m; |
|
|
1291 | } |
|
|
1292 | |
|
|
1293 | /*******************************************************************************/ |
|
|
1294 | /* floating point stuff, can be disabled by defining ECB_NO_LIBM */ |
|
|
1295 | |
|
|
1296 | /* basically, everything uses "ieee pure-endian" floating point numbers */ |
|
|
1297 | /* the only noteworthy exception is ancient armle, which uses order 43218765 */ |
|
|
1298 | #if 0 \ |
|
|
1299 | || __i386 || __i386__ \ |
|
|
1300 | || ECB_GCC_AMD64 \ |
|
|
1301 | || __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ \ |
|
|
1302 | || defined __s390__ || defined __s390x__ \ |
|
|
1303 | || defined __mips__ \ |
|
|
1304 | || defined __alpha__ \ |
|
|
1305 | || defined __hppa__ \ |
|
|
1306 | || defined __ia64__ \ |
|
|
1307 | || defined __m68k__ \ |
|
|
1308 | || defined __m88k__ \ |
|
|
1309 | || defined __sh__ \ |
|
|
1310 | || defined _M_IX86 || defined ECB_MSVC_AMD64 || defined _M_IA64 \ |
|
|
1311 | || (defined __arm__ && (defined __ARM_EABI__ || defined __EABI__ || defined __VFP_FP__ || defined _WIN32_WCE || defined __ANDROID__)) \ |
|
|
1312 | || defined __aarch64__ |
|
|
1313 | #define ECB_STDFP 1 |
|
|
1314 | #include <string.h> /* for memcpy */ |
|
|
1315 | #else |
|
|
1316 | #define ECB_STDFP 0 |
|
|
1317 | #endif |
|
|
1318 | |
|
|
1319 | #ifndef ECB_NO_LIBM |
|
|
1320 | |
|
|
1321 | #include <math.h> /* for frexp*, ldexp*, INFINITY, NAN */ |
|
|
1322 | |
|
|
1323 | /* only the oldest of old doesn't have this one. solaris. */ |
|
|
1324 | #ifdef INFINITY |
|
|
1325 | #define ECB_INFINITY INFINITY |
|
|
1326 | #else |
|
|
1327 | #define ECB_INFINITY HUGE_VAL |
|
|
1328 | #endif |
|
|
1329 | |
|
|
1330 | #ifdef NAN |
|
|
1331 | #define ECB_NAN NAN |
|
|
1332 | #else |
|
|
1333 | #define ECB_NAN ECB_INFINITY |
|
|
1334 | #endif |
|
|
1335 | |
|
|
1336 | #if ECB_C99 || _XOPEN_VERSION >= 600 || _POSIX_VERSION >= 200112L |
|
|
1337 | #define ecb_ldexpf(x,e) ldexpf ((x), (e)) |
|
|
1338 | #define ecb_frexpf(x,e) frexpf ((x), (e)) |
|
|
1339 | #else |
|
|
1340 | #define ecb_ldexpf(x,e) (float) ldexp ((double) (x), (e)) |
|
|
1341 | #define ecb_frexpf(x,e) (float) frexp ((double) (x), (e)) |
|
|
1342 | #endif |
|
|
1343 | |
|
|
1344 | /* convert a float to ieee single/binary32 */ |
|
|
1345 | ecb_function_ ecb_const uint32_t ecb_float_to_binary32 (float x); |
|
|
1346 | ecb_function_ ecb_const uint32_t |
|
|
1347 | ecb_float_to_binary32 (float x) |
|
|
1348 | { |
|
|
1349 | uint32_t r; |
|
|
1350 | |
|
|
1351 | #if ECB_STDFP |
|
|
1352 | memcpy (&r, &x, 4); |
|
|
1353 | #else |
|
|
1354 | /* slow emulation, works for anything but -0 */ |
|
|
1355 | uint32_t m; |
|
|
1356 | int e; |
|
|
1357 | |
|
|
1358 | if (x == 0e0f ) return 0x00000000U; |
|
|
1359 | if (x > +3.40282346638528860e+38f) return 0x7f800000U; |
|
|
1360 | if (x < -3.40282346638528860e+38f) return 0xff800000U; |
|
|
1361 | if (x != x ) return 0x7fbfffffU; |
|
|
1362 | |
|
|
1363 | m = ecb_frexpf (x, &e) * 0x1000000U; |
|
|
1364 | |
|
|
1365 | r = m & 0x80000000U; |
|
|
1366 | |
|
|
1367 | if (r) |
|
|
1368 | m = -m; |
|
|
1369 | |
|
|
1370 | if (e <= -126) |
|
|
1371 | { |
|
|
1372 | m &= 0xffffffU; |
|
|
1373 | m >>= (-125 - e); |
|
|
1374 | e = -126; |
|
|
1375 | } |
|
|
1376 | |
|
|
1377 | r |= (e + 126) << 23; |
|
|
1378 | r |= m & 0x7fffffU; |
|
|
1379 | #endif |
|
|
1380 | |
|
|
1381 | return r; |
|
|
1382 | } |
|
|
1383 | |
|
|
1384 | /* converts an ieee single/binary32 to a float */ |
|
|
1385 | ecb_function_ ecb_const float ecb_binary32_to_float (uint32_t x); |
|
|
1386 | ecb_function_ ecb_const float |
|
|
1387 | ecb_binary32_to_float (uint32_t x) |
|
|
1388 | { |
|
|
1389 | float r; |
|
|
1390 | |
|
|
1391 | #if ECB_STDFP |
|
|
1392 | memcpy (&r, &x, 4); |
|
|
1393 | #else |
|
|
1394 | /* emulation, only works for normals and subnormals and +0 */ |
|
|
1395 | int neg = x >> 31; |
|
|
1396 | int e = (x >> 23) & 0xffU; |
|
|
1397 | |
|
|
1398 | x &= 0x7fffffU; |
|
|
1399 | |
|
|
1400 | if (e) |
|
|
1401 | x |= 0x800000U; |
|
|
1402 | else |
|
|
1403 | e = 1; |
|
|
1404 | |
|
|
1405 | /* we distrust ldexpf a bit and do the 2**-24 scaling by an extra multiply */ |
|
|
1406 | r = ecb_ldexpf (x * (0.5f / 0x800000U), e - 126); |
|
|
1407 | |
|
|
1408 | r = neg ? -r : r; |
|
|
1409 | #endif |
|
|
1410 | |
|
|
1411 | return r; |
|
|
1412 | } |
|
|
1413 | |
|
|
1414 | /* convert a double to ieee double/binary64 */ |
|
|
1415 | ecb_function_ ecb_const uint64_t ecb_double_to_binary64 (double x); |
|
|
1416 | ecb_function_ ecb_const uint64_t |
|
|
1417 | ecb_double_to_binary64 (double x) |
|
|
1418 | { |
|
|
1419 | uint64_t r; |
|
|
1420 | |
|
|
1421 | #if ECB_STDFP |
|
|
1422 | memcpy (&r, &x, 8); |
|
|
1423 | #else |
|
|
1424 | /* slow emulation, works for anything but -0 */ |
|
|
1425 | uint64_t m; |
|
|
1426 | int e; |
|
|
1427 | |
|
|
1428 | if (x == 0e0 ) return 0x0000000000000000U; |
|
|
1429 | if (x > +1.79769313486231470e+308) return 0x7ff0000000000000U; |
|
|
1430 | if (x < -1.79769313486231470e+308) return 0xfff0000000000000U; |
|
|
1431 | if (x != x ) return 0X7ff7ffffffffffffU; |
|
|
1432 | |
|
|
1433 | m = frexp (x, &e) * 0x20000000000000U; |
|
|
1434 | |
|
|
1435 | r = m & 0x8000000000000000;; |
|
|
1436 | |
|
|
1437 | if (r) |
|
|
1438 | m = -m; |
|
|
1439 | |
|
|
1440 | if (e <= -1022) |
|
|
1441 | { |
|
|
1442 | m &= 0x1fffffffffffffU; |
|
|
1443 | m >>= (-1021 - e); |
|
|
1444 | e = -1022; |
|
|
1445 | } |
|
|
1446 | |
|
|
1447 | r |= ((uint64_t)(e + 1022)) << 52; |
|
|
1448 | r |= m & 0xfffffffffffffU; |
|
|
1449 | #endif |
|
|
1450 | |
|
|
1451 | return r; |
|
|
1452 | } |
|
|
1453 | |
|
|
1454 | /* converts an ieee double/binary64 to a double */ |
|
|
1455 | ecb_function_ ecb_const double ecb_binary64_to_double (uint64_t x); |
|
|
1456 | ecb_function_ ecb_const double |
|
|
1457 | ecb_binary64_to_double (uint64_t x) |
|
|
1458 | { |
|
|
1459 | double r; |
|
|
1460 | |
|
|
1461 | #if ECB_STDFP |
|
|
1462 | memcpy (&r, &x, 8); |
|
|
1463 | #else |
|
|
1464 | /* emulation, only works for normals and subnormals and +0 */ |
|
|
1465 | int neg = x >> 63; |
|
|
1466 | int e = (x >> 52) & 0x7ffU; |
|
|
1467 | |
|
|
1468 | x &= 0xfffffffffffffU; |
|
|
1469 | |
|
|
1470 | if (e) |
|
|
1471 | x |= 0x10000000000000U; |
|
|
1472 | else |
|
|
1473 | e = 1; |
|
|
1474 | |
|
|
1475 | /* we distrust ldexp a bit and do the 2**-53 scaling by an extra multiply */ |
|
|
1476 | r = ldexp (x * (0.5 / 0x10000000000000U), e - 1022); |
|
|
1477 | |
|
|
1478 | r = neg ? -r : r; |
|
|
1479 | #endif |
|
|
1480 | |
|
|
1481 | return r; |
|
|
1482 | } |
|
|
1483 | |
|
|
1484 | /* convert a float to ieee half/binary16 */ |
|
|
1485 | ecb_function_ ecb_const uint16_t ecb_float_to_binary16 (float x); |
|
|
1486 | ecb_function_ ecb_const uint16_t |
|
|
1487 | ecb_float_to_binary16 (float x) |
|
|
1488 | { |
|
|
1489 | return ecb_binary32_to_binary16 (ecb_float_to_binary32 (x)); |
|
|
1490 | } |
|
|
1491 | |
|
|
1492 | /* convert an ieee half/binary16 to float */ |
|
|
1493 | ecb_function_ ecb_const float ecb_binary16_to_float (uint16_t x); |
|
|
1494 | ecb_function_ ecb_const float |
|
|
1495 | ecb_binary16_to_float (uint16_t x) |
|
|
1496 | { |
|
|
1497 | return ecb_binary32_to_float (ecb_binary16_to_binary32 (x)); |
|
|
1498 | } |
|
|
1499 | |
|
|
1500 | #endif |
|
|
1501 | |
988 | #endif |
1502 | #endif |
989 | |
1503 | |
990 | /* ECB.H END */ |
1504 | /* ECB.H END */ |
991 | |
1505 | |
992 | #if ECB_MEMORY_FENCE_NEEDS_PTHREADS |
1506 | #if ECB_MEMORY_FENCE_NEEDS_PTHREADS |
… | |
… | |
1191 | } |
1705 | } |
1192 | |
1706 | |
1193 | static void * |
1707 | static void * |
1194 | ev_realloc_emul (void *ptr, long size) EV_THROW |
1708 | ev_realloc_emul (void *ptr, long size) EV_THROW |
1195 | { |
1709 | { |
1196 | #if __GLIBC__ |
|
|
1197 | return realloc (ptr, size); |
|
|
1198 | #else |
|
|
1199 | /* some systems, notably openbsd and darwin, fail to properly |
1710 | /* some systems, notably openbsd and darwin, fail to properly |
1200 | * implement realloc (x, 0) (as required by both ansi c-89 and |
1711 | * implement realloc (x, 0) (as required by both ansi c-89 and |
1201 | * the single unix specification, so work around them here. |
1712 | * the single unix specification, so work around them here. |
|
|
1713 | * recently, also (at least) fedora and debian started breaking it, |
|
|
1714 | * despite documenting it otherwise. |
1202 | */ |
1715 | */ |
1203 | |
1716 | |
1204 | if (size) |
1717 | if (size) |
1205 | return realloc (ptr, size); |
1718 | return realloc (ptr, size); |
1206 | |
1719 | |
1207 | free (ptr); |
1720 | free (ptr); |
1208 | return 0; |
1721 | return 0; |
1209 | #endif |
|
|
1210 | } |
1722 | } |
1211 | |
1723 | |
1212 | static void *(*alloc)(void *ptr, long size) EV_THROW = ev_realloc_emul; |
1724 | static void *(*alloc)(void *ptr, long size) EV_THROW = ev_realloc_emul; |
1213 | |
1725 | |
1214 | void ecb_cold |
1726 | void ecb_cold |
… | |
… | |
1869 | static void noinline ecb_cold |
2381 | static void noinline ecb_cold |
1870 | evpipe_init (EV_P) |
2382 | evpipe_init (EV_P) |
1871 | { |
2383 | { |
1872 | if (!ev_is_active (&pipe_w)) |
2384 | if (!ev_is_active (&pipe_w)) |
1873 | { |
2385 | { |
|
|
2386 | int fds [2]; |
|
|
2387 | |
1874 | # if EV_USE_EVENTFD |
2388 | # if EV_USE_EVENTFD |
|
|
2389 | fds [0] = -1; |
1875 | evfd = eventfd (0, EFD_NONBLOCK | EFD_CLOEXEC); |
2390 | fds [1] = eventfd (0, EFD_NONBLOCK | EFD_CLOEXEC); |
1876 | if (evfd < 0 && errno == EINVAL) |
2391 | if (fds [1] < 0 && errno == EINVAL) |
1877 | evfd = eventfd (0, 0); |
2392 | fds [1] = eventfd (0, 0); |
1878 | |
2393 | |
1879 | if (evfd >= 0) |
2394 | if (fds [1] < 0) |
1880 | { |
|
|
1881 | evpipe [0] = -1; |
|
|
1882 | fd_intern (evfd); /* doing it twice doesn't hurt */ |
|
|
1883 | ev_io_set (&pipe_w, evfd, EV_READ); |
|
|
1884 | } |
|
|
1885 | else |
|
|
1886 | # endif |
2395 | # endif |
1887 | { |
2396 | { |
1888 | while (pipe (evpipe)) |
2397 | while (pipe (fds)) |
1889 | ev_syserr ("(libev) error creating signal/async pipe"); |
2398 | ev_syserr ("(libev) error creating signal/async pipe"); |
1890 | |
2399 | |
1891 | fd_intern (evpipe [0]); |
2400 | fd_intern (fds [0]); |
1892 | fd_intern (evpipe [1]); |
|
|
1893 | ev_io_set (&pipe_w, evpipe [0], EV_READ); |
|
|
1894 | } |
2401 | } |
1895 | |
2402 | |
|
|
2403 | evpipe [0] = fds [0]; |
|
|
2404 | |
|
|
2405 | if (evpipe [1] < 0) |
|
|
2406 | evpipe [1] = fds [1]; /* first call, set write fd */ |
|
|
2407 | else |
|
|
2408 | { |
|
|
2409 | /* on subsequent calls, do not change evpipe [1] */ |
|
|
2410 | /* so that evpipe_write can always rely on its value. */ |
|
|
2411 | /* this branch does not do anything sensible on windows, */ |
|
|
2412 | /* so must not be executed on windows */ |
|
|
2413 | |
|
|
2414 | dup2 (fds [1], evpipe [1]); |
|
|
2415 | close (fds [1]); |
|
|
2416 | } |
|
|
2417 | |
|
|
2418 | fd_intern (evpipe [1]); |
|
|
2419 | |
|
|
2420 | ev_io_set (&pipe_w, evpipe [0] < 0 ? evpipe [1] : evpipe [0], EV_READ); |
1896 | ev_io_start (EV_A_ &pipe_w); |
2421 | ev_io_start (EV_A_ &pipe_w); |
1897 | ev_unref (EV_A); /* watcher should not keep loop alive */ |
2422 | ev_unref (EV_A); /* watcher should not keep loop alive */ |
1898 | } |
2423 | } |
1899 | } |
2424 | } |
1900 | |
2425 | |
… | |
… | |
1921 | ECB_MEMORY_FENCE_RELEASE; |
2446 | ECB_MEMORY_FENCE_RELEASE; |
1922 | |
2447 | |
1923 | old_errno = errno; /* save errno because write will clobber it */ |
2448 | old_errno = errno; /* save errno because write will clobber it */ |
1924 | |
2449 | |
1925 | #if EV_USE_EVENTFD |
2450 | #if EV_USE_EVENTFD |
1926 | if (evfd >= 0) |
2451 | if (evpipe [0] < 0) |
1927 | { |
2452 | { |
1928 | uint64_t counter = 1; |
2453 | uint64_t counter = 1; |
1929 | write (evfd, &counter, sizeof (uint64_t)); |
2454 | write (evpipe [1], &counter, sizeof (uint64_t)); |
1930 | } |
2455 | } |
1931 | else |
2456 | else |
1932 | #endif |
2457 | #endif |
1933 | { |
2458 | { |
1934 | #ifdef _WIN32 |
2459 | #ifdef _WIN32 |
… | |
… | |
1954 | int i; |
2479 | int i; |
1955 | |
2480 | |
1956 | if (revents & EV_READ) |
2481 | if (revents & EV_READ) |
1957 | { |
2482 | { |
1958 | #if EV_USE_EVENTFD |
2483 | #if EV_USE_EVENTFD |
1959 | if (evfd >= 0) |
2484 | if (evpipe [0] < 0) |
1960 | { |
2485 | { |
1961 | uint64_t counter; |
2486 | uint64_t counter; |
1962 | read (evfd, &counter, sizeof (uint64_t)); |
2487 | read (evpipe [1], &counter, sizeof (uint64_t)); |
1963 | } |
2488 | } |
1964 | else |
2489 | else |
1965 | #endif |
2490 | #endif |
1966 | { |
2491 | { |
1967 | char dummy[4]; |
2492 | char dummy[4]; |
… | |
… | |
2017 | |
2542 | |
2018 | void |
2543 | void |
2019 | ev_feed_signal (int signum) EV_THROW |
2544 | ev_feed_signal (int signum) EV_THROW |
2020 | { |
2545 | { |
2021 | #if EV_MULTIPLICITY |
2546 | #if EV_MULTIPLICITY |
|
|
2547 | EV_P; |
|
|
2548 | ECB_MEMORY_FENCE_ACQUIRE; |
2022 | EV_P = signals [signum - 1].loop; |
2549 | EV_A = signals [signum - 1].loop; |
2023 | |
2550 | |
2024 | if (!EV_A) |
2551 | if (!EV_A) |
2025 | return; |
2552 | return; |
2026 | #endif |
2553 | #endif |
2027 | |
2554 | |
2028 | if (!ev_active (&pipe_w)) |
|
|
2029 | return; |
|
|
2030 | |
|
|
2031 | signals [signum - 1].pending = 1; |
2555 | signals [signum - 1].pending = 1; |
2032 | evpipe_write (EV_A_ &sig_pending); |
2556 | evpipe_write (EV_A_ &sig_pending); |
2033 | } |
2557 | } |
2034 | |
2558 | |
2035 | static void |
2559 | static void |
… | |
… | |
2045 | void noinline |
2569 | void noinline |
2046 | ev_feed_signal_event (EV_P_ int signum) EV_THROW |
2570 | ev_feed_signal_event (EV_P_ int signum) EV_THROW |
2047 | { |
2571 | { |
2048 | WL w; |
2572 | WL w; |
2049 | |
2573 | |
2050 | if (expect_false (signum <= 0 || signum > EV_NSIG)) |
2574 | if (expect_false (signum <= 0 || signum >= EV_NSIG)) |
2051 | return; |
2575 | return; |
2052 | |
2576 | |
2053 | --signum; |
2577 | --signum; |
2054 | |
2578 | |
2055 | #if EV_MULTIPLICITY |
2579 | #if EV_MULTIPLICITY |
… | |
… | |
2283 | { |
2807 | { |
2284 | return userdata; |
2808 | return userdata; |
2285 | } |
2809 | } |
2286 | |
2810 | |
2287 | void |
2811 | void |
2288 | ev_set_invoke_pending_cb (EV_P_ void (*invoke_pending_cb)(EV_P)) EV_THROW |
2812 | ev_set_invoke_pending_cb (EV_P_ ev_loop_callback invoke_pending_cb) EV_THROW |
2289 | { |
2813 | { |
2290 | invoke_cb = invoke_pending_cb; |
2814 | invoke_cb = invoke_pending_cb; |
2291 | } |
2815 | } |
2292 | |
2816 | |
2293 | void |
2817 | void |
… | |
… | |
2353 | #if EV_ASYNC_ENABLE |
2877 | #if EV_ASYNC_ENABLE |
2354 | async_pending = 0; |
2878 | async_pending = 0; |
2355 | #endif |
2879 | #endif |
2356 | pipe_write_skipped = 0; |
2880 | pipe_write_skipped = 0; |
2357 | pipe_write_wanted = 0; |
2881 | pipe_write_wanted = 0; |
|
|
2882 | evpipe [0] = -1; |
|
|
2883 | evpipe [1] = -1; |
2358 | #if EV_USE_INOTIFY |
2884 | #if EV_USE_INOTIFY |
2359 | fs_fd = flags & EVFLAG_NOINOTIFY ? -1 : -2; |
2885 | fs_fd = flags & EVFLAG_NOINOTIFY ? -1 : -2; |
2360 | #endif |
2886 | #endif |
2361 | #if EV_USE_SIGNALFD |
2887 | #if EV_USE_SIGNALFD |
2362 | sigfd = flags & EVFLAG_SIGNALFD ? -2 : -1; |
2888 | sigfd = flags & EVFLAG_SIGNALFD ? -2 : -1; |
… | |
… | |
2425 | if (ev_is_active (&pipe_w)) |
2951 | if (ev_is_active (&pipe_w)) |
2426 | { |
2952 | { |
2427 | /*ev_ref (EV_A);*/ |
2953 | /*ev_ref (EV_A);*/ |
2428 | /*ev_io_stop (EV_A_ &pipe_w);*/ |
2954 | /*ev_io_stop (EV_A_ &pipe_w);*/ |
2429 | |
2955 | |
2430 | #if EV_USE_EVENTFD |
|
|
2431 | if (evfd >= 0) |
|
|
2432 | close (evfd); |
|
|
2433 | #endif |
|
|
2434 | |
|
|
2435 | if (evpipe [0] >= 0) |
|
|
2436 | { |
|
|
2437 | EV_WIN32_CLOSE_FD (evpipe [0]); |
2956 | if (evpipe [0] >= 0) EV_WIN32_CLOSE_FD (evpipe [0]); |
2438 | EV_WIN32_CLOSE_FD (evpipe [1]); |
2957 | if (evpipe [1] >= 0) EV_WIN32_CLOSE_FD (evpipe [1]); |
2439 | } |
|
|
2440 | } |
2958 | } |
2441 | |
2959 | |
2442 | #if EV_USE_SIGNALFD |
2960 | #if EV_USE_SIGNALFD |
2443 | if (ev_is_active (&sigfd_w)) |
2961 | if (ev_is_active (&sigfd_w)) |
2444 | close (sigfd); |
2962 | close (sigfd); |
… | |
… | |
2530 | #endif |
3048 | #endif |
2531 | #if EV_USE_INOTIFY |
3049 | #if EV_USE_INOTIFY |
2532 | infy_fork (EV_A); |
3050 | infy_fork (EV_A); |
2533 | #endif |
3051 | #endif |
2534 | |
3052 | |
|
|
3053 | #if EV_SIGNAL_ENABLE || EV_ASYNC_ENABLE |
2535 | if (ev_is_active (&pipe_w)) |
3054 | if (ev_is_active (&pipe_w) && postfork != 2) |
2536 | { |
3055 | { |
2537 | /* pipe_write_wanted must be false now, so modifying fd vars should be safe */ |
3056 | /* pipe_write_wanted must be false now, so modifying fd vars should be safe */ |
2538 | |
3057 | |
2539 | ev_ref (EV_A); |
3058 | ev_ref (EV_A); |
2540 | ev_io_stop (EV_A_ &pipe_w); |
3059 | ev_io_stop (EV_A_ &pipe_w); |
2541 | |
3060 | |
2542 | #if EV_USE_EVENTFD |
|
|
2543 | if (evfd >= 0) |
|
|
2544 | close (evfd); |
|
|
2545 | #endif |
|
|
2546 | |
|
|
2547 | if (evpipe [0] >= 0) |
3061 | if (evpipe [0] >= 0) |
2548 | { |
|
|
2549 | EV_WIN32_CLOSE_FD (evpipe [0]); |
3062 | EV_WIN32_CLOSE_FD (evpipe [0]); |
2550 | EV_WIN32_CLOSE_FD (evpipe [1]); |
|
|
2551 | } |
|
|
2552 | |
3063 | |
2553 | #if EV_SIGNAL_ENABLE || EV_ASYNC_ENABLE |
|
|
2554 | evpipe_init (EV_A); |
3064 | evpipe_init (EV_A); |
2555 | /* now iterate over everything, in case we missed something */ |
3065 | /* iterate over everything, in case we missed something before */ |
2556 | pipecb (EV_A_ &pipe_w, EV_READ); |
3066 | ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); |
2557 | #endif |
|
|
2558 | } |
3067 | } |
|
|
3068 | #endif |
2559 | |
3069 | |
2560 | postfork = 0; |
3070 | postfork = 0; |
2561 | } |
3071 | } |
2562 | |
3072 | |
2563 | #if EV_MULTIPLICITY |
3073 | #if EV_MULTIPLICITY |
… | |
… | |
2736 | } |
3246 | } |
2737 | |
3247 | |
2738 | void |
3248 | void |
2739 | ev_loop_fork (EV_P) EV_THROW |
3249 | ev_loop_fork (EV_P) EV_THROW |
2740 | { |
3250 | { |
2741 | postfork = 1; /* must be in line with ev_default_fork */ |
3251 | postfork = 1; |
2742 | } |
3252 | } |
2743 | |
3253 | |
2744 | /*****************************************************************************/ |
3254 | /*****************************************************************************/ |
2745 | |
3255 | |
2746 | void |
3256 | void |
… | |
… | |
2762 | } |
3272 | } |
2763 | |
3273 | |
2764 | void noinline |
3274 | void noinline |
2765 | ev_invoke_pending (EV_P) |
3275 | ev_invoke_pending (EV_P) |
2766 | { |
3276 | { |
2767 | for (pendingpri = NUMPRI; pendingpri--; ) /* pendingpri is modified during the loop */ |
3277 | pendingpri = NUMPRI; |
|
|
3278 | |
|
|
3279 | while (pendingpri) /* pendingpri possibly gets modified in the inner loop */ |
|
|
3280 | { |
|
|
3281 | --pendingpri; |
|
|
3282 | |
2768 | while (pendingcnt [pendingpri]) |
3283 | while (pendingcnt [pendingpri]) |
2769 | { |
3284 | { |
2770 | ANPENDING *p = pendings [pendingpri] + --pendingcnt [pendingpri]; |
3285 | ANPENDING *p = pendings [pendingpri] + --pendingcnt [pendingpri]; |
2771 | |
3286 | |
2772 | p->w->pending = 0; |
3287 | p->w->pending = 0; |
2773 | EV_CB_INVOKE (p->w, p->events); |
3288 | EV_CB_INVOKE (p->w, p->events); |
2774 | EV_FREQUENT_CHECK; |
3289 | EV_FREQUENT_CHECK; |
2775 | } |
3290 | } |
|
|
3291 | } |
2776 | } |
3292 | } |
2777 | |
3293 | |
2778 | #if EV_IDLE_ENABLE |
3294 | #if EV_IDLE_ENABLE |
2779 | /* make idle watchers pending. this handles the "call-idle */ |
3295 | /* make idle watchers pending. this handles the "call-idle */ |
2780 | /* only when higher priorities are idle" logic */ |
3296 | /* only when higher priorities are idle" logic */ |
… | |
… | |
3138 | backend_poll (EV_A_ waittime); |
3654 | backend_poll (EV_A_ waittime); |
3139 | assert ((loop_done = EVBREAK_CANCEL, 1)); /* assert for side effect */ |
3655 | assert ((loop_done = EVBREAK_CANCEL, 1)); /* assert for side effect */ |
3140 | |
3656 | |
3141 | pipe_write_wanted = 0; /* just an optimisation, no fence needed */ |
3657 | pipe_write_wanted = 0; /* just an optimisation, no fence needed */ |
3142 | |
3658 | |
|
|
3659 | ECB_MEMORY_FENCE_ACQUIRE; |
3143 | if (pipe_write_skipped) |
3660 | if (pipe_write_skipped) |
3144 | { |
3661 | { |
3145 | assert (("libev: pipe_w not active, but pipe not written", ev_is_active (&pipe_w))); |
3662 | assert (("libev: pipe_w not active, but pipe not written", ev_is_active (&pipe_w))); |
3146 | ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); |
3663 | ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); |
3147 | } |
3664 | } |
… | |
… | |
3525 | #if EV_MULTIPLICITY |
4042 | #if EV_MULTIPLICITY |
3526 | assert (("libev: a signal must not be attached to two different loops", |
4043 | assert (("libev: a signal must not be attached to two different loops", |
3527 | !signals [w->signum - 1].loop || signals [w->signum - 1].loop == loop)); |
4044 | !signals [w->signum - 1].loop || signals [w->signum - 1].loop == loop)); |
3528 | |
4045 | |
3529 | signals [w->signum - 1].loop = EV_A; |
4046 | signals [w->signum - 1].loop = EV_A; |
|
|
4047 | ECB_MEMORY_FENCE_RELEASE; |
3530 | #endif |
4048 | #endif |
3531 | |
4049 | |
3532 | EV_FREQUENT_CHECK; |
4050 | EV_FREQUENT_CHECK; |
3533 | |
4051 | |
3534 | #if EV_USE_SIGNALFD |
4052 | #if EV_USE_SIGNALFD |
… | |
… | |
3689 | # define EV_INOTIFY_BUFSIZE (sizeof (struct inotify_event) * 2 + NAME_MAX) |
4207 | # define EV_INOTIFY_BUFSIZE (sizeof (struct inotify_event) * 2 + NAME_MAX) |
3690 | |
4208 | |
3691 | static void noinline |
4209 | static void noinline |
3692 | infy_add (EV_P_ ev_stat *w) |
4210 | infy_add (EV_P_ ev_stat *w) |
3693 | { |
4211 | { |
3694 | w->wd = inotify_add_watch (fs_fd, w->path, IN_ATTRIB | IN_DELETE_SELF | IN_MOVE_SELF | IN_MODIFY | IN_DONT_FOLLOW | IN_MASK_ADD); |
4212 | w->wd = inotify_add_watch (fs_fd, w->path, |
|
|
4213 | IN_ATTRIB | IN_DELETE_SELF | IN_MOVE_SELF | IN_MODIFY |
|
|
4214 | | IN_CREATE | IN_DELETE | IN_MOVED_FROM | IN_MOVED_TO |
|
|
4215 | | IN_DONT_FOLLOW | IN_MASK_ADD); |
3695 | |
4216 | |
3696 | if (w->wd >= 0) |
4217 | if (w->wd >= 0) |
3697 | { |
4218 | { |
3698 | struct statfs sfs; |
4219 | struct statfs sfs; |
3699 | |
4220 | |
… | |
… | |
3703 | |
4224 | |
3704 | if (!fs_2625) |
4225 | if (!fs_2625) |
3705 | w->timer.repeat = w->interval ? w->interval : DEF_STAT_INTERVAL; |
4226 | w->timer.repeat = w->interval ? w->interval : DEF_STAT_INTERVAL; |
3706 | else if (!statfs (w->path, &sfs) |
4227 | else if (!statfs (w->path, &sfs) |
3707 | && (sfs.f_type == 0x1373 /* devfs */ |
4228 | && (sfs.f_type == 0x1373 /* devfs */ |
|
|
4229 | || sfs.f_type == 0x4006 /* fat */ |
|
|
4230 | || sfs.f_type == 0x4d44 /* msdos */ |
3708 | || sfs.f_type == 0xEF53 /* ext2/3 */ |
4231 | || sfs.f_type == 0xEF53 /* ext2/3 */ |
|
|
4232 | || sfs.f_type == 0x72b6 /* jffs2 */ |
|
|
4233 | || sfs.f_type == 0x858458f6 /* ramfs */ |
|
|
4234 | || sfs.f_type == 0x5346544e /* ntfs */ |
3709 | || sfs.f_type == 0x3153464a /* jfs */ |
4235 | || sfs.f_type == 0x3153464a /* jfs */ |
|
|
4236 | || sfs.f_type == 0x9123683e /* btrfs */ |
3710 | || sfs.f_type == 0x52654973 /* reiser3 */ |
4237 | || sfs.f_type == 0x52654973 /* reiser3 */ |
3711 | || sfs.f_type == 0x01021994 /* tempfs */ |
4238 | || sfs.f_type == 0x01021994 /* tmpfs */ |
3712 | || sfs.f_type == 0x58465342 /* xfs */)) |
4239 | || sfs.f_type == 0x58465342 /* xfs */)) |
3713 | w->timer.repeat = 0.; /* filesystem is local, kernel new enough */ |
4240 | w->timer.repeat = 0.; /* filesystem is local, kernel new enough */ |
3714 | else |
4241 | else |
3715 | w->timer.repeat = w->interval ? w->interval : NFS_STAT_INTERVAL; /* remote, use reduced frequency */ |
4242 | w->timer.repeat = w->interval ? w->interval : NFS_STAT_INTERVAL; /* remote, use reduced frequency */ |
3716 | } |
4243 | } |