… | |
… | |
241 | #elif defined SIGARRAYSIZE |
241 | #elif defined SIGARRAYSIZE |
242 | # define EV_NSIG (SIGARRAYSIZE) /* Assume ary[SIGARRAYSIZE] */ |
242 | # define EV_NSIG (SIGARRAYSIZE) /* Assume ary[SIGARRAYSIZE] */ |
243 | #elif defined _sys_nsig |
243 | #elif defined _sys_nsig |
244 | # define EV_NSIG (_sys_nsig) /* Solaris 2.5 */ |
244 | # define EV_NSIG (_sys_nsig) /* Solaris 2.5 */ |
245 | #else |
245 | #else |
246 | # error "unable to find value for NSIG, please report" |
246 | # define EV_NSIG (8 * sizeof (sigset_t) + 1) |
247 | /* to make it compile regardless, just remove the above line, */ |
|
|
248 | /* but consider reporting it, too! :) */ |
|
|
249 | # define EV_NSIG 65 |
|
|
250 | #endif |
247 | #endif |
251 | |
248 | |
252 | #ifndef EV_USE_FLOOR |
249 | #ifndef EV_USE_FLOOR |
253 | # define EV_USE_FLOOR 0 |
250 | # define EV_USE_FLOOR 0 |
254 | #endif |
251 | #endif |
255 | |
252 | |
256 | #ifndef EV_USE_CLOCK_SYSCALL |
253 | #ifndef EV_USE_CLOCK_SYSCALL |
257 | # if __linux && __GLIBC__ >= 2 |
254 | # if __linux && __GLIBC__ == 2 && __GLIBC_MINOR__ < 17 |
258 | # define EV_USE_CLOCK_SYSCALL EV_FEATURE_OS |
255 | # define EV_USE_CLOCK_SYSCALL EV_FEATURE_OS |
259 | # else |
256 | # else |
260 | # define EV_USE_CLOCK_SYSCALL 0 |
257 | # define EV_USE_CLOCK_SYSCALL 0 |
261 | # endif |
258 | # endif |
262 | #endif |
259 | #endif |
… | |
… | |
355 | # define EV_USE_4HEAP EV_FEATURE_DATA |
352 | # define EV_USE_4HEAP EV_FEATURE_DATA |
356 | #endif |
353 | #endif |
357 | |
354 | |
358 | #ifndef EV_HEAP_CACHE_AT |
355 | #ifndef EV_HEAP_CACHE_AT |
359 | # define EV_HEAP_CACHE_AT EV_FEATURE_DATA |
356 | # define EV_HEAP_CACHE_AT EV_FEATURE_DATA |
|
|
357 | #endif |
|
|
358 | |
|
|
359 | #ifdef ANDROID |
|
|
360 | /* supposedly, android doesn't typedef fd_mask */ |
|
|
361 | # undef EV_USE_SELECT |
|
|
362 | # define EV_USE_SELECT 0 |
|
|
363 | /* supposedly, we need to include syscall.h, not sys/syscall.h, so just disable */ |
|
|
364 | # undef EV_USE_CLOCK_SYSCALL |
|
|
365 | # define EV_USE_CLOCK_SYSCALL 0 |
|
|
366 | #endif |
|
|
367 | |
|
|
368 | /* aix's poll.h seems to cause lots of trouble */ |
|
|
369 | #ifdef _AIX |
|
|
370 | /* AIX has a completely broken poll.h header */ |
|
|
371 | # undef EV_USE_POLL |
|
|
372 | # define EV_USE_POLL 0 |
360 | #endif |
373 | #endif |
361 | |
374 | |
362 | /* on linux, we can use a (slow) syscall to avoid a dependency on pthread, */ |
375 | /* on linux, we can use a (slow) syscall to avoid a dependency on pthread, */ |
363 | /* which makes programs even slower. might work on other unices, too. */ |
376 | /* which makes programs even slower. might work on other unices, too. */ |
364 | #if EV_USE_CLOCK_SYSCALL |
377 | #if EV_USE_CLOCK_SYSCALL |
… | |
… | |
372 | # define EV_USE_CLOCK_SYSCALL 0 |
385 | # define EV_USE_CLOCK_SYSCALL 0 |
373 | # endif |
386 | # endif |
374 | #endif |
387 | #endif |
375 | |
388 | |
376 | /* this block fixes any misconfiguration where we know we run into trouble otherwise */ |
389 | /* this block fixes any misconfiguration where we know we run into trouble otherwise */ |
377 | |
|
|
378 | #ifdef _AIX |
|
|
379 | /* AIX has a completely broken poll.h header */ |
|
|
380 | # undef EV_USE_POLL |
|
|
381 | # define EV_USE_POLL 0 |
|
|
382 | #endif |
|
|
383 | |
390 | |
384 | #ifndef CLOCK_MONOTONIC |
391 | #ifndef CLOCK_MONOTONIC |
385 | # undef EV_USE_MONOTONIC |
392 | # undef EV_USE_MONOTONIC |
386 | # define EV_USE_MONOTONIC 0 |
393 | # define EV_USE_MONOTONIC 0 |
387 | #endif |
394 | #endif |
… | |
… | |
504 | */ |
511 | */ |
505 | |
512 | |
506 | #ifndef ECB_H |
513 | #ifndef ECB_H |
507 | #define ECB_H |
514 | #define ECB_H |
508 | |
515 | |
|
|
516 | /* 16 bits major, 16 bits minor */ |
|
|
517 | #define ECB_VERSION 0x00010003 |
|
|
518 | |
509 | #ifdef _WIN32 |
519 | #ifdef _WIN32 |
510 | typedef signed char int8_t; |
520 | typedef signed char int8_t; |
511 | typedef unsigned char uint8_t; |
521 | typedef unsigned char uint8_t; |
512 | typedef signed short int16_t; |
522 | typedef signed short int16_t; |
513 | typedef unsigned short uint16_t; |
523 | typedef unsigned short uint16_t; |
… | |
… | |
518 | typedef unsigned long long uint64_t; |
528 | typedef unsigned long long uint64_t; |
519 | #else /* _MSC_VER || __BORLANDC__ */ |
529 | #else /* _MSC_VER || __BORLANDC__ */ |
520 | typedef signed __int64 int64_t; |
530 | typedef signed __int64 int64_t; |
521 | typedef unsigned __int64 uint64_t; |
531 | typedef unsigned __int64 uint64_t; |
522 | #endif |
532 | #endif |
|
|
533 | #ifdef _WIN64 |
|
|
534 | #define ECB_PTRSIZE 8 |
|
|
535 | typedef uint64_t uintptr_t; |
|
|
536 | typedef int64_t intptr_t; |
|
|
537 | #else |
|
|
538 | #define ECB_PTRSIZE 4 |
|
|
539 | typedef uint32_t uintptr_t; |
|
|
540 | typedef int32_t intptr_t; |
|
|
541 | #endif |
523 | #else |
542 | #else |
524 | #include <inttypes.h> |
543 | #include <inttypes.h> |
|
|
544 | #if UINTMAX_MAX > 0xffffffffU |
|
|
545 | #define ECB_PTRSIZE 8 |
|
|
546 | #else |
|
|
547 | #define ECB_PTRSIZE 4 |
|
|
548 | #endif |
|
|
549 | #endif |
|
|
550 | |
|
|
551 | /* work around x32 idiocy by defining proper macros */ |
|
|
552 | #if __x86_64 || _M_AMD64 |
|
|
553 | #if _ILP32 |
|
|
554 | #define ECB_AMD64_X32 1 |
|
|
555 | #else |
|
|
556 | #define ECB_AMD64 1 |
|
|
557 | #endif |
525 | #endif |
558 | #endif |
526 | |
559 | |
527 | /* many compilers define _GNUC_ to some versions but then only implement |
560 | /* many compilers define _GNUC_ to some versions but then only implement |
528 | * what their idiot authors think are the "more important" extensions, |
561 | * what their idiot authors think are the "more important" extensions, |
529 | * causing enormous grief in return for some better fake benchmark numbers. |
562 | * causing enormous grief in return for some better fake benchmark numbers. |
… | |
… | |
537 | #else |
570 | #else |
538 | #define ECB_GCC_VERSION(major,minor) (__GNUC__ > (major) || (__GNUC__ == (major) && __GNUC_MINOR__ >= (minor))) |
571 | #define ECB_GCC_VERSION(major,minor) (__GNUC__ > (major) || (__GNUC__ == (major) && __GNUC_MINOR__ >= (minor))) |
539 | #endif |
572 | #endif |
540 | #endif |
573 | #endif |
541 | |
574 | |
|
|
575 | #define ECB_C (__STDC__+0) /* this assumes that __STDC__ is either empty or a number */ |
|
|
576 | #define ECB_C99 (__STDC_VERSION__ >= 199901L) |
|
|
577 | #define ECB_C11 (__STDC_VERSION__ >= 201112L) |
|
|
578 | #define ECB_CPP (__cplusplus+0) |
|
|
579 | #define ECB_CPP11 (__cplusplus >= 201103L) |
|
|
580 | |
|
|
581 | #if ECB_CPP |
|
|
582 | #define ECB_EXTERN_C extern "C" |
|
|
583 | #define ECB_EXTERN_C_BEG ECB_EXTERN_C { |
|
|
584 | #define ECB_EXTERN_C_END } |
|
|
585 | #else |
|
|
586 | #define ECB_EXTERN_C extern |
|
|
587 | #define ECB_EXTERN_C_BEG |
|
|
588 | #define ECB_EXTERN_C_END |
|
|
589 | #endif |
|
|
590 | |
542 | /*****************************************************************************/ |
591 | /*****************************************************************************/ |
543 | |
592 | |
544 | /* ECB_NO_THREADS - ecb is not used by multiple threads, ever */ |
593 | /* ECB_NO_THREADS - ecb is not used by multiple threads, ever */ |
545 | /* ECB_NO_SMP - ecb might be used in multiple threads, but only on a single cpu */ |
594 | /* ECB_NO_SMP - ecb might be used in multiple threads, but only on a single cpu */ |
546 | |
595 | |
547 | #if ECB_NO_THREADS |
596 | #if ECB_NO_THREADS |
548 | # define ECB_NO_SMP 1 |
597 | #define ECB_NO_SMP 1 |
549 | #endif |
598 | #endif |
550 | |
599 | |
551 | #if ECB_NO_THREADS || ECB_NO_SMP |
600 | #if ECB_NO_SMP |
552 | #define ECB_MEMORY_FENCE do { } while (0) |
601 | #define ECB_MEMORY_FENCE do { } while (0) |
553 | #endif |
602 | #endif |
554 | |
603 | |
555 | #ifndef ECB_MEMORY_FENCE |
604 | #ifndef ECB_MEMORY_FENCE |
556 | #if ECB_GCC_VERSION(2,5) || defined __INTEL_COMPILER || (__llvm__ && __GNUC__) || __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
605 | #if ECB_GCC_VERSION(2,5) || defined __INTEL_COMPILER || (__llvm__ && __GNUC__) || __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
557 | #if __i386 || __i386__ |
606 | #if __i386 || __i386__ |
558 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
607 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
559 | #define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE /* non-lock xchg might be enough */ |
608 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory") |
560 | #define ECB_MEMORY_FENCE_RELEASE do { } while (0) /* unlikely to change in future cpus */ |
609 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
561 | #elif __amd64 || __amd64__ || __x86_64 || __x86_64__ |
610 | #elif __amd64 || __amd64__ || __x86_64 || __x86_64__ |
562 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
611 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
563 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("lfence" : : : "memory") |
612 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory") |
564 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("sfence") /* play safe - not needed in any current cpu */ |
613 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
565 | #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ |
614 | #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ |
566 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
615 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
567 | #elif defined __ARM_ARCH_6__ || defined __ARM_ARCH_6J__ \ |
616 | #elif defined __ARM_ARCH_6__ || defined __ARM_ARCH_6J__ \ |
568 | || defined __ARM_ARCH_6K__ || defined __ARM_ARCH_6ZK__ |
617 | || defined __ARM_ARCH_6K__ || defined __ARM_ARCH_6ZK__ |
569 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,5" : : "r" (0) : "memory") |
618 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,5" : : "r" (0) : "memory") |
570 | #elif defined __ARM_ARCH_7__ || defined __ARM_ARCH_7A__ \ |
619 | #elif defined __ARM_ARCH_7__ || defined __ARM_ARCH_7A__ \ |
571 | || defined __ARM_ARCH_7M__ || defined __ARM_ARCH_7R__ |
620 | || defined __ARM_ARCH_7M__ || defined __ARM_ARCH_7R__ |
572 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") |
621 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") |
573 | #elif __sparc || __sparc__ |
622 | #elif __sparc || __sparc__ |
574 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad | " : : : "memory") |
623 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad" : : : "memory") |
575 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory") |
624 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory") |
576 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore") |
625 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore") |
577 | #elif defined __s390__ || defined __s390x__ |
626 | #elif defined __s390__ || defined __s390x__ |
578 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory") |
627 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory") |
579 | #elif defined __mips__ |
628 | #elif defined __mips__ |
|
|
629 | /* GNU/Linux emulates sync on mips1 architectures, so we force its use */ |
|
|
630 | /* anybody else who still uses mips1 is supposed to send in their version, with detection code. */ |
580 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
631 | #define ECB_MEMORY_FENCE __asm__ __volatile__ (".set mips2; sync; .set mips0" : : : "memory") |
581 | #elif defined __alpha__ |
632 | #elif defined __alpha__ |
582 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mb" : : : "memory") |
633 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mb" : : : "memory") |
|
|
634 | #elif defined __hppa__ |
|
|
635 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("" : : : "memory") |
|
|
636 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
|
|
637 | #elif defined __ia64__ |
|
|
638 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mf" : : : "memory") |
|
|
639 | #elif defined __m68k__ |
|
|
640 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("" : : : "memory") |
|
|
641 | #elif defined __m88k__ |
|
|
642 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("tb1 0,%%r0,128" : : : "memory") |
|
|
643 | #elif defined __sh__ |
|
|
644 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("" : : : "memory") |
583 | #endif |
645 | #endif |
584 | #endif |
646 | #endif |
585 | #endif |
647 | #endif |
586 | |
648 | |
587 | #ifndef ECB_MEMORY_FENCE |
649 | #ifndef ECB_MEMORY_FENCE |
|
|
650 | #if ECB_GCC_VERSION(4,7) |
|
|
651 | /* see comment below (stdatomic.h) about the C11 memory model. */ |
|
|
652 | #define ECB_MEMORY_FENCE __atomic_thread_fence (__ATOMIC_SEQ_CST) |
|
|
653 | |
|
|
654 | /* The __has_feature syntax from clang is so misdesigned that we cannot use it |
|
|
655 | * without risking compile time errors with other compilers. We *could* |
|
|
656 | * define our own ecb_clang_has_feature, but I just can't be bothered to work |
|
|
657 | * around this shit time and again. |
|
|
658 | * #elif defined __clang && __has_feature (cxx_atomic) |
|
|
659 | * // see comment below (stdatomic.h) about the C11 memory model. |
|
|
660 | * #define ECB_MEMORY_FENCE __c11_atomic_thread_fence (__ATOMIC_SEQ_CST) |
|
|
661 | */ |
|
|
662 | |
588 | #if ECB_GCC_VERSION(4,4) || defined __INTEL_COMPILER || defined __clang__ |
663 | #elif ECB_GCC_VERSION(4,4) || defined __INTEL_COMPILER || defined __clang__ |
589 | #define ECB_MEMORY_FENCE __sync_synchronize () |
664 | #define ECB_MEMORY_FENCE __sync_synchronize () |
590 | /*#define ECB_MEMORY_FENCE_ACQUIRE ({ char dummy = 0; __sync_lock_test_and_set (&dummy, 1); }) */ |
|
|
591 | /*#define ECB_MEMORY_FENCE_RELEASE ({ char dummy = 1; __sync_lock_release (&dummy ); }) */ |
|
|
592 | #elif _MSC_VER >= 1400 /* VC++ 2005 */ |
665 | #elif _MSC_VER >= 1400 /* VC++ 2005 */ |
593 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
666 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
594 | #define ECB_MEMORY_FENCE _ReadWriteBarrier () |
667 | #define ECB_MEMORY_FENCE _ReadWriteBarrier () |
595 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ |
668 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ |
596 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () |
669 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () |
… | |
… | |
606 | #define ECB_MEMORY_FENCE __sync () |
679 | #define ECB_MEMORY_FENCE __sync () |
607 | #endif |
680 | #endif |
608 | #endif |
681 | #endif |
609 | |
682 | |
610 | #ifndef ECB_MEMORY_FENCE |
683 | #ifndef ECB_MEMORY_FENCE |
|
|
684 | #if ECB_C11 && !defined __STDC_NO_ATOMICS__ |
|
|
685 | /* we assume that these memory fences work on all variables/all memory accesses, */ |
|
|
686 | /* not just C11 atomics and atomic accesses */ |
|
|
687 | #include <stdatomic.h> |
|
|
688 | /* Unfortunately, neither gcc 4.7 nor clang 3.1 generate any instructions for */ |
|
|
689 | /* any fence other than seq_cst, which isn't very efficient for us. */ |
|
|
690 | /* Why that is, we don't know - either the C11 memory model is quite useless */ |
|
|
691 | /* for most usages, or gcc and clang have a bug */ |
|
|
692 | /* I *currently* lean towards the latter, and inefficiently implement */ |
|
|
693 | /* all three of ecb's fences as a seq_cst fence */ |
|
|
694 | #define ECB_MEMORY_FENCE atomic_thread_fence (memory_order_seq_cst) |
|
|
695 | #endif |
|
|
696 | #endif |
|
|
697 | |
|
|
698 | #ifndef ECB_MEMORY_FENCE |
611 | #if !ECB_AVOID_PTHREADS |
699 | #if !ECB_AVOID_PTHREADS |
612 | /* |
700 | /* |
613 | * if you get undefined symbol references to pthread_mutex_lock, |
701 | * if you get undefined symbol references to pthread_mutex_lock, |
614 | * or failure to find pthread.h, then you should implement |
702 | * or failure to find pthread.h, then you should implement |
615 | * the ECB_MEMORY_FENCE operations for your cpu/compiler |
703 | * the ECB_MEMORY_FENCE operations for your cpu/compiler |
… | |
… | |
633 | #define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE |
721 | #define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE |
634 | #endif |
722 | #endif |
635 | |
723 | |
636 | /*****************************************************************************/ |
724 | /*****************************************************************************/ |
637 | |
725 | |
638 | #define ECB_C99 (__STDC_VERSION__ >= 199901L) |
|
|
639 | |
|
|
640 | #if __cplusplus |
726 | #if __cplusplus |
641 | #define ecb_inline static inline |
727 | #define ecb_inline static inline |
642 | #elif ECB_GCC_VERSION(2,5) |
728 | #elif ECB_GCC_VERSION(2,5) |
643 | #define ecb_inline static __inline__ |
729 | #define ecb_inline static __inline__ |
644 | #elif ECB_C99 |
730 | #elif ECB_C99 |
… | |
… | |
682 | #elif ECB_GCC_VERSION(3,0) |
768 | #elif ECB_GCC_VERSION(3,0) |
683 | #define ecb_decltype(x) __typeof(x) |
769 | #define ecb_decltype(x) __typeof(x) |
684 | #endif |
770 | #endif |
685 | |
771 | |
686 | #define ecb_noinline ecb_attribute ((__noinline__)) |
772 | #define ecb_noinline ecb_attribute ((__noinline__)) |
687 | #define ecb_noreturn ecb_attribute ((__noreturn__)) |
|
|
688 | #define ecb_unused ecb_attribute ((__unused__)) |
773 | #define ecb_unused ecb_attribute ((__unused__)) |
689 | #define ecb_const ecb_attribute ((__const__)) |
774 | #define ecb_const ecb_attribute ((__const__)) |
690 | #define ecb_pure ecb_attribute ((__pure__)) |
775 | #define ecb_pure ecb_attribute ((__pure__)) |
|
|
776 | |
|
|
777 | #if ECB_C11 |
|
|
778 | #define ecb_noreturn _Noreturn |
|
|
779 | #else |
|
|
780 | #define ecb_noreturn ecb_attribute ((__noreturn__)) |
|
|
781 | #endif |
691 | |
782 | |
692 | #if ECB_GCC_VERSION(4,3) |
783 | #if ECB_GCC_VERSION(4,3) |
693 | #define ecb_artificial ecb_attribute ((__artificial__)) |
784 | #define ecb_artificial ecb_attribute ((__artificial__)) |
694 | #define ecb_hot ecb_attribute ((__hot__)) |
785 | #define ecb_hot ecb_attribute ((__hot__)) |
695 | #define ecb_cold ecb_attribute ((__cold__)) |
786 | #define ecb_cold ecb_attribute ((__cold__)) |
… | |
… | |
786 | |
877 | |
787 | return r + ecb_ld32 (x); |
878 | return r + ecb_ld32 (x); |
788 | } |
879 | } |
789 | #endif |
880 | #endif |
790 | |
881 | |
|
|
882 | ecb_function_ ecb_bool ecb_is_pot32 (uint32_t x) ecb_const; |
|
|
883 | ecb_function_ ecb_bool ecb_is_pot32 (uint32_t x) { return !(x & (x - 1)); } |
|
|
884 | ecb_function_ ecb_bool ecb_is_pot64 (uint64_t x) ecb_const; |
|
|
885 | ecb_function_ ecb_bool ecb_is_pot64 (uint64_t x) { return !(x & (x - 1)); } |
|
|
886 | |
791 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) ecb_const; |
887 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) ecb_const; |
792 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) |
888 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) |
793 | { |
889 | { |
794 | return ( (x * 0x0802U & 0x22110U) |
890 | return ( (x * 0x0802U & 0x22110U) |
795 | | (x * 0x8020U & 0x88440U)) * 0x10101U >> 16; |
891 | | (x * 0x8020U & 0x88440U)) * 0x10101U >> 16; |
… | |
… | |
879 | ecb_inline void ecb_unreachable (void) ecb_noreturn; |
975 | ecb_inline void ecb_unreachable (void) ecb_noreturn; |
880 | ecb_inline void ecb_unreachable (void) { } |
976 | ecb_inline void ecb_unreachable (void) { } |
881 | #endif |
977 | #endif |
882 | |
978 | |
883 | /* try to tell the compiler that some condition is definitely true */ |
979 | /* try to tell the compiler that some condition is definitely true */ |
884 | #define ecb_assume(cond) do { if (!(cond)) ecb_unreachable (); } while (0) |
980 | #define ecb_assume(cond) if (!(cond)) ecb_unreachable (); else 0 |
885 | |
981 | |
886 | ecb_inline unsigned char ecb_byteorder_helper (void) ecb_const; |
982 | ecb_inline unsigned char ecb_byteorder_helper (void) ecb_const; |
887 | ecb_inline unsigned char |
983 | ecb_inline unsigned char |
888 | ecb_byteorder_helper (void) |
984 | ecb_byteorder_helper (void) |
889 | { |
985 | { |
890 | const uint32_t u = 0x11223344; |
986 | /* the union code still generates code under pressure in gcc, */ |
891 | return *(unsigned char *)&u; |
987 | /* but less than using pointers, and always seems to */ |
|
|
988 | /* successfully return a constant. */ |
|
|
989 | /* the reason why we have this horrible preprocessor mess */ |
|
|
990 | /* is to avoid it in all cases, at least on common architectures */ |
|
|
991 | /* or when using a recent enough gcc version (>= 4.6) */ |
|
|
992 | #if __i386 || __i386__ || _M_X86 || __amd64 || __amd64__ || _M_X64 |
|
|
993 | return 0x44; |
|
|
994 | #elif __BYTE_ORDER__ && __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ |
|
|
995 | return 0x44; |
|
|
996 | #elif __BYTE_ORDER__ && __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ |
|
|
997 | return 0x11; |
|
|
998 | #else |
|
|
999 | union |
|
|
1000 | { |
|
|
1001 | uint32_t i; |
|
|
1002 | uint8_t c; |
|
|
1003 | } u = { 0x11223344 }; |
|
|
1004 | return u.c; |
|
|
1005 | #endif |
892 | } |
1006 | } |
893 | |
1007 | |
894 | ecb_inline ecb_bool ecb_big_endian (void) ecb_const; |
1008 | ecb_inline ecb_bool ecb_big_endian (void) ecb_const; |
895 | ecb_inline ecb_bool ecb_big_endian (void) { return ecb_byteorder_helper () == 0x11; } |
1009 | ecb_inline ecb_bool ecb_big_endian (void) { return ecb_byteorder_helper () == 0x11; } |
896 | ecb_inline ecb_bool ecb_little_endian (void) ecb_const; |
1010 | ecb_inline ecb_bool ecb_little_endian (void) ecb_const; |
… | |
… | |
927 | } |
1041 | } |
928 | #else |
1042 | #else |
929 | #define ecb_array_length(name) (sizeof (name) / sizeof (name [0])) |
1043 | #define ecb_array_length(name) (sizeof (name) / sizeof (name [0])) |
930 | #endif |
1044 | #endif |
931 | |
1045 | |
|
|
1046 | /*******************************************************************************/ |
|
|
1047 | /* floating point stuff, can be disabled by defining ECB_NO_LIBM */ |
|
|
1048 | |
|
|
1049 | /* basically, everything uses "ieee pure-endian" floating point numbers */ |
|
|
1050 | /* the only noteworthy exception is ancient armle, which uses order 43218765 */ |
|
|
1051 | #if 0 \ |
|
|
1052 | || __i386 || __i386__ \ |
|
|
1053 | || __amd64 || __amd64__ || __x86_64 || __x86_64__ \ |
|
|
1054 | || __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ \ |
|
|
1055 | || defined __arm__ && defined __ARM_EABI__ \ |
|
|
1056 | || defined __s390__ || defined __s390x__ \ |
|
|
1057 | || defined __mips__ \ |
|
|
1058 | || defined __alpha__ \ |
|
|
1059 | || defined __hppa__ \ |
|
|
1060 | || defined __ia64__ \ |
|
|
1061 | || defined __m68k__ \ |
|
|
1062 | || defined __m88k__ \ |
|
|
1063 | || defined __sh__ \ |
|
|
1064 | || defined _M_IX86 || defined _M_AMD64 || defined _M_IA64 |
|
|
1065 | #define ECB_STDFP 1 |
|
|
1066 | #include <string.h> /* for memcpy */ |
|
|
1067 | #else |
|
|
1068 | #define ECB_STDFP 0 |
|
|
1069 | #endif |
|
|
1070 | |
|
|
1071 | #ifndef ECB_NO_LIBM |
|
|
1072 | |
|
|
1073 | #include <math.h> /* for frexp*, ldexp*, INFINITY, NAN */ |
|
|
1074 | |
|
|
1075 | #ifdef NEN |
|
|
1076 | #define ECB_NAN NAN |
|
|
1077 | #else |
|
|
1078 | #define ECB_NAN INFINITY |
|
|
1079 | #endif |
|
|
1080 | |
|
|
1081 | /* converts an ieee half/binary16 to a float */ |
|
|
1082 | ecb_function_ float ecb_binary16_to_float (uint16_t x) ecb_const; |
|
|
1083 | ecb_function_ float |
|
|
1084 | ecb_binary16_to_float (uint16_t x) |
|
|
1085 | { |
|
|
1086 | int e = (x >> 10) & 0x1f; |
|
|
1087 | int m = x & 0x3ff; |
|
|
1088 | float r; |
|
|
1089 | |
|
|
1090 | if (!e ) r = ldexpf (m , -24); |
|
|
1091 | else if (e != 31) r = ldexpf (m + 0x400, e - 25); |
|
|
1092 | else if (m ) r = ECB_NAN; |
|
|
1093 | else r = INFINITY; |
|
|
1094 | |
|
|
1095 | return x & 0x8000 ? -r : r; |
|
|
1096 | } |
|
|
1097 | |
|
|
1098 | /* convert a float to ieee single/binary32 */ |
|
|
1099 | ecb_function_ uint32_t ecb_float_to_binary32 (float x) ecb_const; |
|
|
1100 | ecb_function_ uint32_t |
|
|
1101 | ecb_float_to_binary32 (float x) |
|
|
1102 | { |
|
|
1103 | uint32_t r; |
|
|
1104 | |
|
|
1105 | #if ECB_STDFP |
|
|
1106 | memcpy (&r, &x, 4); |
|
|
1107 | #else |
|
|
1108 | /* slow emulation, works for anything but -0 */ |
|
|
1109 | uint32_t m; |
|
|
1110 | int e; |
|
|
1111 | |
|
|
1112 | if (x == 0e0f ) return 0x00000000U; |
|
|
1113 | if (x > +3.40282346638528860e+38f) return 0x7f800000U; |
|
|
1114 | if (x < -3.40282346638528860e+38f) return 0xff800000U; |
|
|
1115 | if (x != x ) return 0x7fbfffffU; |
|
|
1116 | |
|
|
1117 | m = frexpf (x, &e) * 0x1000000U; |
|
|
1118 | |
|
|
1119 | r = m & 0x80000000U; |
|
|
1120 | |
|
|
1121 | if (r) |
|
|
1122 | m = -m; |
|
|
1123 | |
|
|
1124 | if (e <= -126) |
|
|
1125 | { |
|
|
1126 | m &= 0xffffffU; |
|
|
1127 | m >>= (-125 - e); |
|
|
1128 | e = -126; |
|
|
1129 | } |
|
|
1130 | |
|
|
1131 | r |= (e + 126) << 23; |
|
|
1132 | r |= m & 0x7fffffU; |
|
|
1133 | #endif |
|
|
1134 | |
|
|
1135 | return r; |
|
|
1136 | } |
|
|
1137 | |
|
|
1138 | /* converts an ieee single/binary32 to a float */ |
|
|
1139 | ecb_function_ float ecb_binary32_to_float (uint32_t x) ecb_const; |
|
|
1140 | ecb_function_ float |
|
|
1141 | ecb_binary32_to_float (uint32_t x) |
|
|
1142 | { |
|
|
1143 | float r; |
|
|
1144 | |
|
|
1145 | #if ECB_STDFP |
|
|
1146 | memcpy (&r, &x, 4); |
|
|
1147 | #else |
|
|
1148 | /* emulation, only works for normals and subnormals and +0 */ |
|
|
1149 | int neg = x >> 31; |
|
|
1150 | int e = (x >> 23) & 0xffU; |
|
|
1151 | |
|
|
1152 | x &= 0x7fffffU; |
|
|
1153 | |
|
|
1154 | if (e) |
|
|
1155 | x |= 0x800000U; |
|
|
1156 | else |
|
|
1157 | e = 1; |
|
|
1158 | |
|
|
1159 | /* we distrust ldexpf a bit and do the 2**-24 scaling by an extra multiply */ |
|
|
1160 | r = ldexpf (x * (0.5f / 0x800000U), e - 126); |
|
|
1161 | |
|
|
1162 | r = neg ? -r : r; |
|
|
1163 | #endif |
|
|
1164 | |
|
|
1165 | return r; |
|
|
1166 | } |
|
|
1167 | |
|
|
1168 | /* convert a double to ieee double/binary64 */ |
|
|
1169 | ecb_function_ uint64_t ecb_double_to_binary64 (double x) ecb_const; |
|
|
1170 | ecb_function_ uint64_t |
|
|
1171 | ecb_double_to_binary64 (double x) |
|
|
1172 | { |
|
|
1173 | uint64_t r; |
|
|
1174 | |
|
|
1175 | #if ECB_STDFP |
|
|
1176 | memcpy (&r, &x, 8); |
|
|
1177 | #else |
|
|
1178 | /* slow emulation, works for anything but -0 */ |
|
|
1179 | uint64_t m; |
|
|
1180 | int e; |
|
|
1181 | |
|
|
1182 | if (x == 0e0 ) return 0x0000000000000000U; |
|
|
1183 | if (x > +1.79769313486231470e+308) return 0x7ff0000000000000U; |
|
|
1184 | if (x < -1.79769313486231470e+308) return 0xfff0000000000000U; |
|
|
1185 | if (x != x ) return 0X7ff7ffffffffffffU; |
|
|
1186 | |
|
|
1187 | m = frexp (x, &e) * 0x20000000000000U; |
|
|
1188 | |
|
|
1189 | r = m & 0x8000000000000000;; |
|
|
1190 | |
|
|
1191 | if (r) |
|
|
1192 | m = -m; |
|
|
1193 | |
|
|
1194 | if (e <= -1022) |
|
|
1195 | { |
|
|
1196 | m &= 0x1fffffffffffffU; |
|
|
1197 | m >>= (-1021 - e); |
|
|
1198 | e = -1022; |
|
|
1199 | } |
|
|
1200 | |
|
|
1201 | r |= ((uint64_t)(e + 1022)) << 52; |
|
|
1202 | r |= m & 0xfffffffffffffU; |
|
|
1203 | #endif |
|
|
1204 | |
|
|
1205 | return r; |
|
|
1206 | } |
|
|
1207 | |
|
|
1208 | /* converts an ieee double/binary64 to a double */ |
|
|
1209 | ecb_function_ double ecb_binary64_to_double (uint64_t x) ecb_const; |
|
|
1210 | ecb_function_ double |
|
|
1211 | ecb_binary64_to_double (uint64_t x) |
|
|
1212 | { |
|
|
1213 | double r; |
|
|
1214 | |
|
|
1215 | #if ECB_STDFP |
|
|
1216 | memcpy (&r, &x, 8); |
|
|
1217 | #else |
|
|
1218 | /* emulation, only works for normals and subnormals and +0 */ |
|
|
1219 | int neg = x >> 63; |
|
|
1220 | int e = (x >> 52) & 0x7ffU; |
|
|
1221 | |
|
|
1222 | x &= 0xfffffffffffffU; |
|
|
1223 | |
|
|
1224 | if (e) |
|
|
1225 | x |= 0x10000000000000U; |
|
|
1226 | else |
|
|
1227 | e = 1; |
|
|
1228 | |
|
|
1229 | /* we distrust ldexp a bit and do the 2**-53 scaling by an extra multiply */ |
|
|
1230 | r = ldexp (x * (0.5 / 0x10000000000000U), e - 1022); |
|
|
1231 | |
|
|
1232 | r = neg ? -r : r; |
|
|
1233 | #endif |
|
|
1234 | |
|
|
1235 | return r; |
|
|
1236 | } |
|
|
1237 | |
|
|
1238 | #endif |
|
|
1239 | |
932 | #endif |
1240 | #endif |
933 | |
1241 | |
934 | /* ECB.H END */ |
1242 | /* ECB.H END */ |
935 | |
1243 | |
936 | #if ECB_MEMORY_FENCE_NEEDS_PTHREADS |
1244 | #if ECB_MEMORY_FENCE_NEEDS_PTHREADS |
… | |
… | |
1135 | } |
1443 | } |
1136 | |
1444 | |
1137 | static void * |
1445 | static void * |
1138 | ev_realloc_emul (void *ptr, long size) EV_THROW |
1446 | ev_realloc_emul (void *ptr, long size) EV_THROW |
1139 | { |
1447 | { |
1140 | #if __GLIBC__ |
|
|
1141 | return realloc (ptr, size); |
|
|
1142 | #else |
|
|
1143 | /* some systems, notably openbsd and darwin, fail to properly |
1448 | /* some systems, notably openbsd and darwin, fail to properly |
1144 | * implement realloc (x, 0) (as required by both ansi c-89 and |
1449 | * implement realloc (x, 0) (as required by both ansi c-89 and |
1145 | * the single unix specification, so work around them here. |
1450 | * the single unix specification, so work around them here. |
|
|
1451 | * recently, also (at least) fedora and debian started breaking it, |
|
|
1452 | * despite documenting it otherwise. |
1146 | */ |
1453 | */ |
1147 | |
1454 | |
1148 | if (size) |
1455 | if (size) |
1149 | return realloc (ptr, size); |
1456 | return realloc (ptr, size); |
1150 | |
1457 | |
1151 | free (ptr); |
1458 | free (ptr); |
1152 | return 0; |
1459 | return 0; |
1153 | #endif |
|
|
1154 | } |
1460 | } |
1155 | |
1461 | |
1156 | static void *(*alloc)(void *ptr, long size) EV_THROW = ev_realloc_emul; |
1462 | static void *(*alloc)(void *ptr, long size) EV_THROW = ev_realloc_emul; |
1157 | |
1463 | |
1158 | void ecb_cold |
1464 | void ecb_cold |
… | |
… | |
1813 | static void noinline ecb_cold |
2119 | static void noinline ecb_cold |
1814 | evpipe_init (EV_P) |
2120 | evpipe_init (EV_P) |
1815 | { |
2121 | { |
1816 | if (!ev_is_active (&pipe_w)) |
2122 | if (!ev_is_active (&pipe_w)) |
1817 | { |
2123 | { |
|
|
2124 | int fds [2]; |
|
|
2125 | |
1818 | # if EV_USE_EVENTFD |
2126 | # if EV_USE_EVENTFD |
|
|
2127 | fds [0] = -1; |
1819 | evfd = eventfd (0, EFD_NONBLOCK | EFD_CLOEXEC); |
2128 | fds [1] = eventfd (0, EFD_NONBLOCK | EFD_CLOEXEC); |
1820 | if (evfd < 0 && errno == EINVAL) |
2129 | if (fds [1] < 0 && errno == EINVAL) |
1821 | evfd = eventfd (0, 0); |
2130 | fds [1] = eventfd (0, 0); |
1822 | |
2131 | |
1823 | if (evfd >= 0) |
2132 | if (fds [1] < 0) |
1824 | { |
|
|
1825 | evpipe [0] = -1; |
|
|
1826 | fd_intern (evfd); /* doing it twice doesn't hurt */ |
|
|
1827 | ev_io_set (&pipe_w, evfd, EV_READ); |
|
|
1828 | } |
|
|
1829 | else |
|
|
1830 | # endif |
2133 | # endif |
1831 | { |
2134 | { |
1832 | while (pipe (evpipe)) |
2135 | while (pipe (fds)) |
1833 | ev_syserr ("(libev) error creating signal/async pipe"); |
2136 | ev_syserr ("(libev) error creating signal/async pipe"); |
1834 | |
2137 | |
1835 | fd_intern (evpipe [0]); |
2138 | fd_intern (fds [0]); |
1836 | fd_intern (evpipe [1]); |
|
|
1837 | ev_io_set (&pipe_w, evpipe [0], EV_READ); |
|
|
1838 | } |
2139 | } |
1839 | |
2140 | |
|
|
2141 | evpipe [0] = fds [0]; |
|
|
2142 | |
|
|
2143 | if (evpipe [1] < 0) |
|
|
2144 | evpipe [1] = fds [1]; /* first call, set write fd */ |
|
|
2145 | else |
|
|
2146 | { |
|
|
2147 | /* on subsequent calls, do not change evpipe [1] */ |
|
|
2148 | /* so that evpipe_write can always rely on its value. */ |
|
|
2149 | /* this branch does not do anything sensible on windows, */ |
|
|
2150 | /* so must not be executed on windows */ |
|
|
2151 | |
|
|
2152 | dup2 (fds [1], evpipe [1]); |
|
|
2153 | close (fds [1]); |
|
|
2154 | } |
|
|
2155 | |
|
|
2156 | fd_intern (evpipe [1]); |
|
|
2157 | |
|
|
2158 | ev_io_set (&pipe_w, evpipe [0] < 0 ? evpipe [1] : evpipe [0], EV_READ); |
1840 | ev_io_start (EV_A_ &pipe_w); |
2159 | ev_io_start (EV_A_ &pipe_w); |
1841 | ev_unref (EV_A); /* watcher should not keep loop alive */ |
2160 | ev_unref (EV_A); /* watcher should not keep loop alive */ |
1842 | } |
2161 | } |
1843 | } |
2162 | } |
1844 | |
2163 | |
… | |
… | |
1849 | |
2168 | |
1850 | if (expect_true (*flag)) |
2169 | if (expect_true (*flag)) |
1851 | return; |
2170 | return; |
1852 | |
2171 | |
1853 | *flag = 1; |
2172 | *flag = 1; |
1854 | |
|
|
1855 | ECB_MEMORY_FENCE_RELEASE; /* make sure flag is visible before the wakeup */ |
2173 | ECB_MEMORY_FENCE_RELEASE; /* make sure flag is visible before the wakeup */ |
1856 | |
2174 | |
1857 | pipe_write_skipped = 1; |
2175 | pipe_write_skipped = 1; |
1858 | |
2176 | |
1859 | ECB_MEMORY_FENCE; /* make sure pipe_write_skipped is visible before we check pipe_write_wanted */ |
2177 | ECB_MEMORY_FENCE; /* make sure pipe_write_skipped is visible before we check pipe_write_wanted */ |
1860 | |
2178 | |
1861 | if (pipe_write_wanted) |
2179 | if (pipe_write_wanted) |
1862 | { |
2180 | { |
1863 | int old_errno; |
2181 | int old_errno; |
1864 | |
2182 | |
1865 | pipe_write_skipped = 0; /* just an optimisation, no fence needed */ |
2183 | pipe_write_skipped = 0; |
|
|
2184 | ECB_MEMORY_FENCE_RELEASE; |
1866 | |
2185 | |
1867 | old_errno = errno; /* save errno because write will clobber it */ |
2186 | old_errno = errno; /* save errno because write will clobber it */ |
1868 | |
2187 | |
1869 | #if EV_USE_EVENTFD |
2188 | #if EV_USE_EVENTFD |
1870 | if (evfd >= 0) |
2189 | if (evpipe [0] < 0) |
1871 | { |
2190 | { |
1872 | uint64_t counter = 1; |
2191 | uint64_t counter = 1; |
1873 | write (evfd, &counter, sizeof (uint64_t)); |
2192 | write (evpipe [1], &counter, sizeof (uint64_t)); |
1874 | } |
2193 | } |
1875 | else |
2194 | else |
1876 | #endif |
2195 | #endif |
1877 | { |
2196 | { |
1878 | #ifdef _WIN32 |
2197 | #ifdef _WIN32 |
… | |
… | |
1898 | int i; |
2217 | int i; |
1899 | |
2218 | |
1900 | if (revents & EV_READ) |
2219 | if (revents & EV_READ) |
1901 | { |
2220 | { |
1902 | #if EV_USE_EVENTFD |
2221 | #if EV_USE_EVENTFD |
1903 | if (evfd >= 0) |
2222 | if (evpipe [0] < 0) |
1904 | { |
2223 | { |
1905 | uint64_t counter; |
2224 | uint64_t counter; |
1906 | read (evfd, &counter, sizeof (uint64_t)); |
2225 | read (evpipe [1], &counter, sizeof (uint64_t)); |
1907 | } |
2226 | } |
1908 | else |
2227 | else |
1909 | #endif |
2228 | #endif |
1910 | { |
2229 | { |
1911 | char dummy[4]; |
2230 | char dummy[4]; |
… | |
… | |
1929 | #if EV_SIGNAL_ENABLE |
2248 | #if EV_SIGNAL_ENABLE |
1930 | if (sig_pending) |
2249 | if (sig_pending) |
1931 | { |
2250 | { |
1932 | sig_pending = 0; |
2251 | sig_pending = 0; |
1933 | |
2252 | |
1934 | ECB_MEMORY_FENCE_RELEASE; |
2253 | ECB_MEMORY_FENCE; |
1935 | |
2254 | |
1936 | for (i = EV_NSIG - 1; i--; ) |
2255 | for (i = EV_NSIG - 1; i--; ) |
1937 | if (expect_false (signals [i].pending)) |
2256 | if (expect_false (signals [i].pending)) |
1938 | ev_feed_signal_event (EV_A_ i + 1); |
2257 | ev_feed_signal_event (EV_A_ i + 1); |
1939 | } |
2258 | } |
… | |
… | |
1942 | #if EV_ASYNC_ENABLE |
2261 | #if EV_ASYNC_ENABLE |
1943 | if (async_pending) |
2262 | if (async_pending) |
1944 | { |
2263 | { |
1945 | async_pending = 0; |
2264 | async_pending = 0; |
1946 | |
2265 | |
1947 | ECB_MEMORY_FENCE_RELEASE; |
2266 | ECB_MEMORY_FENCE; |
1948 | |
2267 | |
1949 | for (i = asynccnt; i--; ) |
2268 | for (i = asynccnt; i--; ) |
1950 | if (asyncs [i]->sent) |
2269 | if (asyncs [i]->sent) |
1951 | { |
2270 | { |
1952 | asyncs [i]->sent = 0; |
2271 | asyncs [i]->sent = 0; |
|
|
2272 | ECB_MEMORY_FENCE_RELEASE; |
1953 | ev_feed_event (EV_A_ asyncs [i], EV_ASYNC); |
2273 | ev_feed_event (EV_A_ asyncs [i], EV_ASYNC); |
1954 | } |
2274 | } |
1955 | } |
2275 | } |
1956 | #endif |
2276 | #endif |
1957 | } |
2277 | } |
… | |
… | |
1960 | |
2280 | |
1961 | void |
2281 | void |
1962 | ev_feed_signal (int signum) EV_THROW |
2282 | ev_feed_signal (int signum) EV_THROW |
1963 | { |
2283 | { |
1964 | #if EV_MULTIPLICITY |
2284 | #if EV_MULTIPLICITY |
|
|
2285 | EV_P; |
|
|
2286 | ECB_MEMORY_FENCE_ACQUIRE; |
1965 | EV_P = signals [signum - 1].loop; |
2287 | EV_A = signals [signum - 1].loop; |
1966 | |
2288 | |
1967 | if (!EV_A) |
2289 | if (!EV_A) |
1968 | return; |
2290 | return; |
1969 | #endif |
2291 | #endif |
1970 | |
2292 | |
1971 | if (!ev_active (&pipe_w)) |
|
|
1972 | return; |
|
|
1973 | |
|
|
1974 | signals [signum - 1].pending = 1; |
2293 | signals [signum - 1].pending = 1; |
1975 | evpipe_write (EV_A_ &sig_pending); |
2294 | evpipe_write (EV_A_ &sig_pending); |
1976 | } |
2295 | } |
1977 | |
2296 | |
1978 | static void |
2297 | static void |
… | |
… | |
1988 | void noinline |
2307 | void noinline |
1989 | ev_feed_signal_event (EV_P_ int signum) EV_THROW |
2308 | ev_feed_signal_event (EV_P_ int signum) EV_THROW |
1990 | { |
2309 | { |
1991 | WL w; |
2310 | WL w; |
1992 | |
2311 | |
1993 | if (expect_false (signum <= 0 || signum > EV_NSIG)) |
2312 | if (expect_false (signum <= 0 || signum >= EV_NSIG)) |
1994 | return; |
2313 | return; |
1995 | |
2314 | |
1996 | --signum; |
2315 | --signum; |
1997 | |
2316 | |
1998 | #if EV_MULTIPLICITY |
2317 | #if EV_MULTIPLICITY |
… | |
… | |
2002 | if (expect_false (signals [signum].loop != EV_A)) |
2321 | if (expect_false (signals [signum].loop != EV_A)) |
2003 | return; |
2322 | return; |
2004 | #endif |
2323 | #endif |
2005 | |
2324 | |
2006 | signals [signum].pending = 0; |
2325 | signals [signum].pending = 0; |
|
|
2326 | ECB_MEMORY_FENCE_RELEASE; |
2007 | |
2327 | |
2008 | for (w = signals [signum].head; w; w = w->next) |
2328 | for (w = signals [signum].head; w; w = w->next) |
2009 | ev_feed_event (EV_A_ (W)w, EV_SIGNAL); |
2329 | ev_feed_event (EV_A_ (W)w, EV_SIGNAL); |
2010 | } |
2330 | } |
2011 | |
2331 | |
… | |
… | |
2225 | { |
2545 | { |
2226 | return userdata; |
2546 | return userdata; |
2227 | } |
2547 | } |
2228 | |
2548 | |
2229 | void |
2549 | void |
2230 | ev_set_invoke_pending_cb (EV_P_ void (*invoke_pending_cb)(EV_P)) EV_THROW |
2550 | ev_set_invoke_pending_cb (EV_P_ ev_loop_callback invoke_pending_cb) EV_THROW |
2231 | { |
2551 | { |
2232 | invoke_cb = invoke_pending_cb; |
2552 | invoke_cb = invoke_pending_cb; |
2233 | } |
2553 | } |
2234 | |
2554 | |
2235 | void |
2555 | void |
2236 | ev_set_loop_release_cb (EV_P_ void (*release)(EV_P) EV_THROW, void (*acquire)(EV_P) EV_THROW) EV_THROW |
2556 | ev_set_loop_release_cb (EV_P_ ev_loop_callback_nothrow release, ev_loop_callback_nothrow acquire) EV_THROW |
2237 | { |
2557 | { |
2238 | release_cb = release; |
2558 | release_cb = release; |
2239 | acquire_cb = acquire; |
2559 | acquire_cb = acquire; |
2240 | } |
2560 | } |
2241 | #endif |
2561 | #endif |
… | |
… | |
2295 | #if EV_ASYNC_ENABLE |
2615 | #if EV_ASYNC_ENABLE |
2296 | async_pending = 0; |
2616 | async_pending = 0; |
2297 | #endif |
2617 | #endif |
2298 | pipe_write_skipped = 0; |
2618 | pipe_write_skipped = 0; |
2299 | pipe_write_wanted = 0; |
2619 | pipe_write_wanted = 0; |
|
|
2620 | evpipe [0] = -1; |
|
|
2621 | evpipe [1] = -1; |
2300 | #if EV_USE_INOTIFY |
2622 | #if EV_USE_INOTIFY |
2301 | fs_fd = flags & EVFLAG_NOINOTIFY ? -1 : -2; |
2623 | fs_fd = flags & EVFLAG_NOINOTIFY ? -1 : -2; |
2302 | #endif |
2624 | #endif |
2303 | #if EV_USE_SIGNALFD |
2625 | #if EV_USE_SIGNALFD |
2304 | sigfd = flags & EVFLAG_SIGNALFD ? -2 : -1; |
2626 | sigfd = flags & EVFLAG_SIGNALFD ? -2 : -1; |
… | |
… | |
2367 | if (ev_is_active (&pipe_w)) |
2689 | if (ev_is_active (&pipe_w)) |
2368 | { |
2690 | { |
2369 | /*ev_ref (EV_A);*/ |
2691 | /*ev_ref (EV_A);*/ |
2370 | /*ev_io_stop (EV_A_ &pipe_w);*/ |
2692 | /*ev_io_stop (EV_A_ &pipe_w);*/ |
2371 | |
2693 | |
2372 | #if EV_USE_EVENTFD |
|
|
2373 | if (evfd >= 0) |
|
|
2374 | close (evfd); |
|
|
2375 | #endif |
|
|
2376 | |
|
|
2377 | if (evpipe [0] >= 0) |
|
|
2378 | { |
|
|
2379 | EV_WIN32_CLOSE_FD (evpipe [0]); |
2694 | if (evpipe [0] >= 0) EV_WIN32_CLOSE_FD (evpipe [0]); |
2380 | EV_WIN32_CLOSE_FD (evpipe [1]); |
2695 | if (evpipe [1] >= 0) EV_WIN32_CLOSE_FD (evpipe [1]); |
2381 | } |
|
|
2382 | } |
2696 | } |
2383 | |
2697 | |
2384 | #if EV_USE_SIGNALFD |
2698 | #if EV_USE_SIGNALFD |
2385 | if (ev_is_active (&sigfd_w)) |
2699 | if (ev_is_active (&sigfd_w)) |
2386 | close (sigfd); |
2700 | close (sigfd); |
… | |
… | |
2472 | #endif |
2786 | #endif |
2473 | #if EV_USE_INOTIFY |
2787 | #if EV_USE_INOTIFY |
2474 | infy_fork (EV_A); |
2788 | infy_fork (EV_A); |
2475 | #endif |
2789 | #endif |
2476 | |
2790 | |
|
|
2791 | #if EV_SIGNAL_ENABLE || EV_ASYNC_ENABLE |
2477 | if (ev_is_active (&pipe_w)) |
2792 | if (ev_is_active (&pipe_w)) |
2478 | { |
2793 | { |
2479 | /* pipe_write_wanted must be false now, so modifying fd vars should be safe */ |
2794 | /* pipe_write_wanted must be false now, so modifying fd vars should be safe */ |
2480 | |
2795 | |
2481 | ev_ref (EV_A); |
2796 | ev_ref (EV_A); |
2482 | ev_io_stop (EV_A_ &pipe_w); |
2797 | ev_io_stop (EV_A_ &pipe_w); |
2483 | |
2798 | |
2484 | #if EV_USE_EVENTFD |
|
|
2485 | if (evfd >= 0) |
|
|
2486 | close (evfd); |
|
|
2487 | #endif |
|
|
2488 | |
|
|
2489 | if (evpipe [0] >= 0) |
2799 | if (evpipe [0] >= 0) |
2490 | { |
|
|
2491 | EV_WIN32_CLOSE_FD (evpipe [0]); |
2800 | EV_WIN32_CLOSE_FD (evpipe [0]); |
2492 | EV_WIN32_CLOSE_FD (evpipe [1]); |
|
|
2493 | } |
|
|
2494 | |
2801 | |
2495 | #if EV_SIGNAL_ENABLE || EV_ASYNC_ENABLE |
|
|
2496 | evpipe_init (EV_A); |
2802 | evpipe_init (EV_A); |
2497 | /* now iterate over everything, in case we missed something */ |
2803 | /* iterate over everything, in case we missed something before */ |
2498 | pipecb (EV_A_ &pipe_w, EV_READ); |
2804 | ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); |
2499 | #endif |
|
|
2500 | } |
2805 | } |
|
|
2806 | #endif |
2501 | |
2807 | |
2502 | postfork = 0; |
2808 | postfork = 0; |
2503 | } |
2809 | } |
2504 | |
2810 | |
2505 | #if EV_MULTIPLICITY |
2811 | #if EV_MULTIPLICITY |
… | |
… | |
2678 | } |
2984 | } |
2679 | |
2985 | |
2680 | void |
2986 | void |
2681 | ev_loop_fork (EV_P) EV_THROW |
2987 | ev_loop_fork (EV_P) EV_THROW |
2682 | { |
2988 | { |
2683 | postfork = 1; /* must be in line with ev_default_fork */ |
2989 | postfork = 1; |
2684 | } |
2990 | } |
2685 | |
2991 | |
2686 | /*****************************************************************************/ |
2992 | /*****************************************************************************/ |
2687 | |
2993 | |
2688 | void |
2994 | void |
… | |
… | |
2704 | } |
3010 | } |
2705 | |
3011 | |
2706 | void noinline |
3012 | void noinline |
2707 | ev_invoke_pending (EV_P) |
3013 | ev_invoke_pending (EV_P) |
2708 | { |
3014 | { |
2709 | for (pendingpri = NUMPRI; pendingpri--; ) /* pendingpri is modified during the loop */ |
3015 | pendingpri = NUMPRI; |
|
|
3016 | |
|
|
3017 | while (pendingpri) /* pendingpri possibly gets modified in the inner loop */ |
|
|
3018 | { |
|
|
3019 | --pendingpri; |
|
|
3020 | |
2710 | while (pendingcnt [pendingpri]) |
3021 | while (pendingcnt [pendingpri]) |
2711 | { |
3022 | { |
2712 | ANPENDING *p = pendings [pendingpri] + --pendingcnt [pendingpri]; |
3023 | ANPENDING *p = pendings [pendingpri] + --pendingcnt [pendingpri]; |
2713 | |
3024 | |
2714 | p->w->pending = 0; |
3025 | p->w->pending = 0; |
2715 | EV_CB_INVOKE (p->w, p->events); |
3026 | EV_CB_INVOKE (p->w, p->events); |
2716 | EV_FREQUENT_CHECK; |
3027 | EV_FREQUENT_CHECK; |
2717 | } |
3028 | } |
|
|
3029 | } |
2718 | } |
3030 | } |
2719 | |
3031 | |
2720 | #if EV_IDLE_ENABLE |
3032 | #if EV_IDLE_ENABLE |
2721 | /* make idle watchers pending. this handles the "call-idle */ |
3033 | /* make idle watchers pending. this handles the "call-idle */ |
2722 | /* only when higher priorities are idle" logic */ |
3034 | /* only when higher priorities are idle" logic */ |
… | |
… | |
3080 | backend_poll (EV_A_ waittime); |
3392 | backend_poll (EV_A_ waittime); |
3081 | assert ((loop_done = EVBREAK_CANCEL, 1)); /* assert for side effect */ |
3393 | assert ((loop_done = EVBREAK_CANCEL, 1)); /* assert for side effect */ |
3082 | |
3394 | |
3083 | pipe_write_wanted = 0; /* just an optimisation, no fence needed */ |
3395 | pipe_write_wanted = 0; /* just an optimisation, no fence needed */ |
3084 | |
3396 | |
|
|
3397 | ECB_MEMORY_FENCE_ACQUIRE; |
3085 | if (pipe_write_skipped) |
3398 | if (pipe_write_skipped) |
3086 | { |
3399 | { |
3087 | assert (("libev: pipe_w not active, but pipe not written", ev_is_active (&pipe_w))); |
3400 | assert (("libev: pipe_w not active, but pipe not written", ev_is_active (&pipe_w))); |
3088 | ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); |
3401 | ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); |
3089 | } |
3402 | } |
… | |
… | |
3467 | #if EV_MULTIPLICITY |
3780 | #if EV_MULTIPLICITY |
3468 | assert (("libev: a signal must not be attached to two different loops", |
3781 | assert (("libev: a signal must not be attached to two different loops", |
3469 | !signals [w->signum - 1].loop || signals [w->signum - 1].loop == loop)); |
3782 | !signals [w->signum - 1].loop || signals [w->signum - 1].loop == loop)); |
3470 | |
3783 | |
3471 | signals [w->signum - 1].loop = EV_A; |
3784 | signals [w->signum - 1].loop = EV_A; |
|
|
3785 | ECB_MEMORY_FENCE_RELEASE; |
3472 | #endif |
3786 | #endif |
3473 | |
3787 | |
3474 | EV_FREQUENT_CHECK; |
3788 | EV_FREQUENT_CHECK; |
3475 | |
3789 | |
3476 | #if EV_USE_SIGNALFD |
3790 | #if EV_USE_SIGNALFD |
… | |
… | |
3631 | # define EV_INOTIFY_BUFSIZE (sizeof (struct inotify_event) * 2 + NAME_MAX) |
3945 | # define EV_INOTIFY_BUFSIZE (sizeof (struct inotify_event) * 2 + NAME_MAX) |
3632 | |
3946 | |
3633 | static void noinline |
3947 | static void noinline |
3634 | infy_add (EV_P_ ev_stat *w) |
3948 | infy_add (EV_P_ ev_stat *w) |
3635 | { |
3949 | { |
3636 | w->wd = inotify_add_watch (fs_fd, w->path, IN_ATTRIB | IN_DELETE_SELF | IN_MOVE_SELF | IN_MODIFY | IN_DONT_FOLLOW | IN_MASK_ADD); |
3950 | w->wd = inotify_add_watch (fs_fd, w->path, |
|
|
3951 | IN_ATTRIB | IN_DELETE_SELF | IN_MOVE_SELF | IN_MODIFY |
|
|
3952 | | IN_CREATE | IN_DELETE | IN_MOVED_FROM | IN_MOVED_TO |
|
|
3953 | | IN_DONT_FOLLOW | IN_MASK_ADD); |
3637 | |
3954 | |
3638 | if (w->wd >= 0) |
3955 | if (w->wd >= 0) |
3639 | { |
3956 | { |
3640 | struct statfs sfs; |
3957 | struct statfs sfs; |
3641 | |
3958 | |
… | |
… | |
3645 | |
3962 | |
3646 | if (!fs_2625) |
3963 | if (!fs_2625) |
3647 | w->timer.repeat = w->interval ? w->interval : DEF_STAT_INTERVAL; |
3964 | w->timer.repeat = w->interval ? w->interval : DEF_STAT_INTERVAL; |
3648 | else if (!statfs (w->path, &sfs) |
3965 | else if (!statfs (w->path, &sfs) |
3649 | && (sfs.f_type == 0x1373 /* devfs */ |
3966 | && (sfs.f_type == 0x1373 /* devfs */ |
|
|
3967 | || sfs.f_type == 0x4006 /* fat */ |
|
|
3968 | || sfs.f_type == 0x4d44 /* msdos */ |
3650 | || sfs.f_type == 0xEF53 /* ext2/3 */ |
3969 | || sfs.f_type == 0xEF53 /* ext2/3 */ |
|
|
3970 | || sfs.f_type == 0x72b6 /* jffs2 */ |
|
|
3971 | || sfs.f_type == 0x858458f6 /* ramfs */ |
|
|
3972 | || sfs.f_type == 0x5346544e /* ntfs */ |
3651 | || sfs.f_type == 0x3153464a /* jfs */ |
3973 | || sfs.f_type == 0x3153464a /* jfs */ |
|
|
3974 | || sfs.f_type == 0x9123683e /* btrfs */ |
3652 | || sfs.f_type == 0x52654973 /* reiser3 */ |
3975 | || sfs.f_type == 0x52654973 /* reiser3 */ |
3653 | || sfs.f_type == 0x01021994 /* tempfs */ |
3976 | || sfs.f_type == 0x01021994 /* tmpfs */ |
3654 | || sfs.f_type == 0x58465342 /* xfs */)) |
3977 | || sfs.f_type == 0x58465342 /* xfs */)) |
3655 | w->timer.repeat = 0.; /* filesystem is local, kernel new enough */ |
3978 | w->timer.repeat = 0.; /* filesystem is local, kernel new enough */ |
3656 | else |
3979 | else |
3657 | w->timer.repeat = w->interval ? w->interval : NFS_STAT_INTERVAL; /* remote, use reduced frequency */ |
3980 | w->timer.repeat = w->interval ? w->interval : NFS_STAT_INTERVAL; /* remote, use reduced frequency */ |
3658 | } |
3981 | } |