… | |
… | |
201 | # include <sys/wait.h> |
201 | # include <sys/wait.h> |
202 | # include <unistd.h> |
202 | # include <unistd.h> |
203 | #else |
203 | #else |
204 | # include <io.h> |
204 | # include <io.h> |
205 | # define WIN32_LEAN_AND_MEAN |
205 | # define WIN32_LEAN_AND_MEAN |
|
|
206 | # include <winsock2.h> |
206 | # include <windows.h> |
207 | # include <windows.h> |
207 | # ifndef EV_SELECT_IS_WINSOCKET |
208 | # ifndef EV_SELECT_IS_WINSOCKET |
208 | # define EV_SELECT_IS_WINSOCKET 1 |
209 | # define EV_SELECT_IS_WINSOCKET 1 |
209 | # endif |
210 | # endif |
210 | # undef EV_AVOID_STDIO |
211 | # undef EV_AVOID_STDIO |
… | |
… | |
356 | |
357 | |
357 | #ifndef EV_HEAP_CACHE_AT |
358 | #ifndef EV_HEAP_CACHE_AT |
358 | # define EV_HEAP_CACHE_AT EV_FEATURE_DATA |
359 | # define EV_HEAP_CACHE_AT EV_FEATURE_DATA |
359 | #endif |
360 | #endif |
360 | |
361 | |
|
|
362 | #ifdef ANDROID |
|
|
363 | /* supposedly, android doesn't typedef fd_mask */ |
|
|
364 | # undef EV_USE_SELECT |
|
|
365 | # define EV_USE_SELECT 0 |
|
|
366 | /* supposedly, we need to include syscall.h, not sys/syscall.h, so just disable */ |
|
|
367 | # undef EV_USE_CLOCK_SYSCALL |
|
|
368 | # define EV_USE_CLOCK_SYSCALL 0 |
|
|
369 | #endif |
|
|
370 | |
|
|
371 | /* aix's poll.h seems to cause lots of trouble */ |
|
|
372 | #ifdef _AIX |
|
|
373 | /* AIX has a completely broken poll.h header */ |
|
|
374 | # undef EV_USE_POLL |
|
|
375 | # define EV_USE_POLL 0 |
|
|
376 | #endif |
|
|
377 | |
361 | /* on linux, we can use a (slow) syscall to avoid a dependency on pthread, */ |
378 | /* on linux, we can use a (slow) syscall to avoid a dependency on pthread, */ |
362 | /* which makes programs even slower. might work on other unices, too. */ |
379 | /* which makes programs even slower. might work on other unices, too. */ |
363 | #if EV_USE_CLOCK_SYSCALL |
380 | #if EV_USE_CLOCK_SYSCALL |
364 | # include <syscall.h> |
381 | # include <sys/syscall.h> |
365 | # ifdef SYS_clock_gettime |
382 | # ifdef SYS_clock_gettime |
366 | # define clock_gettime(id, ts) syscall (SYS_clock_gettime, (id), (ts)) |
383 | # define clock_gettime(id, ts) syscall (SYS_clock_gettime, (id), (ts)) |
367 | # undef EV_USE_MONOTONIC |
384 | # undef EV_USE_MONOTONIC |
368 | # define EV_USE_MONOTONIC 1 |
385 | # define EV_USE_MONOTONIC 1 |
369 | # else |
386 | # else |
… | |
… | |
372 | # endif |
389 | # endif |
373 | #endif |
390 | #endif |
374 | |
391 | |
375 | /* this block fixes any misconfiguration where we know we run into trouble otherwise */ |
392 | /* this block fixes any misconfiguration where we know we run into trouble otherwise */ |
376 | |
393 | |
377 | #ifdef _AIX |
|
|
378 | /* AIX has a completely broken poll.h header */ |
|
|
379 | # undef EV_USE_POLL |
|
|
380 | # define EV_USE_POLL 0 |
|
|
381 | #endif |
|
|
382 | |
|
|
383 | #ifndef CLOCK_MONOTONIC |
394 | #ifndef CLOCK_MONOTONIC |
384 | # undef EV_USE_MONOTONIC |
395 | # undef EV_USE_MONOTONIC |
385 | # define EV_USE_MONOTONIC 0 |
396 | # define EV_USE_MONOTONIC 0 |
386 | #endif |
397 | #endif |
387 | |
398 | |
… | |
… | |
408 | /* some very old inotify.h headers don't have IN_DONT_FOLLOW */ |
419 | /* some very old inotify.h headers don't have IN_DONT_FOLLOW */ |
409 | # ifndef IN_DONT_FOLLOW |
420 | # ifndef IN_DONT_FOLLOW |
410 | # undef EV_USE_INOTIFY |
421 | # undef EV_USE_INOTIFY |
411 | # define EV_USE_INOTIFY 0 |
422 | # define EV_USE_INOTIFY 0 |
412 | # endif |
423 | # endif |
413 | #endif |
|
|
414 | |
|
|
415 | #if EV_SELECT_IS_WINSOCKET |
|
|
416 | # include <winsock.h> |
|
|
417 | #endif |
424 | #endif |
418 | |
425 | |
419 | #if EV_USE_EVENTFD |
426 | #if EV_USE_EVENTFD |
420 | /* our minimum requirement is glibc 2.7 which has the stub, but not the header */ |
427 | /* our minimum requirement is glibc 2.7 which has the stub, but not the header */ |
421 | # include <stdint.h> |
428 | # include <stdint.h> |
… | |
… | |
507 | */ |
514 | */ |
508 | |
515 | |
509 | #ifndef ECB_H |
516 | #ifndef ECB_H |
510 | #define ECB_H |
517 | #define ECB_H |
511 | |
518 | |
|
|
519 | /* 16 bits major, 16 bits minor */ |
|
|
520 | #define ECB_VERSION 0x00010003 |
|
|
521 | |
512 | #ifdef _WIN32 |
522 | #ifdef _WIN32 |
513 | typedef signed char int8_t; |
523 | typedef signed char int8_t; |
514 | typedef unsigned char uint8_t; |
524 | typedef unsigned char uint8_t; |
515 | typedef signed short int16_t; |
525 | typedef signed short int16_t; |
516 | typedef unsigned short uint16_t; |
526 | typedef unsigned short uint16_t; |
… | |
… | |
521 | typedef unsigned long long uint64_t; |
531 | typedef unsigned long long uint64_t; |
522 | #else /* _MSC_VER || __BORLANDC__ */ |
532 | #else /* _MSC_VER || __BORLANDC__ */ |
523 | typedef signed __int64 int64_t; |
533 | typedef signed __int64 int64_t; |
524 | typedef unsigned __int64 uint64_t; |
534 | typedef unsigned __int64 uint64_t; |
525 | #endif |
535 | #endif |
|
|
536 | #ifdef _WIN64 |
|
|
537 | #define ECB_PTRSIZE 8 |
|
|
538 | typedef uint64_t uintptr_t; |
|
|
539 | typedef int64_t intptr_t; |
|
|
540 | #else |
|
|
541 | #define ECB_PTRSIZE 4 |
|
|
542 | typedef uint32_t uintptr_t; |
|
|
543 | typedef int32_t intptr_t; |
|
|
544 | #endif |
526 | #else |
545 | #else |
527 | #include <inttypes.h> |
546 | #include <inttypes.h> |
|
|
547 | #if UINTMAX_MAX > 0xffffffffU |
|
|
548 | #define ECB_PTRSIZE 8 |
|
|
549 | #else |
|
|
550 | #define ECB_PTRSIZE 4 |
|
|
551 | #endif |
|
|
552 | #endif |
|
|
553 | |
|
|
554 | /* work around x32 idiocy by defining proper macros */ |
|
|
555 | #if __x86_64 || _M_AMD64 |
|
|
556 | #if __ILP32 |
|
|
557 | #define ECB_AMD64_X32 1 |
|
|
558 | #else |
|
|
559 | #define ECB_AMD64 1 |
|
|
560 | #endif |
528 | #endif |
561 | #endif |
529 | |
562 | |
530 | /* many compilers define _GNUC_ to some versions but then only implement |
563 | /* many compilers define _GNUC_ to some versions but then only implement |
531 | * what their idiot authors think are the "more important" extensions, |
564 | * what their idiot authors think are the "more important" extensions, |
532 | * causing enormous grief in return for some better fake benchmark numbers. |
565 | * causing enormous grief in return for some better fake benchmark numbers. |
… | |
… | |
540 | #else |
573 | #else |
541 | #define ECB_GCC_VERSION(major,minor) (__GNUC__ > (major) || (__GNUC__ == (major) && __GNUC_MINOR__ >= (minor))) |
574 | #define ECB_GCC_VERSION(major,minor) (__GNUC__ > (major) || (__GNUC__ == (major) && __GNUC_MINOR__ >= (minor))) |
542 | #endif |
575 | #endif |
543 | #endif |
576 | #endif |
544 | |
577 | |
|
|
578 | #define ECB_C (__STDC__+0) /* this assumes that __STDC__ is either empty or a number */ |
|
|
579 | #define ECB_C99 (__STDC_VERSION__ >= 199901L) |
|
|
580 | #define ECB_C11 (__STDC_VERSION__ >= 201112L) |
|
|
581 | #define ECB_CPP (__cplusplus+0) |
|
|
582 | #define ECB_CPP11 (__cplusplus >= 201103L) |
|
|
583 | |
|
|
584 | #if ECB_CPP |
|
|
585 | #define ECB_EXTERN_C extern "C" |
|
|
586 | #define ECB_EXTERN_C_BEG ECB_EXTERN_C { |
|
|
587 | #define ECB_EXTERN_C_END } |
|
|
588 | #else |
|
|
589 | #define ECB_EXTERN_C extern |
|
|
590 | #define ECB_EXTERN_C_BEG |
|
|
591 | #define ECB_EXTERN_C_END |
|
|
592 | #endif |
|
|
593 | |
545 | /*****************************************************************************/ |
594 | /*****************************************************************************/ |
546 | |
595 | |
547 | /* ECB_NO_THREADS - ecb is not used by multiple threads, ever */ |
596 | /* ECB_NO_THREADS - ecb is not used by multiple threads, ever */ |
548 | /* ECB_NO_SMP - ecb might be used in multiple threads, but only on a single cpu */ |
597 | /* ECB_NO_SMP - ecb might be used in multiple threads, but only on a single cpu */ |
549 | |
598 | |
550 | #if ECB_NO_THREADS |
599 | #if ECB_NO_THREADS |
551 | # define ECB_NO_SMP 1 |
600 | #define ECB_NO_SMP 1 |
552 | #endif |
601 | #endif |
553 | |
602 | |
554 | #if ECB_NO_THREADS || ECB_NO_SMP |
603 | #if ECB_NO_SMP |
555 | #define ECB_MEMORY_FENCE do { } while (0) |
604 | #define ECB_MEMORY_FENCE do { } while (0) |
556 | #endif |
605 | #endif |
557 | |
606 | |
558 | #ifndef ECB_MEMORY_FENCE |
607 | #ifndef ECB_MEMORY_FENCE |
559 | #if ECB_GCC_VERSION(2,5) || defined __INTEL_COMPILER || (__llvm__ && __GNUC__) || __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
608 | #if ECB_GCC_VERSION(2,5) || defined __INTEL_COMPILER || (__llvm__ && __GNUC__) || __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
560 | #if __i386 || __i386__ |
609 | #if __i386 || __i386__ |
561 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
610 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
562 | #define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE /* non-lock xchg might be enough */ |
611 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory") |
563 | #define ECB_MEMORY_FENCE_RELEASE do { } while (0) /* unlikely to change in future cpus */ |
612 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
564 | #elif __amd64 || __amd64__ || __x86_64 || __x86_64__ |
613 | #elif __amd64 || __amd64__ || __x86_64 || __x86_64__ |
565 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
614 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
566 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("lfence" : : : "memory") |
615 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory") |
567 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("sfence") /* play safe - not needed in any current cpu */ |
616 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
568 | #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ |
617 | #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ |
569 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
618 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
570 | #elif defined __ARM_ARCH_6__ || defined __ARM_ARCH_6J__ \ |
619 | #elif defined __ARM_ARCH_6__ || defined __ARM_ARCH_6J__ \ |
571 | || defined __ARM_ARCH_6K__ || defined __ARM_ARCH_6ZK__ |
620 | || defined __ARM_ARCH_6K__ || defined __ARM_ARCH_6ZK__ |
572 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,5" : : "r" (0) : "memory") |
621 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,5" : : "r" (0) : "memory") |
573 | #elif defined __ARM_ARCH_7__ || defined __ARM_ARCH_7A__ \ |
622 | #elif defined __ARM_ARCH_7__ || defined __ARM_ARCH_7A__ \ |
574 | || defined __ARM_ARCH_7M__ || defined __ARM_ARCH_7R__ |
623 | || defined __ARM_ARCH_7M__ || defined __ARM_ARCH_7R__ |
575 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") |
624 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") |
576 | #elif __sparc || __sparc__ |
625 | #elif __sparc || __sparc__ |
577 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad | " : : : "memory") |
626 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad" : : : "memory") |
578 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory") |
627 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory") |
579 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore") |
628 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore") |
580 | #elif defined __s390__ || defined __s390x__ |
629 | #elif defined __s390__ || defined __s390x__ |
581 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory") |
630 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory") |
582 | #elif defined __mips__ |
631 | #elif defined __mips__ |
|
|
632 | /* GNU/Linux emulates sync on mips1 architectures, so we force it's use */ |
|
|
633 | /* anybody else who still uses mips1 is supposed to send in their version, with detection code. */ |
583 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
634 | #define ECB_MEMORY_FENCE __asm__ __volatile__ (".set mips2; sync; .set mips0" : : : "memory") |
584 | #elif defined __alpha__ |
635 | #elif defined __alpha__ |
585 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mb" : : : "memory") |
636 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mb" : : : "memory") |
|
|
637 | #elif defined __hppa__ |
|
|
638 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("" : : : "memory") |
|
|
639 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("") |
|
|
640 | #elif defined __ia64__ |
|
|
641 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mf" : : : "memory") |
|
|
642 | #elif defined __m68k__ |
|
|
643 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("" : : : "memory") |
|
|
644 | #elif defined __m88k__ |
|
|
645 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("tb1 0,%%r0,128" : : : "memory") |
|
|
646 | #elif defined __sh__ |
|
|
647 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("" : : : "memory") |
586 | #endif |
648 | #endif |
587 | #endif |
649 | #endif |
588 | #endif |
650 | #endif |
589 | |
651 | |
590 | #ifndef ECB_MEMORY_FENCE |
652 | #ifndef ECB_MEMORY_FENCE |
|
|
653 | #if ECB_GCC_VERSION(4,7) |
|
|
654 | /* see comment below (stdatomic.h) about the C11 memory model. */ |
|
|
655 | #define ECB_MEMORY_FENCE __atomic_thread_fence (__ATOMIC_SEQ_CST) |
|
|
656 | |
|
|
657 | /* The __has_feature syntax from clang is so misdesigned that we cannot use it |
|
|
658 | * without risking compile time errors with other compilers. We *could* |
|
|
659 | * define our own ecb_clang_has_feature, but I just can't be bothered to work |
|
|
660 | * around this shit time and again. |
|
|
661 | * #elif defined __clang && __has_feature (cxx_atomic) |
|
|
662 | * // see comment below (stdatomic.h) about the C11 memory model. |
|
|
663 | * #define ECB_MEMORY_FENCE __c11_atomic_thread_fence (__ATOMIC_SEQ_CST) |
|
|
664 | */ |
|
|
665 | |
591 | #if ECB_GCC_VERSION(4,4) || defined __INTEL_COMPILER || defined __clang__ |
666 | #elif ECB_GCC_VERSION(4,4) || defined __INTEL_COMPILER || defined __clang__ |
592 | #define ECB_MEMORY_FENCE __sync_synchronize () |
667 | #define ECB_MEMORY_FENCE __sync_synchronize () |
593 | /*#define ECB_MEMORY_FENCE_ACQUIRE ({ char dummy = 0; __sync_lock_test_and_set (&dummy, 1); }) */ |
|
|
594 | /*#define ECB_MEMORY_FENCE_RELEASE ({ char dummy = 1; __sync_lock_release (&dummy ); }) */ |
|
|
595 | #elif _MSC_VER >= 1400 /* VC++ 2005 */ |
668 | #elif _MSC_VER >= 1400 /* VC++ 2005 */ |
596 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
669 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
597 | #define ECB_MEMORY_FENCE _ReadWriteBarrier () |
670 | #define ECB_MEMORY_FENCE _ReadWriteBarrier () |
598 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ |
671 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ |
599 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () |
672 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () |
… | |
… | |
609 | #define ECB_MEMORY_FENCE __sync () |
682 | #define ECB_MEMORY_FENCE __sync () |
610 | #endif |
683 | #endif |
611 | #endif |
684 | #endif |
612 | |
685 | |
613 | #ifndef ECB_MEMORY_FENCE |
686 | #ifndef ECB_MEMORY_FENCE |
|
|
687 | #if ECB_C11 && !defined __STDC_NO_ATOMICS__ |
|
|
688 | /* we assume that these memory fences work on all variables/all memory accesses, */ |
|
|
689 | /* not just C11 atomics and atomic accesses */ |
|
|
690 | #include <stdatomic.h> |
|
|
691 | /* Unfortunately, neither gcc 4.7 nor clang 3.1 generate any instructions for */ |
|
|
692 | /* any fence other than seq_cst, which isn't very efficient for us. */ |
|
|
693 | /* Why that is, we don't know - either the C11 memory model is quite useless */ |
|
|
694 | /* for most usages, or gcc and clang have a bug */ |
|
|
695 | /* I *currently* lean towards the latter, and inefficiently implement */ |
|
|
696 | /* all three of ecb's fences as a seq_cst fence */ |
|
|
697 | #define ECB_MEMORY_FENCE atomic_thread_fence (memory_order_seq_cst) |
|
|
698 | #endif |
|
|
699 | #endif |
|
|
700 | |
|
|
701 | #ifndef ECB_MEMORY_FENCE |
614 | #if !ECB_AVOID_PTHREADS |
702 | #if !ECB_AVOID_PTHREADS |
615 | /* |
703 | /* |
616 | * if you get undefined symbol references to pthread_mutex_lock, |
704 | * if you get undefined symbol references to pthread_mutex_lock, |
617 | * or failure to find pthread.h, then you should implement |
705 | * or failure to find pthread.h, then you should implement |
618 | * the ECB_MEMORY_FENCE operations for your cpu/compiler |
706 | * the ECB_MEMORY_FENCE operations for your cpu/compiler |
… | |
… | |
636 | #define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE |
724 | #define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE |
637 | #endif |
725 | #endif |
638 | |
726 | |
639 | /*****************************************************************************/ |
727 | /*****************************************************************************/ |
640 | |
728 | |
641 | #define ECB_C99 (__STDC_VERSION__ >= 199901L) |
|
|
642 | |
|
|
643 | #if __cplusplus |
729 | #if __cplusplus |
644 | #define ecb_inline static inline |
730 | #define ecb_inline static inline |
645 | #elif ECB_GCC_VERSION(2,5) |
731 | #elif ECB_GCC_VERSION(2,5) |
646 | #define ecb_inline static __inline__ |
732 | #define ecb_inline static __inline__ |
647 | #elif ECB_C99 |
733 | #elif ECB_C99 |
… | |
… | |
685 | #elif ECB_GCC_VERSION(3,0) |
771 | #elif ECB_GCC_VERSION(3,0) |
686 | #define ecb_decltype(x) __typeof(x) |
772 | #define ecb_decltype(x) __typeof(x) |
687 | #endif |
773 | #endif |
688 | |
774 | |
689 | #define ecb_noinline ecb_attribute ((__noinline__)) |
775 | #define ecb_noinline ecb_attribute ((__noinline__)) |
690 | #define ecb_noreturn ecb_attribute ((__noreturn__)) |
|
|
691 | #define ecb_unused ecb_attribute ((__unused__)) |
776 | #define ecb_unused ecb_attribute ((__unused__)) |
692 | #define ecb_const ecb_attribute ((__const__)) |
777 | #define ecb_const ecb_attribute ((__const__)) |
693 | #define ecb_pure ecb_attribute ((__pure__)) |
778 | #define ecb_pure ecb_attribute ((__pure__)) |
|
|
779 | |
|
|
780 | #if ECB_C11 |
|
|
781 | #define ecb_noreturn _Noreturn |
|
|
782 | #else |
|
|
783 | #define ecb_noreturn ecb_attribute ((__noreturn__)) |
|
|
784 | #endif |
694 | |
785 | |
695 | #if ECB_GCC_VERSION(4,3) |
786 | #if ECB_GCC_VERSION(4,3) |
696 | #define ecb_artificial ecb_attribute ((__artificial__)) |
787 | #define ecb_artificial ecb_attribute ((__artificial__)) |
697 | #define ecb_hot ecb_attribute ((__hot__)) |
788 | #define ecb_hot ecb_attribute ((__hot__)) |
698 | #define ecb_cold ecb_attribute ((__cold__)) |
789 | #define ecb_cold ecb_attribute ((__cold__)) |
… | |
… | |
789 | |
880 | |
790 | return r + ecb_ld32 (x); |
881 | return r + ecb_ld32 (x); |
791 | } |
882 | } |
792 | #endif |
883 | #endif |
793 | |
884 | |
|
|
885 | ecb_function_ ecb_bool ecb_is_pot32 (uint32_t x) ecb_const; |
|
|
886 | ecb_function_ ecb_bool ecb_is_pot32 (uint32_t x) { return !(x & (x - 1)); } |
|
|
887 | ecb_function_ ecb_bool ecb_is_pot64 (uint64_t x) ecb_const; |
|
|
888 | ecb_function_ ecb_bool ecb_is_pot64 (uint64_t x) { return !(x & (x - 1)); } |
|
|
889 | |
794 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) ecb_const; |
890 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) ecb_const; |
795 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) |
891 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) |
796 | { |
892 | { |
797 | return ( (x * 0x0802U & 0x22110U) |
893 | return ( (x * 0x0802U & 0x22110U) |
798 | | (x * 0x8020U & 0x88440U)) * 0x10101U >> 16; |
894 | | (x * 0x8020U & 0x88440U)) * 0x10101U >> 16; |
… | |
… | |
882 | ecb_inline void ecb_unreachable (void) ecb_noreturn; |
978 | ecb_inline void ecb_unreachable (void) ecb_noreturn; |
883 | ecb_inline void ecb_unreachable (void) { } |
979 | ecb_inline void ecb_unreachable (void) { } |
884 | #endif |
980 | #endif |
885 | |
981 | |
886 | /* try to tell the compiler that some condition is definitely true */ |
982 | /* try to tell the compiler that some condition is definitely true */ |
887 | #define ecb_assume(cond) do { if (!(cond)) ecb_unreachable (); } while (0) |
983 | #define ecb_assume(cond) if (!(cond)) ecb_unreachable (); else 0 |
888 | |
984 | |
889 | ecb_inline unsigned char ecb_byteorder_helper (void) ecb_const; |
985 | ecb_inline unsigned char ecb_byteorder_helper (void) ecb_const; |
890 | ecb_inline unsigned char |
986 | ecb_inline unsigned char |
891 | ecb_byteorder_helper (void) |
987 | ecb_byteorder_helper (void) |
892 | { |
988 | { |
893 | const uint32_t u = 0x11223344; |
989 | /* the union code still generates code under pressure in gcc, */ |
894 | return *(unsigned char *)&u; |
990 | /* but less than using pointers, and always seems to */ |
|
|
991 | /* successfully return a constant. */ |
|
|
992 | /* the reason why we have this horrible preprocessor mess */ |
|
|
993 | /* is to avoid it in all cases, at least on common architectures */ |
|
|
994 | /* or when using a recent enough gcc version (>= 4.6) */ |
|
|
995 | #if __i386 || __i386__ || _M_X86 || __amd64 || __amd64__ || _M_X64 |
|
|
996 | return 0x44; |
|
|
997 | #elif __BYTE_ORDER__ && __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ |
|
|
998 | return 0x44; |
|
|
999 | #elif __BYTE_ORDER__ && __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ |
|
|
1000 | return 0x11; |
|
|
1001 | #else |
|
|
1002 | union |
|
|
1003 | { |
|
|
1004 | uint32_t i; |
|
|
1005 | uint8_t c; |
|
|
1006 | } u = { 0x11223344 }; |
|
|
1007 | return u.c; |
|
|
1008 | #endif |
895 | } |
1009 | } |
896 | |
1010 | |
897 | ecb_inline ecb_bool ecb_big_endian (void) ecb_const; |
1011 | ecb_inline ecb_bool ecb_big_endian (void) ecb_const; |
898 | ecb_inline ecb_bool ecb_big_endian (void) { return ecb_byteorder_helper () == 0x11; } |
1012 | ecb_inline ecb_bool ecb_big_endian (void) { return ecb_byteorder_helper () == 0x11; } |
899 | ecb_inline ecb_bool ecb_little_endian (void) ecb_const; |
1013 | ecb_inline ecb_bool ecb_little_endian (void) ecb_const; |
… | |
… | |
930 | } |
1044 | } |
931 | #else |
1045 | #else |
932 | #define ecb_array_length(name) (sizeof (name) / sizeof (name [0])) |
1046 | #define ecb_array_length(name) (sizeof (name) / sizeof (name [0])) |
933 | #endif |
1047 | #endif |
934 | |
1048 | |
|
|
1049 | /*******************************************************************************/ |
|
|
1050 | /* floating point stuff, can be disabled by defining ECB_NO_LIBM */ |
|
|
1051 | |
|
|
1052 | /* basically, everything uses "ieee pure-endian" floating point numbers */ |
|
|
1053 | /* the only noteworthy exception is ancient armle, which uses order 43218765 */ |
|
|
1054 | #if 0 \ |
|
|
1055 | || __i386 || __i386__ \ |
|
|
1056 | || __amd64 || __amd64__ || __x86_64 || __x86_64__ \ |
|
|
1057 | || __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ \ |
|
|
1058 | || defined __arm__ && defined __ARM_EABI__ \ |
|
|
1059 | || defined __s390__ || defined __s390x__ \ |
|
|
1060 | || defined __mips__ \ |
|
|
1061 | || defined __alpha__ \ |
|
|
1062 | || defined __hppa__ \ |
|
|
1063 | || defined __ia64__ \ |
|
|
1064 | || defined __m68k__ \ |
|
|
1065 | || defined __m88k__ \ |
|
|
1066 | || defined __sh__ \ |
|
|
1067 | || defined _M_IX86 || defined _M_AMD64 || defined _M_IA64 |
|
|
1068 | #define ECB_STDFP 1 |
|
|
1069 | #include <string.h> /* for memcpy */ |
|
|
1070 | #else |
|
|
1071 | #define ECB_STDFP 0 |
|
|
1072 | #include <math.h> /* for frexp*, ldexp* */ |
|
|
1073 | #endif |
|
|
1074 | |
|
|
1075 | #ifndef ECB_NO_LIBM |
|
|
1076 | |
|
|
1077 | /* convert a float to ieee single/binary32 */ |
|
|
1078 | ecb_function_ uint32_t ecb_float_to_binary32 (float x) ecb_const; |
|
|
1079 | ecb_function_ uint32_t |
|
|
1080 | ecb_float_to_binary32 (float x) |
|
|
1081 | { |
|
|
1082 | uint32_t r; |
|
|
1083 | |
|
|
1084 | #if ECB_STDFP |
|
|
1085 | memcpy (&r, &x, 4); |
|
|
1086 | #else |
|
|
1087 | /* slow emulation, works for anything but -0 */ |
|
|
1088 | uint32_t m; |
|
|
1089 | int e; |
|
|
1090 | |
|
|
1091 | if (x == 0e0f ) return 0x00000000U; |
|
|
1092 | if (x > +3.40282346638528860e+38f) return 0x7f800000U; |
|
|
1093 | if (x < -3.40282346638528860e+38f) return 0xff800000U; |
|
|
1094 | if (x != x ) return 0x7fbfffffU; |
|
|
1095 | |
|
|
1096 | m = frexpf (x, &e) * 0x1000000U; |
|
|
1097 | |
|
|
1098 | r = m & 0x80000000U; |
|
|
1099 | |
|
|
1100 | if (r) |
|
|
1101 | m = -m; |
|
|
1102 | |
|
|
1103 | if (e <= -126) |
|
|
1104 | { |
|
|
1105 | m &= 0xffffffU; |
|
|
1106 | m >>= (-125 - e); |
|
|
1107 | e = -126; |
|
|
1108 | } |
|
|
1109 | |
|
|
1110 | r |= (e + 126) << 23; |
|
|
1111 | r |= m & 0x7fffffU; |
|
|
1112 | #endif |
|
|
1113 | |
|
|
1114 | return r; |
|
|
1115 | } |
|
|
1116 | |
|
|
1117 | /* converts an ieee single/binary32 to a float */ |
|
|
1118 | ecb_function_ float ecb_binary32_to_float (uint32_t x) ecb_const; |
|
|
1119 | ecb_function_ float |
|
|
1120 | ecb_binary32_to_float (uint32_t x) |
|
|
1121 | { |
|
|
1122 | float r; |
|
|
1123 | |
|
|
1124 | #if ECB_STDFP |
|
|
1125 | memcpy (&r, &x, 4); |
|
|
1126 | #else |
|
|
1127 | /* emulation, only works for normals and subnormals and +0 */ |
|
|
1128 | int neg = x >> 31; |
|
|
1129 | int e = (x >> 23) & 0xffU; |
|
|
1130 | |
|
|
1131 | x &= 0x7fffffU; |
|
|
1132 | |
|
|
1133 | if (e) |
|
|
1134 | x |= 0x800000U; |
|
|
1135 | else |
|
|
1136 | e = 1; |
|
|
1137 | |
|
|
1138 | /* we distrust ldexpf a bit and do the 2**-24 scaling by an extra multiply */ |
|
|
1139 | r = ldexpf (x * (0.5f / 0x800000U), e - 126); |
|
|
1140 | |
|
|
1141 | r = neg ? -r : r; |
|
|
1142 | #endif |
|
|
1143 | |
|
|
1144 | return r; |
|
|
1145 | } |
|
|
1146 | |
|
|
1147 | /* convert a double to ieee double/binary64 */ |
|
|
1148 | ecb_function_ uint64_t ecb_double_to_binary64 (double x) ecb_const; |
|
|
1149 | ecb_function_ uint64_t |
|
|
1150 | ecb_double_to_binary64 (double x) |
|
|
1151 | { |
|
|
1152 | uint64_t r; |
|
|
1153 | |
|
|
1154 | #if ECB_STDFP |
|
|
1155 | memcpy (&r, &x, 8); |
|
|
1156 | #else |
|
|
1157 | /* slow emulation, works for anything but -0 */ |
|
|
1158 | uint64_t m; |
|
|
1159 | int e; |
|
|
1160 | |
|
|
1161 | if (x == 0e0 ) return 0x0000000000000000U; |
|
|
1162 | if (x > +1.79769313486231470e+308) return 0x7ff0000000000000U; |
|
|
1163 | if (x < -1.79769313486231470e+308) return 0xfff0000000000000U; |
|
|
1164 | if (x != x ) return 0X7ff7ffffffffffffU; |
|
|
1165 | |
|
|
1166 | m = frexp (x, &e) * 0x20000000000000U; |
|
|
1167 | |
|
|
1168 | r = m & 0x8000000000000000;; |
|
|
1169 | |
|
|
1170 | if (r) |
|
|
1171 | m = -m; |
|
|
1172 | |
|
|
1173 | if (e <= -1022) |
|
|
1174 | { |
|
|
1175 | m &= 0x1fffffffffffffU; |
|
|
1176 | m >>= (-1021 - e); |
|
|
1177 | e = -1022; |
|
|
1178 | } |
|
|
1179 | |
|
|
1180 | r |= ((uint64_t)(e + 1022)) << 52; |
|
|
1181 | r |= m & 0xfffffffffffffU; |
|
|
1182 | #endif |
|
|
1183 | |
|
|
1184 | return r; |
|
|
1185 | } |
|
|
1186 | |
|
|
1187 | /* converts an ieee double/binary64 to a double */ |
|
|
1188 | ecb_function_ double ecb_binary64_to_double (uint64_t x) ecb_const; |
|
|
1189 | ecb_function_ double |
|
|
1190 | ecb_binary64_to_double (uint64_t x) |
|
|
1191 | { |
|
|
1192 | double r; |
|
|
1193 | |
|
|
1194 | #if ECB_STDFP |
|
|
1195 | memcpy (&r, &x, 8); |
|
|
1196 | #else |
|
|
1197 | /* emulation, only works for normals and subnormals and +0 */ |
|
|
1198 | int neg = x >> 63; |
|
|
1199 | int e = (x >> 52) & 0x7ffU; |
|
|
1200 | |
|
|
1201 | x &= 0xfffffffffffffU; |
|
|
1202 | |
|
|
1203 | if (e) |
|
|
1204 | x |= 0x10000000000000U; |
|
|
1205 | else |
|
|
1206 | e = 1; |
|
|
1207 | |
|
|
1208 | /* we distrust ldexp a bit and do the 2**-53 scaling by an extra multiply */ |
|
|
1209 | r = ldexp (x * (0.5 / 0x10000000000000U), e - 1022); |
|
|
1210 | |
|
|
1211 | r = neg ? -r : r; |
|
|
1212 | #endif |
|
|
1213 | |
|
|
1214 | return r; |
|
|
1215 | } |
|
|
1216 | |
|
|
1217 | #endif |
|
|
1218 | |
935 | #endif |
1219 | #endif |
936 | |
1220 | |
937 | /* ECB.H END */ |
1221 | /* ECB.H END */ |
938 | |
1222 | |
939 | #if ECB_MEMORY_FENCE_NEEDS_PTHREADS |
1223 | #if ECB_MEMORY_FENCE_NEEDS_PTHREADS |
… | |
… | |
1108 | #endif |
1392 | #endif |
1109 | |
1393 | |
1110 | static void (*syserr_cb)(const char *msg) EV_THROW; |
1394 | static void (*syserr_cb)(const char *msg) EV_THROW; |
1111 | |
1395 | |
1112 | void ecb_cold |
1396 | void ecb_cold |
1113 | ev_set_syserr_cb (void (*cb)(const char *msg)) EV_THROW |
1397 | ev_set_syserr_cb (void (*cb)(const char *msg) EV_THROW) EV_THROW |
1114 | { |
1398 | { |
1115 | syserr_cb = cb; |
1399 | syserr_cb = cb; |
1116 | } |
1400 | } |
1117 | |
1401 | |
1118 | static void noinline ecb_cold |
1402 | static void noinline ecb_cold |
… | |
… | |
1136 | abort (); |
1420 | abort (); |
1137 | } |
1421 | } |
1138 | } |
1422 | } |
1139 | |
1423 | |
1140 | static void * |
1424 | static void * |
1141 | ev_realloc_emul (void *ptr, long size) |
1425 | ev_realloc_emul (void *ptr, long size) EV_THROW |
1142 | { |
1426 | { |
1143 | #if __GLIBC__ |
|
|
1144 | return realloc (ptr, size); |
|
|
1145 | #else |
|
|
1146 | /* some systems, notably openbsd and darwin, fail to properly |
1427 | /* some systems, notably openbsd and darwin, fail to properly |
1147 | * implement realloc (x, 0) (as required by both ansi c-89 and |
1428 | * implement realloc (x, 0) (as required by both ansi c-89 and |
1148 | * the single unix specification, so work around them here. |
1429 | * the single unix specification, so work around them here. |
|
|
1430 | * recently, also (at least) fedora and debian started breaking it, |
|
|
1431 | * despite documenting it otherwise. |
1149 | */ |
1432 | */ |
1150 | |
1433 | |
1151 | if (size) |
1434 | if (size) |
1152 | return realloc (ptr, size); |
1435 | return realloc (ptr, size); |
1153 | |
1436 | |
1154 | free (ptr); |
1437 | free (ptr); |
1155 | return 0; |
1438 | return 0; |
1156 | #endif |
|
|
1157 | } |
1439 | } |
1158 | |
1440 | |
1159 | static void *(*alloc)(void *ptr, long size) EV_THROW = ev_realloc_emul; |
1441 | static void *(*alloc)(void *ptr, long size) EV_THROW = ev_realloc_emul; |
1160 | |
1442 | |
1161 | void ecb_cold |
1443 | void ecb_cold |
1162 | ev_set_allocator (void *(*cb)(void *ptr, long size)) EV_THROW |
1444 | ev_set_allocator (void *(*cb)(void *ptr, long size) EV_THROW) EV_THROW |
1163 | { |
1445 | { |
1164 | alloc = cb; |
1446 | alloc = cb; |
1165 | } |
1447 | } |
1166 | |
1448 | |
1167 | inline_speed void * |
1449 | inline_speed void * |
… | |
… | |
1430 | w_->pending = ++pendingcnt [pri]; |
1712 | w_->pending = ++pendingcnt [pri]; |
1431 | array_needsize (ANPENDING, pendings [pri], pendingmax [pri], w_->pending, EMPTY2); |
1713 | array_needsize (ANPENDING, pendings [pri], pendingmax [pri], w_->pending, EMPTY2); |
1432 | pendings [pri][w_->pending - 1].w = w_; |
1714 | pendings [pri][w_->pending - 1].w = w_; |
1433 | pendings [pri][w_->pending - 1].events = revents; |
1715 | pendings [pri][w_->pending - 1].events = revents; |
1434 | } |
1716 | } |
|
|
1717 | |
|
|
1718 | pendingpri = NUMPRI - 1; |
1435 | } |
1719 | } |
1436 | |
1720 | |
1437 | inline_speed void |
1721 | inline_speed void |
1438 | feed_reverse (EV_P_ W w) |
1722 | feed_reverse (EV_P_ W w) |
1439 | { |
1723 | { |
… | |
… | |
1814 | static void noinline ecb_cold |
2098 | static void noinline ecb_cold |
1815 | evpipe_init (EV_P) |
2099 | evpipe_init (EV_P) |
1816 | { |
2100 | { |
1817 | if (!ev_is_active (&pipe_w)) |
2101 | if (!ev_is_active (&pipe_w)) |
1818 | { |
2102 | { |
|
|
2103 | int fds [2]; |
|
|
2104 | |
1819 | # if EV_USE_EVENTFD |
2105 | # if EV_USE_EVENTFD |
|
|
2106 | fds [0] = -1; |
1820 | evfd = eventfd (0, EFD_NONBLOCK | EFD_CLOEXEC); |
2107 | fds [1] = eventfd (0, EFD_NONBLOCK | EFD_CLOEXEC); |
1821 | if (evfd < 0 && errno == EINVAL) |
2108 | if (fds [1] < 0 && errno == EINVAL) |
1822 | evfd = eventfd (0, 0); |
2109 | fds [1] = eventfd (0, 0); |
1823 | |
2110 | |
1824 | if (evfd >= 0) |
2111 | if (fds [1] < 0) |
|
|
2112 | # endif |
1825 | { |
2113 | { |
|
|
2114 | while (pipe (fds)) |
|
|
2115 | ev_syserr ("(libev) error creating signal/async pipe"); |
|
|
2116 | |
|
|
2117 | fd_intern (fds [0]); |
|
|
2118 | } |
|
|
2119 | |
1826 | evpipe [0] = -1; |
2120 | evpipe [0] = fds [0]; |
1827 | fd_intern (evfd); /* doing it twice doesn't hurt */ |
2121 | |
1828 | ev_io_set (&pipe_w, evfd, EV_READ); |
2122 | if (evpipe [1] < 0) |
|
|
2123 | evpipe [1] = fds [1]; /* first call, set write fd */ |
|
|
2124 | else |
|
|
2125 | { |
|
|
2126 | /* on subsequent calls, do not change evpipe [1] */ |
|
|
2127 | /* so that evpipe_write can always rely on its value. */ |
|
|
2128 | /* this branch does not do anything sensible on windows, */ |
|
|
2129 | /* so must not be executed on windows */ |
|
|
2130 | |
|
|
2131 | dup2 (fds [1], evpipe [1]); |
|
|
2132 | close (fds [1]); |
|
|
2133 | } |
|
|
2134 | |
|
|
2135 | fd_intern (evpipe [1]); |
|
|
2136 | |
|
|
2137 | ev_io_set (&pipe_w, evpipe [0] < 0 ? evpipe [1] : evpipe [0], EV_READ); |
|
|
2138 | ev_io_start (EV_A_ &pipe_w); |
|
|
2139 | ev_unref (EV_A); /* watcher should not keep loop alive */ |
|
|
2140 | } |
|
|
2141 | } |
|
|
2142 | |
|
|
2143 | inline_speed void |
|
|
2144 | evpipe_write (EV_P_ EV_ATOMIC_T *flag) |
|
|
2145 | { |
|
|
2146 | ECB_MEMORY_FENCE; /* push out the write before this function was called, acquire flag */ |
|
|
2147 | |
|
|
2148 | if (expect_true (*flag)) |
|
|
2149 | return; |
|
|
2150 | |
|
|
2151 | *flag = 1; |
|
|
2152 | ECB_MEMORY_FENCE_RELEASE; /* make sure flag is visible before the wakeup */ |
|
|
2153 | |
|
|
2154 | pipe_write_skipped = 1; |
|
|
2155 | |
|
|
2156 | ECB_MEMORY_FENCE; /* make sure pipe_write_skipped is visible before we check pipe_write_wanted */ |
|
|
2157 | |
|
|
2158 | if (pipe_write_wanted) |
|
|
2159 | { |
|
|
2160 | int old_errno; |
|
|
2161 | |
|
|
2162 | pipe_write_skipped = 0; |
|
|
2163 | ECB_MEMORY_FENCE_RELEASE; |
|
|
2164 | |
|
|
2165 | old_errno = errno; /* save errno because write will clobber it */ |
|
|
2166 | |
|
|
2167 | #if EV_USE_EVENTFD |
|
|
2168 | if (evpipe [0] < 0) |
|
|
2169 | { |
|
|
2170 | uint64_t counter = 1; |
|
|
2171 | write (evpipe [1], &counter, sizeof (uint64_t)); |
1829 | } |
2172 | } |
1830 | else |
2173 | else |
1831 | # endif |
2174 | #endif |
1832 | { |
2175 | { |
1833 | while (pipe (evpipe)) |
2176 | #ifdef _WIN32 |
1834 | ev_syserr ("(libev) error creating signal/async pipe"); |
2177 | WSABUF buf; |
1835 | |
2178 | DWORD sent; |
1836 | fd_intern (evpipe [0]); |
2179 | buf.buf = &buf; |
1837 | fd_intern (evpipe [1]); |
2180 | buf.len = 1; |
1838 | ev_io_set (&pipe_w, evpipe [0], EV_READ); |
2181 | WSASend (EV_FD_TO_WIN32_HANDLE (evpipe [1]), &buf, 1, &sent, 0, 0, 0); |
1839 | } |
2182 | #else |
1840 | |
|
|
1841 | ev_io_start (EV_A_ &pipe_w); |
|
|
1842 | ev_unref (EV_A); /* watcher should not keep loop alive */ |
|
|
1843 | } |
|
|
1844 | } |
|
|
1845 | |
|
|
1846 | inline_speed void |
|
|
1847 | evpipe_write (EV_P_ EV_ATOMIC_T *flag) |
|
|
1848 | { |
|
|
1849 | if (expect_true (*flag)) |
|
|
1850 | return; |
|
|
1851 | |
|
|
1852 | *flag = 1; |
|
|
1853 | |
|
|
1854 | ECB_MEMORY_FENCE_RELEASE; /* make sure flag is visible before the wakeup */ |
|
|
1855 | |
|
|
1856 | pipe_write_skipped = 1; |
|
|
1857 | |
|
|
1858 | ECB_MEMORY_FENCE; /* make sure pipe_write_skipped is visible before we check pipe_write_wanted */ |
|
|
1859 | |
|
|
1860 | if (pipe_write_wanted) |
|
|
1861 | { |
|
|
1862 | int old_errno; |
|
|
1863 | |
|
|
1864 | pipe_write_skipped = 0; /* just an optimisation, no fence needed */ |
|
|
1865 | |
|
|
1866 | old_errno = errno; /* save errno because write will clobber it */ |
|
|
1867 | |
|
|
1868 | #if EV_USE_EVENTFD |
|
|
1869 | if (evfd >= 0) |
|
|
1870 | { |
|
|
1871 | uint64_t counter = 1; |
|
|
1872 | write (evfd, &counter, sizeof (uint64_t)); |
|
|
1873 | } |
|
|
1874 | else |
|
|
1875 | #endif |
|
|
1876 | { |
|
|
1877 | /* win32 people keep sending patches that change this write() to send() */ |
|
|
1878 | /* and then run away. but send() is wrong, it wants a socket handle on win32 */ |
|
|
1879 | /* so when you think this write should be a send instead, please find out */ |
|
|
1880 | /* where your send() is from - it's definitely not the microsoft send, and */ |
|
|
1881 | /* tell me. thank you. */ |
|
|
1882 | /* it might be that your problem is that your environment needs EV_USE_WSASOCKET */ |
|
|
1883 | /* check the ev documentation on how to use this flag */ |
|
|
1884 | write (evpipe [1], &(evpipe [1]), 1); |
2183 | write (evpipe [1], &(evpipe [1]), 1); |
|
|
2184 | #endif |
1885 | } |
2185 | } |
1886 | |
2186 | |
1887 | errno = old_errno; |
2187 | errno = old_errno; |
1888 | } |
2188 | } |
1889 | } |
2189 | } |
… | |
… | |
1896 | int i; |
2196 | int i; |
1897 | |
2197 | |
1898 | if (revents & EV_READ) |
2198 | if (revents & EV_READ) |
1899 | { |
2199 | { |
1900 | #if EV_USE_EVENTFD |
2200 | #if EV_USE_EVENTFD |
1901 | if (evfd >= 0) |
2201 | if (evpipe [0] < 0) |
1902 | { |
2202 | { |
1903 | uint64_t counter; |
2203 | uint64_t counter; |
1904 | read (evfd, &counter, sizeof (uint64_t)); |
2204 | read (evpipe [1], &counter, sizeof (uint64_t)); |
1905 | } |
2205 | } |
1906 | else |
2206 | else |
1907 | #endif |
2207 | #endif |
1908 | { |
2208 | { |
1909 | char dummy; |
2209 | char dummy[4]; |
1910 | /* see discussion in evpipe_write when you think this read should be recv in win32 */ |
2210 | #ifdef _WIN32 |
|
|
2211 | WSABUF buf; |
|
|
2212 | DWORD recvd; |
|
|
2213 | DWORD flags = 0; |
|
|
2214 | buf.buf = dummy; |
|
|
2215 | buf.len = sizeof (dummy); |
|
|
2216 | WSARecv (EV_FD_TO_WIN32_HANDLE (evpipe [0]), &buf, 1, &recvd, &flags, 0, 0); |
|
|
2217 | #else |
1911 | read (evpipe [0], &dummy, 1); |
2218 | read (evpipe [0], &dummy, sizeof (dummy)); |
|
|
2219 | #endif |
1912 | } |
2220 | } |
1913 | } |
2221 | } |
1914 | |
2222 | |
1915 | pipe_write_skipped = 0; |
2223 | pipe_write_skipped = 0; |
|
|
2224 | |
|
|
2225 | ECB_MEMORY_FENCE; /* push out skipped, acquire flags */ |
1916 | |
2226 | |
1917 | #if EV_SIGNAL_ENABLE |
2227 | #if EV_SIGNAL_ENABLE |
1918 | if (sig_pending) |
2228 | if (sig_pending) |
1919 | { |
2229 | { |
1920 | sig_pending = 0; |
2230 | sig_pending = 0; |
|
|
2231 | |
|
|
2232 | ECB_MEMORY_FENCE; |
1921 | |
2233 | |
1922 | for (i = EV_NSIG - 1; i--; ) |
2234 | for (i = EV_NSIG - 1; i--; ) |
1923 | if (expect_false (signals [i].pending)) |
2235 | if (expect_false (signals [i].pending)) |
1924 | ev_feed_signal_event (EV_A_ i + 1); |
2236 | ev_feed_signal_event (EV_A_ i + 1); |
1925 | } |
2237 | } |
… | |
… | |
1927 | |
2239 | |
1928 | #if EV_ASYNC_ENABLE |
2240 | #if EV_ASYNC_ENABLE |
1929 | if (async_pending) |
2241 | if (async_pending) |
1930 | { |
2242 | { |
1931 | async_pending = 0; |
2243 | async_pending = 0; |
|
|
2244 | |
|
|
2245 | ECB_MEMORY_FENCE; |
1932 | |
2246 | |
1933 | for (i = asynccnt; i--; ) |
2247 | for (i = asynccnt; i--; ) |
1934 | if (asyncs [i]->sent) |
2248 | if (asyncs [i]->sent) |
1935 | { |
2249 | { |
1936 | asyncs [i]->sent = 0; |
2250 | asyncs [i]->sent = 0; |
|
|
2251 | ECB_MEMORY_FENCE_RELEASE; |
1937 | ev_feed_event (EV_A_ asyncs [i], EV_ASYNC); |
2252 | ev_feed_event (EV_A_ asyncs [i], EV_ASYNC); |
1938 | } |
2253 | } |
1939 | } |
2254 | } |
1940 | #endif |
2255 | #endif |
1941 | } |
2256 | } |
… | |
… | |
1944 | |
2259 | |
1945 | void |
2260 | void |
1946 | ev_feed_signal (int signum) EV_THROW |
2261 | ev_feed_signal (int signum) EV_THROW |
1947 | { |
2262 | { |
1948 | #if EV_MULTIPLICITY |
2263 | #if EV_MULTIPLICITY |
|
|
2264 | EV_P; |
|
|
2265 | ECB_MEMORY_FENCE_ACQUIRE; |
1949 | EV_P = signals [signum - 1].loop; |
2266 | EV_A = signals [signum - 1].loop; |
1950 | |
2267 | |
1951 | if (!EV_A) |
2268 | if (!EV_A) |
1952 | return; |
2269 | return; |
1953 | #endif |
2270 | #endif |
1954 | |
2271 | |
1955 | if (!ev_active (&pipe_w)) |
|
|
1956 | return; |
|
|
1957 | |
|
|
1958 | signals [signum - 1].pending = 1; |
2272 | signals [signum - 1].pending = 1; |
1959 | evpipe_write (EV_A_ &sig_pending); |
2273 | evpipe_write (EV_A_ &sig_pending); |
1960 | } |
2274 | } |
1961 | |
2275 | |
1962 | static void |
2276 | static void |
… | |
… | |
1972 | void noinline |
2286 | void noinline |
1973 | ev_feed_signal_event (EV_P_ int signum) EV_THROW |
2287 | ev_feed_signal_event (EV_P_ int signum) EV_THROW |
1974 | { |
2288 | { |
1975 | WL w; |
2289 | WL w; |
1976 | |
2290 | |
1977 | if (expect_false (signum <= 0 || signum > EV_NSIG)) |
2291 | if (expect_false (signum <= 0 || signum >= EV_NSIG)) |
1978 | return; |
2292 | return; |
1979 | |
2293 | |
1980 | --signum; |
2294 | --signum; |
1981 | |
2295 | |
1982 | #if EV_MULTIPLICITY |
2296 | #if EV_MULTIPLICITY |
… | |
… | |
1986 | if (expect_false (signals [signum].loop != EV_A)) |
2300 | if (expect_false (signals [signum].loop != EV_A)) |
1987 | return; |
2301 | return; |
1988 | #endif |
2302 | #endif |
1989 | |
2303 | |
1990 | signals [signum].pending = 0; |
2304 | signals [signum].pending = 0; |
|
|
2305 | ECB_MEMORY_FENCE_RELEASE; |
1991 | |
2306 | |
1992 | for (w = signals [signum].head; w; w = w->next) |
2307 | for (w = signals [signum].head; w; w = w->next) |
1993 | ev_feed_event (EV_A_ (W)w, EV_SIGNAL); |
2308 | ev_feed_event (EV_A_ (W)w, EV_SIGNAL); |
1994 | } |
2309 | } |
1995 | |
2310 | |
… | |
… | |
2279 | #if EV_ASYNC_ENABLE |
2594 | #if EV_ASYNC_ENABLE |
2280 | async_pending = 0; |
2595 | async_pending = 0; |
2281 | #endif |
2596 | #endif |
2282 | pipe_write_skipped = 0; |
2597 | pipe_write_skipped = 0; |
2283 | pipe_write_wanted = 0; |
2598 | pipe_write_wanted = 0; |
|
|
2599 | evpipe [0] = -1; |
|
|
2600 | evpipe [1] = -1; |
2284 | #if EV_USE_INOTIFY |
2601 | #if EV_USE_INOTIFY |
2285 | fs_fd = flags & EVFLAG_NOINOTIFY ? -1 : -2; |
2602 | fs_fd = flags & EVFLAG_NOINOTIFY ? -1 : -2; |
2286 | #endif |
2603 | #endif |
2287 | #if EV_USE_SIGNALFD |
2604 | #if EV_USE_SIGNALFD |
2288 | sigfd = flags & EVFLAG_SIGNALFD ? -2 : -1; |
2605 | sigfd = flags & EVFLAG_SIGNALFD ? -2 : -1; |
… | |
… | |
2319 | } |
2636 | } |
2320 | } |
2637 | } |
2321 | |
2638 | |
2322 | /* free up a loop structure */ |
2639 | /* free up a loop structure */ |
2323 | void ecb_cold |
2640 | void ecb_cold |
2324 | ev_loop_destroy (EV_P) EV_THROW |
2641 | ev_loop_destroy (EV_P) |
2325 | { |
2642 | { |
2326 | int i; |
2643 | int i; |
2327 | |
2644 | |
2328 | #if EV_MULTIPLICITY |
2645 | #if EV_MULTIPLICITY |
2329 | /* mimic free (0) */ |
2646 | /* mimic free (0) */ |
… | |
… | |
2339 | EV_INVOKE_PENDING; |
2656 | EV_INVOKE_PENDING; |
2340 | } |
2657 | } |
2341 | #endif |
2658 | #endif |
2342 | |
2659 | |
2343 | #if EV_CHILD_ENABLE |
2660 | #if EV_CHILD_ENABLE |
2344 | if (ev_is_active (&childev)) |
2661 | if (ev_is_default_loop (EV_A) && ev_is_active (&childev)) |
2345 | { |
2662 | { |
2346 | ev_ref (EV_A); /* child watcher */ |
2663 | ev_ref (EV_A); /* child watcher */ |
2347 | ev_signal_stop (EV_A_ &childev); |
2664 | ev_signal_stop (EV_A_ &childev); |
2348 | } |
2665 | } |
2349 | #endif |
2666 | #endif |
… | |
… | |
2351 | if (ev_is_active (&pipe_w)) |
2668 | if (ev_is_active (&pipe_w)) |
2352 | { |
2669 | { |
2353 | /*ev_ref (EV_A);*/ |
2670 | /*ev_ref (EV_A);*/ |
2354 | /*ev_io_stop (EV_A_ &pipe_w);*/ |
2671 | /*ev_io_stop (EV_A_ &pipe_w);*/ |
2355 | |
2672 | |
2356 | #if EV_USE_EVENTFD |
|
|
2357 | if (evfd >= 0) |
|
|
2358 | close (evfd); |
|
|
2359 | #endif |
|
|
2360 | |
|
|
2361 | if (evpipe [0] >= 0) |
|
|
2362 | { |
|
|
2363 | EV_WIN32_CLOSE_FD (evpipe [0]); |
2673 | if (evpipe [0] >= 0) EV_WIN32_CLOSE_FD (evpipe [0]); |
2364 | EV_WIN32_CLOSE_FD (evpipe [1]); |
2674 | if (evpipe [1] >= 0) EV_WIN32_CLOSE_FD (evpipe [1]); |
2365 | } |
|
|
2366 | } |
2675 | } |
2367 | |
2676 | |
2368 | #if EV_USE_SIGNALFD |
2677 | #if EV_USE_SIGNALFD |
2369 | if (ev_is_active (&sigfd_w)) |
2678 | if (ev_is_active (&sigfd_w)) |
2370 | close (sigfd); |
2679 | close (sigfd); |
… | |
… | |
2456 | #endif |
2765 | #endif |
2457 | #if EV_USE_INOTIFY |
2766 | #if EV_USE_INOTIFY |
2458 | infy_fork (EV_A); |
2767 | infy_fork (EV_A); |
2459 | #endif |
2768 | #endif |
2460 | |
2769 | |
|
|
2770 | #if EV_SIGNAL_ENABLE || EV_ASYNC_ENABLE |
2461 | if (ev_is_active (&pipe_w)) |
2771 | if (ev_is_active (&pipe_w)) |
2462 | { |
2772 | { |
2463 | /* pipe_write_wanted must be false now, so modifying fd vars should be safe */ |
2773 | /* pipe_write_wanted must be false now, so modifying fd vars should be safe */ |
2464 | |
2774 | |
2465 | ev_ref (EV_A); |
2775 | ev_ref (EV_A); |
2466 | ev_io_stop (EV_A_ &pipe_w); |
2776 | ev_io_stop (EV_A_ &pipe_w); |
2467 | |
2777 | |
2468 | #if EV_USE_EVENTFD |
|
|
2469 | if (evfd >= 0) |
|
|
2470 | close (evfd); |
|
|
2471 | #endif |
|
|
2472 | |
|
|
2473 | if (evpipe [0] >= 0) |
2778 | if (evpipe [0] >= 0) |
2474 | { |
|
|
2475 | EV_WIN32_CLOSE_FD (evpipe [0]); |
2779 | EV_WIN32_CLOSE_FD (evpipe [0]); |
2476 | EV_WIN32_CLOSE_FD (evpipe [1]); |
|
|
2477 | } |
|
|
2478 | |
2780 | |
2479 | #if EV_SIGNAL_ENABLE || EV_ASYNC_ENABLE |
|
|
2480 | evpipe_init (EV_A); |
2781 | evpipe_init (EV_A); |
2481 | /* now iterate over everything, in case we missed something */ |
2782 | /* iterate over everything, in case we missed something before */ |
2482 | pipecb (EV_A_ &pipe_w, EV_READ); |
2783 | ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); |
2483 | #endif |
|
|
2484 | } |
2784 | } |
|
|
2785 | #endif |
2485 | |
2786 | |
2486 | postfork = 0; |
2787 | postfork = 0; |
2487 | } |
2788 | } |
2488 | |
2789 | |
2489 | #if EV_MULTIPLICITY |
2790 | #if EV_MULTIPLICITY |
… | |
… | |
2545 | void ecb_cold |
2846 | void ecb_cold |
2546 | ev_verify (EV_P) EV_THROW |
2847 | ev_verify (EV_P) EV_THROW |
2547 | { |
2848 | { |
2548 | #if EV_VERIFY |
2849 | #if EV_VERIFY |
2549 | int i; |
2850 | int i; |
2550 | WL w; |
2851 | WL w, w2; |
2551 | |
2852 | |
2552 | assert (activecnt >= -1); |
2853 | assert (activecnt >= -1); |
2553 | |
2854 | |
2554 | assert (fdchangemax >= fdchangecnt); |
2855 | assert (fdchangemax >= fdchangecnt); |
2555 | for (i = 0; i < fdchangecnt; ++i) |
2856 | for (i = 0; i < fdchangecnt; ++i) |
2556 | assert (("libev: negative fd in fdchanges", fdchanges [i] >= 0)); |
2857 | assert (("libev: negative fd in fdchanges", fdchanges [i] >= 0)); |
2557 | |
2858 | |
2558 | assert (anfdmax >= 0); |
2859 | assert (anfdmax >= 0); |
2559 | for (i = 0; i < anfdmax; ++i) |
2860 | for (i = 0; i < anfdmax; ++i) |
|
|
2861 | { |
|
|
2862 | int j = 0; |
|
|
2863 | |
2560 | for (w = anfds [i].head; w; w = w->next) |
2864 | for (w = w2 = anfds [i].head; w; w = w->next) |
2561 | { |
2865 | { |
2562 | verify_watcher (EV_A_ (W)w); |
2866 | verify_watcher (EV_A_ (W)w); |
|
|
2867 | |
|
|
2868 | if (j++ & 1) |
|
|
2869 | { |
|
|
2870 | assert (("libev: io watcher list contains a loop", w != w2)); |
|
|
2871 | w2 = w2->next; |
|
|
2872 | } |
|
|
2873 | |
2563 | assert (("libev: inactive fd watcher on anfd list", ev_active (w) == 1)); |
2874 | assert (("libev: inactive fd watcher on anfd list", ev_active (w) == 1)); |
2564 | assert (("libev: fd mismatch between watcher and anfd", ((ev_io *)w)->fd == i)); |
2875 | assert (("libev: fd mismatch between watcher and anfd", ((ev_io *)w)->fd == i)); |
2565 | } |
2876 | } |
|
|
2877 | } |
2566 | |
2878 | |
2567 | assert (timermax >= timercnt); |
2879 | assert (timermax >= timercnt); |
2568 | verify_heap (EV_A_ timers, timercnt); |
2880 | verify_heap (EV_A_ timers, timercnt); |
2569 | |
2881 | |
2570 | #if EV_PERIODIC_ENABLE |
2882 | #if EV_PERIODIC_ENABLE |
… | |
… | |
2651 | } |
2963 | } |
2652 | |
2964 | |
2653 | void |
2965 | void |
2654 | ev_loop_fork (EV_P) EV_THROW |
2966 | ev_loop_fork (EV_P) EV_THROW |
2655 | { |
2967 | { |
2656 | postfork = 1; /* must be in line with ev_default_fork */ |
2968 | postfork = 1; |
2657 | } |
2969 | } |
2658 | |
2970 | |
2659 | /*****************************************************************************/ |
2971 | /*****************************************************************************/ |
2660 | |
2972 | |
2661 | void |
2973 | void |
… | |
… | |
2677 | } |
2989 | } |
2678 | |
2990 | |
2679 | void noinline |
2991 | void noinline |
2680 | ev_invoke_pending (EV_P) |
2992 | ev_invoke_pending (EV_P) |
2681 | { |
2993 | { |
2682 | int pri; |
2994 | pendingpri = NUMPRI; |
2683 | |
2995 | |
2684 | for (pri = NUMPRI; pri--; ) |
2996 | while (pendingpri) /* pendingpri possibly gets modified in the inner loop */ |
|
|
2997 | { |
|
|
2998 | --pendingpri; |
|
|
2999 | |
2685 | while (pendingcnt [pri]) |
3000 | while (pendingcnt [pendingpri]) |
2686 | { |
3001 | { |
2687 | ANPENDING *p = pendings [pri] + --pendingcnt [pri]; |
3002 | ANPENDING *p = pendings [pendingpri] + --pendingcnt [pendingpri]; |
2688 | |
3003 | |
2689 | p->w->pending = 0; |
3004 | p->w->pending = 0; |
2690 | EV_CB_INVOKE (p->w, p->events); |
3005 | EV_CB_INVOKE (p->w, p->events); |
2691 | EV_FREQUENT_CHECK; |
3006 | EV_FREQUENT_CHECK; |
2692 | } |
3007 | } |
|
|
3008 | } |
2693 | } |
3009 | } |
2694 | |
3010 | |
2695 | #if EV_IDLE_ENABLE |
3011 | #if EV_IDLE_ENABLE |
2696 | /* make idle watchers pending. this handles the "call-idle */ |
3012 | /* make idle watchers pending. this handles the "call-idle */ |
2697 | /* only when higher priorities are idle" logic */ |
3013 | /* only when higher priorities are idle" logic */ |
… | |
… | |
2787 | { |
3103 | { |
2788 | EV_FREQUENT_CHECK; |
3104 | EV_FREQUENT_CHECK; |
2789 | |
3105 | |
2790 | while (periodiccnt && ANHE_at (periodics [HEAP0]) < ev_rt_now) |
3106 | while (periodiccnt && ANHE_at (periodics [HEAP0]) < ev_rt_now) |
2791 | { |
3107 | { |
2792 | int feed_count = 0; |
|
|
2793 | |
|
|
2794 | do |
3108 | do |
2795 | { |
3109 | { |
2796 | ev_periodic *w = (ev_periodic *)ANHE_w (periodics [HEAP0]); |
3110 | ev_periodic *w = (ev_periodic *)ANHE_w (periodics [HEAP0]); |
2797 | |
3111 | |
2798 | /*assert (("libev: inactive timer on periodic heap detected", ev_is_active (w)));*/ |
3112 | /*assert (("libev: inactive timer on periodic heap detected", ev_is_active (w)));*/ |
… | |
… | |
3057 | backend_poll (EV_A_ waittime); |
3371 | backend_poll (EV_A_ waittime); |
3058 | assert ((loop_done = EVBREAK_CANCEL, 1)); /* assert for side effect */ |
3372 | assert ((loop_done = EVBREAK_CANCEL, 1)); /* assert for side effect */ |
3059 | |
3373 | |
3060 | pipe_write_wanted = 0; /* just an optimisation, no fence needed */ |
3374 | pipe_write_wanted = 0; /* just an optimisation, no fence needed */ |
3061 | |
3375 | |
|
|
3376 | ECB_MEMORY_FENCE_ACQUIRE; |
3062 | if (pipe_write_skipped) |
3377 | if (pipe_write_skipped) |
3063 | { |
3378 | { |
3064 | assert (("libev: pipe_w not active, but pipe not written", ev_is_active (&pipe_w))); |
3379 | assert (("libev: pipe_w not active, but pipe not written", ev_is_active (&pipe_w))); |
3065 | ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); |
3380 | ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); |
3066 | } |
3381 | } |
… | |
… | |
3241 | EV_FREQUENT_CHECK; |
3556 | EV_FREQUENT_CHECK; |
3242 | |
3557 | |
3243 | ev_start (EV_A_ (W)w, 1); |
3558 | ev_start (EV_A_ (W)w, 1); |
3244 | array_needsize (ANFD, anfds, anfdmax, fd + 1, array_init_zero); |
3559 | array_needsize (ANFD, anfds, anfdmax, fd + 1, array_init_zero); |
3245 | wlist_add (&anfds[fd].head, (WL)w); |
3560 | wlist_add (&anfds[fd].head, (WL)w); |
|
|
3561 | |
|
|
3562 | /* common bug, apparently */ |
|
|
3563 | assert (("libev: ev_io_start called with corrupted watcher", ((WL)w)->next != (WL)w)); |
3246 | |
3564 | |
3247 | fd_change (EV_A_ fd, w->events & EV__IOFDSET | EV_ANFD_REIFY); |
3565 | fd_change (EV_A_ fd, w->events & EV__IOFDSET | EV_ANFD_REIFY); |
3248 | w->events &= ~EV__IOFDSET; |
3566 | w->events &= ~EV__IOFDSET; |
3249 | |
3567 | |
3250 | EV_FREQUENT_CHECK; |
3568 | EV_FREQUENT_CHECK; |
… | |
… | |
3441 | #if EV_MULTIPLICITY |
3759 | #if EV_MULTIPLICITY |
3442 | assert (("libev: a signal must not be attached to two different loops", |
3760 | assert (("libev: a signal must not be attached to two different loops", |
3443 | !signals [w->signum - 1].loop || signals [w->signum - 1].loop == loop)); |
3761 | !signals [w->signum - 1].loop || signals [w->signum - 1].loop == loop)); |
3444 | |
3762 | |
3445 | signals [w->signum - 1].loop = EV_A; |
3763 | signals [w->signum - 1].loop = EV_A; |
|
|
3764 | ECB_MEMORY_FENCE_RELEASE; |
3446 | #endif |
3765 | #endif |
3447 | |
3766 | |
3448 | EV_FREQUENT_CHECK; |
3767 | EV_FREQUENT_CHECK; |
3449 | |
3768 | |
3450 | #if EV_USE_SIGNALFD |
3769 | #if EV_USE_SIGNALFD |
… | |
… | |
3605 | # define EV_INOTIFY_BUFSIZE (sizeof (struct inotify_event) * 2 + NAME_MAX) |
3924 | # define EV_INOTIFY_BUFSIZE (sizeof (struct inotify_event) * 2 + NAME_MAX) |
3606 | |
3925 | |
3607 | static void noinline |
3926 | static void noinline |
3608 | infy_add (EV_P_ ev_stat *w) |
3927 | infy_add (EV_P_ ev_stat *w) |
3609 | { |
3928 | { |
3610 | w->wd = inotify_add_watch (fs_fd, w->path, IN_ATTRIB | IN_DELETE_SELF | IN_MOVE_SELF | IN_MODIFY | IN_DONT_FOLLOW | IN_MASK_ADD); |
3929 | w->wd = inotify_add_watch (fs_fd, w->path, |
|
|
3930 | IN_ATTRIB | IN_DELETE_SELF | IN_MOVE_SELF | IN_MODIFY |
|
|
3931 | | IN_CREATE | IN_DELETE | IN_MOVED_FROM | IN_MOVED_TO |
|
|
3932 | | IN_DONT_FOLLOW | IN_MASK_ADD); |
3611 | |
3933 | |
3612 | if (w->wd >= 0) |
3934 | if (w->wd >= 0) |
3613 | { |
3935 | { |
3614 | struct statfs sfs; |
3936 | struct statfs sfs; |
3615 | |
3937 | |
… | |
… | |
3619 | |
3941 | |
3620 | if (!fs_2625) |
3942 | if (!fs_2625) |
3621 | w->timer.repeat = w->interval ? w->interval : DEF_STAT_INTERVAL; |
3943 | w->timer.repeat = w->interval ? w->interval : DEF_STAT_INTERVAL; |
3622 | else if (!statfs (w->path, &sfs) |
3944 | else if (!statfs (w->path, &sfs) |
3623 | && (sfs.f_type == 0x1373 /* devfs */ |
3945 | && (sfs.f_type == 0x1373 /* devfs */ |
|
|
3946 | || sfs.f_type == 0x4006 /* fat */ |
|
|
3947 | || sfs.f_type == 0x4d44 /* msdos */ |
3624 | || sfs.f_type == 0xEF53 /* ext2/3 */ |
3948 | || sfs.f_type == 0xEF53 /* ext2/3 */ |
|
|
3949 | || sfs.f_type == 0x72b6 /* jffs2 */ |
|
|
3950 | || sfs.f_type == 0x858458f6 /* ramfs */ |
|
|
3951 | || sfs.f_type == 0x5346544e /* ntfs */ |
3625 | || sfs.f_type == 0x3153464a /* jfs */ |
3952 | || sfs.f_type == 0x3153464a /* jfs */ |
|
|
3953 | || sfs.f_type == 0x9123683e /* btrfs */ |
3626 | || sfs.f_type == 0x52654973 /* reiser3 */ |
3954 | || sfs.f_type == 0x52654973 /* reiser3 */ |
3627 | || sfs.f_type == 0x01021994 /* tempfs */ |
3955 | || sfs.f_type == 0x01021994 /* tmpfs */ |
3628 | || sfs.f_type == 0x58465342 /* xfs */)) |
3956 | || sfs.f_type == 0x58465342 /* xfs */)) |
3629 | w->timer.repeat = 0.; /* filesystem is local, kernel new enough */ |
3957 | w->timer.repeat = 0.; /* filesystem is local, kernel new enough */ |
3630 | else |
3958 | else |
3631 | w->timer.repeat = w->interval ? w->interval : NFS_STAT_INTERVAL; /* remote, use reduced frequency */ |
3959 | w->timer.repeat = w->interval ? w->interval : NFS_STAT_INTERVAL; /* remote, use reduced frequency */ |
3632 | } |
3960 | } |