… | |
… | |
183 | # include EV_H |
183 | # include EV_H |
184 | #else |
184 | #else |
185 | # include "ev.h" |
185 | # include "ev.h" |
186 | #endif |
186 | #endif |
187 | |
187 | |
188 | EV_CPP(extern "C" {) |
|
|
189 | |
|
|
190 | #ifndef _WIN32 |
188 | #ifndef _WIN32 |
191 | # include <sys/time.h> |
189 | # include <sys/time.h> |
192 | # include <sys/wait.h> |
190 | # include <sys/wait.h> |
193 | # include <unistd.h> |
191 | # include <unistd.h> |
194 | #else |
192 | #else |
… | |
… | |
464 | #define MAX_BLOCKTIME 59.743 /* never wait longer than this time (to detect time jumps) */ |
462 | #define MAX_BLOCKTIME 59.743 /* never wait longer than this time (to detect time jumps) */ |
465 | |
463 | |
466 | #define EV_TV_SET(tv,t) do { tv.tv_sec = (long)t; tv.tv_usec = (long)((t - tv.tv_sec) * 1e6); } while (0) |
464 | #define EV_TV_SET(tv,t) do { tv.tv_sec = (long)t; tv.tv_usec = (long)((t - tv.tv_sec) * 1e6); } while (0) |
467 | #define EV_TS_SET(ts,t) do { ts.tv_sec = (long)t; ts.tv_nsec = (long)((t - ts.tv_sec) * 1e9); } while (0) |
465 | #define EV_TS_SET(ts,t) do { ts.tv_sec = (long)t; ts.tv_nsec = (long)((t - ts.tv_sec) * 1e9); } while (0) |
468 | |
466 | |
469 | /* the following are taken from libecb */ |
467 | /* the following is ecb.h embedded into libev - use update_ev_c to update from an external copy */ |
470 | /* ecb.h start */ |
468 | /* ECB.H BEGIN */ |
|
|
469 | /* |
|
|
470 | * libecb - http://software.schmorp.de/pkg/libecb |
|
|
471 | * |
|
|
472 | * Copyright (©) 2009-2012 Marc Alexander Lehmann <libecb@schmorp.de> |
|
|
473 | * Copyright (©) 2011 Emanuele Giaquinta |
|
|
474 | * All rights reserved. |
|
|
475 | * |
|
|
476 | * Redistribution and use in source and binary forms, with or without modifica- |
|
|
477 | * tion, are permitted provided that the following conditions are met: |
|
|
478 | * |
|
|
479 | * 1. Redistributions of source code must retain the above copyright notice, |
|
|
480 | * this list of conditions and the following disclaimer. |
|
|
481 | * |
|
|
482 | * 2. Redistributions in binary form must reproduce the above copyright |
|
|
483 | * notice, this list of conditions and the following disclaimer in the |
|
|
484 | * documentation and/or other materials provided with the distribution. |
|
|
485 | * |
|
|
486 | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED |
|
|
487 | * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MER- |
|
|
488 | * CHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO |
|
|
489 | * EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPE- |
|
|
490 | * CIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
|
|
491 | * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; |
|
|
492 | * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, |
|
|
493 | * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTH- |
|
|
494 | * ERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED |
|
|
495 | * OF THE POSSIBILITY OF SUCH DAMAGE. |
|
|
496 | */ |
|
|
497 | |
|
|
498 | #ifndef ECB_H |
|
|
499 | #define ECB_H |
|
|
500 | |
|
|
501 | #ifdef _WIN32 |
|
|
502 | typedef signed char int8_t; |
|
|
503 | typedef unsigned char uint8_t; |
|
|
504 | typedef signed short int16_t; |
|
|
505 | typedef unsigned short uint16_t; |
|
|
506 | typedef signed int int32_t; |
|
|
507 | typedef unsigned int uint32_t; |
|
|
508 | #if __GNUC__ |
|
|
509 | typedef signed long long int64_t; |
|
|
510 | typedef unsigned long long uint64_t; |
|
|
511 | #else /* _MSC_VER || __BORLANDC__ */ |
|
|
512 | typedef signed __int64 int64_t; |
|
|
513 | typedef unsigned __int64 uint64_t; |
|
|
514 | #endif |
|
|
515 | #else |
|
|
516 | #include <inttypes.h> |
|
|
517 | #endif |
471 | |
518 | |
472 | /* many compilers define _GNUC_ to some versions but then only implement |
519 | /* many compilers define _GNUC_ to some versions but then only implement |
473 | * what their idiot authors think are the "more important" extensions, |
520 | * what their idiot authors think are the "more important" extensions, |
474 | * causing enourmous grief in return for some better fake benchmark numbers. |
521 | * causing enormous grief in return for some better fake benchmark numbers. |
475 | * or so. |
522 | * or so. |
476 | * we try to detect these and simply assume they are not gcc - if they have |
523 | * we try to detect these and simply assume they are not gcc - if they have |
477 | * an issue with that they should have done it right in the first place. |
524 | * an issue with that they should have done it right in the first place. |
478 | */ |
525 | */ |
479 | #ifndef ECB_GCC_VERSION |
526 | #ifndef ECB_GCC_VERSION |
… | |
… | |
482 | #else |
529 | #else |
483 | #define ECB_GCC_VERSION(major,minor) (__GNUC__ > (major) || (__GNUC__ == (major) && __GNUC_MINOR__ >= (minor))) |
530 | #define ECB_GCC_VERSION(major,minor) (__GNUC__ > (major) || (__GNUC__ == (major) && __GNUC_MINOR__ >= (minor))) |
484 | #endif |
531 | #endif |
485 | #endif |
532 | #endif |
486 | |
533 | |
|
|
534 | /*****************************************************************************/ |
|
|
535 | |
|
|
536 | /* ECB_NO_THREADS - ecb is not used by multiple threads, ever */ |
|
|
537 | /* ECB_NO_SMP - ecb might be used in multiple threads, but only on a single cpu */ |
|
|
538 | |
|
|
539 | #if ECB_NO_THREADS || ECB_NO_SMP |
|
|
540 | #define ECB_MEMORY_FENCE do { } while (0) |
|
|
541 | #endif |
|
|
542 | |
|
|
543 | #ifndef ECB_MEMORY_FENCE |
|
|
544 | #if ECB_GCC_VERSION(2,5) || defined(__INTEL_COMPILER) || defined(__clang__) || __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
|
|
545 | #if __i386 || __i386__ |
|
|
546 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
|
|
547 | #define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE /* non-lock xchg might be enough */ |
|
|
548 | #define ECB_MEMORY_FENCE_RELEASE do { } while (0) /* unlikely to change in future cpus */ |
|
|
549 | #elif __amd64 || __amd64__ || __x86_64 || __x86_64__ |
|
|
550 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
|
|
551 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("lfence" : : : "memory") |
|
|
552 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("sfence") /* play safe - not needed in any current cpu */ |
|
|
553 | #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ |
|
|
554 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
|
|
555 | #elif defined(__ARM_ARCH_6__ ) || defined(__ARM_ARCH_6J__ ) \ |
|
|
556 | || defined(__ARM_ARCH_6K__) || defined(__ARM_ARCH_6ZK__) |
|
|
557 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,5" : : "r" (0) : "memory") |
|
|
558 | #elif defined(__ARM_ARCH_7__ ) || defined(__ARM_ARCH_7A__ ) \ |
|
|
559 | || defined(__ARM_ARCH_7M__) || defined(__ARM_ARCH_7R__ ) |
|
|
560 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") |
|
|
561 | #elif __sparc || __sparc__ |
|
|
562 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #StoreLoad | #LoadLoad | #StoreStore" : : : "memory") |
|
|
563 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadLoad" : : : "memory") |
|
|
564 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #StoreStore") |
|
|
565 | #endif |
|
|
566 | #endif |
|
|
567 | #endif |
|
|
568 | |
|
|
569 | #ifndef ECB_MEMORY_FENCE |
|
|
570 | #if ECB_GCC_VERSION(4,4) || defined(__INTEL_COMPILER) || defined(__clang__) |
|
|
571 | #define ECB_MEMORY_FENCE __sync_synchronize () |
|
|
572 | /*#define ECB_MEMORY_FENCE_ACQUIRE ({ char dummy = 0; __sync_lock_test_and_set (&dummy, 1); }) */ |
|
|
573 | /*#define ECB_MEMORY_FENCE_RELEASE ({ char dummy = 1; __sync_lock_release (&dummy ); }) */ |
|
|
574 | #elif _MSC_VER >= 1400 /* VC++ 2005 */ |
|
|
575 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
|
|
576 | #define ECB_MEMORY_FENCE _ReadWriteBarrier () |
|
|
577 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ |
|
|
578 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () |
|
|
579 | #elif defined(_WIN32) |
|
|
580 | #include <WinNT.h> |
|
|
581 | #define ECB_MEMORY_FENCE MemoryBarrier () /* actually just xchg on x86... scary */ |
|
|
582 | #elif __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
|
|
583 | #include <mbarrier.h> |
|
|
584 | #define ECB_MEMORY_FENCE __machine_rw_barrier () |
|
|
585 | #define ECB_MEMORY_FENCE_ACQUIRE __machine_r_barrier () |
|
|
586 | #define ECB_MEMORY_FENCE_RELEASE __machine_w_barrier () |
|
|
587 | #endif |
|
|
588 | #endif |
|
|
589 | |
|
|
590 | #ifndef ECB_MEMORY_FENCE |
|
|
591 | #if !ECB_AVOID_PTHREADS |
|
|
592 | /* |
|
|
593 | * if you get undefined symbol references to pthread_mutex_lock, |
|
|
594 | * or failure to find pthread.h, then you should implement |
|
|
595 | * the ECB_MEMORY_FENCE operations for your cpu/compiler |
|
|
596 | * OR provide pthread.h and link against the posix thread library |
|
|
597 | * of your system. |
|
|
598 | */ |
|
|
599 | #include <pthread.h> |
|
|
600 | #define ECB_NEEDS_PTHREADS 1 |
|
|
601 | #define ECB_MEMORY_FENCE_NEEDS_PTHREADS 1 |
|
|
602 | |
|
|
603 | static pthread_mutex_t ecb_mf_lock = PTHREAD_MUTEX_INITIALIZER; |
|
|
604 | #define ECB_MEMORY_FENCE do { pthread_mutex_lock (&ecb_mf_lock); pthread_mutex_unlock (&ecb_mf_lock); } while (0) |
|
|
605 | #endif |
|
|
606 | #endif |
|
|
607 | |
|
|
608 | #if !defined(ECB_MEMORY_FENCE_ACQUIRE) && defined(ECB_MEMORY_FENCE) |
|
|
609 | #define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE |
|
|
610 | #endif |
|
|
611 | |
|
|
612 | #if !defined(ECB_MEMORY_FENCE_RELEASE) && defined(ECB_MEMORY_FENCE) |
|
|
613 | #define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE |
|
|
614 | #endif |
|
|
615 | |
|
|
616 | /*****************************************************************************/ |
|
|
617 | |
|
|
618 | #define ECB_C99 (__STDC_VERSION__ >= 199901L) |
|
|
619 | |
487 | #if __cplusplus |
620 | #if __cplusplus |
488 | #define ecb_inline static inline |
621 | #define ecb_inline static inline |
489 | #elif ECB_GCC_VERSION(2,5) |
622 | #elif ECB_GCC_VERSION(2,5) |
490 | #define ecb_inline static __inline__ |
623 | #define ecb_inline static __inline__ |
491 | #elif ECB_C99 |
624 | #elif ECB_C99 |
492 | #define ecb_inline static inline |
625 | #define ecb_inline static inline |
493 | #else |
626 | #else |
494 | #define ecb_inline static |
627 | #define ecb_inline static |
495 | #endif |
628 | #endif |
496 | |
629 | |
497 | #ifndef ECB_MEMORY_FENCE |
|
|
498 | #if ECB_GCC_VERSION(2,5) |
630 | #if ECB_GCC_VERSION(3,3) |
499 | #if __x86 |
631 | #define ecb_restrict __restrict__ |
500 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
632 | #elif ECB_C99 |
501 | #define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE |
633 | #define ecb_restrict restrict |
502 | #define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE /* better be safe than sorry */ |
634 | #else |
503 | #elif __amd64 |
635 | #define ecb_restrict |
504 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
|
|
505 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("lfence" : : : "memory") |
|
|
506 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("sfence") |
|
|
507 | #endif |
|
|
508 | #endif |
636 | #endif |
509 | #endif |
|
|
510 | |
637 | |
511 | #ifndef ECB_MEMORY_FENCE |
638 | typedef int ecb_bool; |
512 | #if ECB_GCC_VERSION(4,4) |
|
|
513 | #define ECB_MEMORY_FENCE __sync_synchronize () |
|
|
514 | #define ECB_MEMORY_FENCE_ACQUIRE ({ char dummy = 0; __sync_lock_test_and_set (&dummy, 1); }) |
|
|
515 | #define ECB_MEMORY_FENCE_RELEASE ({ char dummy = 1; __sync_lock_release (&dummy ); }) |
|
|
516 | #elif _MSC_VER >= 1400 |
|
|
517 | #define ECB_MEMORY_FENCE do { } while (0) |
|
|
518 | #define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE |
|
|
519 | #define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE |
|
|
520 | #elif defined(_WIN32) && defined(MemoryBarrier) |
|
|
521 | #define ECB_MEMORY_FENCE MemoryBarrier () |
|
|
522 | #define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE |
|
|
523 | #define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE |
|
|
524 | #endif |
|
|
525 | #endif |
|
|
526 | |
639 | |
527 | #ifndef ECB_MEMORY_FENCE |
640 | #define ECB_CONCAT_(a, b) a ## b |
528 | #include <pthread.h> |
641 | #define ECB_CONCAT(a, b) ECB_CONCAT_(a, b) |
|
|
642 | #define ECB_STRINGIFY_(a) # a |
|
|
643 | #define ECB_STRINGIFY(a) ECB_STRINGIFY_(a) |
529 | |
644 | |
530 | static pthread_mutex_t ecb_mf_lock = PTHREAD_MUTEX_INITIALIZER; |
645 | #define ecb_function_ ecb_inline |
531 | #define ECB_MEMORY_FENCE do { pthread_mutex_lock (&ecb_mf_lock); pthread_mutex_unlock (&ecb_mf_lock); } while (0) |
|
|
532 | #define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE |
|
|
533 | #define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE |
|
|
534 | #endif |
|
|
535 | |
646 | |
536 | #if ECB_GCC_VERSION(3,1) |
647 | #if ECB_GCC_VERSION(3,1) |
537 | #define ecb_attribute(attrlist) __attribute__(attrlist) |
648 | #define ecb_attribute(attrlist) __attribute__(attrlist) |
538 | #define ecb_is_constant(expr) __builtin_constant_p (expr) |
649 | #define ecb_is_constant(expr) __builtin_constant_p (expr) |
539 | #define ecb_expect(expr,value) __builtin_expect ((expr),(value)) |
650 | #define ecb_expect(expr,value) __builtin_expect ((expr),(value)) |
… | |
… | |
543 | #define ecb_is_constant(expr) 0 |
654 | #define ecb_is_constant(expr) 0 |
544 | #define ecb_expect(expr,value) (expr) |
655 | #define ecb_expect(expr,value) (expr) |
545 | #define ecb_prefetch(addr,rw,locality) |
656 | #define ecb_prefetch(addr,rw,locality) |
546 | #endif |
657 | #endif |
547 | |
658 | |
|
|
659 | /* no emulation for ecb_decltype */ |
|
|
660 | #if ECB_GCC_VERSION(4,5) |
|
|
661 | #define ecb_decltype(x) __decltype(x) |
|
|
662 | #elif ECB_GCC_VERSION(3,0) |
|
|
663 | #define ecb_decltype(x) __typeof(x) |
|
|
664 | #endif |
|
|
665 | |
548 | #define ecb_noinline ecb_attribute ((__noinline__)) |
666 | #define ecb_noinline ecb_attribute ((__noinline__)) |
549 | #define ecb_noreturn ecb_attribute ((__noreturn__)) |
667 | #define ecb_noreturn ecb_attribute ((__noreturn__)) |
550 | #define ecb_unused ecb_attribute ((__unused__)) |
668 | #define ecb_unused ecb_attribute ((__unused__)) |
551 | #define ecb_const ecb_attribute ((__const__)) |
669 | #define ecb_const ecb_attribute ((__const__)) |
552 | #define ecb_pure ecb_attribute ((__pure__)) |
670 | #define ecb_pure ecb_attribute ((__pure__)) |
… | |
… | |
564 | /* put around conditional expressions if you are very sure that the */ |
682 | /* put around conditional expressions if you are very sure that the */ |
565 | /* expression is mostly true or mostly false. note that these return */ |
683 | /* expression is mostly true or mostly false. note that these return */ |
566 | /* booleans, not the expression. */ |
684 | /* booleans, not the expression. */ |
567 | #define ecb_expect_false(expr) ecb_expect (!!(expr), 0) |
685 | #define ecb_expect_false(expr) ecb_expect (!!(expr), 0) |
568 | #define ecb_expect_true(expr) ecb_expect (!!(expr), 1) |
686 | #define ecb_expect_true(expr) ecb_expect (!!(expr), 1) |
569 | /* ecb.h end */ |
687 | /* for compatibility to the rest of the world */ |
|
|
688 | #define ecb_likely(expr) ecb_expect_true (expr) |
|
|
689 | #define ecb_unlikely(expr) ecb_expect_false (expr) |
|
|
690 | |
|
|
691 | /* count trailing zero bits and count # of one bits */ |
|
|
692 | #if ECB_GCC_VERSION(3,4) |
|
|
693 | /* we assume int == 32 bit, long == 32 or 64 bit and long long == 64 bit */ |
|
|
694 | #define ecb_ld32(x) (__builtin_clz (x) ^ 31) |
|
|
695 | #define ecb_ld64(x) (__builtin_clzll (x) ^ 63) |
|
|
696 | #define ecb_ctz32(x) __builtin_ctz (x) |
|
|
697 | #define ecb_ctz64(x) __builtin_ctzll (x) |
|
|
698 | #define ecb_popcount32(x) __builtin_popcount (x) |
|
|
699 | /* no popcountll */ |
|
|
700 | #else |
|
|
701 | ecb_function_ int ecb_ctz32 (uint32_t x) ecb_const; |
|
|
702 | ecb_function_ int |
|
|
703 | ecb_ctz32 (uint32_t x) |
|
|
704 | { |
|
|
705 | int r = 0; |
|
|
706 | |
|
|
707 | x &= ~x + 1; /* this isolates the lowest bit */ |
|
|
708 | |
|
|
709 | #if ECB_branchless_on_i386 |
|
|
710 | r += !!(x & 0xaaaaaaaa) << 0; |
|
|
711 | r += !!(x & 0xcccccccc) << 1; |
|
|
712 | r += !!(x & 0xf0f0f0f0) << 2; |
|
|
713 | r += !!(x & 0xff00ff00) << 3; |
|
|
714 | r += !!(x & 0xffff0000) << 4; |
|
|
715 | #else |
|
|
716 | if (x & 0xaaaaaaaa) r += 1; |
|
|
717 | if (x & 0xcccccccc) r += 2; |
|
|
718 | if (x & 0xf0f0f0f0) r += 4; |
|
|
719 | if (x & 0xff00ff00) r += 8; |
|
|
720 | if (x & 0xffff0000) r += 16; |
|
|
721 | #endif |
|
|
722 | |
|
|
723 | return r; |
|
|
724 | } |
|
|
725 | |
|
|
726 | ecb_function_ int ecb_ctz64 (uint64_t x) ecb_const; |
|
|
727 | ecb_function_ int |
|
|
728 | ecb_ctz64 (uint64_t x) |
|
|
729 | { |
|
|
730 | int shift = x & 0xffffffffU ? 0 : 32; |
|
|
731 | return ecb_ctz32 (x >> shift) + shift; |
|
|
732 | } |
|
|
733 | |
|
|
734 | ecb_function_ int ecb_popcount32 (uint32_t x) ecb_const; |
|
|
735 | ecb_function_ int |
|
|
736 | ecb_popcount32 (uint32_t x) |
|
|
737 | { |
|
|
738 | x -= (x >> 1) & 0x55555555; |
|
|
739 | x = ((x >> 2) & 0x33333333) + (x & 0x33333333); |
|
|
740 | x = ((x >> 4) + x) & 0x0f0f0f0f; |
|
|
741 | x *= 0x01010101; |
|
|
742 | |
|
|
743 | return x >> 24; |
|
|
744 | } |
|
|
745 | |
|
|
746 | ecb_function_ int ecb_ld32 (uint32_t x) ecb_const; |
|
|
747 | ecb_function_ int ecb_ld32 (uint32_t x) |
|
|
748 | { |
|
|
749 | int r = 0; |
|
|
750 | |
|
|
751 | if (x >> 16) { x >>= 16; r += 16; } |
|
|
752 | if (x >> 8) { x >>= 8; r += 8; } |
|
|
753 | if (x >> 4) { x >>= 4; r += 4; } |
|
|
754 | if (x >> 2) { x >>= 2; r += 2; } |
|
|
755 | if (x >> 1) { r += 1; } |
|
|
756 | |
|
|
757 | return r; |
|
|
758 | } |
|
|
759 | |
|
|
760 | ecb_function_ int ecb_ld64 (uint64_t x) ecb_const; |
|
|
761 | ecb_function_ int ecb_ld64 (uint64_t x) |
|
|
762 | { |
|
|
763 | int r = 0; |
|
|
764 | |
|
|
765 | if (x >> 32) { x >>= 32; r += 32; } |
|
|
766 | |
|
|
767 | return r + ecb_ld32 (x); |
|
|
768 | } |
|
|
769 | #endif |
|
|
770 | |
|
|
771 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) ecb_const; |
|
|
772 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) |
|
|
773 | { |
|
|
774 | return ( (x * 0x0802U & 0x22110U) |
|
|
775 | | (x * 0x8020U & 0x88440U)) * 0x10101U >> 16; |
|
|
776 | } |
|
|
777 | |
|
|
778 | ecb_function_ uint16_t ecb_bitrev16 (uint16_t x) ecb_const; |
|
|
779 | ecb_function_ uint16_t ecb_bitrev16 (uint16_t x) |
|
|
780 | { |
|
|
781 | x = ((x >> 1) & 0x5555) | ((x & 0x5555) << 1); |
|
|
782 | x = ((x >> 2) & 0x3333) | ((x & 0x3333) << 2); |
|
|
783 | x = ((x >> 4) & 0x0f0f) | ((x & 0x0f0f) << 4); |
|
|
784 | x = ( x >> 8 ) | ( x << 8); |
|
|
785 | |
|
|
786 | return x; |
|
|
787 | } |
|
|
788 | |
|
|
789 | ecb_function_ uint32_t ecb_bitrev32 (uint32_t x) ecb_const; |
|
|
790 | ecb_function_ uint32_t ecb_bitrev32 (uint32_t x) |
|
|
791 | { |
|
|
792 | x = ((x >> 1) & 0x55555555) | ((x & 0x55555555) << 1); |
|
|
793 | x = ((x >> 2) & 0x33333333) | ((x & 0x33333333) << 2); |
|
|
794 | x = ((x >> 4) & 0x0f0f0f0f) | ((x & 0x0f0f0f0f) << 4); |
|
|
795 | x = ((x >> 8) & 0x00ff00ff) | ((x & 0x00ff00ff) << 8); |
|
|
796 | x = ( x >> 16 ) | ( x << 16); |
|
|
797 | |
|
|
798 | return x; |
|
|
799 | } |
|
|
800 | |
|
|
801 | /* popcount64 is only available on 64 bit cpus as gcc builtin */ |
|
|
802 | /* so for this version we are lazy */ |
|
|
803 | ecb_function_ int ecb_popcount64 (uint64_t x) ecb_const; |
|
|
804 | ecb_function_ int |
|
|
805 | ecb_popcount64 (uint64_t x) |
|
|
806 | { |
|
|
807 | return ecb_popcount32 (x) + ecb_popcount32 (x >> 32); |
|
|
808 | } |
|
|
809 | |
|
|
810 | ecb_inline uint8_t ecb_rotl8 (uint8_t x, unsigned int count) ecb_const; |
|
|
811 | ecb_inline uint8_t ecb_rotr8 (uint8_t x, unsigned int count) ecb_const; |
|
|
812 | ecb_inline uint16_t ecb_rotl16 (uint16_t x, unsigned int count) ecb_const; |
|
|
813 | ecb_inline uint16_t ecb_rotr16 (uint16_t x, unsigned int count) ecb_const; |
|
|
814 | ecb_inline uint32_t ecb_rotl32 (uint32_t x, unsigned int count) ecb_const; |
|
|
815 | ecb_inline uint32_t ecb_rotr32 (uint32_t x, unsigned int count) ecb_const; |
|
|
816 | ecb_inline uint64_t ecb_rotl64 (uint64_t x, unsigned int count) ecb_const; |
|
|
817 | ecb_inline uint64_t ecb_rotr64 (uint64_t x, unsigned int count) ecb_const; |
|
|
818 | |
|
|
819 | ecb_inline uint8_t ecb_rotl8 (uint8_t x, unsigned int count) { return (x >> ( 8 - count)) | (x << count); } |
|
|
820 | ecb_inline uint8_t ecb_rotr8 (uint8_t x, unsigned int count) { return (x << ( 8 - count)) | (x >> count); } |
|
|
821 | ecb_inline uint16_t ecb_rotl16 (uint16_t x, unsigned int count) { return (x >> (16 - count)) | (x << count); } |
|
|
822 | ecb_inline uint16_t ecb_rotr16 (uint16_t x, unsigned int count) { return (x << (16 - count)) | (x >> count); } |
|
|
823 | ecb_inline uint32_t ecb_rotl32 (uint32_t x, unsigned int count) { return (x >> (32 - count)) | (x << count); } |
|
|
824 | ecb_inline uint32_t ecb_rotr32 (uint32_t x, unsigned int count) { return (x << (32 - count)) | (x >> count); } |
|
|
825 | ecb_inline uint64_t ecb_rotl64 (uint64_t x, unsigned int count) { return (x >> (64 - count)) | (x << count); } |
|
|
826 | ecb_inline uint64_t ecb_rotr64 (uint64_t x, unsigned int count) { return (x << (64 - count)) | (x >> count); } |
|
|
827 | |
|
|
828 | #if ECB_GCC_VERSION(4,3) |
|
|
829 | #define ecb_bswap16(x) (__builtin_bswap32 (x) >> 16) |
|
|
830 | #define ecb_bswap32(x) __builtin_bswap32 (x) |
|
|
831 | #define ecb_bswap64(x) __builtin_bswap64 (x) |
|
|
832 | #else |
|
|
833 | ecb_function_ uint16_t ecb_bswap16 (uint16_t x) ecb_const; |
|
|
834 | ecb_function_ uint16_t |
|
|
835 | ecb_bswap16 (uint16_t x) |
|
|
836 | { |
|
|
837 | return ecb_rotl16 (x, 8); |
|
|
838 | } |
|
|
839 | |
|
|
840 | ecb_function_ uint32_t ecb_bswap32 (uint32_t x) ecb_const; |
|
|
841 | ecb_function_ uint32_t |
|
|
842 | ecb_bswap32 (uint32_t x) |
|
|
843 | { |
|
|
844 | return (((uint32_t)ecb_bswap16 (x)) << 16) | ecb_bswap16 (x >> 16); |
|
|
845 | } |
|
|
846 | |
|
|
847 | ecb_function_ uint64_t ecb_bswap64 (uint64_t x) ecb_const; |
|
|
848 | ecb_function_ uint64_t |
|
|
849 | ecb_bswap64 (uint64_t x) |
|
|
850 | { |
|
|
851 | return (((uint64_t)ecb_bswap32 (x)) << 32) | ecb_bswap32 (x >> 32); |
|
|
852 | } |
|
|
853 | #endif |
|
|
854 | |
|
|
855 | #if ECB_GCC_VERSION(4,5) |
|
|
856 | #define ecb_unreachable() __builtin_unreachable () |
|
|
857 | #else |
|
|
858 | /* this seems to work fine, but gcc always emits a warning for it :/ */ |
|
|
859 | ecb_function_ void ecb_unreachable (void) ecb_noreturn; |
|
|
860 | ecb_function_ void ecb_unreachable (void) { } |
|
|
861 | #endif |
|
|
862 | |
|
|
863 | /* try to tell the compiler that some condition is definitely true */ |
|
|
864 | #define ecb_assume(cond) do { if (!(cond)) ecb_unreachable (); } while (0) |
|
|
865 | |
|
|
866 | ecb_function_ unsigned char ecb_byteorder_helper (void) ecb_const; |
|
|
867 | ecb_function_ unsigned char |
|
|
868 | ecb_byteorder_helper (void) |
|
|
869 | { |
|
|
870 | const uint32_t u = 0x11223344; |
|
|
871 | return *(unsigned char *)&u; |
|
|
872 | } |
|
|
873 | |
|
|
874 | ecb_function_ ecb_bool ecb_big_endian (void) ecb_const; |
|
|
875 | ecb_function_ ecb_bool ecb_big_endian (void) { return ecb_byteorder_helper () == 0x11; } |
|
|
876 | ecb_function_ ecb_bool ecb_little_endian (void) ecb_const; |
|
|
877 | ecb_function_ ecb_bool ecb_little_endian (void) { return ecb_byteorder_helper () == 0x44; } |
|
|
878 | |
|
|
879 | #if ECB_GCC_VERSION(3,0) || ECB_C99 |
|
|
880 | #define ecb_mod(m,n) ((m) % (n) + ((m) % (n) < 0 ? (n) : 0)) |
|
|
881 | #else |
|
|
882 | #define ecb_mod(m,n) ((m) < 0 ? ((n) - 1 - ((-1 - (m)) % (n))) : ((m) % (n))) |
|
|
883 | #endif |
|
|
884 | |
|
|
885 | #if __cplusplus |
|
|
886 | template<typename T> |
|
|
887 | static inline T ecb_div_rd (T val, T div) |
|
|
888 | { |
|
|
889 | return val < 0 ? - ((-val + div - 1) / div) : (val ) / div; |
|
|
890 | } |
|
|
891 | template<typename T> |
|
|
892 | static inline T ecb_div_ru (T val, T div) |
|
|
893 | { |
|
|
894 | return val < 0 ? - ((-val ) / div) : (val + div - 1) / div; |
|
|
895 | } |
|
|
896 | #else |
|
|
897 | #define ecb_div_rd(val,div) ((val) < 0 ? - ((-(val) + (div) - 1) / (div)) : ((val) ) / (div)) |
|
|
898 | #define ecb_div_ru(val,div) ((val) < 0 ? - ((-(val) ) / (div)) : ((val) + (div) - 1) / (div)) |
|
|
899 | #endif |
|
|
900 | |
|
|
901 | #if ecb_cplusplus_does_not_suck |
|
|
902 | /* does not work for local types (http://www.open-std.org/jtc1/sc22/wg21/docs/papers/2008/n2657.htm) */ |
|
|
903 | template<typename T, int N> |
|
|
904 | static inline int ecb_array_length (const T (&arr)[N]) |
|
|
905 | { |
|
|
906 | return N; |
|
|
907 | } |
|
|
908 | #else |
|
|
909 | #define ecb_array_length(name) (sizeof (name) / sizeof (name [0])) |
|
|
910 | #endif |
|
|
911 | |
|
|
912 | #endif |
|
|
913 | |
|
|
914 | /* ECB.H END */ |
|
|
915 | |
|
|
916 | #if ECB_MEMORY_FENCE_NEEDS_PTHREADS |
|
|
917 | /* if your architecture doesn't need memory fences, e.g. because it is |
|
|
918 | * single-cpu/core, or if you use libev in a project that doesn't use libev |
|
|
919 | * from multiple threads, then you can define ECB_AVOID_PTHREADS when compiling |
|
|
920 | * libev, in which cases the memory fences become nops. |
|
|
921 | * alternatively, you can remove this #error and link against libpthread, |
|
|
922 | * which will then provide the memory fences. |
|
|
923 | */ |
|
|
924 | # error "memory fences not defined for your architecture, please report" |
|
|
925 | #endif |
|
|
926 | |
|
|
927 | #ifndef ECB_MEMORY_FENCE |
|
|
928 | # define ECB_MEMORY_FENCE do { } while (0) |
|
|
929 | # define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE |
|
|
930 | # define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE |
|
|
931 | #endif |
570 | |
932 | |
571 | #define expect_false(cond) ecb_expect_false (cond) |
933 | #define expect_false(cond) ecb_expect_false (cond) |
572 | #define expect_true(cond) ecb_expect_true (cond) |
934 | #define expect_true(cond) ecb_expect_true (cond) |
573 | #define noinline ecb_noinline |
935 | #define noinline ecb_noinline |
574 | |
936 | |
… | |
… | |
870 | #undef VAR |
1232 | #undef VAR |
871 | }; |
1233 | }; |
872 | #include "ev_wrap.h" |
1234 | #include "ev_wrap.h" |
873 | |
1235 | |
874 | static struct ev_loop default_loop_struct; |
1236 | static struct ev_loop default_loop_struct; |
875 | struct ev_loop *ev_default_loop_ptr; |
1237 | EV_API_DECL struct ev_loop *ev_default_loop_ptr = 0; /* needs to be initialised to make it a definition despite extern */ |
876 | |
1238 | |
877 | #else |
1239 | #else |
878 | |
1240 | |
879 | ev_tstamp ev_rt_now; |
1241 | EV_API_DECL ev_tstamp ev_rt_now = 0; /* needs to be initialised to make it a definition despite extern */ |
880 | #define VAR(name,decl) static decl; |
1242 | #define VAR(name,decl) static decl; |
881 | #include "ev_vars.h" |
1243 | #include "ev_vars.h" |
882 | #undef VAR |
1244 | #undef VAR |
883 | |
1245 | |
884 | static int ev_default_loop_ptr; |
1246 | static int ev_default_loop_ptr; |
… | |
… | |
978 | |
1340 | |
979 | do |
1341 | do |
980 | ncur <<= 1; |
1342 | ncur <<= 1; |
981 | while (cnt > ncur); |
1343 | while (cnt > ncur); |
982 | |
1344 | |
983 | /* if size is large, round to MALLOC_ROUND - 4 * longs to accomodate malloc overhead */ |
1345 | /* if size is large, round to MALLOC_ROUND - 4 * longs to accommodate malloc overhead */ |
984 | if (elem * ncur > MALLOC_ROUND - sizeof (void *) * 4) |
1346 | if (elem * ncur > MALLOC_ROUND - sizeof (void *) * 4) |
985 | { |
1347 | { |
986 | ncur *= elem; |
1348 | ncur *= elem; |
987 | ncur = (ncur + elem + (MALLOC_ROUND - 1) + sizeof (void *) * 4) & ~(MALLOC_ROUND - 1); |
1349 | ncur = (ncur + elem + (MALLOC_ROUND - 1) + sizeof (void *) * 4) & ~(MALLOC_ROUND - 1); |
988 | ncur = ncur - sizeof (void *) * 4; |
1350 | ncur = ncur - sizeof (void *) * 4; |
… | |
… | |
1474 | |
1836 | |
1475 | if (pipe_write_wanted) |
1837 | if (pipe_write_wanted) |
1476 | { |
1838 | { |
1477 | int old_errno; |
1839 | int old_errno; |
1478 | |
1840 | |
1479 | pipe_write_skipped = 0; /* just an optimsiation, no fence needed */ |
1841 | pipe_write_skipped = 0; /* just an optimisation, no fence needed */ |
1480 | |
1842 | |
1481 | old_errno = errno; /* save errno because write will clobber it */ |
1843 | old_errno = errno; /* save errno because write will clobber it */ |
1482 | |
1844 | |
1483 | #if EV_USE_EVENTFD |
1845 | #if EV_USE_EVENTFD |
1484 | if (evfd >= 0) |
1846 | if (evfd >= 0) |
… | |
… | |
2616 | time_update (EV_A_ 1e100); |
2978 | time_update (EV_A_ 1e100); |
2617 | |
2979 | |
2618 | /* from now on, we want a pipe-wake-up */ |
2980 | /* from now on, we want a pipe-wake-up */ |
2619 | pipe_write_wanted = 1; |
2981 | pipe_write_wanted = 1; |
2620 | |
2982 | |
2621 | ECB_MEMORY_FENCE; /* amke sure pipe_write_wanted is visible before we check for potential skips */ |
2983 | ECB_MEMORY_FENCE; /* make sure pipe_write_wanted is visible before we check for potential skips */ |
2622 | |
2984 | |
2623 | if (expect_true (!(flags & EVRUN_NOWAIT || idleall || !activecnt || pipe_write_skipped))) |
2985 | if (expect_true (!(flags & EVRUN_NOWAIT || idleall || !activecnt || pipe_write_skipped))) |
2624 | { |
2986 | { |
2625 | waittime = MAX_BLOCKTIME; |
2987 | waittime = MAX_BLOCKTIME; |
2626 | |
2988 | |
… | |
… | |
2668 | #endif |
3030 | #endif |
2669 | assert ((loop_done = EVBREAK_RECURSE, 1)); /* assert for side effect */ |
3031 | assert ((loop_done = EVBREAK_RECURSE, 1)); /* assert for side effect */ |
2670 | backend_poll (EV_A_ waittime); |
3032 | backend_poll (EV_A_ waittime); |
2671 | assert ((loop_done = EVBREAK_CANCEL, 1)); /* assert for side effect */ |
3033 | assert ((loop_done = EVBREAK_CANCEL, 1)); /* assert for side effect */ |
2672 | |
3034 | |
2673 | pipe_write_wanted = 0; /* just an optimsiation, no fence needed */ |
3035 | pipe_write_wanted = 0; /* just an optimisation, no fence needed */ |
2674 | |
3036 | |
2675 | if (pipe_write_skipped) |
3037 | if (pipe_write_skipped) |
2676 | { |
3038 | { |
2677 | assert (("libev: pipe_w not active, but pipe not written", ev_is_active (&pipe_w))); |
3039 | assert (("libev: pipe_w not active, but pipe not written", ev_is_active (&pipe_w))); |
2678 | ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); |
3040 | ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); |
… | |
… | |
4028 | cb (EV_A_ EV_PERIODIC, ANHE_w (periodics [i])); |
4390 | cb (EV_A_ EV_PERIODIC, ANHE_w (periodics [i])); |
4029 | #endif |
4391 | #endif |
4030 | |
4392 | |
4031 | #if EV_IDLE_ENABLE |
4393 | #if EV_IDLE_ENABLE |
4032 | if (types & EV_IDLE) |
4394 | if (types & EV_IDLE) |
4033 | for (j = NUMPRI; i--; ) |
4395 | for (j = NUMPRI; j--; ) |
4034 | for (i = idlecnt [j]; i--; ) |
4396 | for (i = idlecnt [j]; i--; ) |
4035 | cb (EV_A_ EV_IDLE, idles [j][i]); |
4397 | cb (EV_A_ EV_IDLE, idles [j][i]); |
4036 | #endif |
4398 | #endif |
4037 | |
4399 | |
4038 | #if EV_FORK_ENABLE |
4400 | #if EV_FORK_ENABLE |
… | |
… | |
4091 | |
4453 | |
4092 | #if EV_MULTIPLICITY |
4454 | #if EV_MULTIPLICITY |
4093 | #include "ev_wrap.h" |
4455 | #include "ev_wrap.h" |
4094 | #endif |
4456 | #endif |
4095 | |
4457 | |
4096 | EV_CPP(}) |
|
|
4097 | |
|
|