… | |
… | |
183 | # include EV_H |
183 | # include EV_H |
184 | #else |
184 | #else |
185 | # include "ev.h" |
185 | # include "ev.h" |
186 | #endif |
186 | #endif |
187 | |
187 | |
188 | EV_CPP(extern "C" {) |
188 | #if EV_NO_THREADS |
|
|
189 | # undef EV_NO_SMP |
|
|
190 | # define EV_NO_SMP 1 |
|
|
191 | # undef ECB_NO_THREADS |
|
|
192 | # define ECB_NO_THREADS 1 |
|
|
193 | #endif |
|
|
194 | #if EV_NO_SMP |
|
|
195 | # undef EV_NO_SMP |
|
|
196 | # define ECB_NO_SMP 1 |
|
|
197 | #endif |
189 | |
198 | |
190 | #ifndef _WIN32 |
199 | #ifndef _WIN32 |
191 | # include <sys/time.h> |
200 | # include <sys/time.h> |
192 | # include <sys/wait.h> |
201 | # include <sys/wait.h> |
193 | # include <unistd.h> |
202 | # include <unistd.h> |
… | |
… | |
469 | /* the following is ecb.h embedded into libev - use update_ev_c to update from an external copy */ |
478 | /* the following is ecb.h embedded into libev - use update_ev_c to update from an external copy */ |
470 | /* ECB.H BEGIN */ |
479 | /* ECB.H BEGIN */ |
471 | /* |
480 | /* |
472 | * libecb - http://software.schmorp.de/pkg/libecb |
481 | * libecb - http://software.schmorp.de/pkg/libecb |
473 | * |
482 | * |
474 | * Copyright (©) 2009-2011 Marc Alexander Lehmann <libecb@schmorp.de> |
483 | * Copyright (©) 2009-2012 Marc Alexander Lehmann <libecb@schmorp.de> |
475 | * Copyright (©) 2011 Emanuele Giaquinta |
484 | * Copyright (©) 2011 Emanuele Giaquinta |
476 | * All rights reserved. |
485 | * All rights reserved. |
477 | * |
486 | * |
478 | * Redistribution and use in source and binary forms, with or without modifica- |
487 | * Redistribution and use in source and binary forms, with or without modifica- |
479 | * tion, are permitted provided that the following conditions are met: |
488 | * tion, are permitted provided that the following conditions are met: |
… | |
… | |
536 | /*****************************************************************************/ |
545 | /*****************************************************************************/ |
537 | |
546 | |
538 | /* ECB_NO_THREADS - ecb is not used by multiple threads, ever */ |
547 | /* ECB_NO_THREADS - ecb is not used by multiple threads, ever */ |
539 | /* ECB_NO_SMP - ecb might be used in multiple threads, but only on a single cpu */ |
548 | /* ECB_NO_SMP - ecb might be used in multiple threads, but only on a single cpu */ |
540 | |
549 | |
|
|
550 | #if ECB_NO_THREADS |
|
|
551 | # define ECB_NO_SMP 1 |
|
|
552 | #endif |
|
|
553 | |
541 | #if ECB_NO_THREADS || ECB_NO_SMP |
554 | #if ECB_NO_THREADS || ECB_NO_SMP |
542 | #define ECB_MEMORY_FENCE do { } while (0) |
555 | #define ECB_MEMORY_FENCE do { } while (0) |
543 | #endif |
556 | #endif |
544 | |
557 | |
545 | #ifndef ECB_MEMORY_FENCE |
558 | #ifndef ECB_MEMORY_FENCE |
546 | #if ECB_GCC_VERSION(2,5) |
559 | #if ECB_GCC_VERSION(2,5) || defined(__INTEL_COMPILER) || (__llvm__ && __GNUC__) || __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
547 | #if __x86 |
560 | #if __i386 || __i386__ |
548 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
561 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
549 | #define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE /* non-lock xchg might be enough */ |
562 | #define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE /* non-lock xchg might be enough */ |
550 | #define ECB_MEMORY_FENCE_RELEASE do { } while (0) /* unlikely to change in future cpus */ |
563 | #define ECB_MEMORY_FENCE_RELEASE do { } while (0) /* unlikely to change in future cpus */ |
551 | #elif __amd64 |
564 | #elif __amd64 || __amd64__ || __x86_64 || __x86_64__ |
552 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
565 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
553 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("lfence" : : : "memory") |
566 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("lfence" : : : "memory") |
554 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("sfence") /* play safe - not needed in any current cpu */ |
567 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("sfence") /* play safe - not needed in any current cpu */ |
555 | #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ |
568 | #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ |
556 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
569 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
557 | #elif defined(__ARM_ARCH_6__ ) || defined(__ARM_ARCH_6J__ ) \ |
570 | #elif defined(__ARM_ARCH_6__ ) || defined(__ARM_ARCH_6J__ ) \ |
558 | || defined(__ARM_ARCH_6K__) || defined(__ARM_ARCH_6ZK__) |
571 | || defined(__ARM_ARCH_6K__) || defined(__ARM_ARCH_6ZK__) |
559 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,4" : : "r" (0) : "memory") |
572 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,5" : : "r" (0) : "memory") |
560 | #elif defined(__ARM_ARCH_7__ ) || defined(__ARM_ARCH_7A__ ) \ |
573 | #elif defined(__ARM_ARCH_7__ ) || defined(__ARM_ARCH_7A__ ) \ |
561 | || defined(__ARM_ARCH_7M__) || defined(__ARM_ARCH_7R__ ) |
574 | || defined(__ARM_ARCH_7M__) || defined(__ARM_ARCH_7R__ ) |
562 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dsb" : : : "memory") |
575 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") |
|
|
576 | #elif __sparc || __sparc__ |
|
|
577 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad | " : : : "memory") |
|
|
578 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory") |
|
|
579 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore") |
|
|
580 | #elif defined(__s390__) || defined(__s390x__) |
|
|
581 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory") |
|
|
582 | #elif defined(__mips__) |
|
|
583 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
563 | #endif |
584 | #endif |
564 | #endif |
585 | #endif |
565 | #endif |
586 | #endif |
566 | |
587 | |
567 | #ifndef ECB_MEMORY_FENCE |
588 | #ifndef ECB_MEMORY_FENCE |
568 | #if ECB_GCC_VERSION(4,4) || defined(__INTEL_COMPILER) |
589 | #if ECB_GCC_VERSION(4,4) || defined(__INTEL_COMPILER) || defined(__clang__) |
569 | #define ECB_MEMORY_FENCE __sync_synchronize () |
590 | #define ECB_MEMORY_FENCE __sync_synchronize () |
570 | /*#define ECB_MEMORY_FENCE_ACQUIRE ({ char dummy = 0; __sync_lock_test_and_set (&dummy, 1); }) */ |
591 | /*#define ECB_MEMORY_FENCE_ACQUIRE ({ char dummy = 0; __sync_lock_test_and_set (&dummy, 1); }) */ |
571 | /*#define ECB_MEMORY_FENCE_RELEASE ({ char dummy = 1; __sync_lock_release (&dummy ); }) */ |
592 | /*#define ECB_MEMORY_FENCE_RELEASE ({ char dummy = 1; __sync_lock_release (&dummy ); }) */ |
572 | #elif _MSC_VER >= 1400 /* VC++ 2005 */ |
593 | #elif _MSC_VER >= 1400 /* VC++ 2005 */ |
573 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
594 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
… | |
… | |
575 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ |
596 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ |
576 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () |
597 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () |
577 | #elif defined(_WIN32) |
598 | #elif defined(_WIN32) |
578 | #include <WinNT.h> |
599 | #include <WinNT.h> |
579 | #define ECB_MEMORY_FENCE MemoryBarrier () /* actually just xchg on x86... scary */ |
600 | #define ECB_MEMORY_FENCE MemoryBarrier () /* actually just xchg on x86... scary */ |
|
|
601 | #elif __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
|
|
602 | #include <mbarrier.h> |
|
|
603 | #define ECB_MEMORY_FENCE __machine_rw_barrier () |
|
|
604 | #define ECB_MEMORY_FENCE_ACQUIRE __machine_r_barrier () |
|
|
605 | #define ECB_MEMORY_FENCE_RELEASE __machine_w_barrier () |
|
|
606 | #elif __xlC__ |
|
|
607 | #define ECB_MEMORY_FENCE __lwsync () |
580 | #endif |
608 | #endif |
581 | #endif |
609 | #endif |
582 | |
610 | |
583 | #ifndef ECB_MEMORY_FENCE |
611 | #ifndef ECB_MEMORY_FENCE |
584 | #if !ECB_AVOID_PTHREADS |
612 | #if !ECB_AVOID_PTHREADS |
… | |
… | |
759 | |
787 | |
760 | return r + ecb_ld32 (x); |
788 | return r + ecb_ld32 (x); |
761 | } |
789 | } |
762 | #endif |
790 | #endif |
763 | |
791 | |
|
|
792 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) ecb_const; |
|
|
793 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) |
|
|
794 | { |
|
|
795 | return ( (x * 0x0802U & 0x22110U) |
|
|
796 | | (x * 0x8020U & 0x88440U)) * 0x10101U >> 16; |
|
|
797 | } |
|
|
798 | |
|
|
799 | ecb_function_ uint16_t ecb_bitrev16 (uint16_t x) ecb_const; |
|
|
800 | ecb_function_ uint16_t ecb_bitrev16 (uint16_t x) |
|
|
801 | { |
|
|
802 | x = ((x >> 1) & 0x5555) | ((x & 0x5555) << 1); |
|
|
803 | x = ((x >> 2) & 0x3333) | ((x & 0x3333) << 2); |
|
|
804 | x = ((x >> 4) & 0x0f0f) | ((x & 0x0f0f) << 4); |
|
|
805 | x = ( x >> 8 ) | ( x << 8); |
|
|
806 | |
|
|
807 | return x; |
|
|
808 | } |
|
|
809 | |
|
|
810 | ecb_function_ uint32_t ecb_bitrev32 (uint32_t x) ecb_const; |
|
|
811 | ecb_function_ uint32_t ecb_bitrev32 (uint32_t x) |
|
|
812 | { |
|
|
813 | x = ((x >> 1) & 0x55555555) | ((x & 0x55555555) << 1); |
|
|
814 | x = ((x >> 2) & 0x33333333) | ((x & 0x33333333) << 2); |
|
|
815 | x = ((x >> 4) & 0x0f0f0f0f) | ((x & 0x0f0f0f0f) << 4); |
|
|
816 | x = ((x >> 8) & 0x00ff00ff) | ((x & 0x00ff00ff) << 8); |
|
|
817 | x = ( x >> 16 ) | ( x << 16); |
|
|
818 | |
|
|
819 | return x; |
|
|
820 | } |
|
|
821 | |
764 | /* popcount64 is only available on 64 bit cpus as gcc builtin */ |
822 | /* popcount64 is only available on 64 bit cpus as gcc builtin */ |
765 | /* so for this version we are lazy */ |
823 | /* so for this version we are lazy */ |
766 | ecb_function_ int ecb_popcount64 (uint64_t x) ecb_const; |
824 | ecb_function_ int ecb_popcount64 (uint64_t x) ecb_const; |
767 | ecb_function_ int |
825 | ecb_function_ int |
768 | ecb_popcount64 (uint64_t x) |
826 | ecb_popcount64 (uint64_t x) |
… | |
… | |
817 | |
875 | |
818 | #if ECB_GCC_VERSION(4,5) |
876 | #if ECB_GCC_VERSION(4,5) |
819 | #define ecb_unreachable() __builtin_unreachable () |
877 | #define ecb_unreachable() __builtin_unreachable () |
820 | #else |
878 | #else |
821 | /* this seems to work fine, but gcc always emits a warning for it :/ */ |
879 | /* this seems to work fine, but gcc always emits a warning for it :/ */ |
822 | ecb_function_ void ecb_unreachable (void) ecb_noreturn; |
880 | ecb_inline void ecb_unreachable (void) ecb_noreturn; |
823 | ecb_function_ void ecb_unreachable (void) { } |
881 | ecb_inline void ecb_unreachable (void) { } |
824 | #endif |
882 | #endif |
825 | |
883 | |
826 | /* try to tell the compiler that some condition is definitely true */ |
884 | /* try to tell the compiler that some condition is definitely true */ |
827 | #define ecb_assume(cond) do { if (!(cond)) ecb_unreachable (); } while (0) |
885 | #define ecb_assume(cond) do { if (!(cond)) ecb_unreachable (); } while (0) |
828 | |
886 | |
829 | ecb_function_ unsigned char ecb_byteorder_helper (void) ecb_const; |
887 | ecb_inline unsigned char ecb_byteorder_helper (void) ecb_const; |
830 | ecb_function_ unsigned char |
888 | ecb_inline unsigned char |
831 | ecb_byteorder_helper (void) |
889 | ecb_byteorder_helper (void) |
832 | { |
890 | { |
833 | const uint32_t u = 0x11223344; |
891 | const uint32_t u = 0x11223344; |
834 | return *(unsigned char *)&u; |
892 | return *(unsigned char *)&u; |
835 | } |
893 | } |
836 | |
894 | |
837 | ecb_function_ ecb_bool ecb_big_endian (void) ecb_const; |
895 | ecb_inline ecb_bool ecb_big_endian (void) ecb_const; |
838 | ecb_function_ ecb_bool ecb_big_endian (void) { return ecb_byteorder_helper () == 0x11; } |
896 | ecb_inline ecb_bool ecb_big_endian (void) { return ecb_byteorder_helper () == 0x11; } |
839 | ecb_function_ ecb_bool ecb_little_endian (void) ecb_const; |
897 | ecb_inline ecb_bool ecb_little_endian (void) ecb_const; |
840 | ecb_function_ ecb_bool ecb_little_endian (void) { return ecb_byteorder_helper () == 0x44; } |
898 | ecb_inline ecb_bool ecb_little_endian (void) { return ecb_byteorder_helper () == 0x44; } |
841 | |
899 | |
842 | #if ECB_GCC_VERSION(3,0) || ECB_C99 |
900 | #if ECB_GCC_VERSION(3,0) || ECB_C99 |
843 | #define ecb_mod(m,n) ((m) % (n) + ((m) % (n) < 0 ? (n) : 0)) |
901 | #define ecb_mod(m,n) ((m) % (n) + ((m) % (n) < 0 ? (n) : 0)) |
844 | #else |
902 | #else |
845 | #define ecb_mod(m,n) ((m) < 0 ? ((n) - 1 - ((-1 - (m)) % (n))) : ((m) % (n))) |
903 | #define ecb_mod(m,n) ((m) < 0 ? ((n) - 1 - ((-1 - (m)) % (n))) : ((m) % (n))) |
|
|
904 | #endif |
|
|
905 | |
|
|
906 | #if __cplusplus |
|
|
907 | template<typename T> |
|
|
908 | static inline T ecb_div_rd (T val, T div) |
|
|
909 | { |
|
|
910 | return val < 0 ? - ((-val + div - 1) / div) : (val ) / div; |
|
|
911 | } |
|
|
912 | template<typename T> |
|
|
913 | static inline T ecb_div_ru (T val, T div) |
|
|
914 | { |
|
|
915 | return val < 0 ? - ((-val ) / div) : (val + div - 1) / div; |
|
|
916 | } |
|
|
917 | #else |
|
|
918 | #define ecb_div_rd(val,div) ((val) < 0 ? - ((-(val) + (div) - 1) / (div)) : ((val) ) / (div)) |
|
|
919 | #define ecb_div_ru(val,div) ((val) < 0 ? - ((-(val) ) / (div)) : ((val) + (div) - 1) / (div)) |
846 | #endif |
920 | #endif |
847 | |
921 | |
848 | #if ecb_cplusplus_does_not_suck |
922 | #if ecb_cplusplus_does_not_suck |
849 | /* does not work for local types (http://www.open-std.org/jtc1/sc22/wg21/docs/papers/2008/n2657.htm) */ |
923 | /* does not work for local types (http://www.open-std.org/jtc1/sc22/wg21/docs/papers/2008/n2657.htm) */ |
850 | template<typename T, int N> |
924 | template<typename T, int N> |
… | |
… | |
859 | #endif |
933 | #endif |
860 | |
934 | |
861 | /* ECB.H END */ |
935 | /* ECB.H END */ |
862 | |
936 | |
863 | #if ECB_MEMORY_FENCE_NEEDS_PTHREADS |
937 | #if ECB_MEMORY_FENCE_NEEDS_PTHREADS |
|
|
938 | /* if your architecture doesn't need memory fences, e.g. because it is |
|
|
939 | * single-cpu/core, or if you use libev in a project that doesn't use libev |
|
|
940 | * from multiple threads, then you can define ECB_AVOID_PTHREADS when compiling |
|
|
941 | * libev, in which cases the memory fences become nops. |
|
|
942 | * alternatively, you can remove this #error and link against libpthread, |
|
|
943 | * which will then provide the memory fences. |
|
|
944 | */ |
|
|
945 | # error "memory fences not defined for your architecture, please report" |
|
|
946 | #endif |
|
|
947 | |
864 | # undef ECB_MEMORY_FENCE |
948 | #ifndef ECB_MEMORY_FENCE |
865 | # undef ECB_MEMORY_FENCE_ACQUIRE |
949 | # define ECB_MEMORY_FENCE do { } while (0) |
866 | # undef ECB_MEMORY_FENCE_RELEASE |
950 | # define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE |
|
|
951 | # define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE |
867 | #endif |
952 | #endif |
868 | |
953 | |
869 | #define expect_false(cond) ecb_expect_false (cond) |
954 | #define expect_false(cond) ecb_expect_false (cond) |
870 | #define expect_true(cond) ecb_expect_true (cond) |
955 | #define expect_true(cond) ecb_expect_true (cond) |
871 | #define noinline ecb_noinline |
956 | #define noinline ecb_noinline |
… | |
… | |
1168 | #undef VAR |
1253 | #undef VAR |
1169 | }; |
1254 | }; |
1170 | #include "ev_wrap.h" |
1255 | #include "ev_wrap.h" |
1171 | |
1256 | |
1172 | static struct ev_loop default_loop_struct; |
1257 | static struct ev_loop default_loop_struct; |
1173 | struct ev_loop *ev_default_loop_ptr; |
1258 | EV_API_DECL struct ev_loop *ev_default_loop_ptr = 0; /* needs to be initialised to make it a definition despite extern */ |
1174 | |
1259 | |
1175 | #else |
1260 | #else |
1176 | |
1261 | |
1177 | ev_tstamp ev_rt_now; |
1262 | EV_API_DECL ev_tstamp ev_rt_now = 0; /* needs to be initialised to make it a definition despite extern */ |
1178 | #define VAR(name,decl) static decl; |
1263 | #define VAR(name,decl) static decl; |
1179 | #include "ev_vars.h" |
1264 | #include "ev_vars.h" |
1180 | #undef VAR |
1265 | #undef VAR |
1181 | |
1266 | |
1182 | static int ev_default_loop_ptr; |
1267 | static int ev_default_loop_ptr; |
… | |
… | |
1276 | |
1361 | |
1277 | do |
1362 | do |
1278 | ncur <<= 1; |
1363 | ncur <<= 1; |
1279 | while (cnt > ncur); |
1364 | while (cnt > ncur); |
1280 | |
1365 | |
1281 | /* if size is large, round to MALLOC_ROUND - 4 * longs to accomodate malloc overhead */ |
1366 | /* if size is large, round to MALLOC_ROUND - 4 * longs to accommodate malloc overhead */ |
1282 | if (elem * ncur > MALLOC_ROUND - sizeof (void *) * 4) |
1367 | if (elem * ncur > MALLOC_ROUND - sizeof (void *) * 4) |
1283 | { |
1368 | { |
1284 | ncur *= elem; |
1369 | ncur *= elem; |
1285 | ncur = (ncur + elem + (MALLOC_ROUND - 1) + sizeof (void *) * 4) & ~(MALLOC_ROUND - 1); |
1370 | ncur = (ncur + elem + (MALLOC_ROUND - 1) + sizeof (void *) * 4) & ~(MALLOC_ROUND - 1); |
1286 | ncur = ncur - sizeof (void *) * 4; |
1371 | ncur = ncur - sizeof (void *) * 4; |
… | |
… | |
1790 | /* win32 people keep sending patches that change this write() to send() */ |
1875 | /* win32 people keep sending patches that change this write() to send() */ |
1791 | /* and then run away. but send() is wrong, it wants a socket handle on win32 */ |
1876 | /* and then run away. but send() is wrong, it wants a socket handle on win32 */ |
1792 | /* so when you think this write should be a send instead, please find out */ |
1877 | /* so when you think this write should be a send instead, please find out */ |
1793 | /* where your send() is from - it's definitely not the microsoft send, and */ |
1878 | /* where your send() is from - it's definitely not the microsoft send, and */ |
1794 | /* tell me. thank you. */ |
1879 | /* tell me. thank you. */ |
|
|
1880 | /* it might be that your problem is that your environment needs EV_USE_WSASOCKET */ |
|
|
1881 | /* check the ev documentation on how to use this flag */ |
1795 | write (evpipe [1], &(evpipe [1]), 1); |
1882 | write (evpipe [1], &(evpipe [1]), 1); |
1796 | } |
1883 | } |
1797 | |
1884 | |
1798 | errno = old_errno; |
1885 | errno = old_errno; |
1799 | } |
1886 | } |
… | |
… | |
2966 | #endif |
3053 | #endif |
2967 | assert ((loop_done = EVBREAK_RECURSE, 1)); /* assert for side effect */ |
3054 | assert ((loop_done = EVBREAK_RECURSE, 1)); /* assert for side effect */ |
2968 | backend_poll (EV_A_ waittime); |
3055 | backend_poll (EV_A_ waittime); |
2969 | assert ((loop_done = EVBREAK_CANCEL, 1)); /* assert for side effect */ |
3056 | assert ((loop_done = EVBREAK_CANCEL, 1)); /* assert for side effect */ |
2970 | |
3057 | |
2971 | pipe_write_wanted = 0; /* just an optimsiation, no fence needed */ |
3058 | pipe_write_wanted = 0; /* just an optimisation, no fence needed */ |
2972 | |
3059 | |
2973 | if (pipe_write_skipped) |
3060 | if (pipe_write_skipped) |
2974 | { |
3061 | { |
2975 | assert (("libev: pipe_w not active, but pipe not written", ev_is_active (&pipe_w))); |
3062 | assert (("libev: pipe_w not active, but pipe not written", ev_is_active (&pipe_w))); |
2976 | ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); |
3063 | ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); |
… | |
… | |
3234 | |
3321 | |
3235 | void noinline |
3322 | void noinline |
3236 | ev_timer_again (EV_P_ ev_timer *w) |
3323 | ev_timer_again (EV_P_ ev_timer *w) |
3237 | { |
3324 | { |
3238 | EV_FREQUENT_CHECK; |
3325 | EV_FREQUENT_CHECK; |
|
|
3326 | |
|
|
3327 | clear_pending (EV_A_ (W)w); |
3239 | |
3328 | |
3240 | if (ev_is_active (w)) |
3329 | if (ev_is_active (w)) |
3241 | { |
3330 | { |
3242 | if (w->repeat) |
3331 | if (w->repeat) |
3243 | { |
3332 | { |
… | |
… | |
4389 | |
4478 | |
4390 | #if EV_MULTIPLICITY |
4479 | #if EV_MULTIPLICITY |
4391 | #include "ev_wrap.h" |
4480 | #include "ev_wrap.h" |
4392 | #endif |
4481 | #endif |
4393 | |
4482 | |
4394 | EV_CPP(}) |
|
|
4395 | |
|
|