… | |
… | |
183 | # include EV_H |
183 | # include EV_H |
184 | #else |
184 | #else |
185 | # include "ev.h" |
185 | # include "ev.h" |
186 | #endif |
186 | #endif |
187 | |
187 | |
188 | EV_CPP(extern "C" {) |
188 | #if EV_NO_THREADS |
|
|
189 | # undef EV_NO_SMP |
|
|
190 | # define EV_NO_SMP 1 |
|
|
191 | # undef ECB_NO_THREADS |
|
|
192 | # define ECB_NO_THREADS 1 |
|
|
193 | #endif |
|
|
194 | #if EV_NO_SMP |
|
|
195 | # undef EV_NO_SMP |
|
|
196 | # define ECB_NO_SMP 1 |
|
|
197 | #endif |
189 | |
198 | |
190 | #ifndef _WIN32 |
199 | #ifndef _WIN32 |
191 | # include <sys/time.h> |
200 | # include <sys/time.h> |
192 | # include <sys/wait.h> |
201 | # include <sys/wait.h> |
193 | # include <unistd.h> |
202 | # include <unistd.h> |
… | |
… | |
469 | /* the following is ecb.h embedded into libev - use update_ev_c to update from an external copy */ |
478 | /* the following is ecb.h embedded into libev - use update_ev_c to update from an external copy */ |
470 | /* ECB.H BEGIN */ |
479 | /* ECB.H BEGIN */ |
471 | /* |
480 | /* |
472 | * libecb - http://software.schmorp.de/pkg/libecb |
481 | * libecb - http://software.schmorp.de/pkg/libecb |
473 | * |
482 | * |
474 | * Copyright (©) 2009-2011 Marc Alexander Lehmann <libecb@schmorp.de> |
483 | * Copyright (©) 2009-2012 Marc Alexander Lehmann <libecb@schmorp.de> |
475 | * Copyright (©) 2011 Emanuele Giaquinta |
484 | * Copyright (©) 2011 Emanuele Giaquinta |
476 | * All rights reserved. |
485 | * All rights reserved. |
477 | * |
486 | * |
478 | * Redistribution and use in source and binary forms, with or without modifica- |
487 | * Redistribution and use in source and binary forms, with or without modifica- |
479 | * tion, are permitted provided that the following conditions are met: |
488 | * tion, are permitted provided that the following conditions are met: |
… | |
… | |
536 | /*****************************************************************************/ |
545 | /*****************************************************************************/ |
537 | |
546 | |
538 | /* ECB_NO_THREADS - ecb is not used by multiple threads, ever */ |
547 | /* ECB_NO_THREADS - ecb is not used by multiple threads, ever */ |
539 | /* ECB_NO_SMP - ecb might be used in multiple threads, but only on a single cpu */ |
548 | /* ECB_NO_SMP - ecb might be used in multiple threads, but only on a single cpu */ |
540 | |
549 | |
|
|
550 | #if ECB_NO_THREADS |
|
|
551 | # define ECB_NO_SMP 1 |
|
|
552 | #endif |
|
|
553 | |
541 | #if ECB_NO_THREADS || ECB_NO_SMP |
554 | #if ECB_NO_THREADS || ECB_NO_SMP |
542 | #define ECB_MEMORY_FENCE do { } while (0) |
555 | #define ECB_MEMORY_FENCE do { } while (0) |
543 | #define ECB_MEMORY_FENCE_ACQUIRE do { } while (0) |
|
|
544 | #define ECB_MEMORY_FENCE_RELEASE do { } while (0) |
|
|
545 | #endif |
556 | #endif |
546 | |
557 | |
547 | #ifndef ECB_MEMORY_FENCE |
558 | #ifndef ECB_MEMORY_FENCE |
548 | #if ECB_GCC_VERSION(2,5) |
559 | #if ECB_GCC_VERSION(2,5) || defined(__INTEL_COMPILER) || (__llvm__ && __GNUC__) || __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
549 | #if __x86 |
560 | #if __i386 || __i386__ |
550 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
561 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") |
551 | #define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE /* non-lock xchg might be enough */ |
562 | #define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE /* non-lock xchg might be enough */ |
552 | #define ECB_MEMORY_FENCE_RELEASE do { } while (0) /* unlikely to change in future cpus */ |
563 | #define ECB_MEMORY_FENCE_RELEASE do { } while (0) /* unlikely to change in future cpus */ |
553 | #elif __amd64 |
564 | #elif __amd64 || __amd64__ || __x86_64 || __x86_64__ |
554 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
565 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") |
555 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("lfence" : : : "memory") |
566 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("lfence" : : : "memory") |
556 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("sfence") /* play safe - not needed in any current cpu */ |
567 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("sfence") /* play safe - not needed in any current cpu */ |
557 | #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ |
568 | #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ |
558 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
569 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") |
559 | #elif defined(__ARM_ARCH_6__ ) || defined(__ARM_ARCH_6J__ ) \ |
570 | #elif defined(__ARM_ARCH_6__ ) || defined(__ARM_ARCH_6J__ ) \ |
560 | || defined(__ARM_ARCH_6K__) || defined(__ARM_ARCH_6ZK__) \ |
571 | || defined(__ARM_ARCH_6K__) || defined(__ARM_ARCH_6ZK__) |
|
|
572 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,5" : : "r" (0) : "memory") |
561 | || defined(__ARM_ARCH_7__ ) || defined(__ARM_ARCH_7A__ ) \ |
573 | #elif defined(__ARM_ARCH_7__ ) || defined(__ARM_ARCH_7A__ ) \ |
562 | || defined(__ARM_ARCH_7M__) || defined(__ARM_ARCH_7R__ ) |
574 | || defined(__ARM_ARCH_7M__) || defined(__ARM_ARCH_7R__ ) |
563 | #define ECB_MEMORY_FENCE \ |
575 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") |
564 | do { \ |
576 | #elif __sparc || __sparc__ |
565 | int null = 0; \ |
577 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad | " : : : "memory") |
566 | __asm__ __volatile__ ("mcr p15,0,%0,c6,c10,5", : "=&r" (null) : : "memory"); \ |
578 | #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory") |
567 | while (0) |
579 | #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore") |
|
|
580 | #elif defined(__s390__) || defined(__s390x__) |
|
|
581 | #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory") |
568 | #endif |
582 | #endif |
569 | #endif |
583 | #endif |
570 | #endif |
584 | #endif |
571 | |
585 | |
572 | #ifndef ECB_MEMORY_FENCE |
586 | #ifndef ECB_MEMORY_FENCE |
573 | #if ECB_GCC_VERSION(4,4) || defined(__INTEL_COMPILER) |
587 | #if ECB_GCC_VERSION(4,4) || defined(__INTEL_COMPILER) || defined(__clang__) |
574 | #define ECB_MEMORY_FENCE __sync_synchronize () |
588 | #define ECB_MEMORY_FENCE __sync_synchronize () |
575 | /*#define ECB_MEMORY_FENCE_ACQUIRE ({ char dummy = 0; __sync_lock_test_and_set (&dummy, 1); }) */ |
589 | /*#define ECB_MEMORY_FENCE_ACQUIRE ({ char dummy = 0; __sync_lock_test_and_set (&dummy, 1); }) */ |
576 | /*#define ECB_MEMORY_FENCE_RELEASE ({ char dummy = 1; __sync_lock_release (&dummy ); }) */ |
590 | /*#define ECB_MEMORY_FENCE_RELEASE ({ char dummy = 1; __sync_lock_release (&dummy ); }) */ |
577 | #elif _MSC_VER >= 1400 /* VC++ 2005 */ |
591 | #elif _MSC_VER >= 1400 /* VC++ 2005 */ |
578 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
592 | #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) |
… | |
… | |
580 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ |
594 | #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ |
581 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () |
595 | #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () |
582 | #elif defined(_WIN32) |
596 | #elif defined(_WIN32) |
583 | #include <WinNT.h> |
597 | #include <WinNT.h> |
584 | #define ECB_MEMORY_FENCE MemoryBarrier () /* actually just xchg on x86... scary */ |
598 | #define ECB_MEMORY_FENCE MemoryBarrier () /* actually just xchg on x86... scary */ |
|
|
599 | #elif __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 |
|
|
600 | #include <mbarrier.h> |
|
|
601 | #define ECB_MEMORY_FENCE __machine_rw_barrier () |
|
|
602 | #define ECB_MEMORY_FENCE_ACQUIRE __machine_r_barrier () |
|
|
603 | #define ECB_MEMORY_FENCE_RELEASE __machine_w_barrier () |
585 | #endif |
604 | #endif |
586 | #endif |
605 | #endif |
587 | |
606 | |
588 | #ifndef ECB_MEMORY_FENCE |
607 | #ifndef ECB_MEMORY_FENCE |
589 | #if !ECB_AVOID_PTHREADS |
608 | #if !ECB_AVOID_PTHREADS |
… | |
… | |
764 | |
783 | |
765 | return r + ecb_ld32 (x); |
784 | return r + ecb_ld32 (x); |
766 | } |
785 | } |
767 | #endif |
786 | #endif |
768 | |
787 | |
|
|
788 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) ecb_const; |
|
|
789 | ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) |
|
|
790 | { |
|
|
791 | return ( (x * 0x0802U & 0x22110U) |
|
|
792 | | (x * 0x8020U & 0x88440U)) * 0x10101U >> 16; |
|
|
793 | } |
|
|
794 | |
|
|
795 | ecb_function_ uint16_t ecb_bitrev16 (uint16_t x) ecb_const; |
|
|
796 | ecb_function_ uint16_t ecb_bitrev16 (uint16_t x) |
|
|
797 | { |
|
|
798 | x = ((x >> 1) & 0x5555) | ((x & 0x5555) << 1); |
|
|
799 | x = ((x >> 2) & 0x3333) | ((x & 0x3333) << 2); |
|
|
800 | x = ((x >> 4) & 0x0f0f) | ((x & 0x0f0f) << 4); |
|
|
801 | x = ( x >> 8 ) | ( x << 8); |
|
|
802 | |
|
|
803 | return x; |
|
|
804 | } |
|
|
805 | |
|
|
806 | ecb_function_ uint32_t ecb_bitrev32 (uint32_t x) ecb_const; |
|
|
807 | ecb_function_ uint32_t ecb_bitrev32 (uint32_t x) |
|
|
808 | { |
|
|
809 | x = ((x >> 1) & 0x55555555) | ((x & 0x55555555) << 1); |
|
|
810 | x = ((x >> 2) & 0x33333333) | ((x & 0x33333333) << 2); |
|
|
811 | x = ((x >> 4) & 0x0f0f0f0f) | ((x & 0x0f0f0f0f) << 4); |
|
|
812 | x = ((x >> 8) & 0x00ff00ff) | ((x & 0x00ff00ff) << 8); |
|
|
813 | x = ( x >> 16 ) | ( x << 16); |
|
|
814 | |
|
|
815 | return x; |
|
|
816 | } |
|
|
817 | |
769 | /* popcount64 is only available on 64 bit cpus as gcc builtin */ |
818 | /* popcount64 is only available on 64 bit cpus as gcc builtin */ |
770 | /* so for this version we are lazy */ |
819 | /* so for this version we are lazy */ |
771 | ecb_function_ int ecb_popcount64 (uint64_t x) ecb_const; |
820 | ecb_function_ int ecb_popcount64 (uint64_t x) ecb_const; |
772 | ecb_function_ int |
821 | ecb_function_ int |
773 | ecb_popcount64 (uint64_t x) |
822 | ecb_popcount64 (uint64_t x) |
… | |
… | |
822 | |
871 | |
823 | #if ECB_GCC_VERSION(4,5) |
872 | #if ECB_GCC_VERSION(4,5) |
824 | #define ecb_unreachable() __builtin_unreachable () |
873 | #define ecb_unreachable() __builtin_unreachable () |
825 | #else |
874 | #else |
826 | /* this seems to work fine, but gcc always emits a warning for it :/ */ |
875 | /* this seems to work fine, but gcc always emits a warning for it :/ */ |
827 | ecb_function_ void ecb_unreachable (void) ecb_noreturn; |
876 | ecb_inline void ecb_unreachable (void) ecb_noreturn; |
828 | ecb_function_ void ecb_unreachable (void) { } |
877 | ecb_inline void ecb_unreachable (void) { } |
829 | #endif |
878 | #endif |
830 | |
879 | |
831 | /* try to tell the compiler that some condition is definitely true */ |
880 | /* try to tell the compiler that some condition is definitely true */ |
832 | #define ecb_assume(cond) do { if (!(cond)) ecb_unreachable (); } while (0) |
881 | #define ecb_assume(cond) do { if (!(cond)) ecb_unreachable (); } while (0) |
833 | |
882 | |
834 | ecb_function_ unsigned char ecb_byteorder_helper (void) ecb_const; |
883 | ecb_inline unsigned char ecb_byteorder_helper (void) ecb_const; |
835 | ecb_function_ unsigned char |
884 | ecb_inline unsigned char |
836 | ecb_byteorder_helper (void) |
885 | ecb_byteorder_helper (void) |
837 | { |
886 | { |
838 | const uint32_t u = 0x11223344; |
887 | const uint32_t u = 0x11223344; |
839 | return *(unsigned char *)&u; |
888 | return *(unsigned char *)&u; |
840 | } |
889 | } |
841 | |
890 | |
842 | ecb_function_ ecb_bool ecb_big_endian (void) ecb_const; |
891 | ecb_inline ecb_bool ecb_big_endian (void) ecb_const; |
843 | ecb_function_ ecb_bool ecb_big_endian (void) { return ecb_byteorder_helper () == 0x11; } |
892 | ecb_inline ecb_bool ecb_big_endian (void) { return ecb_byteorder_helper () == 0x11; } |
844 | ecb_function_ ecb_bool ecb_little_endian (void) ecb_const; |
893 | ecb_inline ecb_bool ecb_little_endian (void) ecb_const; |
845 | ecb_function_ ecb_bool ecb_little_endian (void) { return ecb_byteorder_helper () == 0x44; } |
894 | ecb_inline ecb_bool ecb_little_endian (void) { return ecb_byteorder_helper () == 0x44; } |
846 | |
895 | |
847 | #if ECB_GCC_VERSION(3,0) || ECB_C99 |
896 | #if ECB_GCC_VERSION(3,0) || ECB_C99 |
848 | #define ecb_mod(m,n) ((m) % (n) + ((m) % (n) < 0 ? (n) : 0)) |
897 | #define ecb_mod(m,n) ((m) % (n) + ((m) % (n) < 0 ? (n) : 0)) |
849 | #else |
898 | #else |
850 | #define ecb_mod(m,n) ((m) < 0 ? ((n) - 1 - ((-1 - (m)) % (n))) : ((m) % (n))) |
899 | #define ecb_mod(m,n) ((m) < 0 ? ((n) - 1 - ((-1 - (m)) % (n))) : ((m) % (n))) |
|
|
900 | #endif |
|
|
901 | |
|
|
902 | #if __cplusplus |
|
|
903 | template<typename T> |
|
|
904 | static inline T ecb_div_rd (T val, T div) |
|
|
905 | { |
|
|
906 | return val < 0 ? - ((-val + div - 1) / div) : (val ) / div; |
|
|
907 | } |
|
|
908 | template<typename T> |
|
|
909 | static inline T ecb_div_ru (T val, T div) |
|
|
910 | { |
|
|
911 | return val < 0 ? - ((-val ) / div) : (val + div - 1) / div; |
|
|
912 | } |
|
|
913 | #else |
|
|
914 | #define ecb_div_rd(val,div) ((val) < 0 ? - ((-(val) + (div) - 1) / (div)) : ((val) ) / (div)) |
|
|
915 | #define ecb_div_ru(val,div) ((val) < 0 ? - ((-(val) ) / (div)) : ((val) + (div) - 1) / (div)) |
851 | #endif |
916 | #endif |
852 | |
917 | |
853 | #if ecb_cplusplus_does_not_suck |
918 | #if ecb_cplusplus_does_not_suck |
854 | /* does not work for local types (http://www.open-std.org/jtc1/sc22/wg21/docs/papers/2008/n2657.htm) */ |
919 | /* does not work for local types (http://www.open-std.org/jtc1/sc22/wg21/docs/papers/2008/n2657.htm) */ |
855 | template<typename T, int N> |
920 | template<typename T, int N> |
… | |
… | |
864 | #endif |
929 | #endif |
865 | |
930 | |
866 | /* ECB.H END */ |
931 | /* ECB.H END */ |
867 | |
932 | |
868 | #if ECB_MEMORY_FENCE_NEEDS_PTHREADS |
933 | #if ECB_MEMORY_FENCE_NEEDS_PTHREADS |
|
|
934 | /* if your architecture doesn't need memory fences, e.g. because it is |
|
|
935 | * single-cpu/core, or if you use libev in a project that doesn't use libev |
|
|
936 | * from multiple threads, then you can define ECB_AVOID_PTHREADS when compiling |
|
|
937 | * libev, in which cases the memory fences become nops. |
|
|
938 | * alternatively, you can remove this #error and link against libpthread, |
|
|
939 | * which will then provide the memory fences. |
|
|
940 | */ |
|
|
941 | # error "memory fences not defined for your architecture, please report" |
|
|
942 | #endif |
|
|
943 | |
869 | # undef ECB_MEMORY_FENCE |
944 | #ifndef ECB_MEMORY_FENCE |
870 | # undef ECB_MEMORY_FENCE_ACQUIRE |
945 | # define ECB_MEMORY_FENCE do { } while (0) |
871 | # undef ECB_MEMORY_FENCE_RELEASE |
946 | # define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE |
|
|
947 | # define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE |
872 | #endif |
948 | #endif |
873 | |
949 | |
874 | #define expect_false(cond) ecb_expect_false (cond) |
950 | #define expect_false(cond) ecb_expect_false (cond) |
875 | #define expect_true(cond) ecb_expect_true (cond) |
951 | #define expect_true(cond) ecb_expect_true (cond) |
876 | #define noinline ecb_noinline |
952 | #define noinline ecb_noinline |
… | |
… | |
1173 | #undef VAR |
1249 | #undef VAR |
1174 | }; |
1250 | }; |
1175 | #include "ev_wrap.h" |
1251 | #include "ev_wrap.h" |
1176 | |
1252 | |
1177 | static struct ev_loop default_loop_struct; |
1253 | static struct ev_loop default_loop_struct; |
1178 | struct ev_loop *ev_default_loop_ptr; |
1254 | EV_API_DECL struct ev_loop *ev_default_loop_ptr = 0; /* needs to be initialised to make it a definition despite extern */ |
1179 | |
1255 | |
1180 | #else |
1256 | #else |
1181 | |
1257 | |
1182 | ev_tstamp ev_rt_now; |
1258 | EV_API_DECL ev_tstamp ev_rt_now = 0; /* needs to be initialised to make it a definition despite extern */ |
1183 | #define VAR(name,decl) static decl; |
1259 | #define VAR(name,decl) static decl; |
1184 | #include "ev_vars.h" |
1260 | #include "ev_vars.h" |
1185 | #undef VAR |
1261 | #undef VAR |
1186 | |
1262 | |
1187 | static int ev_default_loop_ptr; |
1263 | static int ev_default_loop_ptr; |
… | |
… | |
1281 | |
1357 | |
1282 | do |
1358 | do |
1283 | ncur <<= 1; |
1359 | ncur <<= 1; |
1284 | while (cnt > ncur); |
1360 | while (cnt > ncur); |
1285 | |
1361 | |
1286 | /* if size is large, round to MALLOC_ROUND - 4 * longs to accomodate malloc overhead */ |
1362 | /* if size is large, round to MALLOC_ROUND - 4 * longs to accommodate malloc overhead */ |
1287 | if (elem * ncur > MALLOC_ROUND - sizeof (void *) * 4) |
1363 | if (elem * ncur > MALLOC_ROUND - sizeof (void *) * 4) |
1288 | { |
1364 | { |
1289 | ncur *= elem; |
1365 | ncur *= elem; |
1290 | ncur = (ncur + elem + (MALLOC_ROUND - 1) + sizeof (void *) * 4) & ~(MALLOC_ROUND - 1); |
1366 | ncur = (ncur + elem + (MALLOC_ROUND - 1) + sizeof (void *) * 4) & ~(MALLOC_ROUND - 1); |
1291 | ncur = ncur - sizeof (void *) * 4; |
1367 | ncur = ncur - sizeof (void *) * 4; |
… | |
… | |
1795 | /* win32 people keep sending patches that change this write() to send() */ |
1871 | /* win32 people keep sending patches that change this write() to send() */ |
1796 | /* and then run away. but send() is wrong, it wants a socket handle on win32 */ |
1872 | /* and then run away. but send() is wrong, it wants a socket handle on win32 */ |
1797 | /* so when you think this write should be a send instead, please find out */ |
1873 | /* so when you think this write should be a send instead, please find out */ |
1798 | /* where your send() is from - it's definitely not the microsoft send, and */ |
1874 | /* where your send() is from - it's definitely not the microsoft send, and */ |
1799 | /* tell me. thank you. */ |
1875 | /* tell me. thank you. */ |
|
|
1876 | /* it might be that your problem is that your environment needs EV_USE_WSASOCKET */ |
|
|
1877 | /* check the ev documentation on how to use this flag */ |
1800 | write (evpipe [1], &(evpipe [1]), 1); |
1878 | write (evpipe [1], &(evpipe [1]), 1); |
1801 | } |
1879 | } |
1802 | |
1880 | |
1803 | errno = old_errno; |
1881 | errno = old_errno; |
1804 | } |
1882 | } |
… | |
… | |
2971 | #endif |
3049 | #endif |
2972 | assert ((loop_done = EVBREAK_RECURSE, 1)); /* assert for side effect */ |
3050 | assert ((loop_done = EVBREAK_RECURSE, 1)); /* assert for side effect */ |
2973 | backend_poll (EV_A_ waittime); |
3051 | backend_poll (EV_A_ waittime); |
2974 | assert ((loop_done = EVBREAK_CANCEL, 1)); /* assert for side effect */ |
3052 | assert ((loop_done = EVBREAK_CANCEL, 1)); /* assert for side effect */ |
2975 | |
3053 | |
2976 | pipe_write_wanted = 0; /* just an optimsiation, no fence needed */ |
3054 | pipe_write_wanted = 0; /* just an optimisation, no fence needed */ |
2977 | |
3055 | |
2978 | if (pipe_write_skipped) |
3056 | if (pipe_write_skipped) |
2979 | { |
3057 | { |
2980 | assert (("libev: pipe_w not active, but pipe not written", ev_is_active (&pipe_w))); |
3058 | assert (("libev: pipe_w not active, but pipe not written", ev_is_active (&pipe_w))); |
2981 | ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); |
3059 | ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); |
… | |
… | |
3239 | |
3317 | |
3240 | void noinline |
3318 | void noinline |
3241 | ev_timer_again (EV_P_ ev_timer *w) |
3319 | ev_timer_again (EV_P_ ev_timer *w) |
3242 | { |
3320 | { |
3243 | EV_FREQUENT_CHECK; |
3321 | EV_FREQUENT_CHECK; |
|
|
3322 | |
|
|
3323 | clear_pending (EV_A_ (W)w); |
3244 | |
3324 | |
3245 | if (ev_is_active (w)) |
3325 | if (ev_is_active (w)) |
3246 | { |
3326 | { |
3247 | if (w->repeat) |
3327 | if (w->repeat) |
3248 | { |
3328 | { |
… | |
… | |
4394 | |
4474 | |
4395 | #if EV_MULTIPLICITY |
4475 | #if EV_MULTIPLICITY |
4396 | #include "ev_wrap.h" |
4476 | #include "ev_wrap.h" |
4397 | #endif |
4477 | #endif |
4398 | |
4478 | |
4399 | EV_CPP(}) |
|
|
4400 | |
|
|