ViewVC Help
View File | Revision Log | Show Annotations | Download File
/cvs/libev/ev.c
(Generate patch)

Comparing libev/ev.c (file contents):
Revision 1.435 by root, Sat May 26 08:52:09 2012 UTC vs.
Revision 1.446 by root, Mon Jun 11 12:50:50 2012 UTC

504 */ 504 */
505 505
506#ifndef ECB_H 506#ifndef ECB_H
507#define ECB_H 507#define ECB_H
508 508
509/* 16 bits major, 16 bits minor */
510#define ECB_VERSION 0x00010001
511
509#ifdef _WIN32 512#ifdef _WIN32
510 typedef signed char int8_t; 513 typedef signed char int8_t;
511 typedef unsigned char uint8_t; 514 typedef unsigned char uint8_t;
512 typedef signed short int16_t; 515 typedef signed short int16_t;
513 typedef unsigned short uint16_t; 516 typedef unsigned short uint16_t;
518 typedef unsigned long long uint64_t; 521 typedef unsigned long long uint64_t;
519 #else /* _MSC_VER || __BORLANDC__ */ 522 #else /* _MSC_VER || __BORLANDC__ */
520 typedef signed __int64 int64_t; 523 typedef signed __int64 int64_t;
521 typedef unsigned __int64 uint64_t; 524 typedef unsigned __int64 uint64_t;
522 #endif 525 #endif
526 #ifdef _WIN64
527 #define ECB_PTRSIZE 8
528 typedef uint64_t uintptr_t;
529 typedef int64_t intptr_t;
530 #else
531 #define ECB_PTRSIZE 4
532 typedef uint32_t uintptr_t;
533 typedef int32_t intptr_t;
534 #endif
535 typedef intptr_t ptrdiff_t;
523#else 536#else
524 #include <inttypes.h> 537 #include <inttypes.h>
538 #if UINTMAX_MAX > 0xffffffffU
539 #define ECB_PTRSIZE 8
540 #else
541 #define ECB_PTRSIZE 4
542 #endif
525#endif 543#endif
526 544
527/* many compilers define _GNUC_ to some versions but then only implement 545/* many compilers define _GNUC_ to some versions but then only implement
528 * what their idiot authors think are the "more important" extensions, 546 * what their idiot authors think are the "more important" extensions,
529 * causing enormous grief in return for some better fake benchmark numbers. 547 * causing enormous grief in return for some better fake benchmark numbers.
537 #else 555 #else
538 #define ECB_GCC_VERSION(major,minor) (__GNUC__ > (major) || (__GNUC__ == (major) && __GNUC_MINOR__ >= (minor))) 556 #define ECB_GCC_VERSION(major,minor) (__GNUC__ > (major) || (__GNUC__ == (major) && __GNUC_MINOR__ >= (minor)))
539 #endif 557 #endif
540#endif 558#endif
541 559
560#define ECB_C (__STDC__+0) /* this assumes that __STDC__ is either empty or a number */
561#define ECB_C99 (__STDC_VERSION__ >= 199901L)
562#define ECB_C11 (__STDC_VERSION__ >= 201112L)
563#define ECB_CPP (__cplusplus+0)
564#define ECB_CPP11 (__cplusplus >= 201103L)
565
542/*****************************************************************************/ 566/*****************************************************************************/
543 567
544/* ECB_NO_THREADS - ecb is not used by multiple threads, ever */ 568/* ECB_NO_THREADS - ecb is not used by multiple threads, ever */
545/* ECB_NO_SMP - ecb might be used in multiple threads, but only on a single cpu */ 569/* ECB_NO_SMP - ecb might be used in multiple threads, but only on a single cpu */
546 570
547#if ECB_NO_THREADS 571#if ECB_NO_THREADS
548# define ECB_NO_SMP 1 572 #define ECB_NO_SMP 1
549#endif 573#endif
550 574
551#if ECB_NO_THREADS || ECB_NO_SMP 575#if ECB_NO_SMP
552 #define ECB_MEMORY_FENCE do { } while (0) 576 #define ECB_MEMORY_FENCE do { } while (0)
553#endif 577#endif
554 578
555#ifndef ECB_MEMORY_FENCE 579#ifndef ECB_MEMORY_FENCE
556 #if ECB_GCC_VERSION(2,5) || defined __INTEL_COMPILER || (__llvm__ && __GNUC__) || __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110 580 #if ECB_GCC_VERSION(2,5) || defined __INTEL_COMPILER || (__llvm__ && __GNUC__) || __SUNPRO_C >= 0x5110 || __SUNPRO_CC >= 0x5110
557 #if __i386 || __i386__ 581 #if __i386 || __i386__
558 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory") 582 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("lock; orb $0, -1(%%esp)" : : : "memory")
559 #define ECB_MEMORY_FENCE_ACQUIRE ECB_MEMORY_FENCE /* non-lock xchg might be enough */ 583 #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory")
560 #define ECB_MEMORY_FENCE_RELEASE do { } while (0) /* unlikely to change in future cpus */ 584 #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("")
561 #elif __amd64 || __amd64__ || __x86_64 || __x86_64__ 585 #elif __amd64 || __amd64__ || __x86_64 || __x86_64__
562 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory") 586 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mfence" : : : "memory")
563 #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("lfence" : : : "memory") 587 #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("" : : : "memory")
564 #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("sfence") /* play safe - not needed in any current cpu */ 588 #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("")
565 #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__ 589 #elif __powerpc__ || __ppc__ || __powerpc64__ || __ppc64__
566 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") 590 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory")
567 #elif defined __ARM_ARCH_6__ || defined __ARM_ARCH_6J__ \ 591 #elif defined __ARM_ARCH_6__ || defined __ARM_ARCH_6J__ \
568 || defined __ARM_ARCH_6K__ || defined __ARM_ARCH_6ZK__ 592 || defined __ARM_ARCH_6K__ || defined __ARM_ARCH_6ZK__
569 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,5" : : "r" (0) : "memory") 593 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mcr p15,0,%0,c7,c10,5" : : "r" (0) : "memory")
570 #elif defined __ARM_ARCH_7__ || defined __ARM_ARCH_7A__ \ 594 #elif defined __ARM_ARCH_7__ || defined __ARM_ARCH_7A__ \
571 || defined __ARM_ARCH_7M__ || defined __ARM_ARCH_7R__ 595 || defined __ARM_ARCH_7M__ || defined __ARM_ARCH_7R__
572 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory") 596 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("dmb" : : : "memory")
573 #elif __sparc || __sparc__ 597 #elif __sparc || __sparc__
574 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad | " : : : "memory") 598 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad | #StoreStore | #StoreLoad" : : : "memory")
575 #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory") 599 #define ECB_MEMORY_FENCE_ACQUIRE __asm__ __volatile__ ("membar #LoadStore | #LoadLoad" : : : "memory")
576 #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore") 600 #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("membar #LoadStore | #StoreStore")
577 #elif defined __s390__ || defined __s390x__ 601 #elif defined __s390__ || defined __s390x__
578 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory") 602 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("bcr 15,0" : : : "memory")
579 #elif defined __mips__ 603 #elif defined __mips__
580 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory") 604 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("sync" : : : "memory")
581 #elif defined __alpha__ 605 #elif defined __alpha__
582 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mb" : : : "memory") 606 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mb" : : : "memory")
607 #elif defined __hppa__
608 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("" : : : "memory")
609 #define ECB_MEMORY_FENCE_RELEASE __asm__ __volatile__ ("")
610 #elif defined __ia64__
611 #define ECB_MEMORY_FENCE __asm__ __volatile__ ("mf" : : : "memory")
583 #endif 612 #endif
584 #endif 613 #endif
585#endif 614#endif
586 615
587#ifndef ECB_MEMORY_FENCE 616#ifndef ECB_MEMORY_FENCE
617 #if ECB_GCC_VERSION(4,7)
618 /* see comment below (stdatomic.h) about the C11 memory model. */
619 #define ECB_MEMORY_FENCE __atomic_thread_fence (__ATOMIC_SEQ_CST)
620 #elif defined __clang && __has_feature (cxx_atomic)
621 /* see comment below (stdatomic.h) about the C11 memory model. */
622 #define ECB_MEMORY_FENCE __c11_atomic_thread_fence (__ATOMIC_SEQ_CST)
588 #if ECB_GCC_VERSION(4,4) || defined __INTEL_COMPILER || defined __clang__ 623 #elif ECB_GCC_VERSION(4,4) || defined __INTEL_COMPILER || defined __clang__
589 #define ECB_MEMORY_FENCE __sync_synchronize () 624 #define ECB_MEMORY_FENCE __sync_synchronize ()
590 /*#define ECB_MEMORY_FENCE_ACQUIRE ({ char dummy = 0; __sync_lock_test_and_set (&dummy, 1); }) */
591 /*#define ECB_MEMORY_FENCE_RELEASE ({ char dummy = 1; __sync_lock_release (&dummy ); }) */
592 #elif _MSC_VER >= 1400 /* VC++ 2005 */ 625 #elif _MSC_VER >= 1400 /* VC++ 2005 */
593 #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier) 626 #pragma intrinsic(_ReadBarrier,_WriteBarrier,_ReadWriteBarrier)
594 #define ECB_MEMORY_FENCE _ReadWriteBarrier () 627 #define ECB_MEMORY_FENCE _ReadWriteBarrier ()
595 #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */ 628 #define ECB_MEMORY_FENCE_ACQUIRE _ReadWriteBarrier () /* according to msdn, _ReadBarrier is not a load fence */
596 #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier () 629 #define ECB_MEMORY_FENCE_RELEASE _WriteBarrier ()
606 #define ECB_MEMORY_FENCE __sync () 639 #define ECB_MEMORY_FENCE __sync ()
607 #endif 640 #endif
608#endif 641#endif
609 642
610#ifndef ECB_MEMORY_FENCE 643#ifndef ECB_MEMORY_FENCE
644 #if ECB_C11 && !defined __STDC_NO_ATOMICS__
645 /* we assume that these memory fences work on all variables/all memory accesses, */
646 /* not just C11 atomics and atomic accesses */
647 #include <stdatomic.h>
648 /* Unfortunately, neither gcc 4.7 nor clang 3.1 generate any instructions for */
649 /* any fence other than seq_cst, which isn't very efficient for us. */
650 /* Why that is, we don't know - either the C11 memory model is quite useless */
651 /* for most usages, or gcc and clang have a bug */
652 /* I *currently* lean towards the latter, and inefficiently implement */
653 /* all three of ecb's fences as a seq_cst fence */
654 #define ECB_MEMORY_FENCE atomic_thread_fence (memory_order_seq_cst)
655 #endif
656#endif
657
658#ifndef ECB_MEMORY_FENCE
611 #if !ECB_AVOID_PTHREADS 659 #if !ECB_AVOID_PTHREADS
612 /* 660 /*
613 * if you get undefined symbol references to pthread_mutex_lock, 661 * if you get undefined symbol references to pthread_mutex_lock,
614 * or failure to find pthread.h, then you should implement 662 * or failure to find pthread.h, then you should implement
615 * the ECB_MEMORY_FENCE operations for your cpu/compiler 663 * the ECB_MEMORY_FENCE operations for your cpu/compiler
633 #define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE 681 #define ECB_MEMORY_FENCE_RELEASE ECB_MEMORY_FENCE
634#endif 682#endif
635 683
636/*****************************************************************************/ 684/*****************************************************************************/
637 685
638#define ECB_C99 (__STDC_VERSION__ >= 199901L)
639
640#if __cplusplus 686#if __cplusplus
641 #define ecb_inline static inline 687 #define ecb_inline static inline
642#elif ECB_GCC_VERSION(2,5) 688#elif ECB_GCC_VERSION(2,5)
643 #define ecb_inline static __inline__ 689 #define ecb_inline static __inline__
644#elif ECB_C99 690#elif ECB_C99
682#elif ECB_GCC_VERSION(3,0) 728#elif ECB_GCC_VERSION(3,0)
683 #define ecb_decltype(x) __typeof(x) 729 #define ecb_decltype(x) __typeof(x)
684#endif 730#endif
685 731
686#define ecb_noinline ecb_attribute ((__noinline__)) 732#define ecb_noinline ecb_attribute ((__noinline__))
687#define ecb_noreturn ecb_attribute ((__noreturn__))
688#define ecb_unused ecb_attribute ((__unused__)) 733#define ecb_unused ecb_attribute ((__unused__))
689#define ecb_const ecb_attribute ((__const__)) 734#define ecb_const ecb_attribute ((__const__))
690#define ecb_pure ecb_attribute ((__pure__)) 735#define ecb_pure ecb_attribute ((__pure__))
736
737#if ECB_C11
738 #define ecb_noreturn _Noreturn
739#else
740 #define ecb_noreturn ecb_attribute ((__noreturn__))
741#endif
691 742
692#if ECB_GCC_VERSION(4,3) 743#if ECB_GCC_VERSION(4,3)
693 #define ecb_artificial ecb_attribute ((__artificial__)) 744 #define ecb_artificial ecb_attribute ((__artificial__))
694 #define ecb_hot ecb_attribute ((__hot__)) 745 #define ecb_hot ecb_attribute ((__hot__))
695 #define ecb_cold ecb_attribute ((__cold__)) 746 #define ecb_cold ecb_attribute ((__cold__))
785 if (x >> 32) { x >>= 32; r += 32; } 836 if (x >> 32) { x >>= 32; r += 32; }
786 837
787 return r + ecb_ld32 (x); 838 return r + ecb_ld32 (x);
788 } 839 }
789#endif 840#endif
841
842ecb_function_ ecb_bool ecb_is_pot32 (uint32_t x) ecb_const;
843ecb_function_ ecb_bool ecb_is_pot32 (uint32_t x) { return !(x & (x - 1)); }
844ecb_function_ ecb_bool ecb_is_pot64 (uint64_t x) ecb_const;
845ecb_function_ ecb_bool ecb_is_pot64 (uint64_t x) { return !(x & (x - 1)); }
790 846
791ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) ecb_const; 847ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) ecb_const;
792ecb_function_ uint8_t ecb_bitrev8 (uint8_t x) 848ecb_function_ uint8_t ecb_bitrev8 (uint8_t x)
793{ 849{
794 return ( (x * 0x0802U & 0x22110U) 850 return ( (x * 0x0802U & 0x22110U)
1849 1905
1850 if (expect_true (*flag)) 1906 if (expect_true (*flag))
1851 return; 1907 return;
1852 1908
1853 *flag = 1; 1909 *flag = 1;
1854
1855 ECB_MEMORY_FENCE_RELEASE; /* make sure flag is visible before the wakeup */ 1910 ECB_MEMORY_FENCE_RELEASE; /* make sure flag is visible before the wakeup */
1856 1911
1857 pipe_write_skipped = 1; 1912 pipe_write_skipped = 1;
1858 1913
1859 ECB_MEMORY_FENCE; /* make sure pipe_write_skipped is visible before we check pipe_write_wanted */ 1914 ECB_MEMORY_FENCE; /* make sure pipe_write_skipped is visible before we check pipe_write_wanted */
1860 1915
1861 if (pipe_write_wanted) 1916 if (pipe_write_wanted)
1862 { 1917 {
1863 int old_errno; 1918 int old_errno;
1864 1919
1865 pipe_write_skipped = 0; /* just an optimisation, no fence needed */ 1920 pipe_write_skipped = 0;
1921 ECB_MEMORY_FENCE_RELEASE;
1866 1922
1867 old_errno = errno; /* save errno because write will clobber it */ 1923 old_errno = errno; /* save errno because write will clobber it */
1868 1924
1869#if EV_USE_EVENTFD 1925#if EV_USE_EVENTFD
1870 if (evfd >= 0) 1926 if (evfd >= 0)
1929#if EV_SIGNAL_ENABLE 1985#if EV_SIGNAL_ENABLE
1930 if (sig_pending) 1986 if (sig_pending)
1931 { 1987 {
1932 sig_pending = 0; 1988 sig_pending = 0;
1933 1989
1934 ECB_MEMORY_FENCE_RELEASE; 1990 ECB_MEMORY_FENCE;
1935 1991
1936 for (i = EV_NSIG - 1; i--; ) 1992 for (i = EV_NSIG - 1; i--; )
1937 if (expect_false (signals [i].pending)) 1993 if (expect_false (signals [i].pending))
1938 ev_feed_signal_event (EV_A_ i + 1); 1994 ev_feed_signal_event (EV_A_ i + 1);
1939 } 1995 }
1942#if EV_ASYNC_ENABLE 1998#if EV_ASYNC_ENABLE
1943 if (async_pending) 1999 if (async_pending)
1944 { 2000 {
1945 async_pending = 0; 2001 async_pending = 0;
1946 2002
1947 ECB_MEMORY_FENCE_RELEASE; 2003 ECB_MEMORY_FENCE;
1948 2004
1949 for (i = asynccnt; i--; ) 2005 for (i = asynccnt; i--; )
1950 if (asyncs [i]->sent) 2006 if (asyncs [i]->sent)
1951 { 2007 {
1952 asyncs [i]->sent = 0; 2008 asyncs [i]->sent = 0;
2009 ECB_MEMORY_FENCE_RELEASE;
1953 ev_feed_event (EV_A_ asyncs [i], EV_ASYNC); 2010 ev_feed_event (EV_A_ asyncs [i], EV_ASYNC);
1954 } 2011 }
1955 } 2012 }
1956#endif 2013#endif
1957} 2014}
1988void noinline 2045void noinline
1989ev_feed_signal_event (EV_P_ int signum) EV_THROW 2046ev_feed_signal_event (EV_P_ int signum) EV_THROW
1990{ 2047{
1991 WL w; 2048 WL w;
1992 2049
1993 if (expect_false (signum <= 0 || signum > EV_NSIG)) 2050 if (expect_false (signum <= 0 || signum >= EV_NSIG))
1994 return; 2051 return;
1995 2052
1996 --signum; 2053 --signum;
1997 2054
1998#if EV_MULTIPLICITY 2055#if EV_MULTIPLICITY
2002 if (expect_false (signals [signum].loop != EV_A)) 2059 if (expect_false (signals [signum].loop != EV_A))
2003 return; 2060 return;
2004#endif 2061#endif
2005 2062
2006 signals [signum].pending = 0; 2063 signals [signum].pending = 0;
2064 ECB_MEMORY_FENCE_RELEASE;
2007 2065
2008 for (w = signals [signum].head; w; w = w->next) 2066 for (w = signals [signum].head; w; w = w->next)
2009 ev_feed_event (EV_A_ (W)w, EV_SIGNAL); 2067 ev_feed_event (EV_A_ (W)w, EV_SIGNAL);
2010} 2068}
2011 2069
2492 EV_WIN32_CLOSE_FD (evpipe [1]); 2550 EV_WIN32_CLOSE_FD (evpipe [1]);
2493 } 2551 }
2494 2552
2495#if EV_SIGNAL_ENABLE || EV_ASYNC_ENABLE 2553#if EV_SIGNAL_ENABLE || EV_ASYNC_ENABLE
2496 evpipe_init (EV_A); 2554 evpipe_init (EV_A);
2497 /* now iterate over everything, in case we missed something */ 2555 /* iterate over everything, in case we missed something before */
2498 pipecb (EV_A_ &pipe_w, EV_READ); 2556 ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM);
2499#endif 2557#endif
2500 } 2558 }
2501 2559
2502 postfork = 0; 2560 postfork = 0;
2503} 2561}
2678} 2736}
2679 2737
2680void 2738void
2681ev_loop_fork (EV_P) EV_THROW 2739ev_loop_fork (EV_P) EV_THROW
2682{ 2740{
2683 postfork = 1; /* must be in line with ev_default_fork */ 2741 postfork = 1;
2684} 2742}
2685 2743
2686/*****************************************************************************/ 2744/*****************************************************************************/
2687 2745
2688void 2746void
2704} 2762}
2705 2763
2706void noinline 2764void noinline
2707ev_invoke_pending (EV_P) 2765ev_invoke_pending (EV_P)
2708{ 2766{
2709 for (pendingpri = NUMPRI; pendingpri--; ) /* pendingpri is modified during the loop */ 2767 pendingpri = NUMPRI;
2768
2769 while (pendingpri) /* pendingpri possibly gets modified in the inner loop */
2770 {
2771 --pendingpri;
2772
2710 while (pendingcnt [pendingpri]) 2773 while (pendingcnt [pendingpri])
2711 { 2774 {
2712 ANPENDING *p = pendings [pendingpri] + --pendingcnt [pendingpri]; 2775 ANPENDING *p = pendings [pendingpri] + --pendingcnt [pendingpri];
2713 2776
2714 p->w->pending = 0; 2777 p->w->pending = 0;
2715 EV_CB_INVOKE (p->w, p->events); 2778 EV_CB_INVOKE (p->w, p->events);
2716 EV_FREQUENT_CHECK; 2779 EV_FREQUENT_CHECK;
2717 } 2780 }
2781 }
2718} 2782}
2719 2783
2720#if EV_IDLE_ENABLE 2784#if EV_IDLE_ENABLE
2721/* make idle watchers pending. this handles the "call-idle */ 2785/* make idle watchers pending. this handles the "call-idle */
2722/* only when higher priorities are idle" logic */ 2786/* only when higher priorities are idle" logic */
3080 backend_poll (EV_A_ waittime); 3144 backend_poll (EV_A_ waittime);
3081 assert ((loop_done = EVBREAK_CANCEL, 1)); /* assert for side effect */ 3145 assert ((loop_done = EVBREAK_CANCEL, 1)); /* assert for side effect */
3082 3146
3083 pipe_write_wanted = 0; /* just an optimisation, no fence needed */ 3147 pipe_write_wanted = 0; /* just an optimisation, no fence needed */
3084 3148
3149 ECB_MEMORY_FENCE_ACQUIRE;
3085 if (pipe_write_skipped) 3150 if (pipe_write_skipped)
3086 { 3151 {
3087 assert (("libev: pipe_w not active, but pipe not written", ev_is_active (&pipe_w))); 3152 assert (("libev: pipe_w not active, but pipe not written", ev_is_active (&pipe_w)));
3088 ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM); 3153 ev_feed_event (EV_A_ &pipe_w, EV_CUSTOM);
3089 } 3154 }

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines