--- libev/ev.c 2007/11/12 05:40:55 1.108 +++ libev/ev.c 2009/07/10 00:36:21 1.297 @@ -1,66 +1,148 @@ /* * libev event processing core, watcher management * - * Copyright (c) 2007 Marc Alexander Lehmann + * Copyright (c) 2007,2008,2009 Marc Alexander Lehmann * All rights reserved. * - * Redistribution and use in source and binary forms, with or without - * modification, are permitted provided that the following conditions are - * met: + * Redistribution and use in source and binary forms, with or without modifica- + * tion, are permitted provided that the following conditions are met: + * + * 1. Redistributions of source code must retain the above copyright notice, + * this list of conditions and the following disclaimer. + * + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED + * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MER- + * CHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO + * EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPE- + * CIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, + * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; + * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, + * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTH- + * ERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED + * OF THE POSSIBILITY OF SUCH DAMAGE. * - * * Redistributions of source code must retain the above copyright - * notice, this list of conditions and the following disclaimer. - * - * * Redistributions in binary form must reproduce the above - * copyright notice, this list of conditions and the following - * disclaimer in the documentation and/or other materials provided - * with the distribution. - * - * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS - * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT - * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR - * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT - * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, - * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT - * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, - * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY - * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT - * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE - * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + * Alternatively, the contents of this file may be used under the terms of + * the GNU General Public License ("GPL") version 2 or any later version, + * in which case the provisions of the GPL are applicable instead of + * the above. If you wish to allow the use of your version of this file + * only under the terms of the GPL and not to allow others to use your + * version of this file under the BSD license, indicate your decision + * by deleting the provisions above and replace them with the notice + * and other provisions required by the GPL. If you do not delete the + * provisions above, a recipient may use your version of this file under + * either the BSD or the GPL. */ #ifdef __cplusplus extern "C" { #endif +/* this big block deduces configuration from config.h */ #ifndef EV_STANDALONE -# include "config.h" +# ifdef EV_CONFIG_H +# include EV_CONFIG_H +# else +# include "config.h" +# endif + +# if HAVE_CLOCK_SYSCALL +# ifndef EV_USE_CLOCK_SYSCALL +# define EV_USE_CLOCK_SYSCALL 1 +# ifndef EV_USE_REALTIME +# define EV_USE_REALTIME 0 +# endif +# ifndef EV_USE_MONOTONIC +# define EV_USE_MONOTONIC 1 +# endif +# endif +# elif !defined(EV_USE_CLOCK_SYSCALL) +# define EV_USE_CLOCK_SYSCALL 0 +# endif # if HAVE_CLOCK_GETTIME # ifndef EV_USE_MONOTONIC # define EV_USE_MONOTONIC 1 # endif # ifndef EV_USE_REALTIME -# define EV_USE_REALTIME 1 +# define EV_USE_REALTIME 0 +# endif +# else +# ifndef EV_USE_MONOTONIC +# define EV_USE_MONOTONIC 0 +# endif +# ifndef EV_USE_REALTIME +# define EV_USE_REALTIME 0 # endif # endif -# if HAVE_SELECT && HAVE_SYS_SELECT_H && !defined (EV_USE_SELECT) -# define EV_USE_SELECT 1 +# ifndef EV_USE_NANOSLEEP +# if HAVE_NANOSLEEP +# define EV_USE_NANOSLEEP 1 +# else +# define EV_USE_NANOSLEEP 0 +# endif # endif -# if HAVE_POLL && HAVE_POLL_H && !defined (EV_USE_POLL) -# define EV_USE_POLL 1 +# ifndef EV_USE_SELECT +# if HAVE_SELECT && HAVE_SYS_SELECT_H +# define EV_USE_SELECT 1 +# else +# define EV_USE_SELECT 0 +# endif # endif -# if HAVE_EPOLL_CTL && HAVE_SYS_EPOLL_H && !defined (EV_USE_EPOLL) -# define EV_USE_EPOLL 1 +# ifndef EV_USE_POLL +# if HAVE_POLL && HAVE_POLL_H +# define EV_USE_POLL 1 +# else +# define EV_USE_POLL 0 +# endif +# endif + +# ifndef EV_USE_EPOLL +# if HAVE_EPOLL_CTL && HAVE_SYS_EPOLL_H +# define EV_USE_EPOLL 1 +# else +# define EV_USE_EPOLL 0 +# endif +# endif + +# ifndef EV_USE_KQUEUE +# if HAVE_KQUEUE && HAVE_SYS_EVENT_H && HAVE_SYS_QUEUE_H +# define EV_USE_KQUEUE 1 +# else +# define EV_USE_KQUEUE 0 +# endif +# endif + +# ifndef EV_USE_PORT +# if HAVE_PORT_H && HAVE_PORT_CREATE +# define EV_USE_PORT 1 +# else +# define EV_USE_PORT 0 +# endif # endif -# if HAVE_KQUEUE && HAVE_SYS_EVENT_H && HAVE_SYS_QUEUE_H && !defined (EV_USE_KQUEUE) -# define EV_USE_KQUEUE 1 +# ifndef EV_USE_INOTIFY +# if HAVE_INOTIFY_INIT && HAVE_SYS_INOTIFY_H +# define EV_USE_INOTIFY 1 +# else +# define EV_USE_INOTIFY 0 +# endif # endif +# ifndef EV_USE_EVENTFD +# if HAVE_EVENTFD +# define EV_USE_EVENTFD 1 +# else +# define EV_USE_EVENTFD 0 +# endif +# endif + #endif #include @@ -77,11 +159,18 @@ #include +#ifdef EV_H +# include EV_H +#else +# include "ev.h" +#endif + #ifndef _WIN32 -# include # include # include +# include #else +# include # define WIN32_LEAN_AND_MEAN # include # ifndef EV_SELECT_IS_WINSOCKET @@ -89,15 +178,38 @@ # endif #endif -/**/ +/* this block tries to deduce configuration from header-defined symbols and defaults */ + +#ifndef EV_USE_CLOCK_SYSCALL +# if __linux && __GLIBC__ >= 2 +# define EV_USE_CLOCK_SYSCALL 1 +# else +# define EV_USE_CLOCK_SYSCALL 0 +# endif +#endif #ifndef EV_USE_MONOTONIC -# define EV_USE_MONOTONIC 1 +# if defined (_POSIX_MONOTONIC_CLOCK) && _POSIX_MONOTONIC_CLOCK >= 0 +# define EV_USE_MONOTONIC 1 +# else +# define EV_USE_MONOTONIC 0 +# endif +#endif + +#ifndef EV_USE_REALTIME +# define EV_USE_REALTIME !EV_USE_CLOCK_SYSCALL +#endif + +#ifndef EV_USE_NANOSLEEP +# if _POSIX_C_SOURCE >= 199309L +# define EV_USE_NANOSLEEP 1 +# else +# define EV_USE_NANOSLEEP 0 +# endif #endif #ifndef EV_USE_SELECT # define EV_USE_SELECT 1 -# define EV_SELECT_USE_FD_SET 1 #endif #ifndef EV_USE_POLL @@ -109,25 +221,87 @@ #endif #ifndef EV_USE_EPOLL -# define EV_USE_EPOLL 0 +# if __linux && (__GLIBC__ > 2 || (__GLIBC__ == 2 && __GLIBC_MINOR__ >= 4)) +# define EV_USE_EPOLL 1 +# else +# define EV_USE_EPOLL 0 +# endif #endif #ifndef EV_USE_KQUEUE # define EV_USE_KQUEUE 0 #endif -#ifndef EV_USE_REALTIME -# define EV_USE_REALTIME 1 +#ifndef EV_USE_PORT +# define EV_USE_PORT 0 #endif -/**/ +#ifndef EV_USE_INOTIFY +# if __linux && (__GLIBC__ > 2 || (__GLIBC__ == 2 && __GLIBC_MINOR__ >= 4)) +# define EV_USE_INOTIFY 1 +# else +# define EV_USE_INOTIFY 0 +# endif +#endif -/* darwin simply cannot be helped */ -#ifdef __APPLE__ -# undef EV_USE_POLL -# undef EV_USE_KQUEUE +#ifndef EV_PID_HASHSIZE +# if EV_MINIMAL +# define EV_PID_HASHSIZE 1 +# else +# define EV_PID_HASHSIZE 16 +# endif +#endif + +#ifndef EV_INOTIFY_HASHSIZE +# if EV_MINIMAL +# define EV_INOTIFY_HASHSIZE 1 +# else +# define EV_INOTIFY_HASHSIZE 16 +# endif +#endif + +#ifndef EV_USE_EVENTFD +# if __linux && (__GLIBC__ > 2 || (__GLIBC__ == 2 && __GLIBC_MINOR__ >= 7)) +# define EV_USE_EVENTFD 1 +# else +# define EV_USE_EVENTFD 0 +# endif #endif +#if 0 /* debugging */ +# define EV_VERIFY 3 +# define EV_USE_4HEAP 1 +# define EV_HEAP_CACHE_AT 1 +#endif + +#ifndef EV_VERIFY +# define EV_VERIFY !EV_MINIMAL +#endif + +#ifndef EV_USE_4HEAP +# define EV_USE_4HEAP !EV_MINIMAL +#endif + +#ifndef EV_HEAP_CACHE_AT +# define EV_HEAP_CACHE_AT !EV_MINIMAL +#endif + +/* on linux, we can use a (slow) syscall to avoid a dependency on pthread, */ +/* which makes programs even slower. might work on other unices, too. */ +#if EV_USE_CLOCK_SYSCALL +# include +# ifdef SYS_clock_gettime +# define clock_gettime(id, ts) syscall (SYS_clock_gettime, (id), (ts)) +# undef EV_USE_MONOTONIC +# define EV_USE_MONOTONIC 1 +# else +# undef EV_USE_CLOCK_SYSCALL +# define EV_USE_CLOCK_SYSCALL 0 +# endif +#endif + +/* this block fixes any misconfiguration where we know we run into trouble otherwise */ + #ifndef CLOCK_MONOTONIC # undef EV_USE_MONOTONIC # define EV_USE_MONOTONIC 0 @@ -138,44 +312,114 @@ # define EV_USE_REALTIME 0 #endif +#if !EV_STAT_ENABLE +# undef EV_USE_INOTIFY +# define EV_USE_INOTIFY 0 +#endif + +#if !EV_USE_NANOSLEEP +# ifndef _WIN32 +# include +# endif +#endif + +#if EV_USE_INOTIFY +# include +# include +# include +/* some very old inotify.h headers don't have IN_DONT_FOLLOW */ +# ifndef IN_DONT_FOLLOW +# undef EV_USE_INOTIFY +# define EV_USE_INOTIFY 0 +# endif +#endif + #if EV_SELECT_IS_WINSOCKET # include #endif -/**/ +#if EV_USE_EVENTFD +/* our minimum requirement is glibc 2.7 which has the stub, but not the header */ +# include +# ifdef __cplusplus +extern "C" { +# endif +int eventfd (unsigned int initval, int flags); +# ifdef __cplusplus +} +# endif +#endif -#define MIN_TIMEJUMP 1. /* minimum timejump that gets detected (if monotonic clock available) */ -#define MAX_BLOCKTIME 59.731 /* never wait longer than this time (to detect time jumps) */ -#define PID_HASHSIZE 16 /* size of pid hash table, must be power of two */ -/*#define CLEANUP_INTERVAL 300. /* how often to try to free memory and re-check fds */ +/**/ -#ifdef EV_H -# include EV_H +#if EV_VERIFY >= 3 +# define EV_FREQUENT_CHECK ev_loop_verify (EV_A) #else -# include "ev.h" +# define EV_FREQUENT_CHECK do { } while (0) #endif -#if __GNUC__ >= 3 +/* + * This is used to avoid floating point rounding problems. + * It is added to ev_rt_now when scheduling periodics + * to ensure progress, time-wise, even when rounding + * errors are against us. + * This value is good at least till the year 4000. + * Better solutions welcome. + */ +#define TIME_EPSILON 0.0001220703125 /* 1/8192 */ + +#define MIN_TIMEJUMP 1. /* minimum timejump that gets detected (if monotonic clock available) */ +#define MAX_BLOCKTIME 59.743 /* never wait longer than this time (to detect time jumps) */ +/*#define CLEANUP_INTERVAL (MAX_BLOCKTIME * 5.) /* how often to try to free memory and re-check fds, TODO */ + +#if __GNUC__ >= 4 # define expect(expr,value) __builtin_expect ((expr),(value)) -# define inline inline +# define noinline __attribute__ ((noinline)) #else # define expect(expr,value) (expr) -# define inline static +# define noinline +# if __STDC_VERSION__ < 199901L && __GNUC__ < 2 +# define inline +# endif #endif #define expect_false(expr) expect ((expr) != 0, 0) #define expect_true(expr) expect ((expr) != 0, 1) +#define inline_size static inline -#define NUMPRI (EV_MAXPRI - EV_MINPRI + 1) -#define ABSPRI(w) ((w)->priority - EV_MINPRI) +#if EV_MINIMAL +# define inline_speed static noinline +#else +# define inline_speed static inline +#endif + +#define NUMPRI (EV_MAXPRI - EV_MINPRI + 1) -#define EMPTY /* required for microsofts broken pseudo-c compiler */ +#if EV_MINPRI == EV_MAXPRI +# define ABSPRI(w) (((W)w), 0) +#else +# define ABSPRI(w) (((W)w)->priority - EV_MINPRI) +#endif -typedef struct ev_watcher *W; -typedef struct ev_watcher_list *WL; -typedef struct ev_watcher_time *WT; +#define EMPTY /* required for microsofts broken pseudo-c compiler */ +#define EMPTY2(a,b) /* used to suppress some warnings */ -static int have_monotonic; /* did clock_gettime (CLOCK_MONOTONIC) work? */ +typedef ev_watcher *W; +typedef ev_watcher_list *WL; +typedef ev_watcher_time *WT; + +#define ev_active(w) ((W)(w))->active +#define ev_at(w) ((WT)(w))->at + +#if EV_USE_REALTIME +/* sig_atomic_t is used to avoid per-thread variables or locking but still */ +/* giving it a reasonably high chance of working on typical architetcures */ +static EV_ATOMIC_T have_realtime; /* did clock_gettime (CLOCK_REALTIME) work? */ +#endif + +#if EV_USE_MONOTONIC +static EV_ATOMIC_T have_monotonic; /* did clock_gettime (CLOCK_MONOTONIC) work? */ +#endif #ifdef _WIN32 # include "ev_win32.c" @@ -185,13 +429,14 @@ static void (*syserr_cb)(const char *msg); -void ev_set_syserr_cb (void (*cb)(const char *msg)) +void +ev_set_syserr_cb (void (*cb)(const char *msg)) { syserr_cb = cb; } -static void -syserr (const char *msg) +static void noinline +ev_syserr (const char *msg) { if (!msg) msg = "(libev) system error"; @@ -205,17 +450,33 @@ } } -static void *(*alloc)(void *ptr, long size); +static void * +ev_realloc_emul (void *ptr, long size) +{ + /* some systems, notably openbsd and darwin, fail to properly + * implement realloc (x, 0) (as required by both ansi c-98 and + * the single unix specification, so work around them here. + */ + + if (size) + return realloc (ptr, size); + + free (ptr); + return 0; +} + +static void *(*alloc)(void *ptr, long size) = ev_realloc_emul; -void ev_set_allocator (void *(*cb)(void *ptr, long size)) +void +ev_set_allocator (void *(*cb)(void *ptr, long size)) { alloc = cb; } -static void * +inline_speed void * ev_realloc (void *ptr, long size) { - ptr = alloc ? alloc (ptr, size) : realloc (ptr, size); + ptr = alloc (ptr, size); if (!ptr && size) { @@ -231,22 +492,57 @@ /*****************************************************************************/ +/* file descriptor info structure */ typedef struct { WL head; - unsigned char events; - unsigned char reify; + unsigned char events; /* the events watched for */ + unsigned char reify; /* flag set when this ANFD needs reification */ + unsigned char emask; /* the epoll backend stores the actual kernel mask in here */ + unsigned char unused; +#if EV_USE_EPOLL + unsigned int egen; /* generation counter to counter epoll bugs */ +#endif #if EV_SELECT_IS_WINSOCKET SOCKET handle; #endif } ANFD; +/* stores the pending event set for a given watcher */ typedef struct { W w; - int events; + int events; /* the pending event set for the given watcher */ } ANPENDING; +#if EV_USE_INOTIFY +/* hash table entry per inotify-id */ +typedef struct +{ + WL head; +} ANFS; +#endif + +/* Heap Entry */ +#if EV_HEAP_CACHE_AT + /* a heap element */ + typedef struct { + ev_tstamp at; + WT w; + } ANHE; + + #define ANHE_w(he) (he).w /* access watcher, read-write */ + #define ANHE_at(he) (he).at /* access cached at, read-only */ + #define ANHE_at_cache(he) (he).at = (he).w->at /* update at from watcher */ +#else + /* a heap element */ + typedef WT ANHE; + + #define ANHE_w(he) (he) + #define ANHE_at(he) (he)->at + #define ANHE_at_cache(he) +#endif + #if EV_MULTIPLICITY struct ev_loop @@ -259,8 +555,8 @@ }; #include "ev_wrap.h" - struct ev_loop default_loop_struct; - static struct ev_loop *default_loop; + static struct ev_loop default_loop_struct; + struct ev_loop *ev_default_loop_ptr; #else @@ -269,27 +565,42 @@ #include "ev_vars.h" #undef VAR - static int default_loop; + static int ev_default_loop_ptr; #endif +#if EV_MINIMAL < 2 +# define EV_SUSPEND_CB if (expect_false (suspend_cb)) suspend_cb (EV_A) +# define EV_RESUME_CB if (expect_false (resume_cb )) resume_cb (EV_A) +# define EV_INVOKE_PENDING invoke_cb (EV_A) +#else +# define EV_SUSPEND_CB (void)0 +# define EV_RESUME_CB (void)0 +# define EV_INVOKE_PENDING ev_invoke_pending (EV_A) +#endif + /*****************************************************************************/ +#ifndef EV_HAVE_EV_TIME ev_tstamp ev_time (void) { #if EV_USE_REALTIME - struct timespec ts; - clock_gettime (CLOCK_REALTIME, &ts); - return ts.tv_sec + ts.tv_nsec * 1e-9; -#else + if (expect_true (have_realtime)) + { + struct timespec ts; + clock_gettime (CLOCK_REALTIME, &ts); + return ts.tv_sec + ts.tv_nsec * 1e-9; + } +#endif + struct timeval tv; gettimeofday (&tv, 0); return tv.tv_sec + tv.tv_usec * 1e-6; -#endif } +#endif -inline ev_tstamp +inline_size ev_tstamp get_clock (void) { #if EV_USE_MONOTONIC @@ -312,23 +623,81 @@ } #endif -#define array_roundsize(type,n) (((n) | 4) & ~3) +void +ev_sleep (ev_tstamp delay) +{ + if (delay > 0.) + { +#if EV_USE_NANOSLEEP + struct timespec ts; + + ts.tv_sec = (time_t)delay; + ts.tv_nsec = (long)((delay - (ev_tstamp)(ts.tv_sec)) * 1e9); + + nanosleep (&ts, 0); +#elif defined(_WIN32) + Sleep ((unsigned long)(delay * 1e3)); +#else + struct timeval tv; + + tv.tv_sec = (time_t)delay; + tv.tv_usec = (long)((delay - (ev_tstamp)(tv.tv_sec)) * 1e6); + + /* here we rely on sys/time.h + sys/types.h + unistd.h providing select */ + /* somehting not guaranteed by newer posix versions, but guaranteed */ + /* by older ones */ + select (0, 0, 0, 0, &tv); +#endif + } +} + +/*****************************************************************************/ + +#define MALLOC_ROUND 4096 /* prefer to allocate in chunks of this size, must be 2**n and >> 4 longs */ + +/* find a suitable new size for the given array, */ +/* hopefully by rounding to a ncie-to-malloc size */ +inline_size int +array_nextsize (int elem, int cur, int cnt) +{ + int ncur = cur + 1; + + do + ncur <<= 1; + while (cnt > ncur); + + /* if size is large, round to MALLOC_ROUND - 4 * longs to accomodate malloc overhead */ + if (elem * ncur > MALLOC_ROUND - sizeof (void *) * 4) + { + ncur *= elem; + ncur = (ncur + elem + (MALLOC_ROUND - 1) + sizeof (void *) * 4) & ~(MALLOC_ROUND - 1); + ncur = ncur - sizeof (void *) * 4; + ncur /= elem; + } + + return ncur; +} + +static noinline void * +array_realloc (int elem, void *base, int *cur, int cnt) +{ + *cur = array_nextsize (elem, *cur, cnt); + return ev_realloc (base, elem * *cur); +} + +#define array_init_zero(base,count) \ + memset ((void *)(base), 0, sizeof (*(base)) * (count)) #define array_needsize(type,base,cur,cnt,init) \ - if (expect_false ((cnt) > cur)) \ + if (expect_false ((cnt) > (cur))) \ { \ - int newcnt = cur; \ - do \ - { \ - newcnt = array_roundsize (type, newcnt << 1); \ - } \ - while ((cnt) > newcnt); \ - \ - base = (type *)ev_realloc (base, sizeof (type) * (newcnt));\ - init (base + cur, newcnt - cur); \ - cur = newcnt; \ + int ocur_ = (cur); \ + (base) = (type *)array_realloc \ + (sizeof (type), (base), &(cur), (cnt)); \ + init ((base) + (ocur_), (cur) - ocur_); \ } +#if 0 #define array_slim(type,stem) \ if (stem ## max < array_roundsize (stem ## cnt >> 2)) \ { \ @@ -336,43 +705,52 @@ base = (type *)ev_realloc (base, sizeof (type) * (stem ## max));\ fprintf (stderr, "slimmed down " # stem " to %d\n", stem ## max);/*D*/\ } +#endif #define array_free(stem, idx) \ - ev_free (stem ## s idx); stem ## cnt idx = stem ## max idx = 0; + ev_free (stem ## s idx); stem ## cnt idx = stem ## max idx = 0; stem ## s idx = 0 /*****************************************************************************/ -static void -anfds_init (ANFD *base, int count) +/* dummy callback for pending events */ +static void noinline +pendingcb (EV_P_ ev_prepare *w, int revents) { - while (count--) - { - base->head = 0; - base->events = EV_NONE; - base->reify = 0; - - ++base; - } } -void +void noinline ev_feed_event (EV_P_ void *w, int revents) { W w_ = (W)w; + int pri = ABSPRI (w_); - if (w_->pending) + if (expect_false (w_->pending)) + pendings [pri][w_->pending - 1].events |= revents; + else { - pendings [ABSPRI (w_)][w_->pending - 1].events |= revents; - return; + w_->pending = ++pendingcnt [pri]; + array_needsize (ANPENDING, pendings [pri], pendingmax [pri], w_->pending, EMPTY2); + pendings [pri][w_->pending - 1].w = w_; + pendings [pri][w_->pending - 1].events = revents; } +} - w_->pending = ++pendingcnt [ABSPRI (w_)]; - array_needsize (ANPENDING, pendings [ABSPRI (w_)], pendingmax [ABSPRI (w_)], pendingcnt [ABSPRI (w_)], (void)); - pendings [ABSPRI (w_)][w_->pending - 1].w = w_; - pendings [ABSPRI (w_)][w_->pending - 1].events = revents; +inline_speed void +feed_reverse (EV_P_ W w) +{ + array_needsize (W, rfeeds, rfeedmax, rfeedcnt + 1, EMPTY2); + rfeeds [rfeedcnt++] = w; } -static void +inline_size void +feed_reverse_done (EV_P_ int revents) +{ + do + ev_feed_event (EV_A_ rfeeds [--rfeedcnt], revents); + while (rfeedcnt); +} + +inline_speed void queue_events (EV_P_ W *events, int eventcnt, int type) { int i; @@ -381,13 +759,15 @@ ev_feed_event (EV_A_ events [i], type); } -inline void +/*****************************************************************************/ + +inline_speed void fd_event (EV_P_ int fd, int revents) { ANFD *anfd = anfds + fd; - struct ev_io *w; + ev_io *w; - for (w = (struct ev_io *)anfd->head; w; w = (struct ev_io *)((WL)w)->next) + for (w = (ev_io *)anfd->head; w; w = (ev_io *)((WL)w)->next) { int ev = w->events & revents; @@ -399,12 +779,13 @@ void ev_feed_fd_event (EV_P_ int fd, int revents) { - fd_event (EV_A_ fd, revents); + if (fd >= 0 && fd < anfdmax) + fd_event (EV_A_ fd, revents); } -/*****************************************************************************/ - -static void +/* make sure the external fd watch events are in-sync */ +/* with the kernel/libev internal state */ +inline_size void fd_reify (EV_P) { int i; @@ -413,57 +794,71 @@ { int fd = fdchanges [i]; ANFD *anfd = anfds + fd; - struct ev_io *w; + ev_io *w; - int events = 0; + unsigned char events = 0; - for (w = (struct ev_io *)anfd->head; w; w = (struct ev_io *)((WL)w)->next) - events |= w->events; + for (w = (ev_io *)anfd->head; w; w = (ev_io *)((WL)w)->next) + events |= (unsigned char)w->events; #if EV_SELECT_IS_WINSOCKET if (events) { - unsigned long argp; - anfd->handle = _get_osfhandle (fd); - assert (("libev only supports socket fds in this configuration", ioctlsocket (anfd->handle, FIONREAD, &argp) == 0)); + unsigned long arg; + #ifdef EV_FD_TO_WIN32_HANDLE + anfd->handle = EV_FD_TO_WIN32_HANDLE (fd); + #else + anfd->handle = _get_osfhandle (fd); + #endif + assert (("libev: only socket fds supported in this configuration", ioctlsocket (anfd->handle, FIONREAD, &arg) == 0)); } #endif - anfd->reify = 0; + { + unsigned char o_events = anfd->events; + unsigned char o_reify = anfd->reify; + + anfd->reify = 0; + anfd->events = events; - method_modify (EV_A_ fd, anfd->events, events); - anfd->events = events; + if (o_events != events || o_reify & EV__IOFDSET) + backend_modify (EV_A_ fd, o_events, events); + } } fdchangecnt = 0; } -static void -fd_change (EV_P_ int fd) +/* something about the given fd changed */ +inline_size void +fd_change (EV_P_ int fd, int flags) { - if (anfds [fd].reify) - return; + unsigned char reify = anfds [fd].reify; + anfds [fd].reify |= flags; - anfds [fd].reify = 1; - - ++fdchangecnt; - array_needsize (int, fdchanges, fdchangemax, fdchangecnt, (void)); - fdchanges [fdchangecnt - 1] = fd; + if (expect_true (!reify)) + { + ++fdchangecnt; + array_needsize (int, fdchanges, fdchangemax, fdchangecnt, EMPTY2); + fdchanges [fdchangecnt - 1] = fd; + } } -static void +/* the given fd is invalid/unusable, so make sure it doesn't hurt us anymore */ +inline_speed void fd_kill (EV_P_ int fd) { - struct ev_io *w; + ev_io *w; - while ((w = (struct ev_io *)anfds [fd].head)) + while ((w = (ev_io *)anfds [fd].head)) { ev_io_stop (EV_A_ w); ev_feed_event (EV_A_ (W)w, EV_ERROR | EV_READ | EV_WRITE); } } -static int +/* check whether the given fd is atcually valid, for error recovery */ +inline_size int fd_valid (int fd) { #ifdef _WIN32 @@ -474,19 +869,19 @@ } /* called on EBADF to verify fds */ -static void +static void noinline fd_ebadf (EV_P) { int fd; for (fd = 0; fd < anfdmax; ++fd) if (anfds [fd].events) - if (!fd_valid (fd) == -1 && errno == EBADF) + if (!fd_valid (fd) && errno == EBADF) fd_kill (EV_A_ fd); } /* called on ENOMEM in select/poll to kill some fds and retry */ -static void +static void noinline fd_enomem (EV_P) { int fd; @@ -499,123 +894,315 @@ } } -/* usually called after fork if method needs to re-arm all fds from scratch */ -static void +/* usually called after fork if backend needs to re-arm all fds from scratch */ +static void noinline fd_rearm_all (EV_P) { int fd; - /* this should be highly optimised to not do anything but set a flag */ for (fd = 0; fd < anfdmax; ++fd) if (anfds [fd].events) { anfds [fd].events = 0; - fd_change (EV_A_ fd); + anfds [fd].emask = 0; + fd_change (EV_A_ fd, EV__IOFDSET | 1); } } /*****************************************************************************/ -static void -upheap (WT *heap, int k) +/* + * the heap functions want a real array index. array index 0 uis guaranteed to not + * be in-use at any time. the first heap entry is at array [HEAP0]. DHEAP gives + * the branching factor of the d-tree. + */ + +/* + * at the moment we allow libev the luxury of two heaps, + * a small-code-size 2-heap one and a ~1.5kb larger 4-heap + * which is more cache-efficient. + * the difference is about 5% with 50000+ watchers. + */ +#if EV_USE_4HEAP + +#define DHEAP 4 +#define HEAP0 (DHEAP - 1) /* index of first element in heap */ +#define HPARENT(k) ((((k) - HEAP0 - 1) / DHEAP) + HEAP0) +#define UPHEAP_DONE(p,k) ((p) == (k)) + +/* away from the root */ +inline_speed void +downheap (ANHE *heap, int N, int k) { - WT w = heap [k]; + ANHE he = heap [k]; + ANHE *E = heap + N + HEAP0; - while (k && heap [k >> 1]->at > w->at) + for (;;) { - heap [k] = heap [k >> 1]; - ((W)heap [k])->active = k + 1; - k >>= 1; - } + ev_tstamp minat; + ANHE *minpos; + ANHE *pos = heap + DHEAP * (k - HEAP0) + HEAP0 + 1; + + /* find minimum child */ + if (expect_true (pos + DHEAP - 1 < E)) + { + /* fast path */ (minpos = pos + 0), (minat = ANHE_at (*minpos)); + if ( ANHE_at (pos [1]) < minat) (minpos = pos + 1), (minat = ANHE_at (*minpos)); + if ( ANHE_at (pos [2]) < minat) (minpos = pos + 2), (minat = ANHE_at (*minpos)); + if ( ANHE_at (pos [3]) < minat) (minpos = pos + 3), (minat = ANHE_at (*minpos)); + } + else if (pos < E) + { + /* slow path */ (minpos = pos + 0), (minat = ANHE_at (*minpos)); + if (pos + 1 < E && ANHE_at (pos [1]) < minat) (minpos = pos + 1), (minat = ANHE_at (*minpos)); + if (pos + 2 < E && ANHE_at (pos [2]) < minat) (minpos = pos + 2), (minat = ANHE_at (*minpos)); + if (pos + 3 < E && ANHE_at (pos [3]) < minat) (minpos = pos + 3), (minat = ANHE_at (*minpos)); + } + else + break; + + if (ANHE_at (he) <= minat) + break; + + heap [k] = *minpos; + ev_active (ANHE_w (*minpos)) = k; - heap [k] = w; - ((W)heap [k])->active = k + 1; + k = minpos - heap; + } + heap [k] = he; + ev_active (ANHE_w (he)) = k; } -static void -downheap (WT *heap, int N, int k) +#else /* 4HEAP */ + +#define HEAP0 1 +#define HPARENT(k) ((k) >> 1) +#define UPHEAP_DONE(p,k) (!(p)) + +/* away from the root */ +inline_speed void +downheap (ANHE *heap, int N, int k) { - WT w = heap [k]; + ANHE he = heap [k]; - while (k < (N >> 1)) + for (;;) { - int j = k << 1; + int c = k << 1; - if (j + 1 < N && heap [j]->at > heap [j + 1]->at) - ++j; + if (c > N + HEAP0 - 1) + break; + + c += c + 1 < N + HEAP0 && ANHE_at (heap [c]) > ANHE_at (heap [c + 1]) + ? 1 : 0; - if (w->at <= heap [j]->at) + if (ANHE_at (he) <= ANHE_at (heap [c])) break; - heap [k] = heap [j]; - ((W)heap [k])->active = k + 1; - k = j; + heap [k] = heap [c]; + ev_active (ANHE_w (heap [k])) = k; + + k = c; } - heap [k] = w; - ((W)heap [k])->active = k + 1; + heap [k] = he; + ev_active (ANHE_w (he)) = k; } +#endif -inline void -adjustheap (WT *heap, int N, int k) +/* towards the root */ +inline_speed void +upheap (ANHE *heap, int k) { - upheap (heap, k); - downheap (heap, N, k); + ANHE he = heap [k]; + + for (;;) + { + int p = HPARENT (k); + + if (UPHEAP_DONE (p, k) || ANHE_at (heap [p]) <= ANHE_at (he)) + break; + + heap [k] = heap [p]; + ev_active (ANHE_w (heap [k])) = k; + k = p; + } + + heap [k] = he; + ev_active (ANHE_w (he)) = k; +} + +/* move an element suitably so it is in a correct place */ +inline_size void +adjustheap (ANHE *heap, int N, int k) +{ + if (k > HEAP0 && ANHE_at (heap [HPARENT (k)]) >= ANHE_at (heap [k])) + upheap (heap, k); + else + downheap (heap, N, k); +} + +/* rebuild the heap: this function is used only once and executed rarely */ +inline_size void +reheap (ANHE *heap, int N) +{ + int i; + + /* we don't use floyds algorithm, upheap is simpler and is more cache-efficient */ + /* also, this is easy to implement and correct for both 2-heaps and 4-heaps */ + for (i = 0; i < N; ++i) + upheap (heap, i + HEAP0); } /*****************************************************************************/ +/* associate signal watchers to a signal signal */ typedef struct { WL head; - sig_atomic_t volatile gotsig; + EV_ATOMIC_T gotsig; } ANSIG; static ANSIG *signals; static int signalmax; -static int sigpipe [2]; -static sig_atomic_t volatile gotsig; -static struct ev_io sigev; +static EV_ATOMIC_T gotsig; -static void -signals_init (ANSIG *base, int count) +/*****************************************************************************/ + +/* used to prepare libev internal fd's */ +/* this is not fork-safe */ +inline_speed void +fd_intern (int fd) +{ +#ifdef _WIN32 + unsigned long arg = 1; + ioctlsocket (_get_osfhandle (fd), FIONBIO, &arg); +#else + fcntl (fd, F_SETFD, FD_CLOEXEC); + fcntl (fd, F_SETFL, O_NONBLOCK); +#endif +} + +static void noinline +evpipe_init (EV_P) { - while (count--) + if (!ev_is_active (&pipe_w)) { - base->head = 0; - base->gotsig = 0; +#if EV_USE_EVENTFD + if ((evfd = eventfd (0, 0)) >= 0) + { + evpipe [0] = -1; + fd_intern (evfd); + ev_io_set (&pipe_w, evfd, EV_READ); + } + else +#endif + { + while (pipe (evpipe)) + ev_syserr ("(libev) error creating signal/async pipe"); + + fd_intern (evpipe [0]); + fd_intern (evpipe [1]); + ev_io_set (&pipe_w, evpipe [0], EV_READ); + } + + ev_io_start (EV_A_ &pipe_w); + ev_unref (EV_A); /* watcher should not keep loop alive */ + } +} + +inline_size void +evpipe_write (EV_P_ EV_ATOMIC_T *flag) +{ + if (!*flag) + { + int old_errno = errno; /* save errno because write might clobber it */ + + *flag = 1; + +#if EV_USE_EVENTFD + if (evfd >= 0) + { + uint64_t counter = 1; + write (evfd, &counter, sizeof (uint64_t)); + } + else +#endif + write (evpipe [1], &old_errno, 1); - ++base; + errno = old_errno; } } +/* called whenever the libev signal pipe */ +/* got some events (signal, async) */ static void -sighandler (int signum) +pipecb (EV_P_ ev_io *iow, int revents) { -#if _WIN32 - signal (signum, sighandler); +#if EV_USE_EVENTFD + if (evfd >= 0) + { + uint64_t counter; + read (evfd, &counter, sizeof (uint64_t)); + } + else #endif + { + char dummy; + read (evpipe [0], &dummy, 1); + } - signals [signum - 1].gotsig = 1; + if (gotsig && ev_is_default_loop (EV_A)) + { + int signum; + gotsig = 0; - if (!gotsig) + for (signum = signalmax; signum--; ) + if (signals [signum].gotsig) + ev_feed_signal_event (EV_A_ signum + 1); + } + +#if EV_ASYNC_ENABLE + if (gotasync) { - int old_errno = errno; - gotsig = 1; - write (sigpipe [1], &signum, 1); - errno = old_errno; + int i; + gotasync = 0; + + for (i = asynccnt; i--; ) + if (asyncs [i]->sent) + { + asyncs [i]->sent = 0; + ev_feed_event (EV_A_ asyncs [i], EV_ASYNC); + } } +#endif } -void +/*****************************************************************************/ + +static void +ev_sighandler (int signum) +{ +#if EV_MULTIPLICITY + struct ev_loop *loop = &default_loop_struct; +#endif + +#if _WIN32 + signal (signum, ev_sighandler); +#endif + + signals [signum - 1].gotsig = 1; + evpipe_write (EV_A_ &gotsig); +} + +void noinline ev_feed_signal_event (EV_P_ int signum) { WL w; #if EV_MULTIPLICITY - assert (("feeding signal events is only supported in the default loop", loop == default_loop)); + assert (("libev: feeding signal events is only supported in the default loop", loop == ev_default_loop_ptr)); #endif --signum; @@ -629,88 +1216,71 @@ ev_feed_event (EV_A_ (W)w, EV_SIGNAL); } -static void -sigcb (EV_P_ struct ev_io *iow, int revents) -{ - int signum; +/*****************************************************************************/ - read (sigpipe [0], &revents, 1); - gotsig = 0; +static WL childs [EV_PID_HASHSIZE]; - for (signum = signalmax; signum--; ) - if (signals [signum].gotsig) - ev_feed_signal_event (EV_A_ signum + 1); -} +#ifndef _WIN32 -inline void -fd_intern (int fd) -{ -#ifdef _WIN32 - int arg = 1; - ioctlsocket (_get_osfhandle (fd), FIONBIO, &arg); -#else - fcntl (fd, F_SETFD, FD_CLOEXEC); - fcntl (fd, F_SETFL, O_NONBLOCK); +static ev_signal childev; + +#ifndef WIFCONTINUED +# define WIFCONTINUED(status) 0 #endif -} -static void -siginit (EV_P) +/* handle a single child status event */ +inline_speed void +child_reap (EV_P_ int chain, int pid, int status) { - fd_intern (sigpipe [0]); - fd_intern (sigpipe [1]); + ev_child *w; + int traced = WIFSTOPPED (status) || WIFCONTINUED (status); - ev_io_set (&sigev, sigpipe [0], EV_READ); - ev_io_start (EV_A_ &sigev); - ev_unref (EV_A); /* child watcher should not keep loop alive */ + for (w = (ev_child *)childs [chain & (EV_PID_HASHSIZE - 1)]; w; w = (ev_child *)((WL)w)->next) + { + if ((w->pid == pid || !w->pid) + && (!traced || (w->flags & 1))) + { + ev_set_priority (w, EV_MAXPRI); /* need to do it *now*, this *must* be the same prio as the signal watcher itself */ + w->rpid = pid; + w->rstatus = status; + ev_feed_event (EV_A_ (W)w, EV_CHILD); + } + } } -/*****************************************************************************/ - -static struct ev_child *childs [PID_HASHSIZE]; - -#ifndef _WIN32 - -static struct ev_signal childev; - #ifndef WCONTINUED # define WCONTINUED 0 #endif +/* called on sigchld etc., calls waitpid */ static void -child_reap (EV_P_ struct ev_signal *sw, int chain, int pid, int status) -{ - struct ev_child *w; - - for (w = (struct ev_child *)childs [chain & (PID_HASHSIZE - 1)]; w; w = (struct ev_child *)((WL)w)->next) - if (w->pid == pid || !w->pid) - { - ev_priority (w) = ev_priority (sw); /* need to do it *now* */ - w->rpid = pid; - w->rstatus = status; - ev_feed_event (EV_A_ (W)w, EV_CHILD); - } -} - -static void -childcb (EV_P_ struct ev_signal *sw, int revents) +childcb (EV_P_ ev_signal *sw, int revents) { int pid, status; - if (0 < (pid = waitpid (-1, &status, WNOHANG | WUNTRACED | WCONTINUED))) - { - /* make sure we are called again until all childs have been reaped */ - ev_feed_event (EV_A_ (W)sw, EV_SIGNAL); + /* some systems define WCONTINUED but then fail to support it (linux 2.4) */ + if (0 >= (pid = waitpid (-1, &status, WNOHANG | WUNTRACED | WCONTINUED))) + if (!WCONTINUED + || errno != EINVAL + || 0 >= (pid = waitpid (-1, &status, WNOHANG | WUNTRACED))) + return; - child_reap (EV_A_ sw, pid, pid, status); - child_reap (EV_A_ sw, 0, pid, status); /* this might trigger a watcher twice, but event catches that */ - } + /* make sure we are called again until all children have been reaped */ + /* we need to do it this way so that the callback gets called before we continue */ + ev_feed_event (EV_A_ (W)sw, EV_SIGNAL); + + child_reap (EV_A_ pid, pid, status); + if (EV_PID_HASHSIZE > 1) + child_reap (EV_A_ 0, pid, status); /* this might trigger a watcher twice, but feed_event catches that */ } #endif /*****************************************************************************/ +#if EV_USE_PORT +# include "ev_port.c" +#endif #if EV_USE_KQUEUE # include "ev_kqueue.c" #endif @@ -737,7 +1307,7 @@ } /* return true if we are running with elevated privileges and should ignore env variables */ -static int +int inline_size enable_secure (void) { #ifdef _WIN32 @@ -748,118 +1318,315 @@ #endif } -int -ev_method (EV_P) +unsigned int +ev_supported_backends (void) { - return method; + unsigned int flags = 0; + + if (EV_USE_PORT ) flags |= EVBACKEND_PORT; + if (EV_USE_KQUEUE) flags |= EVBACKEND_KQUEUE; + if (EV_USE_EPOLL ) flags |= EVBACKEND_EPOLL; + if (EV_USE_POLL ) flags |= EVBACKEND_POLL; + if (EV_USE_SELECT) flags |= EVBACKEND_SELECT; + + return flags; } -static void +unsigned int +ev_recommended_backends (void) +{ + unsigned int flags = ev_supported_backends (); + +#ifndef __NetBSD__ + /* kqueue is borked on everything but netbsd apparently */ + /* it usually doesn't work correctly on anything but sockets and pipes */ + flags &= ~EVBACKEND_KQUEUE; +#endif +#ifdef __APPLE__ + /* only select works correctly on that "unix-certified" platform */ + flags &= ~EVBACKEND_KQUEUE; /* horribly broken, even for sockets */ + flags &= ~EVBACKEND_POLL; /* poll is based on kqueue from 10.5 onwards */ +#endif + + return flags; +} + +unsigned int +ev_embeddable_backends (void) +{ + int flags = EVBACKEND_EPOLL | EVBACKEND_KQUEUE | EVBACKEND_PORT; + + /* epoll embeddability broken on all linux versions up to at least 2.6.23 */ + /* please fix it and tell me how to detect the fix */ + flags &= ~EVBACKEND_EPOLL; + + return flags; +} + +unsigned int +ev_backend (EV_P) +{ + return backend; +} + +#if EV_MINIMAL < 2 +unsigned int +ev_loop_count (EV_P) +{ + return loop_count; +} + +unsigned int +ev_loop_depth (EV_P) +{ + return loop_depth; +} + +void +ev_set_io_collect_interval (EV_P_ ev_tstamp interval) +{ + io_blocktime = interval; +} + +void +ev_set_timeout_collect_interval (EV_P_ ev_tstamp interval) +{ + timeout_blocktime = interval; +} + +void +ev_set_userdata (EV_P_ void *data) +{ + userdata = data; +} + +void * +ev_userdata (EV_P) +{ + return userdata; +} + +void ev_set_invoke_pending_cb (EV_P_ void (*invoke_pending_cb)(EV_P)) +{ + invoke_cb = invoke_pending_cb; +} + +void ev_set_blocking_cb (EV_P_ void (*suspend_cb_)(EV_P), void (*resume_cb_)(EV_P)) +{ + suspend_cb = suspend_cb_; + resume_cb = resume_cb_; +} +#endif + +/* initialise a loop structure, must be zero-initialised */ +static void noinline loop_init (EV_P_ unsigned int flags) { - if (!method) + if (!backend) { +#if EV_USE_REALTIME + if (!have_realtime) + { + struct timespec ts; + + if (!clock_gettime (CLOCK_REALTIME, &ts)) + have_realtime = 1; + } +#endif + #if EV_USE_MONOTONIC - { - struct timespec ts; - if (!clock_gettime (CLOCK_MONOTONIC, &ts)) - have_monotonic = 1; - } + if (!have_monotonic) + { + struct timespec ts; + + if (!clock_gettime (CLOCK_MONOTONIC, &ts)) + have_monotonic = 1; + } #endif - ev_rt_now = ev_time (); - mn_now = get_clock (); - now_floor = mn_now; - rtmn_diff = ev_rt_now - mn_now; + ev_rt_now = ev_time (); + mn_now = get_clock (); + now_floor = mn_now; + rtmn_diff = ev_rt_now - mn_now; +#if EV_MINIMAL < 2 + invoke_cb = ev_invoke_pending; +#endif + + io_blocktime = 0.; + timeout_blocktime = 0.; + backend = 0; + backend_fd = -1; + gotasync = 0; +#if EV_USE_INOTIFY + fs_fd = -2; +#endif + + /* pid check not overridable via env */ +#ifndef _WIN32 + if (flags & EVFLAG_FORKCHECK) + curpid = getpid (); +#endif - if (!(flags & EVMETHOD_NOENV) && !enable_secure () && getenv ("LIBEV_FLAGS")) + if (!(flags & EVFLAG_NOENV) + && !enable_secure () + && getenv ("LIBEV_FLAGS")) flags = atoi (getenv ("LIBEV_FLAGS")); - if (!(flags & 0x0000ffff)) - flags |= 0x0000ffff; + if (!(flags & 0x0000ffffU)) + flags |= ev_recommended_backends (); - method = 0; +#if EV_USE_PORT + if (!backend && (flags & EVBACKEND_PORT )) backend = port_init (EV_A_ flags); +#endif #if EV_USE_KQUEUE - if (!method && (flags & EVMETHOD_KQUEUE)) method = kqueue_init (EV_A_ flags); + if (!backend && (flags & EVBACKEND_KQUEUE)) backend = kqueue_init (EV_A_ flags); #endif #if EV_USE_EPOLL - if (!method && (flags & EVMETHOD_EPOLL )) method = epoll_init (EV_A_ flags); + if (!backend && (flags & EVBACKEND_EPOLL )) backend = epoll_init (EV_A_ flags); #endif #if EV_USE_POLL - if (!method && (flags & EVMETHOD_POLL )) method = poll_init (EV_A_ flags); + if (!backend && (flags & EVBACKEND_POLL )) backend = poll_init (EV_A_ flags); #endif #if EV_USE_SELECT - if (!method && (flags & EVMETHOD_SELECT)) method = select_init (EV_A_ flags); + if (!backend && (flags & EVBACKEND_SELECT)) backend = select_init (EV_A_ flags); #endif - ev_init (&sigev, sigcb); - ev_set_priority (&sigev, EV_MAXPRI); + ev_prepare_init (&pending_w, pendingcb); + + ev_init (&pipe_w, pipecb); + ev_set_priority (&pipe_w, EV_MAXPRI); } } -void +/* free up a loop structure */ +static void noinline loop_destroy (EV_P) { int i; + if (ev_is_active (&pipe_w)) + { + ev_ref (EV_A); /* signal watcher */ + ev_io_stop (EV_A_ &pipe_w); + +#if EV_USE_EVENTFD + if (evfd >= 0) + close (evfd); +#endif + + if (evpipe [0] >= 0) + { + close (evpipe [0]); + close (evpipe [1]); + } + } + +#if EV_USE_INOTIFY + if (fs_fd >= 0) + close (fs_fd); +#endif + + if (backend_fd >= 0) + close (backend_fd); + +#if EV_USE_PORT + if (backend == EVBACKEND_PORT ) port_destroy (EV_A); +#endif #if EV_USE_KQUEUE - if (method == EVMETHOD_KQUEUE) kqueue_destroy (EV_A); + if (backend == EVBACKEND_KQUEUE) kqueue_destroy (EV_A); #endif #if EV_USE_EPOLL - if (method == EVMETHOD_EPOLL ) epoll_destroy (EV_A); + if (backend == EVBACKEND_EPOLL ) epoll_destroy (EV_A); #endif #if EV_USE_POLL - if (method == EVMETHOD_POLL ) poll_destroy (EV_A); + if (backend == EVBACKEND_POLL ) poll_destroy (EV_A); #endif #if EV_USE_SELECT - if (method == EVMETHOD_SELECT) select_destroy (EV_A); + if (backend == EVBACKEND_SELECT) select_destroy (EV_A); #endif for (i = NUMPRI; i--; ) - array_free (pending, [i]); + { + array_free (pending, [i]); +#if EV_IDLE_ENABLE + array_free (idle, [i]); +#endif + } + + ev_free (anfds); anfdmax = 0; /* have to use the microsoft-never-gets-it-right macro */ + array_free (rfeed, EMPTY); array_free (fdchange, EMPTY); array_free (timer, EMPTY); -#if EV_PERIODICS +#if EV_PERIODIC_ENABLE array_free (periodic, EMPTY); #endif - array_free (idle, EMPTY); +#if EV_FORK_ENABLE + array_free (fork, EMPTY); +#endif array_free (prepare, EMPTY); array_free (check, EMPTY); +#if EV_ASYNC_ENABLE + array_free (async, EMPTY); +#endif - method = 0; + backend = 0; } -static void +#if EV_USE_INOTIFY +inline_size void infy_fork (EV_P); +#endif + +inline_size void loop_fork (EV_P) { -#if EV_USE_EPOLL - if (method == EVMETHOD_EPOLL ) epoll_fork (EV_A); +#if EV_USE_PORT + if (backend == EVBACKEND_PORT ) port_fork (EV_A); #endif #if EV_USE_KQUEUE - if (method == EVMETHOD_KQUEUE) kqueue_fork (EV_A); + if (backend == EVBACKEND_KQUEUE) kqueue_fork (EV_A); +#endif +#if EV_USE_EPOLL + if (backend == EVBACKEND_EPOLL ) epoll_fork (EV_A); +#endif +#if EV_USE_INOTIFY + infy_fork (EV_A); #endif - if (ev_is_active (&sigev)) + if (ev_is_active (&pipe_w)) { - /* default loop */ + /* this "locks" the handlers against writing to the pipe */ + /* while we modify the fd vars */ + gotsig = 1; +#if EV_ASYNC_ENABLE + gotasync = 1; +#endif ev_ref (EV_A); - ev_io_stop (EV_A_ &sigev); - close (sigpipe [0]); - close (sigpipe [1]); + ev_io_stop (EV_A_ &pipe_w); + +#if EV_USE_EVENTFD + if (evfd >= 0) + close (evfd); +#endif - while (pipe (sigpipe)) - syserr ("(libev) error creating pipe"); + if (evpipe [0] >= 0) + { + close (evpipe [0]); + close (evpipe [1]); + } - siginit (EV_A); + evpipe_init (EV_A); + /* now iterate over everything, in case we missed something */ + pipecb (EV_A_ &pipe_w, EV_READ); } postfork = 0; } #if EV_MULTIPLICITY + struct ev_loop * ev_loop_new (unsigned int flags) { @@ -869,7 +1636,7 @@ loop_init (EV_A_ flags); - if (ev_method (EV_A)) + if (ev_backend (EV_A)) return loop; return 0; @@ -885,36 +1652,131 @@ void ev_loop_fork (EV_P) { - postfork = 1; + postfork = 1; /* must be in line with ev_default_fork */ } +#endif /* multiplicity */ +#if EV_VERIFY +static void noinline +verify_watcher (EV_P_ W w) +{ + assert (("libev: watcher has invalid priority", ABSPRI (w) >= 0 && ABSPRI (w) < NUMPRI)); + + if (w->pending) + assert (("libev: pending watcher not on pending queue", pendings [ABSPRI (w)][w->pending - 1].w == w)); +} + +static void noinline +verify_heap (EV_P_ ANHE *heap, int N) +{ + int i; + + for (i = HEAP0; i < N + HEAP0; ++i) + { + assert (("libev: active index mismatch in heap", ev_active (ANHE_w (heap [i])) == i)); + assert (("libev: heap condition violated", i == HEAP0 || ANHE_at (heap [HPARENT (i)]) <= ANHE_at (heap [i]))); + assert (("libev: heap at cache mismatch", ANHE_at (heap [i]) == ev_at (ANHE_w (heap [i])))); + + verify_watcher (EV_A_ (W)ANHE_w (heap [i])); + } +} + +static void noinline +array_verify (EV_P_ W *ws, int cnt) +{ + while (cnt--) + { + assert (("libev: active index mismatch", ev_active (ws [cnt]) == cnt + 1)); + verify_watcher (EV_A_ ws [cnt]); + } +} +#endif + +#if EV_MINIMAL < 2 +void +ev_loop_verify (EV_P) +{ +#if EV_VERIFY + int i; + WL w; + + assert (activecnt >= -1); + + assert (fdchangemax >= fdchangecnt); + for (i = 0; i < fdchangecnt; ++i) + assert (("libev: negative fd in fdchanges", fdchanges [i] >= 0)); + + assert (anfdmax >= 0); + for (i = 0; i < anfdmax; ++i) + for (w = anfds [i].head; w; w = w->next) + { + verify_watcher (EV_A_ (W)w); + assert (("libev: inactive fd watcher on anfd list", ev_active (w) == 1)); + assert (("libev: fd mismatch between watcher and anfd", ((ev_io *)w)->fd == i)); + } + + assert (timermax >= timercnt); + verify_heap (EV_A_ timers, timercnt); + +#if EV_PERIODIC_ENABLE + assert (periodicmax >= periodiccnt); + verify_heap (EV_A_ periodics, periodiccnt); +#endif + + for (i = NUMPRI; i--; ) + { + assert (pendingmax [i] >= pendingcnt [i]); +#if EV_IDLE_ENABLE + assert (idleall >= 0); + assert (idlemax [i] >= idlecnt [i]); + array_verify (EV_A_ (W *)idles [i], idlecnt [i]); +#endif + } + +#if EV_FORK_ENABLE + assert (forkmax >= forkcnt); + array_verify (EV_A_ (W *)forks, forkcnt); +#endif + +#if EV_ASYNC_ENABLE + assert (asyncmax >= asynccnt); + array_verify (EV_A_ (W *)asyncs, asynccnt); +#endif + + assert (preparemax >= preparecnt); + array_verify (EV_A_ (W *)prepares, preparecnt); + + assert (checkmax >= checkcnt); + array_verify (EV_A_ (W *)checks, checkcnt); + +# if 0 + for (w = (ev_child *)childs [chain & (EV_PID_HASHSIZE - 1)]; w; w = (ev_child *)((WL)w)->next) + for (signum = signalmax; signum--; ) if (signals [signum].gotsig) +# endif +#endif +} #endif #if EV_MULTIPLICITY struct ev_loop * +ev_default_loop_init (unsigned int flags) #else int +ev_default_loop (unsigned int flags) #endif -ev_default_loop (unsigned int methods) { - if (sigpipe [0] == sigpipe [1]) - if (pipe (sigpipe)) - return 0; - - if (!default_loop) + if (!ev_default_loop_ptr) { #if EV_MULTIPLICITY - struct ev_loop *loop = default_loop = &default_loop_struct; + struct ev_loop *loop = ev_default_loop_ptr = &default_loop_struct; #else - default_loop = 1; + ev_default_loop_ptr = 1; #endif - loop_init (EV_A_ methods); + loop_init (EV_A_ flags); - if (ev_method (EV_A)) + if (ev_backend (EV_A)) { - siginit (EV_A); - #ifndef _WIN32 ev_signal_init (&childev, childcb, SIGCHLD); ev_set_priority (&childev, EV_MAXPRI); @@ -923,30 +1785,26 @@ #endif } else - default_loop = 0; + ev_default_loop_ptr = 0; } - return default_loop; + return ev_default_loop_ptr; } void ev_default_destroy (void) { #if EV_MULTIPLICITY - struct ev_loop *loop = default_loop; + struct ev_loop *loop = ev_default_loop_ptr; #endif + ev_default_loop_ptr = 0; + #ifndef _WIN32 ev_ref (EV_A); /* child watcher */ ev_signal_stop (EV_A_ &childev); #endif - ev_ref (EV_A); /* signal watcher */ - ev_io_stop (EV_A_ &sigev); - - close (sigpipe [0]); sigpipe [0] = 0; - close (sigpipe [1]); sigpipe [1] = 0; - loop_destroy (EV_A); } @@ -954,29 +1812,22 @@ ev_default_fork (void) { #if EV_MULTIPLICITY - struct ev_loop *loop = default_loop; + struct ev_loop *loop = ev_default_loop_ptr; #endif - if (method) - postfork = 1; + postfork = 1; /* must be in line with ev_loop_fork */ } /*****************************************************************************/ -static int -any_pending (EV_P) +void +ev_invoke (EV_P_ void *w, int revents) { - int pri; - - for (pri = NUMPRI; pri--; ) - if (pendingcnt [pri]) - return 1; - - return 0; + EV_CB_INVOKE ((W)w, revents); } -static void -call_pending (EV_P) +void noinline +ev_invoke_pending (EV_P) { int pri; @@ -985,156 +1836,236 @@ { ANPENDING *p = pendings [pri] + --pendingcnt [pri]; - if (p->w) - { - p->w->pending = 0; - EV_CB_INVOKE (p->w, p->events); - } + /*assert (("libev: non-pending watcher on pending list", p->w->pending));*/ + /* ^ this is no longer true, as pending_w could be here */ + + p->w->pending = 0; + EV_CB_INVOKE (p->w, p->events); + EV_FREQUENT_CHECK; } } -static void -timers_reify (EV_P) +#if EV_IDLE_ENABLE +/* make idle watchers pending. this handles the "call-idle */ +/* only when higher priorities are idle" logic */ +inline_size void +idle_reify (EV_P) { - while (timercnt && ((WT)timers [0])->at <= mn_now) + if (expect_false (idleall)) { - struct ev_timer *w = timers [0]; + int pri; - assert (("inactive timer on timer heap detected", ev_is_active (w))); + for (pri = NUMPRI; pri--; ) + { + if (pendingcnt [pri]) + break; - /* first reschedule or stop timer */ - if (w->repeat) + if (idlecnt [pri]) + { + queue_events (EV_A_ (W *)idles [pri], idlecnt [pri], EV_IDLE); + break; + } + } + } +} +#endif + +/* make timers pending */ +inline_size void +timers_reify (EV_P) +{ + EV_FREQUENT_CHECK; + + if (timercnt && ANHE_at (timers [HEAP0]) < mn_now) + { + do { - assert (("negative ev_timer repeat value found while processing timers", w->repeat > 0.)); + ev_timer *w = (ev_timer *)ANHE_w (timers [HEAP0]); + + /*assert (("libev: inactive timer on timer heap detected", ev_is_active (w)));*/ + + /* first reschedule or stop timer */ + if (w->repeat) + { + ev_at (w) += w->repeat; + if (ev_at (w) < mn_now) + ev_at (w) = mn_now; + + assert (("libev: negative ev_timer repeat value found while processing timers", w->repeat > 0.)); - ((WT)w)->at += w->repeat; - if (((WT)w)->at < mn_now) - ((WT)w)->at = mn_now; + ANHE_at_cache (timers [HEAP0]); + downheap (timers, timercnt, HEAP0); + } + else + ev_timer_stop (EV_A_ w); /* nonrepeating: stop timer */ - downheap ((WT *)timers, timercnt, 0); + EV_FREQUENT_CHECK; + feed_reverse (EV_A_ (W)w); } - else - ev_timer_stop (EV_A_ w); /* nonrepeating: stop timer */ + while (timercnt && ANHE_at (timers [HEAP0]) < mn_now); - ev_feed_event (EV_A_ (W)w, EV_TIMEOUT); + feed_reverse_done (EV_A_ EV_TIMEOUT); } } -#if EV_PERIODICS -static void +#if EV_PERIODIC_ENABLE +/* make periodics pending */ +inline_size void periodics_reify (EV_P) { - while (periodiccnt && ((WT)periodics [0])->at <= ev_rt_now) - { - struct ev_periodic *w = periodics [0]; + EV_FREQUENT_CHECK; - assert (("inactive timer on periodic heap detected", ev_is_active (w))); + while (periodiccnt && ANHE_at (periodics [HEAP0]) < ev_rt_now) + { + int feed_count = 0; - /* first reschedule or stop timer */ - if (w->reschedule_cb) + do { - ((WT)w)->at = w->reschedule_cb (w, ev_rt_now + 0.0001); - assert (("ev_periodic reschedule callback returned time in the past", ((WT)w)->at > ev_rt_now)); - downheap ((WT *)periodics, periodiccnt, 0); - } - else if (w->interval) - { - ((WT)w)->at += floor ((ev_rt_now - ((WT)w)->at) / w->interval + 1.) * w->interval; - assert (("ev_periodic timeout in the past detected while processing timers, negative interval?", ((WT)w)->at > ev_rt_now)); - downheap ((WT *)periodics, periodiccnt, 0); + ev_periodic *w = (ev_periodic *)ANHE_w (periodics [HEAP0]); + + /*assert (("libev: inactive timer on periodic heap detected", ev_is_active (w)));*/ + + /* first reschedule or stop timer */ + if (w->reschedule_cb) + { + ev_at (w) = w->reschedule_cb (w, ev_rt_now); + + assert (("libev: ev_periodic reschedule callback returned time in the past", ev_at (w) >= ev_rt_now)); + + ANHE_at_cache (periodics [HEAP0]); + downheap (periodics, periodiccnt, HEAP0); + } + else if (w->interval) + { + ev_at (w) = w->offset + ceil ((ev_rt_now - w->offset) / w->interval) * w->interval; + /* if next trigger time is not sufficiently in the future, put it there */ + /* this might happen because of floating point inexactness */ + if (ev_at (w) - ev_rt_now < TIME_EPSILON) + { + ev_at (w) += w->interval; + + /* if interval is unreasonably low we might still have a time in the past */ + /* so correct this. this will make the periodic very inexact, but the user */ + /* has effectively asked to get triggered more often than possible */ + if (ev_at (w) < ev_rt_now) + ev_at (w) = ev_rt_now; + } + + ANHE_at_cache (periodics [HEAP0]); + downheap (periodics, periodiccnt, HEAP0); + } + else + ev_periodic_stop (EV_A_ w); /* nonrepeating: stop timer */ + + EV_FREQUENT_CHECK; + feed_reverse (EV_A_ (W)w); } - else - ev_periodic_stop (EV_A_ w); /* nonrepeating: stop timer */ + while (periodiccnt && ANHE_at (periodics [HEAP0]) < ev_rt_now); - ev_feed_event (EV_A_ (W)w, EV_PERIODIC); + feed_reverse_done (EV_A_ EV_PERIODIC); } } -static void +/* simply recalculate all periodics */ +/* TODO: maybe ensure that at leats one event happens when jumping forward? */ +static void noinline periodics_reschedule (EV_P) { int i; /* adjust periodics after time jump */ - for (i = 0; i < periodiccnt; ++i) + for (i = HEAP0; i < periodiccnt + HEAP0; ++i) { - struct ev_periodic *w = periodics [i]; + ev_periodic *w = (ev_periodic *)ANHE_w (periodics [i]); if (w->reschedule_cb) - ((WT)w)->at = w->reschedule_cb (w, ev_rt_now); + ev_at (w) = w->reschedule_cb (w, ev_rt_now); else if (w->interval) - ((WT)w)->at += ceil ((ev_rt_now - ((WT)w)->at) / w->interval) * w->interval; + ev_at (w) = w->offset + ceil ((ev_rt_now - w->offset) / w->interval) * w->interval; + + ANHE_at_cache (periodics [i]); } - /* now rebuild the heap */ - for (i = periodiccnt >> 1; i--; ) - downheap ((WT *)periodics, periodiccnt, i); + reheap (periodics, periodiccnt); } #endif -inline int -time_update_monotonic (EV_P) +/* adjust all timers by a given offset */ +static void noinline +timers_reschedule (EV_P_ ev_tstamp adjust) { - mn_now = get_clock (); + int i; - if (expect_true (mn_now - now_floor < MIN_TIMEJUMP * .5)) + for (i = 0; i < timercnt; ++i) { - ev_rt_now = rtmn_diff + mn_now; - return 0; - } - else - { - now_floor = mn_now; - ev_rt_now = ev_time (); - return 1; + ANHE *he = timers + i + HEAP0; + ANHE_w (*he)->at += adjust; + ANHE_at_cache (*he); } } -static void -time_update (EV_P) +/* fetch new monotonic and realtime times from the kernel */ +/* also detetc if there was a timejump, and act accordingly */ +inline_speed void +time_update (EV_P_ ev_tstamp max_block) { - int i; - #if EV_USE_MONOTONIC if (expect_true (have_monotonic)) { - if (time_update_monotonic (EV_A)) + int i; + ev_tstamp odiff = rtmn_diff; + + mn_now = get_clock (); + + /* only fetch the realtime clock every 0.5*MIN_TIMEJUMP seconds */ + /* interpolate in the meantime */ + if (expect_true (mn_now - now_floor < MIN_TIMEJUMP * .5)) { - ev_tstamp odiff = rtmn_diff; + ev_rt_now = rtmn_diff + mn_now; + return; + } - for (i = 4; --i; ) /* loop a few times, before making important decisions */ - { - rtmn_diff = ev_rt_now - mn_now; + now_floor = mn_now; + ev_rt_now = ev_time (); - if (fabs (odiff - rtmn_diff) < MIN_TIMEJUMP) - return; /* all is well */ + /* loop a few times, before making important decisions. + * on the choice of "4": one iteration isn't enough, + * in case we get preempted during the calls to + * ev_time and get_clock. a second call is almost guaranteed + * to succeed in that case, though. and looping a few more times + * doesn't hurt either as we only do this on time-jumps or + * in the unlikely event of having been preempted here. + */ + for (i = 4; --i; ) + { + rtmn_diff = ev_rt_now - mn_now; - ev_rt_now = ev_time (); - mn_now = get_clock (); - now_floor = mn_now; - } + if (expect_true (fabs (odiff - rtmn_diff) < MIN_TIMEJUMP)) + return; /* all is well */ -# if EV_PERIODICS - periodics_reschedule (EV_A); -# endif - /* no timer adjustment, as the monotonic clock doesn't jump */ - /* timers_reschedule (EV_A_ rtmn_diff - odiff) */ + ev_rt_now = ev_time (); + mn_now = get_clock (); + now_floor = mn_now; } + + /* no timer adjustment, as the monotonic clock doesn't jump */ + /* timers_reschedule (EV_A_ rtmn_diff - odiff) */ +# if EV_PERIODIC_ENABLE + periodics_reschedule (EV_A); +# endif } else #endif { ev_rt_now = ev_time (); - if (expect_false (mn_now > ev_rt_now || mn_now < ev_rt_now - MAX_BLOCKTIME - MIN_TIMEJUMP)) + if (expect_false (mn_now > ev_rt_now || ev_rt_now > mn_now + max_block + MIN_TIMEJUMP)) { -#if EV_PERIODICS + /* adjust timers. this is easy, as the offset is the same for all of them */ + timers_reschedule (EV_A_ ev_rt_now - mn_now); +#if EV_PERIODIC_ENABLE periodics_reschedule (EV_A); #endif - - /* adjust timers. this is easy, as the offset is the same for all */ - for (i = 0; i < timercnt; ++i) - ((WT)timers [i])->at += ev_rt_now - mn_now; } mn_now = ev_rt_now; @@ -1142,32 +2073,46 @@ } void -ev_ref (EV_P) -{ - ++activecnt; -} - -void -ev_unref (EV_P) +ev_loop (EV_P_ int flags) { - --activecnt; -} +#if EV_MINIMAL < 2 + ++loop_depth; +#endif -static int loop_done; + loop_done = EVUNLOOP_CANCEL; -void -ev_loop (EV_P_ int flags) -{ - double block; - loop_done = flags & (EVLOOP_ONESHOT | EVLOOP_NONBLOCK) ? 1 : 0; + EV_INVOKE_PENDING; /* in case we recurse, ensure ordering stays nice and clean */ do { - /* queue check watchers (and execute them) */ +#if EV_VERIFY >= 2 + ev_loop_verify (EV_A); +#endif + +#ifndef _WIN32 + if (expect_false (curpid)) /* penalise the forking check even more */ + if (expect_false (getpid () != curpid)) + { + curpid = getpid (); + postfork = 1; + } +#endif + +#if EV_FORK_ENABLE + /* we might have forked, so queue fork handlers */ + if (expect_false (postfork)) + if (forkcnt) + { + queue_events (EV_A_ (W *)forks, forkcnt, EV_FORK); + EV_INVOKE_PENDING; + } +#endif + + /* queue prepare watchers (and execute them) */ if (expect_false (preparecnt)) { queue_events (EV_A_ (W *)prepares, preparecnt, EV_PREPARE); - call_pending (EV_A); + EV_INVOKE_PENDING; } /* we might have forked, so reify kernel state if necessary */ @@ -1178,67 +2123,92 @@ fd_reify (EV_A); /* calculate blocking time */ + { + ev_tstamp waittime = 0.; + ev_tstamp sleeptime = 0.; - /* we only need this for !monotonic clock or timers, but as we basically - always have timers, we just calculate it always */ -#if EV_USE_MONOTONIC - if (expect_true (have_monotonic)) - time_update_monotonic (EV_A); - else -#endif - { - ev_rt_now = ev_time (); - mn_now = ev_rt_now; - } - - if (flags & EVLOOP_NONBLOCK || idlecnt) - block = 0.; - else - { - block = MAX_BLOCKTIME; + if (expect_true (!(flags & EVLOOP_NONBLOCK || idleall || !activecnt))) + { + /* remember old timestamp for io_blocktime calculation */ + ev_tstamp prev_mn_now = mn_now; - if (timercnt) - { - ev_tstamp to = ((WT)timers [0])->at - mn_now + method_fudge; - if (block > to) block = to; - } + /* update time to cancel out callback processing overhead */ + time_update (EV_A_ 1e100); -#if EV_PERIODICS - if (periodiccnt) - { - ev_tstamp to = ((WT)periodics [0])->at - ev_rt_now + method_fudge; - if (block > to) block = to; - } -#endif + waittime = MAX_BLOCKTIME; - if (block < 0.) block = 0.; - } + if (timercnt) + { + ev_tstamp to = ANHE_at (timers [HEAP0]) - mn_now + backend_fudge; + if (waittime > to) waittime = to; + } + +#if EV_PERIODIC_ENABLE + if (periodiccnt) + { + ev_tstamp to = ANHE_at (periodics [HEAP0]) - ev_rt_now + backend_fudge; + if (waittime > to) waittime = to; + } +#endif + + /* don't let timeouts decrease the waittime below timeout_blocktime */ + if (expect_false (waittime < timeout_blocktime)) + waittime = timeout_blocktime; + + /* extra check because io_blocktime is commonly 0 */ + if (expect_false (io_blocktime)) + { + sleeptime = io_blocktime - (mn_now - prev_mn_now); + + if (sleeptime > waittime - backend_fudge) + sleeptime = waittime - backend_fudge; + + if (expect_true (sleeptime > 0.)) + { + ev_sleep (sleeptime); + waittime -= sleeptime; + } + } + } - method_poll (EV_A_ block); +#if EV_MINIMAL < 2 + ++loop_count; +#endif + backend_poll (EV_A_ waittime); - /* update ev_rt_now, do magic */ - time_update (EV_A); + /* update ev_rt_now, do magic */ + time_update (EV_A_ waittime + sleeptime); + } /* queue pending timers and reschedule them */ timers_reify (EV_A); /* relative timers called last */ -#if EV_PERIODICS +#if EV_PERIODIC_ENABLE periodics_reify (EV_A); /* absolute timers called first */ #endif - /* queue idle watchers unless io or timers are pending */ - if (idlecnt && !any_pending (EV_A)) - queue_events (EV_A_ (W *)idles, idlecnt, EV_IDLE); +#if EV_IDLE_ENABLE + /* queue idle watchers unless other events are pending */ + idle_reify (EV_A); +#endif /* queue check watchers, to be executed first */ - if (checkcnt) + if (expect_false (checkcnt)) queue_events (EV_A_ (W *)checks, checkcnt, EV_CHECK); - call_pending (EV_A); + EV_INVOKE_PENDING; } - while (activecnt && !loop_done); + while (expect_true ( + activecnt + && !loop_done + && !(flags & (EVLOOP_ONESHOT | EVLOOP_NONBLOCK)) + )); - if (loop_done != 2) - loop_done = 0; + if (loop_done == EVUNLOOP_ONE) + loop_done = EVUNLOOP_CANCEL; + +#if EV_MINIMAL < 2 + --loop_depth; +#endif } void @@ -1247,16 +2217,54 @@ loop_done = how; } +void +ev_ref (EV_P) +{ + ++activecnt; +} + +void +ev_unref (EV_P) +{ + --activecnt; +} + +void +ev_now_update (EV_P) +{ + time_update (EV_A_ 1e100); +} + +void +ev_suspend (EV_P) +{ + ev_now_update (EV_A); +} + +void +ev_resume (EV_P) +{ + ev_tstamp mn_prev = mn_now; + + ev_now_update (EV_A); + timers_reschedule (EV_A_ mn_now - mn_prev); +#if EV_PERIODIC_ENABLE + /* TODO: really do this? */ + periodics_reschedule (EV_A); +#endif +} + /*****************************************************************************/ +/* singly-linked list management, used when the expected list length is short */ -inline void +inline_size void wlist_add (WL *head, WL elem) { elem->next = *head; *head = elem; } -inline void +inline_size void wlist_del (WL *head, WL elem) { while (*head) @@ -1271,27 +2279,52 @@ } } -inline void -ev_clear_pending (EV_P_ W w) +/* internal, faster, version of ev_clear_pending */ +inline_speed void +clear_pending (EV_P_ W w) { if (w->pending) { - pendings [ABSPRI (w)][w->pending - 1].w = 0; + pendings [ABSPRI (w)][w->pending - 1].w = (W)&pending_w; w->pending = 0; } } -inline void -ev_start (EV_P_ W w, int active) +int +ev_clear_pending (EV_P_ void *w) +{ + W w_ = (W)w; + int pending = w_->pending; + + if (expect_true (pending)) + { + ANPENDING *p = pendings [ABSPRI (w_)] + pending - 1; + p->w = (W)&pending_w; + w_->pending = 0; + return p->events; + } + else + return 0; +} + +inline_size void +pri_adjust (EV_P_ W w) { - if (w->priority < EV_MINPRI) w->priority = EV_MINPRI; - if (w->priority > EV_MAXPRI) w->priority = EV_MAXPRI; + int pri = ev_priority (w); + pri = pri < EV_MINPRI ? EV_MINPRI : pri; + pri = pri > EV_MAXPRI ? EV_MAXPRI : pri; + ev_set_priority (w, pri); +} +inline_speed void +ev_start (EV_P_ W w, int active) +{ + pri_adjust (EV_A_ w); w->active = active; ev_ref (EV_A); } -inline void +inline_size void ev_stop (EV_P_ W w) { ev_unref (EV_A); @@ -1300,137 +2333,189 @@ /*****************************************************************************/ -void -ev_io_start (EV_P_ struct ev_io *w) +void noinline +ev_io_start (EV_P_ ev_io *w) { int fd = w->fd; - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; - assert (("ev_io_start called with negative fd", fd >= 0)); + assert (("libev: ev_io_start called with negative fd", fd >= 0)); + assert (("libev: ev_io start called with illegal event mask", !(w->events & ~(EV__IOFDSET | EV_READ | EV_WRITE)))); + + EV_FREQUENT_CHECK; ev_start (EV_A_ (W)w, 1); - array_needsize (ANFD, anfds, anfdmax, fd + 1, anfds_init); - wlist_add ((WL *)&anfds[fd].head, (WL)w); + array_needsize (ANFD, anfds, anfdmax, fd + 1, array_init_zero); + wlist_add (&anfds[fd].head, (WL)w); + + fd_change (EV_A_ fd, w->events & EV__IOFDSET | 1); + w->events &= ~EV__IOFDSET; - fd_change (EV_A_ fd); + EV_FREQUENT_CHECK; } -void -ev_io_stop (EV_P_ struct ev_io *w) +void noinline +ev_io_stop (EV_P_ ev_io *w) { - ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) return; - assert (("ev_io_start called with illegal fd (must stay constant after start!)", w->fd >= 0 && w->fd < anfdmax)); + assert (("libev: ev_io_stop called with illegal fd (must stay constant after start!)", w->fd >= 0 && w->fd < anfdmax)); - wlist_del ((WL *)&anfds[w->fd].head, (WL)w); + EV_FREQUENT_CHECK; + + wlist_del (&anfds[w->fd].head, (WL)w); ev_stop (EV_A_ (W)w); - fd_change (EV_A_ w->fd); + fd_change (EV_A_ w->fd, 1); + + EV_FREQUENT_CHECK; } -void -ev_timer_start (EV_P_ struct ev_timer *w) +void noinline +ev_timer_start (EV_P_ ev_timer *w) { - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; - ((WT)w)->at += mn_now; + ev_at (w) += mn_now; + + assert (("libev: ev_timer_start called with negative timer repeat value", w->repeat >= 0.)); + + EV_FREQUENT_CHECK; - assert (("ev_timer_start called with negative timer repeat value", w->repeat >= 0.)); + ++timercnt; + ev_start (EV_A_ (W)w, timercnt + HEAP0 - 1); + array_needsize (ANHE, timers, timermax, ev_active (w) + 1, EMPTY2); + ANHE_w (timers [ev_active (w)]) = (WT)w; + ANHE_at_cache (timers [ev_active (w)]); + upheap (timers, ev_active (w)); - ev_start (EV_A_ (W)w, ++timercnt); - array_needsize (struct ev_timer *, timers, timermax, timercnt, (void)); - timers [timercnt - 1] = w; - upheap ((WT *)timers, timercnt - 1); + EV_FREQUENT_CHECK; - assert (("internal timer heap corruption", timers [((W)w)->active - 1] == w)); + /*assert (("libev: internal timer heap corruption", timers [ev_active (w)] == (WT)w));*/ } -void -ev_timer_stop (EV_P_ struct ev_timer *w) +void noinline +ev_timer_stop (EV_P_ ev_timer *w) { - ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) return; - assert (("internal timer heap corruption", timers [((W)w)->active - 1] == w)); + EV_FREQUENT_CHECK; - if (((W)w)->active < timercnt--) - { - timers [((W)w)->active - 1] = timers [timercnt]; - adjustheap ((WT *)timers, timercnt, ((W)w)->active - 1); - } + { + int active = ev_active (w); + + assert (("libev: internal timer heap corruption", ANHE_w (timers [active]) == (WT)w)); + + --timercnt; + + if (expect_true (active < timercnt + HEAP0)) + { + timers [active] = timers [timercnt + HEAP0]; + adjustheap (timers, timercnt, active); + } + } + + EV_FREQUENT_CHECK; - ((WT)w)->at -= mn_now; + ev_at (w) -= mn_now; ev_stop (EV_A_ (W)w); } -void -ev_timer_again (EV_P_ struct ev_timer *w) +void noinline +ev_timer_again (EV_P_ ev_timer *w) { + EV_FREQUENT_CHECK; + if (ev_is_active (w)) { if (w->repeat) { - ((WT)w)->at = mn_now + w->repeat; - adjustheap ((WT *)timers, timercnt, ((W)w)->active - 1); + ev_at (w) = mn_now + w->repeat; + ANHE_at_cache (timers [ev_active (w)]); + adjustheap (timers, timercnt, ev_active (w)); } else ev_timer_stop (EV_A_ w); } else if (w->repeat) - ev_timer_start (EV_A_ w); + { + ev_at (w) = w->repeat; + ev_timer_start (EV_A_ w); + } + + EV_FREQUENT_CHECK; } -#if EV_PERIODICS -void -ev_periodic_start (EV_P_ struct ev_periodic *w) +#if EV_PERIODIC_ENABLE +void noinline +ev_periodic_start (EV_P_ ev_periodic *w) { - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; if (w->reschedule_cb) - ((WT)w)->at = w->reschedule_cb (w, ev_rt_now); + ev_at (w) = w->reschedule_cb (w, ev_rt_now); else if (w->interval) { - assert (("ev_periodic_start called with negative interval value", w->interval >= 0.)); + assert (("libev: ev_periodic_start called with negative interval value", w->interval >= 0.)); /* this formula differs from the one in periodic_reify because we do not always round up */ - ((WT)w)->at += ceil ((ev_rt_now - ((WT)w)->at) / w->interval) * w->interval; + ev_at (w) = w->offset + ceil ((ev_rt_now - w->offset) / w->interval) * w->interval; } + else + ev_at (w) = w->offset; + + EV_FREQUENT_CHECK; - ev_start (EV_A_ (W)w, ++periodiccnt); - array_needsize (struct ev_periodic *, periodics, periodicmax, periodiccnt, (void)); - periodics [periodiccnt - 1] = w; - upheap ((WT *)periodics, periodiccnt - 1); + ++periodiccnt; + ev_start (EV_A_ (W)w, periodiccnt + HEAP0 - 1); + array_needsize (ANHE, periodics, periodicmax, ev_active (w) + 1, EMPTY2); + ANHE_w (periodics [ev_active (w)]) = (WT)w; + ANHE_at_cache (periodics [ev_active (w)]); + upheap (periodics, ev_active (w)); - assert (("internal periodic heap corruption", periodics [((W)w)->active - 1] == w)); + EV_FREQUENT_CHECK; + + /*assert (("libev: internal periodic heap corruption", ANHE_w (periodics [ev_active (w)]) == (WT)w));*/ } -void -ev_periodic_stop (EV_P_ struct ev_periodic *w) +void noinline +ev_periodic_stop (EV_P_ ev_periodic *w) { - ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) return; - assert (("internal periodic heap corruption", periodics [((W)w)->active - 1] == w)); + EV_FREQUENT_CHECK; - if (((W)w)->active < periodiccnt--) - { - periodics [((W)w)->active - 1] = periodics [periodiccnt]; - adjustheap ((WT *)periodics, periodiccnt, ((W)w)->active - 1); - } + { + int active = ev_active (w); + + assert (("libev: internal periodic heap corruption", ANHE_w (periodics [active]) == (WT)w)); + + --periodiccnt; + + if (expect_true (active < periodiccnt + HEAP0)) + { + periodics [active] = periodics [periodiccnt + HEAP0]; + adjustheap (periodics, periodiccnt, active); + } + } + + EV_FREQUENT_CHECK; ev_stop (EV_A_ (W)w); } -void -ev_periodic_again (EV_P_ struct ev_periodic *w) +void noinline +ev_periodic_again (EV_P_ ev_periodic *w) { /* TODO: use adjustheap and recalculation */ ev_periodic_stop (EV_A_ w); @@ -1438,149 +2523,735 @@ } #endif +#ifndef SA_RESTART +# define SA_RESTART 0 +#endif + +void noinline +ev_signal_start (EV_P_ ev_signal *w) +{ +#if EV_MULTIPLICITY + assert (("libev: signal watchers are only supported in the default loop", loop == ev_default_loop_ptr)); +#endif + if (expect_false (ev_is_active (w))) + return; + + assert (("libev: ev_signal_start called with illegal signal number", w->signum > 0)); + + evpipe_init (EV_A); + + EV_FREQUENT_CHECK; + + { +#ifndef _WIN32 + sigset_t full, prev; + sigfillset (&full); + sigprocmask (SIG_SETMASK, &full, &prev); +#endif + + array_needsize (ANSIG, signals, signalmax, w->signum, array_init_zero); + +#ifndef _WIN32 + sigprocmask (SIG_SETMASK, &prev, 0); +#endif + } + + ev_start (EV_A_ (W)w, 1); + wlist_add (&signals [w->signum - 1].head, (WL)w); + + if (!((WL)w)->next) + { +#if _WIN32 + signal (w->signum, ev_sighandler); +#else + struct sigaction sa; + sa.sa_handler = ev_sighandler; + sigfillset (&sa.sa_mask); + sa.sa_flags = SA_RESTART; /* if restarting works we save one iteration */ + sigaction (w->signum, &sa, 0); +#endif + } + + EV_FREQUENT_CHECK; +} + +void noinline +ev_signal_stop (EV_P_ ev_signal *w) +{ + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) + return; + + EV_FREQUENT_CHECK; + + wlist_del (&signals [w->signum - 1].head, (WL)w); + ev_stop (EV_A_ (W)w); + + if (!signals [w->signum - 1].head) + signal (w->signum, SIG_DFL); + + EV_FREQUENT_CHECK; +} + void -ev_idle_start (EV_P_ struct ev_idle *w) +ev_child_start (EV_P_ ev_child *w) { - if (ev_is_active (w)) +#if EV_MULTIPLICITY + assert (("libev: child watchers are only supported in the default loop", loop == ev_default_loop_ptr)); +#endif + if (expect_false (ev_is_active (w))) return; - ev_start (EV_A_ (W)w, ++idlecnt); - array_needsize (struct ev_idle *, idles, idlemax, idlecnt, (void)); - idles [idlecnt - 1] = w; + EV_FREQUENT_CHECK; + + ev_start (EV_A_ (W)w, 1); + wlist_add (&childs [w->pid & (EV_PID_HASHSIZE - 1)], (WL)w); + + EV_FREQUENT_CHECK; } void -ev_idle_stop (EV_P_ struct ev_idle *w) +ev_child_stop (EV_P_ ev_child *w) { - ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) return; - idles [((W)w)->active - 1] = idles [--idlecnt]; + EV_FREQUENT_CHECK; + + wlist_del (&childs [w->pid & (EV_PID_HASHSIZE - 1)], (WL)w); ev_stop (EV_A_ (W)w); + + EV_FREQUENT_CHECK; +} + +#if EV_STAT_ENABLE + +# ifdef _WIN32 +# undef lstat +# define lstat(a,b) _stati64 (a,b) +# endif + +#define DEF_STAT_INTERVAL 5.0074891 +#define NFS_STAT_INTERVAL 30.1074891 /* for filesystems potentially failing inotify */ +#define MIN_STAT_INTERVAL 0.1074891 + +static void noinline stat_timer_cb (EV_P_ ev_timer *w_, int revents); + +#if EV_USE_INOTIFY +# define EV_INOTIFY_BUFSIZE 8192 + +static void noinline +infy_add (EV_P_ ev_stat *w) +{ + w->wd = inotify_add_watch (fs_fd, w->path, IN_ATTRIB | IN_DELETE_SELF | IN_MOVE_SELF | IN_MODIFY | IN_DONT_FOLLOW | IN_MASK_ADD); + + if (w->wd < 0) + { + w->timer.repeat = w->interval ? w->interval : DEF_STAT_INTERVAL; + ev_timer_again (EV_A_ &w->timer); /* this is not race-free, so we still need to recheck periodically */ + + /* monitor some parent directory for speedup hints */ + /* note that exceeding the hardcoded path limit is not a correctness issue, */ + /* but an efficiency issue only */ + if ((errno == ENOENT || errno == EACCES) && strlen (w->path) < 4096) + { + char path [4096]; + strcpy (path, w->path); + + do + { + int mask = IN_MASK_ADD | IN_DELETE_SELF | IN_MOVE_SELF + | (errno == EACCES ? IN_ATTRIB : IN_CREATE | IN_MOVED_TO); + + char *pend = strrchr (path, '/'); + + if (!pend || pend == path) + break; + + *pend = 0; + w->wd = inotify_add_watch (fs_fd, path, mask); + } + while (w->wd < 0 && (errno == ENOENT || errno == EACCES)); + } + } + + if (w->wd >= 0) + { + wlist_add (&fs_hash [w->wd & (EV_INOTIFY_HASHSIZE - 1)].head, (WL)w); + + /* now local changes will be tracked by inotify, but remote changes won't */ + /* unless the filesystem it known to be local, we therefore still poll */ + /* also do poll on <2.6.25, but with normal frequency */ + struct statfs sfs; + + if (fs_2625 && !statfs (w->path, &sfs)) + if (sfs.f_type == 0x1373 /* devfs */ + || sfs.f_type == 0xEF53 /* ext2/3 */ + || sfs.f_type == 0x3153464a /* jfs */ + || sfs.f_type == 0x52654973 /* reiser3 */ + || sfs.f_type == 0x01021994 /* tempfs */ + || sfs.f_type == 0x58465342 /* xfs */) + return; + + w->timer.repeat = w->interval ? w->interval : fs_2625 ? NFS_STAT_INTERVAL : DEF_STAT_INTERVAL; + ev_timer_again (EV_A_ &w->timer); + } +} + +static void noinline +infy_del (EV_P_ ev_stat *w) +{ + int slot; + int wd = w->wd; + + if (wd < 0) + return; + + w->wd = -2; + slot = wd & (EV_INOTIFY_HASHSIZE - 1); + wlist_del (&fs_hash [slot].head, (WL)w); + + /* remove this watcher, if others are watching it, they will rearm */ + inotify_rm_watch (fs_fd, wd); } +static void noinline +infy_wd (EV_P_ int slot, int wd, struct inotify_event *ev) +{ + if (slot < 0) + /* overflow, need to check for all hash slots */ + for (slot = 0; slot < EV_INOTIFY_HASHSIZE; ++slot) + infy_wd (EV_A_ slot, wd, ev); + else + { + WL w_; + + for (w_ = fs_hash [slot & (EV_INOTIFY_HASHSIZE - 1)].head; w_; ) + { + ev_stat *w = (ev_stat *)w_; + w_ = w_->next; /* lets us remove this watcher and all before it */ + + if (w->wd == wd || wd == -1) + { + if (ev->mask & (IN_IGNORED | IN_UNMOUNT | IN_DELETE_SELF)) + { + wlist_del (&fs_hash [slot & (EV_INOTIFY_HASHSIZE - 1)].head, (WL)w); + w->wd = -1; + infy_add (EV_A_ w); /* re-add, no matter what */ + } + + stat_timer_cb (EV_A_ &w->timer, 0); + } + } + } +} + +static void +infy_cb (EV_P_ ev_io *w, int revents) +{ + char buf [EV_INOTIFY_BUFSIZE]; + struct inotify_event *ev = (struct inotify_event *)buf; + int ofs; + int len = read (fs_fd, buf, sizeof (buf)); + + for (ofs = 0; ofs < len; ofs += sizeof (struct inotify_event) + ev->len) + infy_wd (EV_A_ ev->wd, ev->wd, ev); +} + +inline_size void +check_2625 (EV_P) +{ + /* kernels < 2.6.25 are borked + * http://www.ussg.indiana.edu/hypermail/linux/kernel/0711.3/1208.html + */ + struct utsname buf; + int major, minor, micro; + + if (uname (&buf)) + return; + + if (sscanf (buf.release, "%d.%d.%d", &major, &minor, µ) != 3) + return; + + if (major < 2 + || (major == 2 && minor < 6) + || (major == 2 && minor == 6 && micro < 25)) + return; + + fs_2625 = 1; +} + +inline_size void +infy_init (EV_P) +{ + if (fs_fd != -2) + return; + + fs_fd = -1; + + check_2625 (EV_A); + + fs_fd = inotify_init (); + + if (fs_fd >= 0) + { + ev_io_init (&fs_w, infy_cb, fs_fd, EV_READ); + ev_set_priority (&fs_w, EV_MAXPRI); + ev_io_start (EV_A_ &fs_w); + } +} + +inline_size void +infy_fork (EV_P) +{ + int slot; + + if (fs_fd < 0) + return; + + close (fs_fd); + fs_fd = inotify_init (); + + for (slot = 0; slot < EV_INOTIFY_HASHSIZE; ++slot) + { + WL w_ = fs_hash [slot].head; + fs_hash [slot].head = 0; + + while (w_) + { + ev_stat *w = (ev_stat *)w_; + w_ = w_->next; /* lets us add this watcher */ + + w->wd = -1; + + if (fs_fd >= 0) + infy_add (EV_A_ w); /* re-add, no matter what */ + else + ev_timer_again (EV_A_ &w->timer); + } + } +} + +#endif + +#ifdef _WIN32 +# define EV_LSTAT(p,b) _stati64 (p, b) +#else +# define EV_LSTAT(p,b) lstat (p, b) +#endif + void -ev_prepare_start (EV_P_ struct ev_prepare *w) +ev_stat_stat (EV_P_ ev_stat *w) { - if (ev_is_active (w)) + if (lstat (w->path, &w->attr) < 0) + w->attr.st_nlink = 0; + else if (!w->attr.st_nlink) + w->attr.st_nlink = 1; +} + +static void noinline +stat_timer_cb (EV_P_ ev_timer *w_, int revents) +{ + ev_stat *w = (ev_stat *)(((char *)w_) - offsetof (ev_stat, timer)); + + /* we copy this here each the time so that */ + /* prev has the old value when the callback gets invoked */ + w->prev = w->attr; + ev_stat_stat (EV_A_ w); + + /* memcmp doesn't work on netbsd, they.... do stuff to their struct stat */ + if ( + w->prev.st_dev != w->attr.st_dev + || w->prev.st_ino != w->attr.st_ino + || w->prev.st_mode != w->attr.st_mode + || w->prev.st_nlink != w->attr.st_nlink + || w->prev.st_uid != w->attr.st_uid + || w->prev.st_gid != w->attr.st_gid + || w->prev.st_rdev != w->attr.st_rdev + || w->prev.st_size != w->attr.st_size + || w->prev.st_atime != w->attr.st_atime + || w->prev.st_mtime != w->attr.st_mtime + || w->prev.st_ctime != w->attr.st_ctime + ) { + #if EV_USE_INOTIFY + if (fs_fd >= 0) + { + infy_del (EV_A_ w); + infy_add (EV_A_ w); + ev_stat_stat (EV_A_ w); /* avoid race... */ + } + #endif + + ev_feed_event (EV_A_ w, EV_STAT); + } +} + +void +ev_stat_start (EV_P_ ev_stat *w) +{ + if (expect_false (ev_is_active (w))) return; + ev_stat_stat (EV_A_ w); + + if (w->interval < MIN_STAT_INTERVAL && w->interval) + w->interval = MIN_STAT_INTERVAL; + + ev_timer_init (&w->timer, stat_timer_cb, 0., w->interval ? w->interval : DEF_STAT_INTERVAL); + ev_set_priority (&w->timer, ev_priority (w)); + +#if EV_USE_INOTIFY + infy_init (EV_A); + + if (fs_fd >= 0) + infy_add (EV_A_ w); + else +#endif + ev_timer_again (EV_A_ &w->timer); + + ev_start (EV_A_ (W)w, 1); + + EV_FREQUENT_CHECK; +} + +void +ev_stat_stop (EV_P_ ev_stat *w) +{ + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) + return; + + EV_FREQUENT_CHECK; + +#if EV_USE_INOTIFY + infy_del (EV_A_ w); +#endif + ev_timer_stop (EV_A_ &w->timer); + + ev_stop (EV_A_ (W)w); + + EV_FREQUENT_CHECK; +} +#endif + +#if EV_IDLE_ENABLE +void +ev_idle_start (EV_P_ ev_idle *w) +{ + if (expect_false (ev_is_active (w))) + return; + + pri_adjust (EV_A_ (W)w); + + EV_FREQUENT_CHECK; + + { + int active = ++idlecnt [ABSPRI (w)]; + + ++idleall; + ev_start (EV_A_ (W)w, active); + + array_needsize (ev_idle *, idles [ABSPRI (w)], idlemax [ABSPRI (w)], active, EMPTY2); + idles [ABSPRI (w)][active - 1] = w; + } + + EV_FREQUENT_CHECK; +} + +void +ev_idle_stop (EV_P_ ev_idle *w) +{ + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) + return; + + EV_FREQUENT_CHECK; + + { + int active = ev_active (w); + + idles [ABSPRI (w)][active - 1] = idles [ABSPRI (w)][--idlecnt [ABSPRI (w)]]; + ev_active (idles [ABSPRI (w)][active - 1]) = active; + + ev_stop (EV_A_ (W)w); + --idleall; + } + + EV_FREQUENT_CHECK; +} +#endif + +void +ev_prepare_start (EV_P_ ev_prepare *w) +{ + if (expect_false (ev_is_active (w))) + return; + + EV_FREQUENT_CHECK; + ev_start (EV_A_ (W)w, ++preparecnt); - array_needsize (struct ev_prepare *, prepares, preparemax, preparecnt, (void)); + array_needsize (ev_prepare *, prepares, preparemax, preparecnt, EMPTY2); prepares [preparecnt - 1] = w; + + EV_FREQUENT_CHECK; } void -ev_prepare_stop (EV_P_ struct ev_prepare *w) +ev_prepare_stop (EV_P_ ev_prepare *w) { - ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) return; - prepares [((W)w)->active - 1] = prepares [--preparecnt]; + EV_FREQUENT_CHECK; + + { + int active = ev_active (w); + + prepares [active - 1] = prepares [--preparecnt]; + ev_active (prepares [active - 1]) = active; + } + ev_stop (EV_A_ (W)w); + + EV_FREQUENT_CHECK; } void -ev_check_start (EV_P_ struct ev_check *w) +ev_check_start (EV_P_ ev_check *w) { - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; + EV_FREQUENT_CHECK; + ev_start (EV_A_ (W)w, ++checkcnt); - array_needsize (struct ev_check *, checks, checkmax, checkcnt, (void)); + array_needsize (ev_check *, checks, checkmax, checkcnt, EMPTY2); checks [checkcnt - 1] = w; + + EV_FREQUENT_CHECK; } void -ev_check_stop (EV_P_ struct ev_check *w) +ev_check_stop (EV_P_ ev_check *w) { - ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) return; - checks [((W)w)->active - 1] = checks [--checkcnt]; + EV_FREQUENT_CHECK; + + { + int active = ev_active (w); + + checks [active - 1] = checks [--checkcnt]; + ev_active (checks [active - 1]) = active; + } + ev_stop (EV_A_ (W)w); + + EV_FREQUENT_CHECK; } -#ifndef SA_RESTART -# define SA_RESTART 0 +#if EV_EMBED_ENABLE +void noinline +ev_embed_sweep (EV_P_ ev_embed *w) +{ + ev_loop (w->other, EVLOOP_NONBLOCK); +} + +static void +embed_io_cb (EV_P_ ev_io *io, int revents) +{ + ev_embed *w = (ev_embed *)(((char *)io) - offsetof (ev_embed, io)); + + if (ev_cb (w)) + ev_feed_event (EV_A_ (W)w, EV_EMBED); + else + ev_loop (w->other, EVLOOP_NONBLOCK); +} + +static void +embed_prepare_cb (EV_P_ ev_prepare *prepare, int revents) +{ + ev_embed *w = (ev_embed *)(((char *)prepare) - offsetof (ev_embed, prepare)); + + { + struct ev_loop *loop = w->other; + + while (fdchangecnt) + { + fd_reify (EV_A); + ev_loop (EV_A_ EVLOOP_NONBLOCK); + } + } +} + +static void +embed_fork_cb (EV_P_ ev_fork *fork_w, int revents) +{ + ev_embed *w = (ev_embed *)(((char *)fork_w) - offsetof (ev_embed, fork)); + + ev_embed_stop (EV_A_ w); + + { + struct ev_loop *loop = w->other; + + ev_loop_fork (EV_A); + ev_loop (EV_A_ EVLOOP_NONBLOCK); + } + + ev_embed_start (EV_A_ w); +} + +#if 0 +static void +embed_idle_cb (EV_P_ ev_idle *idle, int revents) +{ + ev_idle_stop (EV_A_ idle); +} #endif void -ev_signal_start (EV_P_ struct ev_signal *w) +ev_embed_start (EV_P_ ev_embed *w) { -#if EV_MULTIPLICITY - assert (("signal watchers are only supported in the default loop", loop == default_loop)); -#endif - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; - assert (("ev_signal_start called with illegal signal number", w->signum > 0)); + { + struct ev_loop *loop = w->other; + assert (("libev: loop to be embedded is not embeddable", backend & ev_embeddable_backends ())); + ev_io_init (&w->io, embed_io_cb, backend_fd, EV_READ); + } + + EV_FREQUENT_CHECK; + + ev_set_priority (&w->io, ev_priority (w)); + ev_io_start (EV_A_ &w->io); + + ev_prepare_init (&w->prepare, embed_prepare_cb); + ev_set_priority (&w->prepare, EV_MINPRI); + ev_prepare_start (EV_A_ &w->prepare); + + ev_fork_init (&w->fork, embed_fork_cb); + ev_fork_start (EV_A_ &w->fork); + + /*ev_idle_init (&w->idle, e,bed_idle_cb);*/ ev_start (EV_A_ (W)w, 1); - array_needsize (ANSIG, signals, signalmax, w->signum, signals_init); - wlist_add ((WL *)&signals [w->signum - 1].head, (WL)w); - if (!((WL)w)->next) - { -#if _WIN32 - signal (w->signum, sighandler); -#else - struct sigaction sa; - sa.sa_handler = sighandler; - sigfillset (&sa.sa_mask); - sa.sa_flags = SA_RESTART; /* if restarting works we save one iteration */ - sigaction (w->signum, &sa, 0); + EV_FREQUENT_CHECK; +} + +void +ev_embed_stop (EV_P_ ev_embed *w) +{ + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) + return; + + EV_FREQUENT_CHECK; + + ev_io_stop (EV_A_ &w->io); + ev_prepare_stop (EV_A_ &w->prepare); + ev_fork_stop (EV_A_ &w->fork); + + EV_FREQUENT_CHECK; +} #endif - } + +#if EV_FORK_ENABLE +void +ev_fork_start (EV_P_ ev_fork *w) +{ + if (expect_false (ev_is_active (w))) + return; + + EV_FREQUENT_CHECK; + + ev_start (EV_A_ (W)w, ++forkcnt); + array_needsize (ev_fork *, forks, forkmax, forkcnt, EMPTY2); + forks [forkcnt - 1] = w; + + EV_FREQUENT_CHECK; } void -ev_signal_stop (EV_P_ struct ev_signal *w) +ev_fork_stop (EV_P_ ev_fork *w) { - ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) return; - wlist_del ((WL *)&signals [w->signum - 1].head, (WL)w); + EV_FREQUENT_CHECK; + + { + int active = ev_active (w); + + forks [active - 1] = forks [--forkcnt]; + ev_active (forks [active - 1]) = active; + } + ev_stop (EV_A_ (W)w); - if (!signals [w->signum - 1].head) - signal (w->signum, SIG_DFL); + EV_FREQUENT_CHECK; } +#endif +#if EV_ASYNC_ENABLE void -ev_child_start (EV_P_ struct ev_child *w) +ev_async_start (EV_P_ ev_async *w) { -#if EV_MULTIPLICITY - assert (("child watchers are only supported in the default loop", loop == default_loop)); -#endif - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; - ev_start (EV_A_ (W)w, 1); - wlist_add ((WL *)&childs [w->pid & (PID_HASHSIZE - 1)], (WL)w); + evpipe_init (EV_A); + + EV_FREQUENT_CHECK; + + ev_start (EV_A_ (W)w, ++asynccnt); + array_needsize (ev_async *, asyncs, asyncmax, asynccnt, EMPTY2); + asyncs [asynccnt - 1] = w; + + EV_FREQUENT_CHECK; } void -ev_child_stop (EV_P_ struct ev_child *w) +ev_async_stop (EV_P_ ev_async *w) { - ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) return; - wlist_del ((WL *)&childs [w->pid & (PID_HASHSIZE - 1)], (WL)w); + EV_FREQUENT_CHECK; + + { + int active = ev_active (w); + + asyncs [active - 1] = asyncs [--asynccnt]; + ev_active (asyncs [active - 1]) = active; + } + ev_stop (EV_A_ (W)w); + + EV_FREQUENT_CHECK; } +void +ev_async_send (EV_P_ ev_async *w) +{ + w->sent = 1; + evpipe_write (EV_A_ &gotasync); +} +#endif + /*****************************************************************************/ struct ev_once { - struct ev_io io; - struct ev_timer to; + ev_io io; + ev_timer to; void (*cb)(int revents, void *arg); void *arg; }; @@ -1591,7 +3262,7 @@ void (*cb)(int revents, void *arg) = once->cb; void *arg = once->arg; - ev_io_stop (EV_A_ &once->io); + ev_io_stop (EV_A_ &once->io); ev_timer_stop (EV_A_ &once->to); ev_free (once); @@ -1599,15 +3270,19 @@ } static void -once_cb_io (EV_P_ struct ev_io *w, int revents) +once_cb_io (EV_P_ ev_io *w, int revents) { - once_cb (EV_A_ (struct ev_once *)(((char *)w) - offsetof (struct ev_once, io)), revents); + struct ev_once *once = (struct ev_once *)(((char *)w) - offsetof (struct ev_once, io)); + + once_cb (EV_A_ once, revents | ev_clear_pending (EV_A_ &once->to)); } static void -once_cb_to (EV_P_ struct ev_timer *w, int revents) +once_cb_to (EV_P_ ev_timer *w, int revents) { - once_cb (EV_A_ (struct ev_once *)(((char *)w) - offsetof (struct ev_once, to)), revents); + struct ev_once *once = (struct ev_once *)(((char *)w) - offsetof (struct ev_once, to)); + + once_cb (EV_A_ once, revents | ev_clear_pending (EV_A_ &once->io)); } void @@ -1615,28 +3290,141 @@ { struct ev_once *once = (struct ev_once *)ev_malloc (sizeof (struct ev_once)); - if (!once) - cb (EV_ERROR | EV_READ | EV_WRITE | EV_TIMEOUT, arg); - else + if (expect_false (!once)) + { + cb (EV_ERROR | EV_READ | EV_WRITE | EV_TIMEOUT, arg); + return; + } + + once->cb = cb; + once->arg = arg; + + ev_init (&once->io, once_cb_io); + if (fd >= 0) + { + ev_io_set (&once->io, fd, events); + ev_io_start (EV_A_ &once->io); + } + + ev_init (&once->to, once_cb_to); + if (timeout >= 0.) { - once->cb = cb; - once->arg = arg; + ev_timer_set (&once->to, timeout, 0.); + ev_timer_start (EV_A_ &once->to); + } +} + +/*****************************************************************************/ - ev_init (&once->io, once_cb_io); - if (fd >= 0) +#if EV_WALK_ENABLE +void +ev_walk (EV_P_ int types, void (*cb)(EV_P_ int type, void *w)) +{ + int i, j; + ev_watcher_list *wl, *wn; + + if (types & (EV_IO | EV_EMBED)) + for (i = 0; i < anfdmax; ++i) + for (wl = anfds [i].head; wl; ) { - ev_io_set (&once->io, fd, events); - ev_io_start (EV_A_ &once->io); + wn = wl->next; + +#if EV_EMBED_ENABLE + if (ev_cb ((ev_io *)wl) == embed_io_cb) + { + if (types & EV_EMBED) + cb (EV_A_ EV_EMBED, ((char *)wl) - offsetof (struct ev_embed, io)); + } + else +#endif +#if EV_USE_INOTIFY + if (ev_cb ((ev_io *)wl) == infy_cb) + ; + else +#endif + if ((ev_io *)wl != &pipe_w) + if (types & EV_IO) + cb (EV_A_ EV_IO, wl); + + wl = wn; } - ev_init (&once->to, once_cb_to); - if (timeout >= 0.) + if (types & (EV_TIMER | EV_STAT)) + for (i = timercnt + HEAP0; i-- > HEAP0; ) +#if EV_STAT_ENABLE + /*TODO: timer is not always active*/ + if (ev_cb ((ev_timer *)ANHE_w (timers [i])) == stat_timer_cb) { - ev_timer_set (&once->to, timeout, 0.); - ev_timer_start (EV_A_ &once->to); + if (types & EV_STAT) + cb (EV_A_ EV_STAT, ((char *)ANHE_w (timers [i])) - offsetof (struct ev_stat, timer)); } - } + else +#endif + if (types & EV_TIMER) + cb (EV_A_ EV_TIMER, ANHE_w (timers [i])); + +#if EV_PERIODIC_ENABLE + if (types & EV_PERIODIC) + for (i = periodiccnt + HEAP0; i-- > HEAP0; ) + cb (EV_A_ EV_PERIODIC, ANHE_w (periodics [i])); +#endif + +#if EV_IDLE_ENABLE + if (types & EV_IDLE) + for (j = NUMPRI; i--; ) + for (i = idlecnt [j]; i--; ) + cb (EV_A_ EV_IDLE, idles [j][i]); +#endif + +#if EV_FORK_ENABLE + if (types & EV_FORK) + for (i = forkcnt; i--; ) + if (ev_cb (forks [i]) != embed_fork_cb) + cb (EV_A_ EV_FORK, forks [i]); +#endif + +#if EV_ASYNC_ENABLE + if (types & EV_ASYNC) + for (i = asynccnt; i--; ) + cb (EV_A_ EV_ASYNC, asyncs [i]); +#endif + + if (types & EV_PREPARE) + for (i = preparecnt; i--; ) +#if EV_EMBED_ENABLE + if (ev_cb (prepares [i]) != embed_prepare_cb) +#endif + cb (EV_A_ EV_PREPARE, prepares [i]); + + if (types & EV_CHECK) + for (i = checkcnt; i--; ) + cb (EV_A_ EV_CHECK, checks [i]); + + if (types & EV_SIGNAL) + for (i = 0; i < signalmax; ++i) + for (wl = signals [i].head; wl; ) + { + wn = wl->next; + cb (EV_A_ EV_SIGNAL, wl); + wl = wn; + } + + if (types & EV_CHILD) + for (i = EV_PID_HASHSIZE; i--; ) + for (wl = childs [i]; wl; ) + { + wn = wl->next; + cb (EV_A_ EV_CHILD, wl); + wl = wn; + } +/* EV_STAT 0x00001000 /* stat data changed */ +/* EV_EMBED 0x00010000 /* embedded event loop needs sweep */ } +#endif + +#if EV_MULTIPLICITY + #include "ev_wrap.h" +#endif #ifdef __cplusplus }