1 |
/* |
2 |
* libev event processing core, watcher management |
3 |
* |
4 |
* Copyright (c) 2007 Marc Alexander Lehmann <libev@schmorp.de> |
5 |
* All rights reserved. |
6 |
* |
7 |
* Redistribution and use in source and binary forms, with or without |
8 |
* modification, are permitted provided that the following conditions are |
9 |
* met: |
10 |
* |
11 |
* * Redistributions of source code must retain the above copyright |
12 |
* notice, this list of conditions and the following disclaimer. |
13 |
* |
14 |
* * Redistributions in binary form must reproduce the above |
15 |
* copyright notice, this list of conditions and the following |
16 |
* disclaimer in the documentation and/or other materials provided |
17 |
* with the distribution. |
18 |
* |
19 |
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
20 |
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
21 |
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
22 |
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
23 |
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
24 |
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
25 |
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
26 |
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
27 |
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
28 |
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
29 |
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
30 |
*/ |
31 |
#if EV_USE_CONFIG_H |
32 |
# include "config.h" |
33 |
#endif |
34 |
|
35 |
#include <math.h> |
36 |
#include <stdlib.h> |
37 |
#include <unistd.h> |
38 |
#include <fcntl.h> |
39 |
#include <signal.h> |
40 |
#include <stddef.h> |
41 |
|
42 |
#include <stdio.h> |
43 |
|
44 |
#include <assert.h> |
45 |
#include <errno.h> |
46 |
#include <sys/types.h> |
47 |
#include <sys/wait.h> |
48 |
#include <sys/time.h> |
49 |
#include <time.h> |
50 |
|
51 |
/**/ |
52 |
|
53 |
#ifndef EV_USE_MONOTONIC |
54 |
# define EV_USE_MONOTONIC 1 |
55 |
#endif |
56 |
|
57 |
#ifndef EV_USE_SELECT |
58 |
# define EV_USE_SELECT 1 |
59 |
#endif |
60 |
|
61 |
#ifndef EV_USE_POLL |
62 |
# define EV_USE_POLL 0 /* poll is usually slower than select, and not as well tested */ |
63 |
#endif |
64 |
|
65 |
#ifndef EV_USE_EPOLL |
66 |
# define EV_USE_EPOLL 0 |
67 |
#endif |
68 |
|
69 |
#ifndef EV_USE_REALTIME |
70 |
# define EV_USE_REALTIME 1 |
71 |
#endif |
72 |
|
73 |
/**/ |
74 |
|
75 |
#ifndef CLOCK_MONOTONIC |
76 |
# undef EV_USE_MONOTONIC |
77 |
# define EV_USE_MONOTONIC 0 |
78 |
#endif |
79 |
|
80 |
#ifndef CLOCK_REALTIME |
81 |
# undef EV_USE_REALTIME |
82 |
# define EV_USE_REALTIME 0 |
83 |
#endif |
84 |
|
85 |
/**/ |
86 |
|
87 |
#define MIN_TIMEJUMP 1. /* minimum timejump that gets detected (if monotonic clock available) */ |
88 |
#define MAX_BLOCKTIME 59.731 /* never wait longer than this time (to detect time jumps) */ |
89 |
#define PID_HASHSIZE 16 /* size of pid hash table, must be power of two */ |
90 |
/*#define CLEANUP_INTERVAL 300. /* how often to try to free memory and re-check fds */ |
91 |
|
92 |
#include "ev.h" |
93 |
|
94 |
#if __GNUC__ >= 3 |
95 |
# define expect(expr,value) __builtin_expect ((expr),(value)) |
96 |
# define inline inline |
97 |
#else |
98 |
# define expect(expr,value) (expr) |
99 |
# define inline static |
100 |
#endif |
101 |
|
102 |
#define expect_false(expr) expect ((expr) != 0, 0) |
103 |
#define expect_true(expr) expect ((expr) != 0, 1) |
104 |
|
105 |
typedef struct ev_watcher *W; |
106 |
typedef struct ev_watcher_list *WL; |
107 |
typedef struct ev_watcher_time *WT; |
108 |
|
109 |
static ev_tstamp now_floor, now, diff; /* monotonic clock */ |
110 |
ev_tstamp ev_now; |
111 |
int ev_method; |
112 |
|
113 |
static int have_monotonic; /* runtime */ |
114 |
|
115 |
static ev_tstamp method_fudge; /* stupid epoll-returns-early bug */ |
116 |
static void (*method_modify)(int fd, int oev, int nev); |
117 |
static void (*method_poll)(ev_tstamp timeout); |
118 |
|
119 |
/*****************************************************************************/ |
120 |
|
121 |
ev_tstamp |
122 |
ev_time (void) |
123 |
{ |
124 |
#if EV_USE_REALTIME |
125 |
struct timespec ts; |
126 |
clock_gettime (CLOCK_REALTIME, &ts); |
127 |
return ts.tv_sec + ts.tv_nsec * 1e-9; |
128 |
#else |
129 |
struct timeval tv; |
130 |
gettimeofday (&tv, 0); |
131 |
return tv.tv_sec + tv.tv_usec * 1e-6; |
132 |
#endif |
133 |
} |
134 |
|
135 |
static ev_tstamp |
136 |
get_clock (void) |
137 |
{ |
138 |
#if EV_USE_MONOTONIC |
139 |
if (expect_true (have_monotonic)) |
140 |
{ |
141 |
struct timespec ts; |
142 |
clock_gettime (CLOCK_MONOTONIC, &ts); |
143 |
return ts.tv_sec + ts.tv_nsec * 1e-9; |
144 |
} |
145 |
#endif |
146 |
|
147 |
return ev_time (); |
148 |
} |
149 |
|
150 |
#define array_roundsize(base,n) ((n) | 4 & ~3) |
151 |
|
152 |
#define array_needsize(base,cur,cnt,init) \ |
153 |
if (expect_false ((cnt) > cur)) \ |
154 |
{ \ |
155 |
int newcnt = cur; \ |
156 |
do \ |
157 |
{ \ |
158 |
newcnt = array_roundsize (base, newcnt << 1); \ |
159 |
} \ |
160 |
while ((cnt) > newcnt); \ |
161 |
\ |
162 |
base = realloc (base, sizeof (*base) * (newcnt)); \ |
163 |
init (base + cur, newcnt - cur); \ |
164 |
cur = newcnt; \ |
165 |
} |
166 |
|
167 |
/*****************************************************************************/ |
168 |
|
169 |
typedef struct |
170 |
{ |
171 |
struct ev_io *head; |
172 |
unsigned char events; |
173 |
unsigned char reify; |
174 |
} ANFD; |
175 |
|
176 |
static ANFD *anfds; |
177 |
static int anfdmax; |
178 |
|
179 |
static void |
180 |
anfds_init (ANFD *base, int count) |
181 |
{ |
182 |
while (count--) |
183 |
{ |
184 |
base->head = 0; |
185 |
base->events = EV_NONE; |
186 |
base->reify = 0; |
187 |
|
188 |
++base; |
189 |
} |
190 |
} |
191 |
|
192 |
typedef struct |
193 |
{ |
194 |
W w; |
195 |
int events; |
196 |
} ANPENDING; |
197 |
|
198 |
static ANPENDING *pendings; |
199 |
static int pendingmax, pendingcnt; |
200 |
|
201 |
static void |
202 |
event (W w, int events) |
203 |
{ |
204 |
if (w->pending) |
205 |
{ |
206 |
pendings [w->pending - 1].events |= events; |
207 |
return; |
208 |
} |
209 |
|
210 |
w->pending = ++pendingcnt; |
211 |
array_needsize (pendings, pendingmax, pendingcnt, ); |
212 |
pendings [pendingcnt - 1].w = w; |
213 |
pendings [pendingcnt - 1].events = events; |
214 |
} |
215 |
|
216 |
static void |
217 |
queue_events (W *events, int eventcnt, int type) |
218 |
{ |
219 |
int i; |
220 |
|
221 |
for (i = 0; i < eventcnt; ++i) |
222 |
event (events [i], type); |
223 |
} |
224 |
|
225 |
static void |
226 |
fd_event (int fd, int events) |
227 |
{ |
228 |
ANFD *anfd = anfds + fd; |
229 |
struct ev_io *w; |
230 |
|
231 |
for (w = anfd->head; w; w = w->next) |
232 |
{ |
233 |
int ev = w->events & events; |
234 |
|
235 |
if (ev) |
236 |
event ((W)w, ev); |
237 |
} |
238 |
} |
239 |
|
240 |
/*****************************************************************************/ |
241 |
|
242 |
static int *fdchanges; |
243 |
static int fdchangemax, fdchangecnt; |
244 |
|
245 |
static void |
246 |
fd_reify (void) |
247 |
{ |
248 |
int i; |
249 |
|
250 |
for (i = 0; i < fdchangecnt; ++i) |
251 |
{ |
252 |
int fd = fdchanges [i]; |
253 |
ANFD *anfd = anfds + fd; |
254 |
struct ev_io *w; |
255 |
|
256 |
int events = 0; |
257 |
|
258 |
for (w = anfd->head; w; w = w->next) |
259 |
events |= w->events; |
260 |
|
261 |
anfd->reify = 0; |
262 |
|
263 |
if (anfd->events != events) |
264 |
{ |
265 |
method_modify (fd, anfd->events, events); |
266 |
anfd->events = events; |
267 |
} |
268 |
} |
269 |
|
270 |
fdchangecnt = 0; |
271 |
} |
272 |
|
273 |
static void |
274 |
fd_change (int fd) |
275 |
{ |
276 |
if (anfds [fd].reify || fdchangecnt < 0) |
277 |
return; |
278 |
|
279 |
anfds [fd].reify = 1; |
280 |
|
281 |
++fdchangecnt; |
282 |
array_needsize (fdchanges, fdchangemax, fdchangecnt, ); |
283 |
fdchanges [fdchangecnt - 1] = fd; |
284 |
} |
285 |
|
286 |
static void |
287 |
fd_kill (int fd) |
288 |
{ |
289 |
struct ev_io *w; |
290 |
|
291 |
printf ("killing fd %d\n", fd);//D |
292 |
while ((w = anfds [fd].head)) |
293 |
{ |
294 |
ev_io_stop (w); |
295 |
event ((W)w, EV_ERROR | EV_READ | EV_WRITE); |
296 |
} |
297 |
} |
298 |
|
299 |
/* called on EBADF to verify fds */ |
300 |
static void |
301 |
fd_ebadf (void) |
302 |
{ |
303 |
int fd; |
304 |
|
305 |
for (fd = 0; fd < anfdmax; ++fd) |
306 |
if (anfds [fd].events) |
307 |
if (fcntl (fd, F_GETFD) == -1 && errno == EBADF) |
308 |
fd_kill (fd); |
309 |
} |
310 |
|
311 |
/* called on ENOMEM in select/poll to kill some fds and retry */ |
312 |
static void |
313 |
fd_enomem (void) |
314 |
{ |
315 |
int fd = anfdmax; |
316 |
|
317 |
while (fd--) |
318 |
if (anfds [fd].events) |
319 |
{ |
320 |
close (fd); |
321 |
fd_kill (fd); |
322 |
return; |
323 |
} |
324 |
} |
325 |
|
326 |
/*****************************************************************************/ |
327 |
|
328 |
static struct ev_timer **timers; |
329 |
static int timermax, timercnt; |
330 |
|
331 |
static struct ev_periodic **periodics; |
332 |
static int periodicmax, periodiccnt; |
333 |
|
334 |
static void |
335 |
upheap (WT *timers, int k) |
336 |
{ |
337 |
WT w = timers [k]; |
338 |
|
339 |
while (k && timers [k >> 1]->at > w->at) |
340 |
{ |
341 |
timers [k] = timers [k >> 1]; |
342 |
timers [k]->active = k + 1; |
343 |
k >>= 1; |
344 |
} |
345 |
|
346 |
timers [k] = w; |
347 |
timers [k]->active = k + 1; |
348 |
|
349 |
} |
350 |
|
351 |
static void |
352 |
downheap (WT *timers, int N, int k) |
353 |
{ |
354 |
WT w = timers [k]; |
355 |
|
356 |
while (k < (N >> 1)) |
357 |
{ |
358 |
int j = k << 1; |
359 |
|
360 |
if (j + 1 < N && timers [j]->at > timers [j + 1]->at) |
361 |
++j; |
362 |
|
363 |
if (w->at <= timers [j]->at) |
364 |
break; |
365 |
|
366 |
timers [k] = timers [j]; |
367 |
timers [k]->active = k + 1; |
368 |
k = j; |
369 |
} |
370 |
|
371 |
timers [k] = w; |
372 |
timers [k]->active = k + 1; |
373 |
} |
374 |
|
375 |
/*****************************************************************************/ |
376 |
|
377 |
typedef struct |
378 |
{ |
379 |
struct ev_signal *head; |
380 |
sig_atomic_t volatile gotsig; |
381 |
} ANSIG; |
382 |
|
383 |
static ANSIG *signals; |
384 |
static int signalmax; |
385 |
|
386 |
static int sigpipe [2]; |
387 |
static sig_atomic_t volatile gotsig; |
388 |
static struct ev_io sigev; |
389 |
|
390 |
static void |
391 |
signals_init (ANSIG *base, int count) |
392 |
{ |
393 |
while (count--) |
394 |
{ |
395 |
base->head = 0; |
396 |
base->gotsig = 0; |
397 |
|
398 |
++base; |
399 |
} |
400 |
} |
401 |
|
402 |
static void |
403 |
sighandler (int signum) |
404 |
{ |
405 |
signals [signum - 1].gotsig = 1; |
406 |
|
407 |
if (!gotsig) |
408 |
{ |
409 |
gotsig = 1; |
410 |
write (sigpipe [1], &signum, 1); |
411 |
} |
412 |
} |
413 |
|
414 |
static void |
415 |
sigcb (struct ev_io *iow, int revents) |
416 |
{ |
417 |
struct ev_signal *w; |
418 |
int signum; |
419 |
|
420 |
read (sigpipe [0], &revents, 1); |
421 |
gotsig = 0; |
422 |
|
423 |
for (signum = signalmax; signum--; ) |
424 |
if (signals [signum].gotsig) |
425 |
{ |
426 |
signals [signum].gotsig = 0; |
427 |
|
428 |
for (w = signals [signum].head; w; w = w->next) |
429 |
event ((W)w, EV_SIGNAL); |
430 |
} |
431 |
} |
432 |
|
433 |
static void |
434 |
siginit (void) |
435 |
{ |
436 |
fcntl (sigpipe [0], F_SETFD, FD_CLOEXEC); |
437 |
fcntl (sigpipe [1], F_SETFD, FD_CLOEXEC); |
438 |
|
439 |
/* rather than sort out wether we really need nb, set it */ |
440 |
fcntl (sigpipe [0], F_SETFL, O_NONBLOCK); |
441 |
fcntl (sigpipe [1], F_SETFL, O_NONBLOCK); |
442 |
|
443 |
ev_io_set (&sigev, sigpipe [0], EV_READ); |
444 |
ev_io_start (&sigev); |
445 |
} |
446 |
|
447 |
/*****************************************************************************/ |
448 |
|
449 |
static struct ev_idle **idles; |
450 |
static int idlemax, idlecnt; |
451 |
|
452 |
static struct ev_prepare **prepares; |
453 |
static int preparemax, preparecnt; |
454 |
|
455 |
static struct ev_check **checks; |
456 |
static int checkmax, checkcnt; |
457 |
|
458 |
/*****************************************************************************/ |
459 |
|
460 |
static struct ev_child *childs [PID_HASHSIZE]; |
461 |
static struct ev_signal childev; |
462 |
|
463 |
#ifndef WCONTINUED |
464 |
# define WCONTINUED 0 |
465 |
#endif |
466 |
|
467 |
static void |
468 |
childcb (struct ev_signal *sw, int revents) |
469 |
{ |
470 |
struct ev_child *w; |
471 |
int pid, status; |
472 |
|
473 |
while ((pid = waitpid (-1, &status, WNOHANG | WUNTRACED | WCONTINUED)) != -1) |
474 |
for (w = childs [pid & (PID_HASHSIZE - 1)]; w; w = w->next) |
475 |
if (w->pid == pid || !w->pid) |
476 |
{ |
477 |
w->status = status; |
478 |
event ((W)w, EV_CHILD); |
479 |
} |
480 |
} |
481 |
|
482 |
/*****************************************************************************/ |
483 |
|
484 |
#if EV_USE_EPOLL |
485 |
# include "ev_epoll.c" |
486 |
#endif |
487 |
#if EV_USE_POLL |
488 |
# include "ev_poll.c" |
489 |
#endif |
490 |
#if EV_USE_SELECT |
491 |
# include "ev_select.c" |
492 |
#endif |
493 |
|
494 |
int |
495 |
ev_version_major (void) |
496 |
{ |
497 |
return EV_VERSION_MAJOR; |
498 |
} |
499 |
|
500 |
int |
501 |
ev_version_minor (void) |
502 |
{ |
503 |
return EV_VERSION_MINOR; |
504 |
} |
505 |
|
506 |
/* return true if we are running with elevated privileges and ignore env variables */ |
507 |
static int |
508 |
enable_secure () |
509 |
{ |
510 |
return getuid () != geteuid () |
511 |
|| getgid () != getegid (); |
512 |
} |
513 |
|
514 |
int ev_init (int methods) |
515 |
{ |
516 |
if (!ev_method) |
517 |
{ |
518 |
#if EV_USE_MONOTONIC |
519 |
{ |
520 |
struct timespec ts; |
521 |
if (!clock_gettime (CLOCK_MONOTONIC, &ts)) |
522 |
have_monotonic = 1; |
523 |
} |
524 |
#endif |
525 |
|
526 |
ev_now = ev_time (); |
527 |
now = get_clock (); |
528 |
now_floor = now; |
529 |
diff = ev_now - now; |
530 |
|
531 |
if (pipe (sigpipe)) |
532 |
return 0; |
533 |
|
534 |
if (methods == EVMETHOD_AUTO) |
535 |
if (!enable_secure () && getenv ("LIBEV_METHODS")) |
536 |
methods = atoi (getenv ("LIBEV_METHODS")); |
537 |
else |
538 |
methods = EVMETHOD_ANY; |
539 |
|
540 |
ev_method = 0; |
541 |
#if EV_USE_EPOLL |
542 |
if (!ev_method && (methods & EVMETHOD_EPOLL )) epoll_init (methods); |
543 |
#endif |
544 |
#if EV_USE_POLL |
545 |
if (!ev_method && (methods & EVMETHOD_POLL )) poll_init (methods); |
546 |
#endif |
547 |
#if EV_USE_SELECT |
548 |
if (!ev_method && (methods & EVMETHOD_SELECT)) select_init (methods); |
549 |
#endif |
550 |
|
551 |
if (ev_method) |
552 |
{ |
553 |
ev_watcher_init (&sigev, sigcb); |
554 |
siginit (); |
555 |
|
556 |
ev_signal_init (&childev, childcb, SIGCHLD); |
557 |
ev_signal_start (&childev); |
558 |
} |
559 |
} |
560 |
|
561 |
return ev_method; |
562 |
} |
563 |
|
564 |
/*****************************************************************************/ |
565 |
|
566 |
void |
567 |
ev_fork_prepare (void) |
568 |
{ |
569 |
/* nop */ |
570 |
} |
571 |
|
572 |
void |
573 |
ev_fork_parent (void) |
574 |
{ |
575 |
/* nop */ |
576 |
} |
577 |
|
578 |
void |
579 |
ev_fork_child (void) |
580 |
{ |
581 |
#if EV_USE_EPOLL |
582 |
if (ev_method == EVMETHOD_EPOLL) |
583 |
epoll_postfork_child (); |
584 |
#endif |
585 |
|
586 |
ev_io_stop (&sigev); |
587 |
close (sigpipe [0]); |
588 |
close (sigpipe [1]); |
589 |
pipe (sigpipe); |
590 |
siginit (); |
591 |
} |
592 |
|
593 |
/*****************************************************************************/ |
594 |
|
595 |
static void |
596 |
call_pending (void) |
597 |
{ |
598 |
while (pendingcnt) |
599 |
{ |
600 |
ANPENDING *p = pendings + --pendingcnt; |
601 |
|
602 |
if (p->w) |
603 |
{ |
604 |
p->w->pending = 0; |
605 |
p->w->cb (p->w, p->events); |
606 |
} |
607 |
} |
608 |
} |
609 |
|
610 |
static void |
611 |
timers_reify (void) |
612 |
{ |
613 |
while (timercnt && timers [0]->at <= now) |
614 |
{ |
615 |
struct ev_timer *w = timers [0]; |
616 |
|
617 |
/* first reschedule or stop timer */ |
618 |
if (w->repeat) |
619 |
{ |
620 |
assert (("negative ev_timer repeat value found while processing timers", w->repeat > 0.)); |
621 |
w->at = now + w->repeat; |
622 |
downheap ((WT *)timers, timercnt, 0); |
623 |
} |
624 |
else |
625 |
ev_timer_stop (w); /* nonrepeating: stop timer */ |
626 |
|
627 |
event ((W)w, EV_TIMEOUT); |
628 |
} |
629 |
} |
630 |
|
631 |
static void |
632 |
periodics_reify (void) |
633 |
{ |
634 |
while (periodiccnt && periodics [0]->at <= ev_now) |
635 |
{ |
636 |
struct ev_periodic *w = periodics [0]; |
637 |
|
638 |
/* first reschedule or stop timer */ |
639 |
if (w->interval) |
640 |
{ |
641 |
w->at += floor ((ev_now - w->at) / w->interval + 1.) * w->interval; |
642 |
assert (("ev_periodic timeout in the past detected while processing timers, negative interval?", w->at > ev_now)); |
643 |
downheap ((WT *)periodics, periodiccnt, 0); |
644 |
} |
645 |
else |
646 |
ev_periodic_stop (w); /* nonrepeating: stop timer */ |
647 |
|
648 |
event ((W)w, EV_PERIODIC); |
649 |
} |
650 |
} |
651 |
|
652 |
static void |
653 |
periodics_reschedule (ev_tstamp diff) |
654 |
{ |
655 |
int i; |
656 |
|
657 |
/* adjust periodics after time jump */ |
658 |
for (i = 0; i < periodiccnt; ++i) |
659 |
{ |
660 |
struct ev_periodic *w = periodics [i]; |
661 |
|
662 |
if (w->interval) |
663 |
{ |
664 |
ev_tstamp diff = ceil ((ev_now - w->at) / w->interval) * w->interval; |
665 |
|
666 |
if (fabs (diff) >= 1e-4) |
667 |
{ |
668 |
ev_periodic_stop (w); |
669 |
ev_periodic_start (w); |
670 |
|
671 |
i = 0; /* restart loop, inefficient, but time jumps should be rare */ |
672 |
} |
673 |
} |
674 |
} |
675 |
} |
676 |
|
677 |
static int |
678 |
time_update_monotonic (void) |
679 |
{ |
680 |
now = get_clock (); |
681 |
|
682 |
if (expect_true (now - now_floor < MIN_TIMEJUMP * .5)) |
683 |
{ |
684 |
ev_now = now + diff; |
685 |
return 0; |
686 |
} |
687 |
else |
688 |
{ |
689 |
now_floor = now; |
690 |
ev_now = ev_time (); |
691 |
return 1; |
692 |
} |
693 |
} |
694 |
|
695 |
static void |
696 |
time_update (void) |
697 |
{ |
698 |
int i; |
699 |
|
700 |
#if EV_USE_MONOTONIC |
701 |
if (expect_true (have_monotonic)) |
702 |
{ |
703 |
if (time_update_monotonic ()) |
704 |
{ |
705 |
ev_tstamp odiff = diff; |
706 |
|
707 |
for (i = 4; --i; ) /* loop a few times, before making important decisions */ |
708 |
{ |
709 |
diff = ev_now - now; |
710 |
|
711 |
if (fabs (odiff - diff) < MIN_TIMEJUMP) |
712 |
return; /* all is well */ |
713 |
|
714 |
ev_now = ev_time (); |
715 |
now = get_clock (); |
716 |
now_floor = now; |
717 |
} |
718 |
|
719 |
periodics_reschedule (diff - odiff); |
720 |
/* no timer adjustment, as the monotonic clock doesn't jump */ |
721 |
} |
722 |
} |
723 |
else |
724 |
#endif |
725 |
{ |
726 |
ev_now = ev_time (); |
727 |
|
728 |
if (expect_false (now > ev_now || now < ev_now - MAX_BLOCKTIME - MIN_TIMEJUMP)) |
729 |
{ |
730 |
periodics_reschedule (ev_now - now); |
731 |
|
732 |
/* adjust timers. this is easy, as the offset is the same for all */ |
733 |
for (i = 0; i < timercnt; ++i) |
734 |
timers [i]->at += diff; |
735 |
} |
736 |
|
737 |
now = ev_now; |
738 |
} |
739 |
} |
740 |
|
741 |
int ev_loop_done; |
742 |
|
743 |
void ev_loop (int flags) |
744 |
{ |
745 |
double block; |
746 |
ev_loop_done = flags & (EVLOOP_ONESHOT | EVLOOP_NONBLOCK) ? 1 : 0; |
747 |
|
748 |
do |
749 |
{ |
750 |
/* queue check watchers (and execute them) */ |
751 |
if (expect_false (preparecnt)) |
752 |
{ |
753 |
queue_events ((W *)prepares, preparecnt, EV_PREPARE); |
754 |
call_pending (); |
755 |
} |
756 |
|
757 |
/* update fd-related kernel structures */ |
758 |
fd_reify (); |
759 |
|
760 |
/* calculate blocking time */ |
761 |
|
762 |
/* we only need this for !monotonic clockor timers, but as we basically |
763 |
always have timers, we just calculate it always */ |
764 |
#if EV_USE_MONOTONIC |
765 |
if (expect_true (have_monotonic)) |
766 |
time_update_monotonic (); |
767 |
else |
768 |
#endif |
769 |
{ |
770 |
ev_now = ev_time (); |
771 |
now = ev_now; |
772 |
} |
773 |
|
774 |
if (flags & EVLOOP_NONBLOCK || idlecnt) |
775 |
block = 0.; |
776 |
else |
777 |
{ |
778 |
block = MAX_BLOCKTIME; |
779 |
|
780 |
if (timercnt) |
781 |
{ |
782 |
ev_tstamp to = timers [0]->at - now + method_fudge; |
783 |
if (block > to) block = to; |
784 |
} |
785 |
|
786 |
if (periodiccnt) |
787 |
{ |
788 |
ev_tstamp to = periodics [0]->at - ev_now + method_fudge; |
789 |
if (block > to) block = to; |
790 |
} |
791 |
|
792 |
if (block < 0.) block = 0.; |
793 |
} |
794 |
|
795 |
method_poll (block); |
796 |
|
797 |
/* update ev_now, do magic */ |
798 |
time_update (); |
799 |
|
800 |
/* queue pending timers and reschedule them */ |
801 |
timers_reify (); /* relative timers called last */ |
802 |
periodics_reify (); /* absolute timers called first */ |
803 |
|
804 |
/* queue idle watchers unless io or timers are pending */ |
805 |
if (!pendingcnt) |
806 |
queue_events ((W *)idles, idlecnt, EV_IDLE); |
807 |
|
808 |
/* queue check watchers, to be executed first */ |
809 |
if (checkcnt) |
810 |
queue_events ((W *)checks, checkcnt, EV_CHECK); |
811 |
|
812 |
call_pending (); |
813 |
} |
814 |
while (!ev_loop_done); |
815 |
|
816 |
if (ev_loop_done != 2) |
817 |
ev_loop_done = 0; |
818 |
} |
819 |
|
820 |
/*****************************************************************************/ |
821 |
|
822 |
static void |
823 |
wlist_add (WL *head, WL elem) |
824 |
{ |
825 |
elem->next = *head; |
826 |
*head = elem; |
827 |
} |
828 |
|
829 |
static void |
830 |
wlist_del (WL *head, WL elem) |
831 |
{ |
832 |
while (*head) |
833 |
{ |
834 |
if (*head == elem) |
835 |
{ |
836 |
*head = elem->next; |
837 |
return; |
838 |
} |
839 |
|
840 |
head = &(*head)->next; |
841 |
} |
842 |
} |
843 |
|
844 |
static void |
845 |
ev_clear_pending (W w) |
846 |
{ |
847 |
if (w->pending) |
848 |
{ |
849 |
pendings [w->pending - 1].w = 0; |
850 |
w->pending = 0; |
851 |
} |
852 |
} |
853 |
|
854 |
static void |
855 |
ev_start (W w, int active) |
856 |
{ |
857 |
w->active = active; |
858 |
} |
859 |
|
860 |
static void |
861 |
ev_stop (W w) |
862 |
{ |
863 |
w->active = 0; |
864 |
} |
865 |
|
866 |
/*****************************************************************************/ |
867 |
|
868 |
void |
869 |
ev_io_start (struct ev_io *w) |
870 |
{ |
871 |
int fd = w->fd; |
872 |
|
873 |
if (ev_is_active (w)) |
874 |
return; |
875 |
|
876 |
assert (("ev_io_start called with negative fd", fd >= 0)); |
877 |
|
878 |
ev_start ((W)w, 1); |
879 |
array_needsize (anfds, anfdmax, fd + 1, anfds_init); |
880 |
wlist_add ((WL *)&anfds[fd].head, (WL)w); |
881 |
|
882 |
fd_change (fd); |
883 |
} |
884 |
|
885 |
void |
886 |
ev_io_stop (struct ev_io *w) |
887 |
{ |
888 |
ev_clear_pending ((W)w); |
889 |
if (!ev_is_active (w)) |
890 |
return; |
891 |
|
892 |
wlist_del ((WL *)&anfds[w->fd].head, (WL)w); |
893 |
ev_stop ((W)w); |
894 |
|
895 |
fd_change (w->fd); |
896 |
} |
897 |
|
898 |
void |
899 |
ev_timer_start (struct ev_timer *w) |
900 |
{ |
901 |
if (ev_is_active (w)) |
902 |
return; |
903 |
|
904 |
w->at += now; |
905 |
|
906 |
assert (("ev_timer_start called with negative timer repeat value", w->repeat >= 0.)); |
907 |
|
908 |
ev_start ((W)w, ++timercnt); |
909 |
array_needsize (timers, timermax, timercnt, ); |
910 |
timers [timercnt - 1] = w; |
911 |
upheap ((WT *)timers, timercnt - 1); |
912 |
} |
913 |
|
914 |
void |
915 |
ev_timer_stop (struct ev_timer *w) |
916 |
{ |
917 |
ev_clear_pending ((W)w); |
918 |
if (!ev_is_active (w)) |
919 |
return; |
920 |
|
921 |
if (w->active < timercnt--) |
922 |
{ |
923 |
timers [w->active - 1] = timers [timercnt]; |
924 |
downheap ((WT *)timers, timercnt, w->active - 1); |
925 |
} |
926 |
|
927 |
w->at = w->repeat; |
928 |
|
929 |
ev_stop ((W)w); |
930 |
} |
931 |
|
932 |
void |
933 |
ev_timer_again (struct ev_timer *w) |
934 |
{ |
935 |
if (ev_is_active (w)) |
936 |
{ |
937 |
if (w->repeat) |
938 |
{ |
939 |
w->at = now + w->repeat; |
940 |
downheap ((WT *)timers, timercnt, w->active - 1); |
941 |
} |
942 |
else |
943 |
ev_timer_stop (w); |
944 |
} |
945 |
else if (w->repeat) |
946 |
ev_timer_start (w); |
947 |
} |
948 |
|
949 |
void |
950 |
ev_periodic_start (struct ev_periodic *w) |
951 |
{ |
952 |
if (ev_is_active (w)) |
953 |
return; |
954 |
|
955 |
assert (("ev_periodic_start called with negative interval value", w->interval >= 0.)); |
956 |
|
957 |
/* this formula differs from the one in periodic_reify because we do not always round up */ |
958 |
if (w->interval) |
959 |
w->at += ceil ((ev_now - w->at) / w->interval) * w->interval; |
960 |
|
961 |
ev_start ((W)w, ++periodiccnt); |
962 |
array_needsize (periodics, periodicmax, periodiccnt, ); |
963 |
periodics [periodiccnt - 1] = w; |
964 |
upheap ((WT *)periodics, periodiccnt - 1); |
965 |
} |
966 |
|
967 |
void |
968 |
ev_periodic_stop (struct ev_periodic *w) |
969 |
{ |
970 |
ev_clear_pending ((W)w); |
971 |
if (!ev_is_active (w)) |
972 |
return; |
973 |
|
974 |
if (w->active < periodiccnt--) |
975 |
{ |
976 |
periodics [w->active - 1] = periodics [periodiccnt]; |
977 |
downheap ((WT *)periodics, periodiccnt, w->active - 1); |
978 |
} |
979 |
|
980 |
ev_stop ((W)w); |
981 |
} |
982 |
|
983 |
void |
984 |
ev_signal_start (struct ev_signal *w) |
985 |
{ |
986 |
if (ev_is_active (w)) |
987 |
return; |
988 |
|
989 |
assert (("ev_signal_start called with illegal signal number", w->signum > 0)); |
990 |
|
991 |
ev_start ((W)w, 1); |
992 |
array_needsize (signals, signalmax, w->signum, signals_init); |
993 |
wlist_add ((WL *)&signals [w->signum - 1].head, (WL)w); |
994 |
|
995 |
if (!w->next) |
996 |
{ |
997 |
struct sigaction sa; |
998 |
sa.sa_handler = sighandler; |
999 |
sigfillset (&sa.sa_mask); |
1000 |
sa.sa_flags = 0; |
1001 |
sigaction (w->signum, &sa, 0); |
1002 |
} |
1003 |
} |
1004 |
|
1005 |
void |
1006 |
ev_signal_stop (struct ev_signal *w) |
1007 |
{ |
1008 |
ev_clear_pending ((W)w); |
1009 |
if (!ev_is_active (w)) |
1010 |
return; |
1011 |
|
1012 |
wlist_del ((WL *)&signals [w->signum - 1].head, (WL)w); |
1013 |
ev_stop ((W)w); |
1014 |
|
1015 |
if (!signals [w->signum - 1].head) |
1016 |
signal (w->signum, SIG_DFL); |
1017 |
} |
1018 |
|
1019 |
void |
1020 |
ev_idle_start (struct ev_idle *w) |
1021 |
{ |
1022 |
if (ev_is_active (w)) |
1023 |
return; |
1024 |
|
1025 |
ev_start ((W)w, ++idlecnt); |
1026 |
array_needsize (idles, idlemax, idlecnt, ); |
1027 |
idles [idlecnt - 1] = w; |
1028 |
} |
1029 |
|
1030 |
void |
1031 |
ev_idle_stop (struct ev_idle *w) |
1032 |
{ |
1033 |
ev_clear_pending ((W)w); |
1034 |
if (ev_is_active (w)) |
1035 |
return; |
1036 |
|
1037 |
idles [w->active - 1] = idles [--idlecnt]; |
1038 |
ev_stop ((W)w); |
1039 |
} |
1040 |
|
1041 |
void |
1042 |
ev_prepare_start (struct ev_prepare *w) |
1043 |
{ |
1044 |
if (ev_is_active (w)) |
1045 |
return; |
1046 |
|
1047 |
ev_start ((W)w, ++preparecnt); |
1048 |
array_needsize (prepares, preparemax, preparecnt, ); |
1049 |
prepares [preparecnt - 1] = w; |
1050 |
} |
1051 |
|
1052 |
void |
1053 |
ev_prepare_stop (struct ev_prepare *w) |
1054 |
{ |
1055 |
ev_clear_pending ((W)w); |
1056 |
if (ev_is_active (w)) |
1057 |
return; |
1058 |
|
1059 |
prepares [w->active - 1] = prepares [--preparecnt]; |
1060 |
ev_stop ((W)w); |
1061 |
} |
1062 |
|
1063 |
void |
1064 |
ev_check_start (struct ev_check *w) |
1065 |
{ |
1066 |
if (ev_is_active (w)) |
1067 |
return; |
1068 |
|
1069 |
ev_start ((W)w, ++checkcnt); |
1070 |
array_needsize (checks, checkmax, checkcnt, ); |
1071 |
checks [checkcnt - 1] = w; |
1072 |
} |
1073 |
|
1074 |
void |
1075 |
ev_check_stop (struct ev_check *w) |
1076 |
{ |
1077 |
ev_clear_pending ((W)w); |
1078 |
if (ev_is_active (w)) |
1079 |
return; |
1080 |
|
1081 |
checks [w->active - 1] = checks [--checkcnt]; |
1082 |
ev_stop ((W)w); |
1083 |
} |
1084 |
|
1085 |
void |
1086 |
ev_child_start (struct ev_child *w) |
1087 |
{ |
1088 |
if (ev_is_active (w)) |
1089 |
return; |
1090 |
|
1091 |
ev_start ((W)w, 1); |
1092 |
wlist_add ((WL *)&childs [w->pid & (PID_HASHSIZE - 1)], (WL)w); |
1093 |
} |
1094 |
|
1095 |
void |
1096 |
ev_child_stop (struct ev_child *w) |
1097 |
{ |
1098 |
ev_clear_pending ((W)w); |
1099 |
if (ev_is_active (w)) |
1100 |
return; |
1101 |
|
1102 |
wlist_del ((WL *)&childs [w->pid & (PID_HASHSIZE - 1)], (WL)w); |
1103 |
ev_stop ((W)w); |
1104 |
} |
1105 |
|
1106 |
/*****************************************************************************/ |
1107 |
|
1108 |
struct ev_once |
1109 |
{ |
1110 |
struct ev_io io; |
1111 |
struct ev_timer to; |
1112 |
void (*cb)(int revents, void *arg); |
1113 |
void *arg; |
1114 |
}; |
1115 |
|
1116 |
static void |
1117 |
once_cb (struct ev_once *once, int revents) |
1118 |
{ |
1119 |
void (*cb)(int revents, void *arg) = once->cb; |
1120 |
void *arg = once->arg; |
1121 |
|
1122 |
ev_io_stop (&once->io); |
1123 |
ev_timer_stop (&once->to); |
1124 |
free (once); |
1125 |
|
1126 |
cb (revents, arg); |
1127 |
} |
1128 |
|
1129 |
static void |
1130 |
once_cb_io (struct ev_io *w, int revents) |
1131 |
{ |
1132 |
once_cb ((struct ev_once *)(((char *)w) - offsetof (struct ev_once, io)), revents); |
1133 |
} |
1134 |
|
1135 |
static void |
1136 |
once_cb_to (struct ev_timer *w, int revents) |
1137 |
{ |
1138 |
once_cb ((struct ev_once *)(((char *)w) - offsetof (struct ev_once, to)), revents); |
1139 |
} |
1140 |
|
1141 |
void |
1142 |
ev_once (int fd, int events, ev_tstamp timeout, void (*cb)(int revents, void *arg), void *arg) |
1143 |
{ |
1144 |
struct ev_once *once = malloc (sizeof (struct ev_once)); |
1145 |
|
1146 |
if (!once) |
1147 |
cb (EV_ERROR | EV_READ | EV_WRITE | EV_TIMEOUT, arg); |
1148 |
else |
1149 |
{ |
1150 |
once->cb = cb; |
1151 |
once->arg = arg; |
1152 |
|
1153 |
ev_watcher_init (&once->io, once_cb_io); |
1154 |
if (fd >= 0) |
1155 |
{ |
1156 |
ev_io_set (&once->io, fd, events); |
1157 |
ev_io_start (&once->io); |
1158 |
} |
1159 |
|
1160 |
ev_watcher_init (&once->to, once_cb_to); |
1161 |
if (timeout >= 0.) |
1162 |
{ |
1163 |
ev_timer_set (&once->to, timeout, 0.); |
1164 |
ev_timer_start (&once->to); |
1165 |
} |
1166 |
} |
1167 |
} |
1168 |
|
1169 |
/*****************************************************************************/ |
1170 |
|
1171 |
#if 0 |
1172 |
|
1173 |
struct ev_io wio; |
1174 |
|
1175 |
static void |
1176 |
sin_cb (struct ev_io *w, int revents) |
1177 |
{ |
1178 |
fprintf (stderr, "sin %d, revents %d\n", w->fd, revents); |
1179 |
} |
1180 |
|
1181 |
static void |
1182 |
ocb (struct ev_timer *w, int revents) |
1183 |
{ |
1184 |
//fprintf (stderr, "timer %f,%f (%x) (%f) d%p\n", w->at, w->repeat, revents, w->at - ev_time (), w->data); |
1185 |
ev_timer_stop (w); |
1186 |
ev_timer_start (w); |
1187 |
} |
1188 |
|
1189 |
static void |
1190 |
scb (struct ev_signal *w, int revents) |
1191 |
{ |
1192 |
fprintf (stderr, "signal %x,%d\n", revents, w->signum); |
1193 |
ev_io_stop (&wio); |
1194 |
ev_io_start (&wio); |
1195 |
} |
1196 |
|
1197 |
static void |
1198 |
gcb (struct ev_signal *w, int revents) |
1199 |
{ |
1200 |
fprintf (stderr, "generic %x\n", revents); |
1201 |
|
1202 |
} |
1203 |
|
1204 |
int main (void) |
1205 |
{ |
1206 |
ev_init (0); |
1207 |
|
1208 |
ev_io_init (&wio, sin_cb, 0, EV_READ); |
1209 |
ev_io_start (&wio); |
1210 |
|
1211 |
struct ev_timer t[10000]; |
1212 |
|
1213 |
#if 0 |
1214 |
int i; |
1215 |
for (i = 0; i < 10000; ++i) |
1216 |
{ |
1217 |
struct ev_timer *w = t + i; |
1218 |
ev_watcher_init (w, ocb, i); |
1219 |
ev_timer_init_abs (w, ocb, drand48 (), 0.99775533); |
1220 |
ev_timer_start (w); |
1221 |
if (drand48 () < 0.5) |
1222 |
ev_timer_stop (w); |
1223 |
} |
1224 |
#endif |
1225 |
|
1226 |
struct ev_timer t1; |
1227 |
ev_timer_init (&t1, ocb, 5, 10); |
1228 |
ev_timer_start (&t1); |
1229 |
|
1230 |
struct ev_signal sig; |
1231 |
ev_signal_init (&sig, scb, SIGQUIT); |
1232 |
ev_signal_start (&sig); |
1233 |
|
1234 |
struct ev_check cw; |
1235 |
ev_check_init (&cw, gcb); |
1236 |
ev_check_start (&cw); |
1237 |
|
1238 |
struct ev_idle iw; |
1239 |
ev_idle_init (&iw, gcb); |
1240 |
ev_idle_start (&iw); |
1241 |
|
1242 |
ev_loop (0); |
1243 |
|
1244 |
return 0; |
1245 |
} |
1246 |
|
1247 |
#endif |
1248 |
|
1249 |
|
1250 |
|
1251 |
|