ViewVC Help
View File | Revision Log | Show Annotations | Download File
/cvs/libev/ev.pod
(Generate patch)

Comparing libev/ev.pod (file contents):
Revision 1.96 by ayin, Fri Dec 21 10:06:50 2007 UTC vs.
Revision 1.107 by root, Mon Dec 24 04:34:00 2007 UTC

4 4
5=head1 SYNOPSIS 5=head1 SYNOPSIS
6 6
7 #include <ev.h> 7 #include <ev.h>
8 8
9=head1 EXAMPLE PROGRAM 9=head2 EXAMPLE PROGRAM
10 10
11 #include <ev.h> 11 #include <ev.h>
12 12
13 ev_io stdin_watcher; 13 ev_io stdin_watcher;
14 ev_timer timeout_watcher; 14 ev_timer timeout_watcher;
65You register interest in certain events by registering so-called I<event 65You register interest in certain events by registering so-called I<event
66watchers>, which are relatively small C structures you initialise with the 66watchers>, which are relatively small C structures you initialise with the
67details of the event, and then hand it over to libev by I<starting> the 67details of the event, and then hand it over to libev by I<starting> the
68watcher. 68watcher.
69 69
70=head1 FEATURES 70=head2 FEATURES
71 71
72Libev supports C<select>, C<poll>, the Linux-specific C<epoll>, the 72Libev supports C<select>, C<poll>, the Linux-specific C<epoll>, the
73BSD-specific C<kqueue> and the Solaris-specific event port mechanisms 73BSD-specific C<kqueue> and the Solaris-specific event port mechanisms
74for file descriptor events (C<ev_io>), the Linux C<inotify> interface 74for file descriptor events (C<ev_io>), the Linux C<inotify> interface
75(for C<ev_stat>), relative timers (C<ev_timer>), absolute timers 75(for C<ev_stat>), relative timers (C<ev_timer>), absolute timers
82 82
83It also is quite fast (see this 83It also is quite fast (see this
84L<benchmark|http://libev.schmorp.de/bench.html> comparing it to libevent 84L<benchmark|http://libev.schmorp.de/bench.html> comparing it to libevent
85for example). 85for example).
86 86
87=head1 CONVENTIONS 87=head2 CONVENTIONS
88 88
89Libev is very configurable. In this manual the default configuration will 89Libev is very configurable. In this manual the default configuration will
90be described, which supports multiple event loops. For more info about 90be described, which supports multiple event loops. For more info about
91various configuration options please have a look at B<EMBED> section in 91various configuration options please have a look at B<EMBED> section in
92this manual. If libev was configured without support for multiple event 92this manual. If libev was configured without support for multiple event
93loops, then all functions taking an initial argument of name C<loop> 93loops, then all functions taking an initial argument of name C<loop>
94(which is always of type C<struct ev_loop *>) will not have this argument. 94(which is always of type C<struct ev_loop *>) will not have this argument.
95 95
96=head1 TIME REPRESENTATION 96=head2 TIME REPRESENTATION
97 97
98Libev represents time as a single floating point number, representing the 98Libev represents time as a single floating point number, representing the
99(fractional) number of seconds since the (POSIX) epoch (somewhere near 99(fractional) number of seconds since the (POSIX) epoch (somewhere near
100the beginning of 1970, details are complicated, don't ask). This type is 100the beginning of 1970, details are complicated, don't ask). This type is
101called C<ev_tstamp>, which is what you should use too. It usually aliases 101called C<ev_tstamp>, which is what you should use too. It usually aliases
115 115
116Returns the current time as libev would use it. Please note that the 116Returns the current time as libev would use it. Please note that the
117C<ev_now> function is usually faster and also often returns the timestamp 117C<ev_now> function is usually faster and also often returns the timestamp
118you actually want to know. 118you actually want to know.
119 119
120=item ev_sleep (ev_tstamp interval)
121
122Sleep for the given interval: The current thread will be blocked until
123either it is interrupted or the given time interval has passed. Basically
124this is a subsecond-resolution C<sleep ()>.
125
120=item int ev_version_major () 126=item int ev_version_major ()
121 127
122=item int ev_version_minor () 128=item int ev_version_minor ()
123 129
124You can find out the major and minor ABI version numbers of the library 130You can find out the major and minor ABI version numbers of the library
300=item C<EVBACKEND_SELECT> (value 1, portable select backend) 306=item C<EVBACKEND_SELECT> (value 1, portable select backend)
301 307
302This is your standard select(2) backend. Not I<completely> standard, as 308This is your standard select(2) backend. Not I<completely> standard, as
303libev tries to roll its own fd_set with no limits on the number of fds, 309libev tries to roll its own fd_set with no limits on the number of fds,
304but if that fails, expect a fairly low limit on the number of fds when 310but if that fails, expect a fairly low limit on the number of fds when
305using this backend. It doesn't scale too well (O(highest_fd)), but its usually 311using this backend. It doesn't scale too well (O(highest_fd)), but its
306the fastest backend for a low number of fds. 312usually the fastest backend for a low number of (low-numbered :) fds.
313
314To get good performance out of this backend you need a high amount of
315parallelity (most of the file descriptors should be busy). If you are
316writing a server, you should C<accept ()> in a loop to accept as many
317connections as possible during one iteration. You might also want to have
318a look at C<ev_set_io_collect_interval ()> to increase the amount of
319readyness notifications you get per iteration.
307 320
308=item C<EVBACKEND_POLL> (value 2, poll backend, available everywhere except on windows) 321=item C<EVBACKEND_POLL> (value 2, poll backend, available everywhere except on windows)
309 322
310And this is your standard poll(2) backend. It's more complicated than 323And this is your standard poll(2) backend. It's more complicated
311select, but handles sparse fds better and has no artificial limit on the 324than select, but handles sparse fds better and has no artificial
312number of fds you can use (except it will slow down considerably with a 325limit on the number of fds you can use (except it will slow down
313lot of inactive fds). It scales similarly to select, i.e. O(total_fds). 326considerably with a lot of inactive fds). It scales similarly to select,
327i.e. O(total_fds). See the entry for C<EVBACKEND_SELECT>, above, for
328performance tips.
314 329
315=item C<EVBACKEND_EPOLL> (value 4, Linux) 330=item C<EVBACKEND_EPOLL> (value 4, Linux)
316 331
317For few fds, this backend is a bit little slower than poll and select, 332For few fds, this backend is a bit little slower than poll and select,
318but it scales phenomenally better. While poll and select usually scale 333but it scales phenomenally better. While poll and select usually scale
319like O(total_fds) where n is the total number of fds (or the highest fd), 334like O(total_fds) where n is the total number of fds (or the highest fd),
320epoll scales either O(1) or O(active_fds). The epoll design has a number 335epoll scales either O(1) or O(active_fds). The epoll design has a number
321of shortcomings, such as silently dropping events in some hard-to-detect 336of shortcomings, such as silently dropping events in some hard-to-detect
322cases and rewiring a syscall per fd change, no fork support and bad 337cases and rewiring a syscall per fd change, no fork support and bad
323support for dup: 338support for dup.
324 339
325While stopping, setting and starting an I/O watcher in the same iteration 340While stopping, setting and starting an I/O watcher in the same iteration
326will result in some caching, there is still a syscall per such incident 341will result in some caching, there is still a syscall per such incident
327(because the fd could point to a different file description now), so its 342(because the fd could point to a different file description now), so its
328best to avoid that. Also, C<dup ()>'ed file descriptors might not work 343best to avoid that. Also, C<dup ()>'ed file descriptors might not work
330 345
331Please note that epoll sometimes generates spurious notifications, so you 346Please note that epoll sometimes generates spurious notifications, so you
332need to use non-blocking I/O or other means to avoid blocking when no data 347need to use non-blocking I/O or other means to avoid blocking when no data
333(or space) is available. 348(or space) is available.
334 349
350Best performance from this backend is achieved by not unregistering all
351watchers for a file descriptor until it has been closed, if possible, i.e.
352keep at least one watcher active per fd at all times.
353
354While nominally embeddeble in other event loops, this feature is broken in
355all kernel versions tested so far.
356
335=item C<EVBACKEND_KQUEUE> (value 8, most BSD clones) 357=item C<EVBACKEND_KQUEUE> (value 8, most BSD clones)
336 358
337Kqueue deserves special mention, as at the time of this writing, it 359Kqueue deserves special mention, as at the time of this writing, it
338was broken on I<all> BSDs (usually it doesn't work with anything but 360was broken on all BSDs except NetBSD (usually it doesn't work reliably
339sockets and pipes, except on Darwin, where of course it's completely 361with anything but sockets and pipes, except on Darwin, where of course
340useless. On NetBSD, it seems to work for all the FD types I tested, so it
341is used by default there). For this reason it's not being "autodetected" 362it's completely useless). For this reason it's not being "autodetected"
342unless you explicitly specify it explicitly in the flags (i.e. using 363unless you explicitly specify it explicitly in the flags (i.e. using
343C<EVBACKEND_KQUEUE>) or libev was compiled on a known-to-be-good (-enough) 364C<EVBACKEND_KQUEUE>) or libev was compiled on a known-to-be-good (-enough)
344system like NetBSD. 365system like NetBSD.
345 366
367You still can embed kqueue into a normal poll or select backend and use it
368only for sockets (after having made sure that sockets work with kqueue on
369the target platform). See C<ev_embed> watchers for more info.
370
346It scales in the same way as the epoll backend, but the interface to the 371It scales in the same way as the epoll backend, but the interface to the
347kernel is more efficient (which says nothing about its actual speed, 372kernel is more efficient (which says nothing about its actual speed, of
348of course). While stopping, setting and starting an I/O watcher does 373course). While stopping, setting and starting an I/O watcher does never
349never cause an extra syscall as with epoll, it still adds up to two event 374cause an extra syscall as with C<EVBACKEND_EPOLL>, it still adds up to
350changes per incident, support for C<fork ()> is very bad and it drops fds 375two event changes per incident, support for C<fork ()> is very bad and it
351silently in similarly hard-to-detetc cases. 376drops fds silently in similarly hard-to-detect cases.
377
378This backend usually performs well under most conditions.
379
380While nominally embeddable in other event loops, this doesn't work
381everywhere, so you might need to test for this. And since it is broken
382almost everywhere, you should only use it when you have a lot of sockets
383(for which it usually works), by embedding it into another event loop
384(e.g. C<EVBACKEND_SELECT> or C<EVBACKEND_POLL>) and using it only for
385sockets.
352 386
353=item C<EVBACKEND_DEVPOLL> (value 16, Solaris 8) 387=item C<EVBACKEND_DEVPOLL> (value 16, Solaris 8)
354 388
355This is not implemented yet (and might never be). 389This is not implemented yet (and might never be, unless you send me an
390implementation). According to reports, C</dev/poll> only supports sockets
391and is not embeddable, which would limit the usefulness of this backend
392immensely.
356 393
357=item C<EVBACKEND_PORT> (value 32, Solaris 10) 394=item C<EVBACKEND_PORT> (value 32, Solaris 10)
358 395
359This uses the Solaris 10 event port mechanism. As with everything on Solaris, 396This uses the Solaris 10 event port mechanism. As with everything on Solaris,
360it's really slow, but it still scales very well (O(active_fds)). 397it's really slow, but it still scales very well (O(active_fds)).
361 398
362Please note that solaris event ports can deliver a lot of spurious 399Please note that solaris event ports can deliver a lot of spurious
363notifications, so you need to use non-blocking I/O or other means to avoid 400notifications, so you need to use non-blocking I/O or other means to avoid
364blocking when no data (or space) is available. 401blocking when no data (or space) is available.
365 402
403While this backend scales well, it requires one system call per active
404file descriptor per loop iteration. For small and medium numbers of file
405descriptors a "slow" C<EVBACKEND_SELECT> or C<EVBACKEND_POLL> backend
406might perform better.
407
366=item C<EVBACKEND_ALL> 408=item C<EVBACKEND_ALL>
367 409
368Try all backends (even potentially broken ones that wouldn't be tried 410Try all backends (even potentially broken ones that wouldn't be tried
369with C<EVFLAG_AUTO>). Since this is a mask, you can do stuff such as 411with C<EVFLAG_AUTO>). Since this is a mask, you can do stuff such as
370C<EVBACKEND_ALL & ~EVBACKEND_KQUEUE>. 412C<EVBACKEND_ALL & ~EVBACKEND_KQUEUE>.
413
414It is definitely not recommended to use this flag.
371 415
372=back 416=back
373 417
374If one or more of these are ored into the flags value, then only these 418If one or more of these are ored into the flags value, then only these
375backends will be tried (in the reverse order as given here). If none are 419backends will be tried (in the reverse order as given here). If none are
569Example: For some weird reason, unregister the above signal handler again. 613Example: For some weird reason, unregister the above signal handler again.
570 614
571 ev_ref (loop); 615 ev_ref (loop);
572 ev_signal_stop (loop, &exitsig); 616 ev_signal_stop (loop, &exitsig);
573 617
618=item ev_set_io_collect_interval (loop, ev_tstamp interval)
619
620=item ev_set_timeout_collect_interval (loop, ev_tstamp interval)
621
622These advanced functions influence the time that libev will spend waiting
623for events. Both are by default C<0>, meaning that libev will try to
624invoke timer/periodic callbacks and I/O callbacks with minimum latency.
625
626Setting these to a higher value (the C<interval> I<must> be >= C<0>)
627allows libev to delay invocation of I/O and timer/periodic callbacks to
628increase efficiency of loop iterations.
629
630The background is that sometimes your program runs just fast enough to
631handle one (or very few) event(s) per loop iteration. While this makes
632the program responsive, it also wastes a lot of CPU time to poll for new
633events, especially with backends like C<select ()> which have a high
634overhead for the actual polling but can deliver many events at once.
635
636By setting a higher I<io collect interval> you allow libev to spend more
637time collecting I/O events, so you can handle more events per iteration,
638at the cost of increasing latency. Timeouts (both C<ev_periodic> and
639C<ev_timer>) will be not affected. Setting this to a non-null value will
640introduce an additional C<ev_sleep ()> call into most loop iterations.
641
642Likewise, by setting a higher I<timeout collect interval> you allow libev
643to spend more time collecting timeouts, at the expense of increased
644latency (the watcher callback will be called later). C<ev_io> watchers
645will not be affected. Setting this to a non-null value will not introduce
646any overhead in libev.
647
648Many (busy) programs can usually benefit by setting the io collect
649interval to a value near C<0.1> or so, which is often enough for
650interactive servers (of course not for games), likewise for timeouts. It
651usually doesn't make much sense to set it to a lower value than C<0.01>,
652as this approsaches the timing granularity of most systems.
653
574=back 654=back
575 655
576 656
577=head1 ANATOMY OF A WATCHER 657=head1 ANATOMY OF A WATCHER
578 658
952optimisations to libev. 1032optimisations to libev.
953 1033
954=head3 The special problem of dup'ed file descriptors 1034=head3 The special problem of dup'ed file descriptors
955 1035
956Some backends (e.g. epoll), cannot register events for file descriptors, 1036Some backends (e.g. epoll), cannot register events for file descriptors,
957but only events for the underlying file descriptions. That menas when you 1037but only events for the underlying file descriptions. That means when you
958have C<dup ()>'ed file descriptors and register events for them, only one 1038have C<dup ()>'ed file descriptors and register events for them, only one
959file descriptor might actually receive events. 1039file descriptor might actually receive events.
960 1040
961There is no workaorund possible except not registering events 1041There is no workaround possible except not registering events
962for potentially C<dup ()>'ed file descriptors or to resort to 1042for potentially C<dup ()>'ed file descriptors, or to resort to
963C<EVBACKEND_SELECT> or C<EVBACKEND_POLL>. 1043C<EVBACKEND_SELECT> or C<EVBACKEND_POLL>.
964 1044
965=head3 The special problem of fork 1045=head3 The special problem of fork
966 1046
967Some backends (epoll, kqueue) do not support C<fork ()> at all or exhibit 1047Some backends (epoll, kqueue) do not support C<fork ()> at all or exhibit
1415reader). Inotify will be used to give hints only and should not change the 1495reader). Inotify will be used to give hints only and should not change the
1416semantics of C<ev_stat> watchers, which means that libev sometimes needs 1496semantics of C<ev_stat> watchers, which means that libev sometimes needs
1417to fall back to regular polling again even with inotify, but changes are 1497to fall back to regular polling again even with inotify, but changes are
1418usually detected immediately, and if the file exists there will be no 1498usually detected immediately, and if the file exists there will be no
1419polling. 1499polling.
1500
1501=head3 The special problem of stat time resolution
1502
1503The C<stat ()> syscall only supports full-second resolution portably, and
1504even on systems where the resolution is higher, many filesystems still
1505only support whole seconds.
1506
1507That means that, if the time is the only thing that changes, you might
1508miss updates: on the first update, C<ev_stat> detects a change and calls
1509your callback, which does something. When there is another update within
1510the same second, C<ev_stat> will be unable to detect it.
1511
1512The solution to this is to delay acting on a change for a second (or till
1513the next second boundary), using a roughly one-second delay C<ev_timer>
1514(C<ev_timer_set (w, 0., 1.01); ev_timer_again (loop, w)>). The C<.01>
1515is added to work around small timing inconsistencies of some operating
1516systems.
1420 1517
1421=head3 Watcher-Specific Functions and Data Members 1518=head3 Watcher-Specific Functions and Data Members
1422 1519
1423=over 4 1520=over 4
1424 1521
1484 } 1581 }
1485 1582
1486 ... 1583 ...
1487 ev_stat passwd; 1584 ev_stat passwd;
1488 1585
1489 ev_stat_init (&passwd, passwd_cb, "/etc/passwd"); 1586 ev_stat_init (&passwd, passwd_cb, "/etc/passwd", 0.);
1490 ev_stat_start (loop, &passwd); 1587 ev_stat_start (loop, &passwd);
1588
1589Example: Like above, but additionally use a one-second delay so we do not
1590miss updates (however, frequent updates will delay processing, too, so
1591one might do the work both on C<ev_stat> callback invocation I<and> on
1592C<ev_timer> callback invocation).
1593
1594 static ev_stat passwd;
1595 static ev_timer timer;
1596
1597 static void
1598 timer_cb (EV_P_ ev_timer *w, int revents)
1599 {
1600 ev_timer_stop (EV_A_ w);
1601
1602 /* now it's one second after the most recent passwd change */
1603 }
1604
1605 static void
1606 stat_cb (EV_P_ ev_stat *w, int revents)
1607 {
1608 /* reset the one-second timer */
1609 ev_timer_again (EV_A_ &timer);
1610 }
1611
1612 ...
1613 ev_stat_init (&passwd, stat_cb, "/etc/passwd", 0.);
1614 ev_stat_start (loop, &passwd);
1615 ev_timer_init (&timer, timer_cb, 0., 1.01);
1491 1616
1492 1617
1493=head2 C<ev_idle> - when you've got nothing better to do... 1618=head2 C<ev_idle> - when you've got nothing better to do...
1494 1619
1495Idle watchers trigger events when no other events of the same or higher 1620Idle watchers trigger events when no other events of the same or higher
1581 1706
1582It is recommended to give C<ev_check> watchers highest (C<EV_MAXPRI>) 1707It is recommended to give C<ev_check> watchers highest (C<EV_MAXPRI>)
1583priority, to ensure that they are being run before any other watchers 1708priority, to ensure that they are being run before any other watchers
1584after the poll. Also, C<ev_check> watchers (and C<ev_prepare> watchers, 1709after the poll. Also, C<ev_check> watchers (and C<ev_prepare> watchers,
1585too) should not activate ("feed") events into libev. While libev fully 1710too) should not activate ("feed") events into libev. While libev fully
1586supports this, they will be called before other C<ev_check> watchers did 1711supports this, they will be called before other C<ev_check> watchers
1587their job. As C<ev_check> watchers are often used to embed other event 1712did their job. As C<ev_check> watchers are often used to embed other
1588loops those other event loops might be in an unusable state until their 1713(non-libev) event loops those other event loops might be in an unusable
1589C<ev_check> watcher ran (always remind yourself to coexist peacefully with 1714state until their C<ev_check> watcher ran (always remind yourself to
1590others). 1715coexist peacefully with others).
1591 1716
1592=head3 Watcher-Specific Functions and Data Members 1717=head3 Watcher-Specific Functions and Data Members
1593 1718
1594=over 4 1719=over 4
1595 1720
1734=head2 C<ev_embed> - when one backend isn't enough... 1859=head2 C<ev_embed> - when one backend isn't enough...
1735 1860
1736This is a rather advanced watcher type that lets you embed one event loop 1861This is a rather advanced watcher type that lets you embed one event loop
1737into another (currently only C<ev_io> events are supported in the embedded 1862into another (currently only C<ev_io> events are supported in the embedded
1738loop, other types of watchers might be handled in a delayed or incorrect 1863loop, other types of watchers might be handled in a delayed or incorrect
1739fashion and must not be used). (See portability notes, below). 1864fashion and must not be used).
1740 1865
1741There are primarily two reasons you would want that: work around bugs and 1866There are primarily two reasons you would want that: work around bugs and
1742prioritise I/O. 1867prioritise I/O.
1743 1868
1744As an example for a bug workaround, the kqueue backend might only support 1869As an example for a bug workaround, the kqueue backend might only support
1799 ev_embed_start (loop_hi, &embed); 1924 ev_embed_start (loop_hi, &embed);
1800 } 1925 }
1801 else 1926 else
1802 loop_lo = loop_hi; 1927 loop_lo = loop_hi;
1803 1928
1804=head2 Portability notes
1805
1806Kqueue is nominally embeddable, but this is broken on all BSDs that I
1807tried, in various ways. Usually the embedded event loop will simply never
1808receive events, sometimes it will only trigger a few times, sometimes in a
1809loop. Epoll is also nominally embeddable, but many Linux kernel versions
1810will always eport the epoll fd as ready, even when no events are pending.
1811
1812While libev allows embedding these backends (they are contained in
1813C<ev_embeddable_backends ()>), take extreme care that it will actually
1814work.
1815
1816When in doubt, create a dynamic event loop forced to use sockets (this
1817usually works) and possibly another thread and a pipe or so to report to
1818your main event loop.
1819
1820=head3 Watcher-Specific Functions and Data Members 1929=head3 Watcher-Specific Functions and Data Members
1821 1930
1822=over 4 1931=over 4
1823 1932
1824=item ev_embed_init (ev_embed *, callback, struct ev_loop *embedded_loop) 1933=item ev_embed_init (ev_embed *, callback, struct ev_loop *embedded_loop)
2296realtime clock option at compiletime (and assume its availability at 2405realtime clock option at compiletime (and assume its availability at
2297runtime if successful). Otherwise no use of the realtime clock option will 2406runtime if successful). Otherwise no use of the realtime clock option will
2298be attempted. This effectively replaces C<gettimeofday> by C<clock_get 2407be attempted. This effectively replaces C<gettimeofday> by C<clock_get
2299(CLOCK_REALTIME, ...)> and will not normally affect correctness. See the 2408(CLOCK_REALTIME, ...)> and will not normally affect correctness. See the
2300note about libraries in the description of C<EV_USE_MONOTONIC>, though. 2409note about libraries in the description of C<EV_USE_MONOTONIC>, though.
2410
2411=item EV_USE_NANOSLEEP
2412
2413If defined to be C<1>, libev will assume that C<nanosleep ()> is available
2414and will use it for delays. Otherwise it will use C<select ()>.
2301 2415
2302=item EV_USE_SELECT 2416=item EV_USE_SELECT
2303 2417
2304If undefined or defined to be C<1>, libev will compile in support for the 2418If undefined or defined to be C<1>, libev will compile in support for the
2305C<select>(2) backend. No attempt at autodetection will be done: if no 2419C<select>(2) backend. No attempt at autodetection will be done: if no
2457than enough. If you need to manage thousands of children you might want to 2571than enough. If you need to manage thousands of children you might want to
2458increase this value (I<must> be a power of two). 2572increase this value (I<must> be a power of two).
2459 2573
2460=item EV_INOTIFY_HASHSIZE 2574=item EV_INOTIFY_HASHSIZE
2461 2575
2462C<ev_staz> watchers use a small hash table to distribute workload by 2576C<ev_stat> watchers use a small hash table to distribute workload by
2463inotify watch id. The default size is C<16> (or C<1> with C<EV_MINIMAL>), 2577inotify watch id. The default size is C<16> (or C<1> with C<EV_MINIMAL>),
2464usually more than enough. If you need to manage thousands of C<ev_stat> 2578usually more than enough. If you need to manage thousands of C<ev_stat>
2465watchers you might want to increase this value (I<must> be a power of 2579watchers you might want to increase this value (I<must> be a power of
2466two). 2580two).
2467 2581
2563 2677
2564=item Starting and stopping timer/periodic watchers: O(log skipped_other_timers) 2678=item Starting and stopping timer/periodic watchers: O(log skipped_other_timers)
2565 2679
2566This means that, when you have a watcher that triggers in one hour and 2680This means that, when you have a watcher that triggers in one hour and
2567there are 100 watchers that would trigger before that then inserting will 2681there are 100 watchers that would trigger before that then inserting will
2568have to skip those 100 watchers. 2682have to skip roughly seven (C<ld 100>) of these watchers.
2569 2683
2570=item Changing timer/periodic watchers (by autorepeat, again): O(log skipped_other_timers) 2684=item Changing timer/periodic watchers (by autorepeat or calling again): O(log skipped_other_timers)
2571 2685
2572That means that for changing a timer costs less than removing/adding them 2686That means that changing a timer costs less than removing/adding them
2573as only the relative motion in the event queue has to be paid for. 2687as only the relative motion in the event queue has to be paid for.
2574 2688
2575=item Starting io/check/prepare/idle/signal/child watchers: O(1) 2689=item Starting io/check/prepare/idle/signal/child watchers: O(1)
2576 2690
2577These just add the watcher into an array or at the head of a list. 2691These just add the watcher into an array or at the head of a list.
2692
2578=item Stopping check/prepare/idle watchers: O(1) 2693=item Stopping check/prepare/idle watchers: O(1)
2579 2694
2580=item Stopping an io/signal/child watcher: O(number_of_watchers_for_this_(fd/signal/pid % EV_PID_HASHSIZE)) 2695=item Stopping an io/signal/child watcher: O(number_of_watchers_for_this_(fd/signal/pid % EV_PID_HASHSIZE))
2581 2696
2582These watchers are stored in lists then need to be walked to find the 2697These watchers are stored in lists then need to be walked to find the
2583correct watcher to remove. The lists are usually short (you don't usually 2698correct watcher to remove. The lists are usually short (you don't usually
2584have many watchers waiting for the same fd or signal). 2699have many watchers waiting for the same fd or signal).
2585 2700
2586=item Finding the next timer per loop iteration: O(1) 2701=item Finding the next timer in each loop iteration: O(1)
2702
2703By virtue of using a binary heap, the next timer is always found at the
2704beginning of the storage array.
2587 2705
2588=item Each change on a file descriptor per loop iteration: O(number_of_watchers_for_this_fd) 2706=item Each change on a file descriptor per loop iteration: O(number_of_watchers_for_this_fd)
2589 2707
2590A change means an I/O watcher gets started or stopped, which requires 2708A change means an I/O watcher gets started or stopped, which requires
2591libev to recalculate its status (and possibly tell the kernel). 2709libev to recalculate its status (and possibly tell the kernel, depending
2710on backend and wether C<ev_io_set> was used).
2592 2711
2593=item Activating one watcher: O(1) 2712=item Activating one watcher (putting it into the pending state): O(1)
2594 2713
2595=item Priority handling: O(number_of_priorities) 2714=item Priority handling: O(number_of_priorities)
2596 2715
2597Priorities are implemented by allocating some space for each 2716Priorities are implemented by allocating some space for each
2598priority. When doing priority-based operations, libev usually has to 2717priority. When doing priority-based operations, libev usually has to
2599linearly search all the priorities. 2718linearly search all the priorities, but starting/stopping and activating
2719watchers becomes O(1) w.r.t. prioritiy handling.
2600 2720
2601=back 2721=back
2602 2722
2603 2723
2604=head1 AUTHOR 2724=head1 AUTHOR

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines