ViewVC Help
View File | Revision Log | Show Annotations | Download File
/cvs/libev/ev.pod
(Generate patch)

Comparing libev/ev.pod (file contents):
Revision 1.98 by root, Sat Dec 22 06:10:25 2007 UTC vs.
Revision 1.104 by root, Sun Dec 23 03:48:02 2007 UTC

306=item C<EVBACKEND_SELECT> (value 1, portable select backend) 306=item C<EVBACKEND_SELECT> (value 1, portable select backend)
307 307
308This is your standard select(2) backend. Not I<completely> standard, as 308This is your standard select(2) backend. Not I<completely> standard, as
309libev tries to roll its own fd_set with no limits on the number of fds, 309libev tries to roll its own fd_set with no limits on the number of fds,
310but if that fails, expect a fairly low limit on the number of fds when 310but if that fails, expect a fairly low limit on the number of fds when
311using this backend. It doesn't scale too well (O(highest_fd)), but its usually 311using this backend. It doesn't scale too well (O(highest_fd)), but its
312the fastest backend for a low number of fds. 312usually the fastest backend for a low number of (low-numbered :) fds.
313
314To get good performance out of this backend you need a high amount of
315parallelity (most of the file descriptors should be busy). If you are
316writing a server, you should C<accept ()> in a loop to accept as many
317connections as possible during one iteration. You might also want to have
318a look at C<ev_set_io_collect_interval ()> to increase the amount of
319readyness notifications you get per iteration.
313 320
314=item C<EVBACKEND_POLL> (value 2, poll backend, available everywhere except on windows) 321=item C<EVBACKEND_POLL> (value 2, poll backend, available everywhere except on windows)
315 322
316And this is your standard poll(2) backend. It's more complicated than 323And this is your standard poll(2) backend. It's more complicated
317select, but handles sparse fds better and has no artificial limit on the 324than select, but handles sparse fds better and has no artificial
318number of fds you can use (except it will slow down considerably with a 325limit on the number of fds you can use (except it will slow down
319lot of inactive fds). It scales similarly to select, i.e. O(total_fds). 326considerably with a lot of inactive fds). It scales similarly to select,
327i.e. O(total_fds). See the entry for C<EVBACKEND_SELECT>, above, for
328performance tips.
320 329
321=item C<EVBACKEND_EPOLL> (value 4, Linux) 330=item C<EVBACKEND_EPOLL> (value 4, Linux)
322 331
323For few fds, this backend is a bit little slower than poll and select, 332For few fds, this backend is a bit little slower than poll and select,
324but it scales phenomenally better. While poll and select usually scale 333but it scales phenomenally better. While poll and select usually scale
325like O(total_fds) where n is the total number of fds (or the highest fd), 334like O(total_fds) where n is the total number of fds (or the highest fd),
326epoll scales either O(1) or O(active_fds). The epoll design has a number 335epoll scales either O(1) or O(active_fds). The epoll design has a number
327of shortcomings, such as silently dropping events in some hard-to-detect 336of shortcomings, such as silently dropping events in some hard-to-detect
328cases and rewiring a syscall per fd change, no fork support and bad 337cases and rewiring a syscall per fd change, no fork support and bad
329support for dup: 338support for dup.
330 339
331While stopping, setting and starting an I/O watcher in the same iteration 340While stopping, setting and starting an I/O watcher in the same iteration
332will result in some caching, there is still a syscall per such incident 341will result in some caching, there is still a syscall per such incident
333(because the fd could point to a different file description now), so its 342(because the fd could point to a different file description now), so its
334best to avoid that. Also, C<dup ()>'ed file descriptors might not work 343best to avoid that. Also, C<dup ()>'ed file descriptors might not work
336 345
337Please note that epoll sometimes generates spurious notifications, so you 346Please note that epoll sometimes generates spurious notifications, so you
338need to use non-blocking I/O or other means to avoid blocking when no data 347need to use non-blocking I/O or other means to avoid blocking when no data
339(or space) is available. 348(or space) is available.
340 349
350Best performance from this backend is achieved by not unregistering all
351watchers for a file descriptor until it has been closed, if possible, i.e.
352keep at least one watcher active per fd at all times.
353
354While nominally embeddeble in other event loops, this feature is broken in
355all kernel versions tested so far.
356
341=item C<EVBACKEND_KQUEUE> (value 8, most BSD clones) 357=item C<EVBACKEND_KQUEUE> (value 8, most BSD clones)
342 358
343Kqueue deserves special mention, as at the time of this writing, it 359Kqueue deserves special mention, as at the time of this writing, it
344was broken on I<all> BSDs (usually it doesn't work with anything but 360was broken on all BSDs except NetBSD (usually it doesn't work reliably
345sockets and pipes, except on Darwin, where of course it's completely 361with anything but sockets and pipes, except on Darwin, where of course
346useless. On NetBSD, it seems to work for all the FD types I tested, so it
347is used by default there). For this reason it's not being "autodetected" 362it's completely useless). For this reason it's not being "autodetected"
348unless you explicitly specify it explicitly in the flags (i.e. using 363unless you explicitly specify it explicitly in the flags (i.e. using
349C<EVBACKEND_KQUEUE>) or libev was compiled on a known-to-be-good (-enough) 364C<EVBACKEND_KQUEUE>) or libev was compiled on a known-to-be-good (-enough)
350system like NetBSD. 365system like NetBSD.
351 366
367You still can embed kqueue into a normal poll or select backend and use it
368only for sockets (after having made sure that sockets work with kqueue on
369the target platform). See C<ev_embed> watchers for more info.
370
352It scales in the same way as the epoll backend, but the interface to the 371It scales in the same way as the epoll backend, but the interface to the
353kernel is more efficient (which says nothing about its actual speed, 372kernel is more efficient (which says nothing about its actual speed, of
354of course). While stopping, setting and starting an I/O watcher does 373course). While stopping, setting and starting an I/O watcher does never
355never cause an extra syscall as with epoll, it still adds up to two event 374cause an extra syscall as with C<EVBACKEND_EPOLL>, it still adds up to
356changes per incident, support for C<fork ()> is very bad and it drops fds 375two event changes per incident, support for C<fork ()> is very bad and it
357silently in similarly hard-to-detetc cases. 376drops fds silently in similarly hard-to-detect cases.
377
378This backend usually performs well under most conditions.
379
380While nominally embeddable in other event loops, this doesn't work
381everywhere, so you might need to test for this. And since it is broken
382almost everywhere, you should only use it when you have a lot of sockets
383(for which it usually works), by embedding it into another event loop
384(e.g. C<EVBACKEND_SELECT> or C<EVBACKEND_POLL>) and using it only for
385sockets.
358 386
359=item C<EVBACKEND_DEVPOLL> (value 16, Solaris 8) 387=item C<EVBACKEND_DEVPOLL> (value 16, Solaris 8)
360 388
361This is not implemented yet (and might never be). 389This is not implemented yet (and might never be, unless you send me an
390implementation). According to reports, C</dev/poll> only supports sockets
391and is not embeddable, which would limit the usefulness of this backend
392immensely.
362 393
363=item C<EVBACKEND_PORT> (value 32, Solaris 10) 394=item C<EVBACKEND_PORT> (value 32, Solaris 10)
364 395
365This uses the Solaris 10 event port mechanism. As with everything on Solaris, 396This uses the Solaris 10 event port mechanism. As with everything on Solaris,
366it's really slow, but it still scales very well (O(active_fds)). 397it's really slow, but it still scales very well (O(active_fds)).
367 398
368Please note that solaris event ports can deliver a lot of spurious 399Please note that solaris event ports can deliver a lot of spurious
369notifications, so you need to use non-blocking I/O or other means to avoid 400notifications, so you need to use non-blocking I/O or other means to avoid
370blocking when no data (or space) is available. 401blocking when no data (or space) is available.
371 402
403While this backend scales well, it requires one system call per active
404file descriptor per loop iteration. For small and medium numbers of file
405descriptors a "slow" C<EVBACKEND_SELECT> or C<EVBACKEND_POLL> backend
406might perform better.
407
372=item C<EVBACKEND_ALL> 408=item C<EVBACKEND_ALL>
373 409
374Try all backends (even potentially broken ones that wouldn't be tried 410Try all backends (even potentially broken ones that wouldn't be tried
375with C<EVFLAG_AUTO>). Since this is a mask, you can do stuff such as 411with C<EVFLAG_AUTO>). Since this is a mask, you can do stuff such as
376C<EVBACKEND_ALL & ~EVBACKEND_KQUEUE>. 412C<EVBACKEND_ALL & ~EVBACKEND_KQUEUE>.
413
414It is definitely not recommended to use this flag.
377 415
378=back 416=back
379 417
380If one or more of these are ored into the flags value, then only these 418If one or more of these are ored into the flags value, then only these
381backends will be tried (in the reverse order as given here). If none are 419backends will be tried (in the reverse order as given here). If none are
596overhead for the actual polling but can deliver many events at once. 634overhead for the actual polling but can deliver many events at once.
597 635
598By setting a higher I<io collect interval> you allow libev to spend more 636By setting a higher I<io collect interval> you allow libev to spend more
599time collecting I/O events, so you can handle more events per iteration, 637time collecting I/O events, so you can handle more events per iteration,
600at the cost of increasing latency. Timeouts (both C<ev_periodic> and 638at the cost of increasing latency. Timeouts (both C<ev_periodic> and
601C<ev_timer>) will be not affected. 639C<ev_timer>) will be not affected. Setting this to a non-null value will
640introduce an additional C<ev_sleep ()> call into most loop iterations.
602 641
603Likewise, by setting a higher I<timeout collect interval> you allow libev 642Likewise, by setting a higher I<timeout collect interval> you allow libev
604to spend more time collecting timeouts, at the expense of increased 643to spend more time collecting timeouts, at the expense of increased
605latency (the watcher callback will be called later). C<ev_io> watchers 644latency (the watcher callback will be called later). C<ev_io> watchers
606will not be affected. 645will not be affected. Setting this to a non-null value will not introduce
646any overhead in libev.
607 647
608Many (busy) programs can usually benefit by setting the io collect 648Many (busy) programs can usually benefit by setting the io collect
609interval to a value near C<0.1> or so, which is often enough for 649interval to a value near C<0.1> or so, which is often enough for
610interactive servers (of course not for games), likewise for timeouts. It 650interactive servers (of course not for games), likewise for timeouts. It
611usually doesn't make much sense to set it to a lower value than C<0.01>, 651usually doesn't make much sense to set it to a lower value than C<0.01>,
992optimisations to libev. 1032optimisations to libev.
993 1033
994=head3 The special problem of dup'ed file descriptors 1034=head3 The special problem of dup'ed file descriptors
995 1035
996Some backends (e.g. epoll), cannot register events for file descriptors, 1036Some backends (e.g. epoll), cannot register events for file descriptors,
997but only events for the underlying file descriptions. That menas when you 1037but only events for the underlying file descriptions. That means when you
998have C<dup ()>'ed file descriptors and register events for them, only one 1038have C<dup ()>'ed file descriptors and register events for them, only one
999file descriptor might actually receive events. 1039file descriptor might actually receive events.
1000 1040
1001There is no workaorund possible except not registering events 1041There is no workaround possible except not registering events
1002for potentially C<dup ()>'ed file descriptors or to resort to 1042for potentially C<dup ()>'ed file descriptors, or to resort to
1003C<EVBACKEND_SELECT> or C<EVBACKEND_POLL>. 1043C<EVBACKEND_SELECT> or C<EVBACKEND_POLL>.
1004 1044
1005=head3 The special problem of fork 1045=head3 The special problem of fork
1006 1046
1007Some backends (epoll, kqueue) do not support C<fork ()> at all or exhibit 1047Some backends (epoll, kqueue) do not support C<fork ()> at all or exhibit
1621 1661
1622It is recommended to give C<ev_check> watchers highest (C<EV_MAXPRI>) 1662It is recommended to give C<ev_check> watchers highest (C<EV_MAXPRI>)
1623priority, to ensure that they are being run before any other watchers 1663priority, to ensure that they are being run before any other watchers
1624after the poll. Also, C<ev_check> watchers (and C<ev_prepare> watchers, 1664after the poll. Also, C<ev_check> watchers (and C<ev_prepare> watchers,
1625too) should not activate ("feed") events into libev. While libev fully 1665too) should not activate ("feed") events into libev. While libev fully
1626supports this, they will be called before other C<ev_check> watchers did 1666supports this, they will be called before other C<ev_check> watchers
1627their job. As C<ev_check> watchers are often used to embed other event 1667did their job. As C<ev_check> watchers are often used to embed other
1628loops those other event loops might be in an unusable state until their 1668(non-libev) event loops those other event loops might be in an unusable
1629C<ev_check> watcher ran (always remind yourself to coexist peacefully with 1669state until their C<ev_check> watcher ran (always remind yourself to
1630others). 1670coexist peacefully with others).
1631 1671
1632=head3 Watcher-Specific Functions and Data Members 1672=head3 Watcher-Specific Functions and Data Members
1633 1673
1634=over 4 1674=over 4
1635 1675
1774=head2 C<ev_embed> - when one backend isn't enough... 1814=head2 C<ev_embed> - when one backend isn't enough...
1775 1815
1776This is a rather advanced watcher type that lets you embed one event loop 1816This is a rather advanced watcher type that lets you embed one event loop
1777into another (currently only C<ev_io> events are supported in the embedded 1817into another (currently only C<ev_io> events are supported in the embedded
1778loop, other types of watchers might be handled in a delayed or incorrect 1818loop, other types of watchers might be handled in a delayed or incorrect
1779fashion and must not be used). (See portability notes, below). 1819fashion and must not be used).
1780 1820
1781There are primarily two reasons you would want that: work around bugs and 1821There are primarily two reasons you would want that: work around bugs and
1782prioritise I/O. 1822prioritise I/O.
1783 1823
1784As an example for a bug workaround, the kqueue backend might only support 1824As an example for a bug workaround, the kqueue backend might only support
1839 ev_embed_start (loop_hi, &embed); 1879 ev_embed_start (loop_hi, &embed);
1840 } 1880 }
1841 else 1881 else
1842 loop_lo = loop_hi; 1882 loop_lo = loop_hi;
1843 1883
1844=head2 Portability notes
1845
1846Kqueue is nominally embeddable, but this is broken on all BSDs that I
1847tried, in various ways. Usually the embedded event loop will simply never
1848receive events, sometimes it will only trigger a few times, sometimes in a
1849loop. Epoll is also nominally embeddable, but many Linux kernel versions
1850will always eport the epoll fd as ready, even when no events are pending.
1851
1852While libev allows embedding these backends (they are contained in
1853C<ev_embeddable_backends ()>), take extreme care that it will actually
1854work.
1855
1856When in doubt, create a dynamic event loop forced to use sockets (this
1857usually works) and possibly another thread and a pipe or so to report to
1858your main event loop.
1859
1860=head3 Watcher-Specific Functions and Data Members 1884=head3 Watcher-Specific Functions and Data Members
1861 1885
1862=over 4 1886=over 4
1863 1887
1864=item ev_embed_init (ev_embed *, callback, struct ev_loop *embedded_loop) 1888=item ev_embed_init (ev_embed *, callback, struct ev_loop *embedded_loop)
2502than enough. If you need to manage thousands of children you might want to 2526than enough. If you need to manage thousands of children you might want to
2503increase this value (I<must> be a power of two). 2527increase this value (I<must> be a power of two).
2504 2528
2505=item EV_INOTIFY_HASHSIZE 2529=item EV_INOTIFY_HASHSIZE
2506 2530
2507C<ev_staz> watchers use a small hash table to distribute workload by 2531C<ev_stat> watchers use a small hash table to distribute workload by
2508inotify watch id. The default size is C<16> (or C<1> with C<EV_MINIMAL>), 2532inotify watch id. The default size is C<16> (or C<1> with C<EV_MINIMAL>),
2509usually more than enough. If you need to manage thousands of C<ev_stat> 2533usually more than enough. If you need to manage thousands of C<ev_stat>
2510watchers you might want to increase this value (I<must> be a power of 2534watchers you might want to increase this value (I<must> be a power of
2511two). 2535two).
2512 2536

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines