ViewVC Help
View File | Revision Log | Show Annotations | Download File
/cvs/AnyEvent/lib/AnyEvent.pm
(Generate patch)

Comparing AnyEvent/lib/AnyEvent.pm (file contents):
Revision 1.84 by root, Fri Apr 25 13:48:42 2008 UTC vs.
Revision 1.98 by root, Sun Apr 27 16:31:48 2008 UTC

66 66
67Of course, if you want lots of policy (this can arguably be somewhat 67Of course, if you want lots of policy (this can arguably be somewhat
68useful) and you want to force your users to use the one and only event 68useful) and you want to force your users to use the one and only event
69model, you should I<not> use this module. 69model, you should I<not> use this module.
70 70
71#TODO#
72
73Net::IRC3
74AnyEvent::HTTPD
75AnyEvent::DNS
76IO::AnyEvent
77Net::FPing
78Net::XMPP2
79Coro
80
81AnyEvent::IRC
82AnyEvent::HTTPD
83AnyEvent::DNS
84AnyEvent::Handle
85AnyEvent::Socket
86AnyEvent::FPing
87AnyEvent::XMPP
88AnyEvent::SNMP
89Coro
71 90
72=head1 DESCRIPTION 91=head1 DESCRIPTION
73 92
74L<AnyEvent> provides an identical interface to multiple event loops. This 93L<AnyEvent> provides an identical interface to multiple event loops. This
75allows module authors to utilise an event loop without forcing module 94allows module authors to utilise an event loop without forcing module
141=head2 I/O WATCHERS 160=head2 I/O WATCHERS
142 161
143You can create an I/O watcher by calling the C<< AnyEvent->io >> method 162You can create an I/O watcher by calling the C<< AnyEvent->io >> method
144with the following mandatory key-value pairs as arguments: 163with the following mandatory key-value pairs as arguments:
145 164
146C<fh> the Perl I<file handle> (I<not> file descriptor) to watch for 165C<fh> the Perl I<file handle> (I<not> file descriptor) to watch
147events. C<poll> must be a string that is either C<r> or C<w>, which 166for events. C<poll> must be a string that is either C<r> or C<w>,
148creates a watcher waiting for "r"eadable or "w"ritable events, 167which creates a watcher waiting for "r"eadable or "w"ritable events,
149respectively. C<cb> is the callback to invoke each time the file handle 168respectively. C<cb> is the callback to invoke each time the file handle
150becomes ready. 169becomes ready.
170
171Although the callback might get passed parameters, their value and
172presence is undefined and you cannot rely on them. Portable AnyEvent
173callbacks cannot use arguments passed to I/O watcher callbacks.
151 174
152The I/O watcher might use the underlying file descriptor or a copy of it. 175The I/O watcher might use the underlying file descriptor or a copy of it.
153You must not close a file handle as long as any watcher is active on the 176You must not close a file handle as long as any watcher is active on the
154underlying file descriptor. 177underlying file descriptor.
155 178
156Some event loops issue spurious readyness notifications, so you should 179Some event loops issue spurious readyness notifications, so you should
157always use non-blocking calls when reading/writing from/to your file 180always use non-blocking calls when reading/writing from/to your file
158handles. 181handles.
159
160Although the callback might get passed parameters, their value and
161presence is undefined and you cannot rely on them. Portable AnyEvent
162callbacks cannot use arguments passed to I/O watcher callbacks.
163 182
164Example: 183Example:
165 184
166 # wait for readability of STDIN, then read a line and disable the watcher 185 # wait for readability of STDIN, then read a line and disable the watcher
167 my $w; $w = AnyEvent->io (fh => \*STDIN, poll => 'r', cb => sub { 186 my $w; $w = AnyEvent->io (fh => \*STDIN, poll => 'r', cb => sub {
174 193
175You can create a time watcher by calling the C<< AnyEvent->timer >> 194You can create a time watcher by calling the C<< AnyEvent->timer >>
176method with the following mandatory arguments: 195method with the following mandatory arguments:
177 196
178C<after> specifies after how many seconds (fractional values are 197C<after> specifies after how many seconds (fractional values are
179supported) should the timer activate. C<cb> the callback to invoke in that 198supported) the callback should be invoked. C<cb> is the callback to invoke
180case. 199in that case.
200
201Although the callback might get passed parameters, their value and
202presence is undefined and you cannot rely on them. Portable AnyEvent
203callbacks cannot use arguments passed to time watcher callbacks.
181 204
182The timer callback will be invoked at most once: if you want a repeating 205The timer callback will be invoked at most once: if you want a repeating
183timer you have to create a new watcher (this is a limitation by both Tk 206timer you have to create a new watcher (this is a limitation by both Tk
184and Glib). 207and Glib).
185
186Although the callback might get passed parameters, their value and
187presence is undefined and you cannot rely on them. Portable AnyEvent
188callbacks cannot use arguments passed to time watcher callbacks.
189 208
190Example: 209Example:
191 210
192 # fire an event after 7.7 seconds 211 # fire an event after 7.7 seconds
193 my $w = AnyEvent->timer (after => 7.7, cb => sub { 212 my $w = AnyEvent->timer (after => 7.7, cb => sub {
234 253
235You can watch for signals using a signal watcher, C<signal> is the signal 254You can watch for signals using a signal watcher, C<signal> is the signal
236I<name> without any C<SIG> prefix, C<cb> is the Perl callback to 255I<name> without any C<SIG> prefix, C<cb> is the Perl callback to
237be invoked whenever a signal occurs. 256be invoked whenever a signal occurs.
238 257
258Although the callback might get passed parameters, their value and
259presence is undefined and you cannot rely on them. Portable AnyEvent
260callbacks cannot use arguments passed to signal watcher callbacks.
261
239Multiple signal occurances can be clumped together into one callback 262Multiple signal occurances can be clumped together into one callback
240invocation, and callback invocation will be synchronous. synchronous means 263invocation, and callback invocation will be synchronous. synchronous means
241that it might take a while until the signal gets handled by the process, 264that it might take a while until the signal gets handled by the process,
242but it is guarenteed not to interrupt any other callbacks. 265but it is guarenteed not to interrupt any other callbacks.
243 266
257 280
258The child process is specified by the C<pid> argument (if set to C<0>, it 281The child process is specified by the C<pid> argument (if set to C<0>, it
259watches for any child process exit). The watcher will trigger as often 282watches for any child process exit). The watcher will trigger as often
260as status change for the child are received. This works by installing a 283as status change for the child are received. This works by installing a
261signal handler for C<SIGCHLD>. The callback will be called with the pid 284signal handler for C<SIGCHLD>. The callback will be called with the pid
262and exit status (as returned by waitpid). 285and exit status (as returned by waitpid), so unlike other watcher types,
286you I<can> rely on child watcher callback arguments.
263 287
264There is a slight catch to child watchers, however: you usually start them 288There is a slight catch to child watchers, however: you usually start them
265I<after> the child process was created, and this means the process could 289I<after> the child process was created, and this means the process could
266have exited already (and no SIGCHLD will be sent anymore). 290have exited already (and no SIGCHLD will be sent anymore).
267 291
889 }); 913 });
890 914
891 $quit->wait; 915 $quit->wait;
892 916
893 917
894=head1 BENCHMARK 918=head1 BENCHMARKS
895 919
896To give you an idea of the performance and overheads that AnyEvent adds 920To give you an idea of the performance and overheads that AnyEvent adds
897over the event loops themselves (and to give you an impression of the 921over the event loops themselves and to give you an impression of the speed
898speed of various event loops), here is a benchmark of various supported 922of various event loops I prepared some benchmarks.
899event models natively and with anyevent. The benchmark creates a lot of 923
900timers (with a zero timeout) and I/O watchers (watching STDOUT, a pty, to 924=head2 BENCHMARKING ANYEVENT OVERHEAD
925
926Here is a benchmark of various supported event models used natively and
927through anyevent. The benchmark creates a lot of timers (with a zero
928timeout) and I/O watchers (watching STDOUT, a pty, to become writable,
901become writable, which it is), lets them fire exactly once and destroys 929which it is), lets them fire exactly once and destroys them again.
902them again.
903 930
904Rewriting the benchmark to use many different sockets instead of using 931Source code for this benchmark is found as F<eg/bench> in the AnyEvent
905the same filehandle for all I/O watchers results in a much longer runtime 932distribution.
906(socket creation is expensive), but qualitatively the same figures, so it
907was not used.
908 933
909=head2 Explanation of the columns 934=head3 Explanation of the columns
910 935
911I<watcher> is the number of event watchers created/destroyed. Since 936I<watcher> is the number of event watchers created/destroyed. Since
912different event models feature vastly different performances, each event 937different event models feature vastly different performances, each event
913loop was given a number of watchers so that overall runtime is acceptable 938loop was given a number of watchers so that overall runtime is acceptable
914and similar between tested event loop (and keep them from crashing): Glib 939and similar between tested event loop (and keep them from crashing): Glib
930signal the end of this phase. 955signal the end of this phase.
931 956
932I<destroy> is the time, in microseconds, that it takes to destroy a single 957I<destroy> is the time, in microseconds, that it takes to destroy a single
933watcher. 958watcher.
934 959
935=head2 Results 960=head3 Results
936 961
937 name watchers bytes create invoke destroy comment 962 name watchers bytes create invoke destroy comment
938 EV/EV 400000 244 0.56 0.46 0.31 EV native interface 963 EV/EV 400000 244 0.56 0.46 0.31 EV native interface
939 EV/Any 100000 244 2.50 0.46 0.29 EV + AnyEvent watchers 964 EV/Any 100000 244 2.50 0.46 0.29 EV + AnyEvent watchers
940 CoroEV/Any 100000 244 2.49 0.44 0.29 coroutines + Coro::Signal 965 CoroEV/Any 100000 244 2.49 0.44 0.29 coroutines + Coro::Signal
941 Perl/Any 100000 513 4.92 0.87 1.12 pure perl implementation 966 Perl/Any 100000 513 4.92 0.87 1.12 pure perl implementation
942 Event/Event 16000 516 31.88 31.30 0.85 Event native interface 967 Event/Event 16000 516 31.88 31.30 0.85 Event native interface
943 Event/Any 16000 936 39.17 33.63 1.43 Event + AnyEvent watchers 968 Event/Any 16000 590 35.75 31.42 1.08 Event + AnyEvent watchers
944 Glib/Any 16000 1357 98.22 12.41 54.00 quadratic behaviour 969 Glib/Any 16000 1357 98.22 12.41 54.00 quadratic behaviour
945 Tk/Any 2000 1860 26.97 67.98 14.00 SEGV with >> 2000 watchers 970 Tk/Any 2000 1860 26.97 67.98 14.00 SEGV with >> 2000 watchers
946 POE/Event 2000 6644 108.64 736.02 14.73 via POE::Loop::Event 971 POE/Event 2000 6644 108.64 736.02 14.73 via POE::Loop::Event
947 POE/Select 2000 6343 94.13 809.12 565.96 via POE::Loop::Select 972 POE/Select 2000 6343 94.13 809.12 565.96 via POE::Loop::Select
948 973
949=head2 Discussion 974=head3 Discussion
950 975
951The benchmark does I<not> measure scalability of the event loop very 976The benchmark does I<not> measure scalability of the event loop very
952well. For example, a select-based event loop (such as the pure perl one) 977well. For example, a select-based event loop (such as the pure perl one)
953can never compete with an event loop that uses epoll when the number of 978can never compete with an event loop that uses epoll when the number of
954file descriptors grows high. In this benchmark, all events become ready at 979file descriptors grows high. In this benchmark, all events become ready at
955the same time, so select/poll-based implementations get an unnatural speed 980the same time, so select/poll-based implementations get an unnatural speed
956boost. 981boost.
957 982
983Also, note that the number of watchers usually has a nonlinear effect on
984overall speed, that is, creating twice as many watchers doesn't take twice
985the time - usually it takes longer. This puts event loops tested with a
986higher number of watchers at a disadvantage.
987
988To put the range of results into perspective, consider that on the
989benchmark machine, handling an event takes roughly 1600 CPU cycles with
990EV, 3100 CPU cycles with AnyEvent's pure perl loop and almost 3000000 CPU
991cycles with POE.
992
958C<EV> is the sole leader regarding speed and memory use, which are both 993C<EV> is the sole leader regarding speed and memory use, which are both
959maximal/minimal, respectively. Even when going through AnyEvent, it uses 994maximal/minimal, respectively. Even when going through AnyEvent, it uses
960far less memory than any other event loop and is still faster than Event 995far less memory than any other event loop and is still faster than Event
961natively. 996natively.
962 997
963The pure perl implementation is hit in a few sweet spots (both the 998The pure perl implementation is hit in a few sweet spots (both the
964zero timeout and the use of a single fd hit optimisations in the perl 999constant timeout and the use of a single fd hit optimisations in the perl
965interpreter and the backend itself, and all watchers become ready at the 1000interpreter and the backend itself). Nevertheless this shows that it
966same time). Nevertheless this shows that it adds very little overhead in 1001adds very little overhead in itself. Like any select-based backend its
967itself. Like any select-based backend its performance becomes really bad 1002performance becomes really bad with lots of file descriptors (and few of
968with lots of file descriptors (and few of them active), of course, but 1003them active), of course, but this was not subject of this benchmark.
969this was not subject of this benchmark.
970 1004
971The C<Event> module has a relatively high setup and callback invocation cost, 1005The C<Event> module has a relatively high setup and callback invocation
972but overall scores on the third place. 1006cost, but overall scores in on the third place.
973 1007
974C<Glib>'s memory usage is quite a bit bit higher, but it features a 1008C<Glib>'s memory usage is quite a bit higher, but it features a
975faster callback invocation and overall ends up in the same class as 1009faster callback invocation and overall ends up in the same class as
976C<Event>. However, Glib scales extremely badly, doubling the number of 1010C<Event>. However, Glib scales extremely badly, doubling the number of
977watchers increases the processing time by more than a factor of four, 1011watchers increases the processing time by more than a factor of four,
978making it completely unusable when using larger numbers of watchers 1012making it completely unusable when using larger numbers of watchers
979(note that only a single file descriptor was used in the benchmark, so 1013(note that only a single file descriptor was used in the benchmark, so
982The C<Tk> adaptor works relatively well. The fact that it crashes with 1016The C<Tk> adaptor works relatively well. The fact that it crashes with
983more than 2000 watchers is a big setback, however, as correctness takes 1017more than 2000 watchers is a big setback, however, as correctness takes
984precedence over speed. Nevertheless, its performance is surprising, as the 1018precedence over speed. Nevertheless, its performance is surprising, as the
985file descriptor is dup()ed for each watcher. This shows that the dup() 1019file descriptor is dup()ed for each watcher. This shows that the dup()
986employed by some adaptors is not a big performance issue (it does incur a 1020employed by some adaptors is not a big performance issue (it does incur a
987hidden memory cost inside the kernel, though, that is not reflected in the 1021hidden memory cost inside the kernel which is not reflected in the figures
988figures above). 1022above).
989 1023
990C<POE>, regardless of underlying event loop (wether using its pure perl 1024C<POE>, regardless of underlying event loop (whether using its pure
991select-based backend or the Event module) shows abysmal performance and 1025perl select-based backend or the Event module, the POE-EV backend
1026couldn't be tested because it wasn't working) shows abysmal performance
992memory usage: Watchers use almost 30 times as much memory as EV watchers, 1027and memory usage: Watchers use almost 30 times as much memory as
993and 10 times as much memory as both Event or EV via AnyEvent. Watcher 1028EV watchers, and 10 times as much memory as Event (the high memory
1029requirements are caused by requiring a session for each watcher). Watcher
994invocation is almost 900 times slower than with AnyEvent's pure perl 1030invocation speed is almost 900 times slower than with AnyEvent's pure perl
995implementation. The design of the POE adaptor class in AnyEvent can not 1031implementation. The design of the POE adaptor class in AnyEvent can not
996really account for this, as session creation overhead is small compared 1032really account for this, as session creation overhead is small compared
997to execution of the state machine, which is coded pretty optimally within 1033to execution of the state machine, which is coded pretty optimally within
998L<AnyEvent::Impl::POE>. POE simply seems to be abysmally slow. 1034L<AnyEvent::Impl::POE>. POE simply seems to be abysmally slow.
999 1035
1000=head2 Summary 1036=head3 Summary
1001 1037
1038=over 4
1039
1002Using EV through AnyEvent is faster than any other event loop, but most 1040=item * Using EV through AnyEvent is faster than any other event loop
1003event loops have acceptable performance with or without AnyEvent. 1041(even when used without AnyEvent), but most event loops have acceptable
1042performance with or without AnyEvent.
1004 1043
1005The overhead AnyEvent adds is usually much smaller than the overhead of 1044=item * The overhead AnyEvent adds is usually much smaller than the overhead of
1006the actual event loop, only with extremely fast event loops such as the EV 1045the actual event loop, only with extremely fast event loops such as EV
1007adds AnyEvent significant overhead. 1046adds AnyEvent significant overhead.
1008 1047
1009And you should simply avoid POE like the plague if you want performance or 1048=item * You should avoid POE like the plague if you want performance or
1010reasonable memory usage. 1049reasonable memory usage.
1050
1051=back
1052
1053=head2 BENCHMARKING THE LARGE SERVER CASE
1054
1055This benchmark atcually benchmarks the event loop itself. It works by
1056creating a number of "servers": each server consists of a socketpair, a
1057timeout watcher that gets reset on activity (but never fires), and an I/O
1058watcher waiting for input on one side of the socket. Each time the socket
1059watcher reads a byte it will write that byte to a random other "server".
1060
1061The effect is that there will be a lot of I/O watchers, only part of which
1062are active at any one point (so there is a constant number of active
1063fds for each loop iterstaion, but which fds these are is random). The
1064timeout is reset each time something is read because that reflects how
1065most timeouts work (and puts extra pressure on the event loops).
1066
1067In this benchmark, we use 10000 socketpairs (20000 sockets), of which 100
1068(1%) are active. This mirrors the activity of large servers with many
1069connections, most of which are idle at any one point in time.
1070
1071Source code for this benchmark is found as F<eg/bench2> in the AnyEvent
1072distribution.
1073
1074=head3 Explanation of the columns
1075
1076I<sockets> is the number of sockets, and twice the number of "servers" (as
1077each server has a read and write socket end).
1078
1079I<create> is the time it takes to create a socketpair (which is
1080nontrivial) and two watchers: an I/O watcher and a timeout watcher.
1081
1082I<request>, the most important value, is the time it takes to handle a
1083single "request", that is, reading the token from the pipe and forwarding
1084it to another server. This includes deleting the old timeout and creating
1085a new one that moves the timeout into the future.
1086
1087=head3 Results
1088
1089 name sockets create request
1090 EV 20000 69.01 11.16
1091 Perl 20000 75.28 112.76
1092 Event 20000 212.62 257.32
1093 Glib 20000 651.16 1896.30
1094 POE 20000 349.67 12317.24 uses POE::Loop::Event
1095
1096=head3 Discussion
1097
1098This benchmark I<does> measure scalability and overall performance of the
1099particular event loop.
1100
1101EV is again fastest. Since it is using epoll on my system, the setup time
1102is relatively high, though.
1103
1104Perl surprisingly comes second. It is much faster than the C-based event
1105loops Event and Glib.
1106
1107Event suffers from high setup time as well (look at its code and you will
1108understand why). Callback invocation also has a high overhead compared to
1109the C<< $_->() for .. >>-style loop that the Perl event loop uses. Event
1110uses select or poll in basically all documented configurations.
1111
1112Glib is hit hard by its quadratic behaviour w.r.t. many watchers. It
1113clearly fails to perform with many filehandles or in busy servers.
1114
1115POE is still completely out of the picture, taking over 1000 times as long
1116as EV, and over 100 times as long as the Perl implementation, even though
1117it uses a C-based event loop in this case.
1118
1119=head3 Summary
1120
1121=over 4
1122
1123=item * The pure perl implementation performs extremely well, considering
1124that it uses select.
1125
1126=item * Avoid Glib or POE in large projects where performance matters.
1127
1128=back
1129
1130=head2 BENCHMARKING SMALL SERVERS
1131
1132While event loops should scale (and select-based ones do not...) even to
1133large servers, most programs we (or I :) actually write have only a few
1134I/O watchers.
1135
1136In this benchmark, I use the same benchmark program as in the large server
1137case, but it uses only eight "servers", of which three are active at any
1138one time. This should reflect performance for a small server relatively
1139well.
1140
1141The columns are identical to the previous table.
1142
1143=head3 Results
1144
1145 name sockets create request
1146 EV 16 20.00 6.54
1147 Event 16 81.27 35.86
1148 Glib 16 32.63 15.48
1149 Perl 16 24.62 162.37
1150 POE 16 261.87 276.28 uses POE::Loop::Event
1151
1152=head3 Discussion
1153
1154The benchmark tries to test the performance of a typical small
1155server. While knowing how various event loops perform is interesting, keep
1156in mind that their overhead in this case is usually not as important, due
1157to the small absolute number of watchers (that is, you need efficiency and
1158speed most when you have lots of watchers, not when you only have a few of
1159them).
1160
1161EV is again fastest.
1162
1163The C-based event loops Event and Glib come in second this time, as the
1164overhead of running an iteration is much smaller in C than in Perl (little
1165code to execute in the inner loop, and perl's function calling overhead is
1166high, and updating all the data structures is costly).
1167
1168The pure perl event loop is much slower, but still competitive.
1169
1170POE also performs much better in this case, but is is still far behind the
1171others.
1172
1173=head3 Summary
1174
1175=over 4
1176
1177=item * C-based event loops perform very well with small number of
1178watchers, as the management overhead dominates.
1179
1180=back
1011 1181
1012 1182
1013=head1 FORK 1183=head1 FORK
1014 1184
1015Most event libraries are not fork-safe. The ones who are usually are 1185Most event libraries are not fork-safe. The ones who are usually are

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines