ViewVC Help
View File | Revision Log | Show Annotations | Download File
/cvs/libev/ev.pod
(Generate patch)

Comparing libev/ev.pod (file contents):
Revision 1.240 by root, Sat Apr 25 14:12:48 2009 UTC vs.
Revision 1.254 by root, Tue Jul 14 19:02:43 2009 UTC

620happily wraps around with enough iterations. 620happily wraps around with enough iterations.
621 621
622This value can sometimes be useful as a generation counter of sorts (it 622This value can sometimes be useful as a generation counter of sorts (it
623"ticks" the number of loop iterations), as it roughly corresponds with 623"ticks" the number of loop iterations), as it roughly corresponds with
624C<ev_prepare> and C<ev_check> calls. 624C<ev_prepare> and C<ev_check> calls.
625
626=item unsigned int ev_loop_depth (loop)
627
628Returns the number of times C<ev_loop> was entered minus the number of
629times C<ev_loop> was exited, in other words, the recursion depth.
630
631Outside C<ev_loop>, this number is zero. In a callback, this number is
632C<1>, unless C<ev_loop> was invoked recursively (or from another thread),
633in which case it is higher.
634
635Leaving C<ev_loop> abnormally (setjmp/longjmp, cancelling the thread
636etc.), doesn't count as exit.
625 637
626=item unsigned int ev_backend (loop) 638=item unsigned int ev_backend (loop)
627 639
628Returns one of the C<EVBACKEND_*> flags indicating the event backend in 640Returns one of the C<EVBACKEND_*> flags indicating the event backend in
629use. 641use.
811 823
812By setting a higher I<io collect interval> you allow libev to spend more 824By setting a higher I<io collect interval> you allow libev to spend more
813time collecting I/O events, so you can handle more events per iteration, 825time collecting I/O events, so you can handle more events per iteration,
814at the cost of increasing latency. Timeouts (both C<ev_periodic> and 826at the cost of increasing latency. Timeouts (both C<ev_periodic> and
815C<ev_timer>) will be not affected. Setting this to a non-null value will 827C<ev_timer>) will be not affected. Setting this to a non-null value will
816introduce an additional C<ev_sleep ()> call into most loop iterations. 828introduce an additional C<ev_sleep ()> call into most loop iterations. The
829sleep time ensures that libev will not poll for I/O events more often then
830once per this interval, on average.
817 831
818Likewise, by setting a higher I<timeout collect interval> you allow libev 832Likewise, by setting a higher I<timeout collect interval> you allow libev
819to spend more time collecting timeouts, at the expense of increased 833to spend more time collecting timeouts, at the expense of increased
820latency/jitter/inexactness (the watcher callback will be called 834latency/jitter/inexactness (the watcher callback will be called
821later). C<ev_io> watchers will not be affected. Setting this to a non-null 835later). C<ev_io> watchers will not be affected. Setting this to a non-null
823 837
824Many (busy) programs can usually benefit by setting the I/O collect 838Many (busy) programs can usually benefit by setting the I/O collect
825interval to a value near C<0.1> or so, which is often enough for 839interval to a value near C<0.1> or so, which is often enough for
826interactive servers (of course not for games), likewise for timeouts. It 840interactive servers (of course not for games), likewise for timeouts. It
827usually doesn't make much sense to set it to a lower value than C<0.01>, 841usually doesn't make much sense to set it to a lower value than C<0.01>,
828as this approaches the timing granularity of most systems. 842as this approaches the timing granularity of most systems. Note that if
843you do transactions with the outside world and you can't increase the
844parallelity, then this setting will limit your transaction rate (if you
845need to poll once per transaction and the I/O collect interval is 0.01,
846then you can't do more than 100 transations per second).
829 847
830Setting the I<timeout collect interval> can improve the opportunity for 848Setting the I<timeout collect interval> can improve the opportunity for
831saving power, as the program will "bundle" timer callback invocations that 849saving power, as the program will "bundle" timer callback invocations that
832are "near" in time together, by delaying some, thus reducing the number of 850are "near" in time together, by delaying some, thus reducing the number of
833times the process sleeps and wakes up again. Another useful technique to 851times the process sleeps and wakes up again. Another useful technique to
834reduce iterations/wake-ups is to use C<ev_periodic> watchers and make sure 852reduce iterations/wake-ups is to use C<ev_periodic> watchers and make sure
835they fire on, say, one-second boundaries only. 853they fire on, say, one-second boundaries only.
854
855Example: we only need 0.1s timeout granularity, and we wish not to poll
856more often than 100 times per second:
857
858 ev_set_timeout_collect_interval (EV_DEFAULT_UC_ 0.1);
859 ev_set_io_collect_interval (EV_DEFAULT_UC_ 0.01);
860
861=item ev_invoke_pending (loop)
862
863This call will simply invoke all pending watchers while resetting their
864pending state. Normally, C<ev_loop> does this automatically when required,
865but when overriding the invoke callback this call comes handy.
866
867=item ev_set_invoke_pending_cb (loop, void (*invoke_pending_cb)(EV_P))
868
869This overrides the invoke pending functionality of the loop: Instead of
870invoking all pending watchers when there are any, C<ev_loop> will call
871this callback instead. This is useful, for example, when you want to
872invoke the actual watchers inside another context (another thread etc.).
873
874If you want to reset the callback, use C<ev_invoke_pending> as new
875callback.
876
877=item ev_set_loop_release_cb (loop, void (*release)(EV_P), void (*acquire)(EV_P))
878
879Sometimes you want to share the same loop between multiple threads. This
880can be done relatively simply by putting mutex_lock/unlock calls around
881each call to a libev function.
882
883However, C<ev_loop> can run an indefinite time, so it is not feasible to
884wait for it to return. One way around this is to wake up the loop via
885C<ev_unloop> and C<av_async_send>, another way is to set these I<release>
886and I<acquire> callbacks on the loop.
887
888When set, then C<release> will be called just before the thread is
889suspended waiting for new events, and C<acquire> is called just
890afterwards.
891
892Ideally, C<release> will just call your mutex_unlock function, and
893C<acquire> will just call the mutex_lock function again.
894
895While event loop modifications are allowed between invocations of
896C<release> and C<acquire> (that's their only purpose after all), no
897modifications done will affect the event loop, i.e. adding watchers will
898have no effect on the set of file descriptors being watched, or the time
899waited. USe an C<ev_async> watcher to wake up C<ev_loop> when you want it
900to take note of any changes you made.
901
902In theory, threads executing C<ev_loop> will be async-cancel safe between
903invocations of C<release> and C<acquire>.
904
905See also the locking example in the C<THREADS> section later in this
906document.
907
908=item ev_set_userdata (loop, void *data)
909
910=item ev_userdata (loop)
911
912Set and retrieve a single C<void *> associated with a loop. When
913C<ev_set_userdata> has never been called, then C<ev_userdata> returns
914C<0.>
915
916These two functions can be used to associate arbitrary data with a loop,
917and are intended solely for the C<invoke_pending_cb>, C<release> and
918C<acquire> callbacks described above, but of course can be (ab-)used for
919any other purpose as well.
836 920
837=item ev_loop_verify (loop) 921=item ev_loop_verify (loop)
838 922
839This function only does something when C<EV_VERIFY> support has been 923This function only does something when C<EV_VERIFY> support has been
840compiled in, which is the default for non-minimal builds. It tries to go 924compiled in, which is the default for non-minimal builds. It tries to go
1184 #include <stddef.h> 1268 #include <stddef.h>
1185 1269
1186 static void 1270 static void
1187 t1_cb (EV_P_ ev_timer *w, int revents) 1271 t1_cb (EV_P_ ev_timer *w, int revents)
1188 { 1272 {
1189 struct my_biggy big = (struct my_biggy * 1273 struct my_biggy big = (struct my_biggy *)
1190 (((char *)w) - offsetof (struct my_biggy, t1)); 1274 (((char *)w) - offsetof (struct my_biggy, t1));
1191 } 1275 }
1192 1276
1193 static void 1277 static void
1194 t2_cb (EV_P_ ev_timer *w, int revents) 1278 t2_cb (EV_P_ ev_timer *w, int revents)
1195 { 1279 {
1196 struct my_biggy big = (struct my_biggy * 1280 struct my_biggy big = (struct my_biggy *)
1197 (((char *)w) - offsetof (struct my_biggy, t2)); 1281 (((char *)w) - offsetof (struct my_biggy, t2));
1198 } 1282 }
1199 1283
1200=head2 WATCHER PRIORITY MODELS 1284=head2 WATCHER PRIORITY MODELS
1201 1285
1277 // with the default priority are receiving events. 1361 // with the default priority are receiving events.
1278 ev_idle_start (EV_A_ &idle); 1362 ev_idle_start (EV_A_ &idle);
1279 } 1363 }
1280 1364
1281 static void 1365 static void
1282 idle-cb (EV_P_ ev_idle *w, int revents) 1366 idle_cb (EV_P_ ev_idle *w, int revents)
1283 { 1367 {
1284 // actual processing 1368 // actual processing
1285 read (STDIN_FILENO, ...); 1369 read (STDIN_FILENO, ...);
1286 1370
1287 // have to start the I/O watcher again, as 1371 // have to start the I/O watcher again, as
1468 1552
1469The callback is guaranteed to be invoked only I<after> its timeout has 1553The callback is guaranteed to be invoked only I<after> its timeout has
1470passed (not I<at>, so on systems with very low-resolution clocks this 1554passed (not I<at>, so on systems with very low-resolution clocks this
1471might introduce a small delay). If multiple timers become ready during the 1555might introduce a small delay). If multiple timers become ready during the
1472same loop iteration then the ones with earlier time-out values are invoked 1556same loop iteration then the ones with earlier time-out values are invoked
1473before ones with later time-out values (but this is no longer true when a 1557before ones of the same priority with later time-out values (but this is
1474callback calls C<ev_loop> recursively). 1558no longer true when a callback calls C<ev_loop> recursively).
1475 1559
1476=head3 Be smart about timeouts 1560=head3 Be smart about timeouts
1477 1561
1478Many real-world problems involve some kind of timeout, usually for error 1562Many real-world problems involve some kind of timeout, usually for error
1479recovery. A typical example is an HTTP request - if the other side hangs, 1563recovery. A typical example is an HTTP request - if the other side hangs,
1523C<after> argument to C<ev_timer_set>, and only ever use the C<repeat> 1607C<after> argument to C<ev_timer_set>, and only ever use the C<repeat>
1524member and C<ev_timer_again>. 1608member and C<ev_timer_again>.
1525 1609
1526At start: 1610At start:
1527 1611
1528 ev_timer_init (timer, callback); 1612 ev_init (timer, callback);
1529 timer->repeat = 60.; 1613 timer->repeat = 60.;
1530 ev_timer_again (loop, timer); 1614 ev_timer_again (loop, timer);
1531 1615
1532Each time there is some activity: 1616Each time there is some activity:
1533 1617
1595 1679
1596To start the timer, simply initialise the watcher and set C<last_activity> 1680To start the timer, simply initialise the watcher and set C<last_activity>
1597to the current time (meaning we just have some activity :), then call the 1681to the current time (meaning we just have some activity :), then call the
1598callback, which will "do the right thing" and start the timer: 1682callback, which will "do the right thing" and start the timer:
1599 1683
1600 ev_timer_init (timer, callback); 1684 ev_init (timer, callback);
1601 last_activity = ev_now (loop); 1685 last_activity = ev_now (loop);
1602 callback (loop, timer, EV_TIMEOUT); 1686 callback (loop, timer, EV_TIMEOUT);
1603 1687
1604And when there is some activity, simply store the current time in 1688And when there is some activity, simply store the current time in
1605C<last_activity>, no libev calls at all: 1689C<last_activity>, no libev calls at all:
2002some child status changes (most typically when a child of yours dies or 2086some child status changes (most typically when a child of yours dies or
2003exits). It is permissible to install a child watcher I<after> the child 2087exits). It is permissible to install a child watcher I<after> the child
2004has been forked (which implies it might have already exited), as long 2088has been forked (which implies it might have already exited), as long
2005as the event loop isn't entered (or is continued from a watcher), i.e., 2089as the event loop isn't entered (or is continued from a watcher), i.e.,
2006forking and then immediately registering a watcher for the child is fine, 2090forking and then immediately registering a watcher for the child is fine,
2007but forking and registering a watcher a few event loop iterations later is 2091but forking and registering a watcher a few event loop iterations later or
2008not. 2092in the next callback invocation is not.
2009 2093
2010Only the default event loop is capable of handling signals, and therefore 2094Only the default event loop is capable of handling signals, and therefore
2011you can only register child watchers in the default event loop. 2095you can only register child watchers in the default event loop.
2096
2097Due to some design glitches inside libev, child watchers will always be
2098handled at maximum priority (their priority is set to C<EV_MAXPRI> by
2099libev)
2012 2100
2013=head3 Process Interaction 2101=head3 Process Interaction
2014 2102
2015Libev grabs C<SIGCHLD> as soon as the default event loop is 2103Libev grabs C<SIGCHLD> as soon as the default event loop is
2016initialised. This is necessary to guarantee proper behaviour even if 2104initialised. This is necessary to guarantee proper behaviour even if
2368 // no longer anything immediate to do. 2456 // no longer anything immediate to do.
2369 } 2457 }
2370 2458
2371 ev_idle *idle_watcher = malloc (sizeof (ev_idle)); 2459 ev_idle *idle_watcher = malloc (sizeof (ev_idle));
2372 ev_idle_init (idle_watcher, idle_cb); 2460 ev_idle_init (idle_watcher, idle_cb);
2373 ev_idle_start (loop, idle_cb); 2461 ev_idle_start (loop, idle_watcher);
2374 2462
2375 2463
2376=head2 C<ev_prepare> and C<ev_check> - customise your event loop! 2464=head2 C<ev_prepare> and C<ev_check> - customise your event loop!
2377 2465
2378Prepare and check watchers are usually (but not always) used in pairs: 2466Prepare and check watchers are usually (but not always) used in pairs:
2471 struct pollfd fds [nfd]; 2559 struct pollfd fds [nfd];
2472 // actual code will need to loop here and realloc etc. 2560 // actual code will need to loop here and realloc etc.
2473 adns_beforepoll (ads, fds, &nfd, &timeout, timeval_from (ev_time ())); 2561 adns_beforepoll (ads, fds, &nfd, &timeout, timeval_from (ev_time ()));
2474 2562
2475 /* the callback is illegal, but won't be called as we stop during check */ 2563 /* the callback is illegal, but won't be called as we stop during check */
2476 ev_timer_init (&tw, 0, timeout * 1e-3); 2564 ev_timer_init (&tw, 0, timeout * 1e-3, 0.);
2477 ev_timer_start (loop, &tw); 2565 ev_timer_start (loop, &tw);
2478 2566
2479 // create one ev_io per pollfd 2567 // create one ev_io per pollfd
2480 for (int i = 0; i < nfd; ++i) 2568 for (int i = 0; i < nfd; ++i)
2481 { 2569 {
3643defined to be C<0>, then they are not. 3731defined to be C<0>, then they are not.
3644 3732
3645=item EV_MINIMAL 3733=item EV_MINIMAL
3646 3734
3647If you need to shave off some kilobytes of code at the expense of some 3735If you need to shave off some kilobytes of code at the expense of some
3648speed, define this symbol to C<1>. Currently this is used to override some 3736speed (but with the full API), define this symbol to C<1>. Currently this
3649inlining decisions, saves roughly 30% code size on amd64. It also selects a 3737is used to override some inlining decisions, saves roughly 30% code size
3650much smaller 2-heap for timer management over the default 4-heap. 3738on amd64. It also selects a much smaller 2-heap for timer management over
3739the default 4-heap.
3740
3741You can save even more by disabling watcher types you do not need
3742and setting C<EV_MAXPRI> == C<EV_MINPRI>. Also, disabling C<assert>
3743(C<-DNDEBUG>) will usually reduce code size a lot.
3744
3745Defining C<EV_MINIMAL> to C<2> will additionally reduce the core API to
3746provide a bare-bones event library. See C<ev.h> for details on what parts
3747of the API are still available, and do not complain if this subset changes
3748over time.
3651 3749
3652=item EV_PID_HASHSIZE 3750=item EV_PID_HASHSIZE
3653 3751
3654C<ev_child> watchers use a small hash table to distribute workload by 3752C<ev_child> watchers use a small hash table to distribute workload by
3655pid. The default size is C<16> (or C<1> with C<EV_MINIMAL>), usually more 3753pid. The default size is C<16> (or C<1> with C<EV_MINIMAL>), usually more
3841default loop and triggering an C<ev_async> watcher from the default loop 3939default loop and triggering an C<ev_async> watcher from the default loop
3842watcher callback into the event loop interested in the signal. 3940watcher callback into the event loop interested in the signal.
3843 3941
3844=back 3942=back
3845 3943
3944=head4 THREAD LOCKING EXAMPLE
3945
3946Here is a fictitious example of how to run an event loop in a different
3947thread than where callbacks are being invoked and watchers are
3948created/added/removed.
3949
3950For a real-world example, see the C<EV::Loop::Async> perl module,
3951which uses exactly this technique (which is suited for many high-level
3952languages).
3953
3954The example uses a pthread mutex to protect the loop data, a condition
3955variable to wait for callback invocations, an async watcher to notify the
3956event loop thread and an unspecified mechanism to wake up the main thread.
3957
3958First, you need to associate some data with the event loop:
3959
3960 typedef struct {
3961 mutex_t lock; /* global loop lock */
3962 ev_async async_w;
3963 thread_t tid;
3964 cond_t invoke_cv;
3965 } userdata;
3966
3967 void prepare_loop (EV_P)
3968 {
3969 // for simplicity, we use a static userdata struct.
3970 static userdata u;
3971
3972 ev_async_init (&u->async_w, async_cb);
3973 ev_async_start (EV_A_ &u->async_w);
3974
3975 pthread_mutex_init (&u->lock, 0);
3976 pthread_cond_init (&u->invoke_cv, 0);
3977
3978 // now associate this with the loop
3979 ev_set_userdata (EV_A_ u);
3980 ev_set_invoke_pending_cb (EV_A_ l_invoke);
3981 ev_set_loop_release_cb (EV_A_ l_release, l_acquire);
3982
3983 // then create the thread running ev_loop
3984 pthread_create (&u->tid, 0, l_run, EV_A);
3985 }
3986
3987The callback for the C<ev_async> watcher does nothing: the watcher is used
3988solely to wake up the event loop so it takes notice of any new watchers
3989that might have been added:
3990
3991 static void
3992 async_cb (EV_P_ ev_async *w, int revents)
3993 {
3994 // just used for the side effects
3995 }
3996
3997The C<l_release> and C<l_acquire> callbacks simply unlock/lock the mutex
3998protecting the loop data, respectively.
3999
4000 static void
4001 l_release (EV_P)
4002 {
4003 udat *u = ev_userdata (EV_A);
4004 pthread_mutex_unlock (&u->lock);
4005 }
4006
4007 static void
4008 l_acquire (EV_P)
4009 {
4010 udat *u = ev_userdata (EV_A);
4011 pthread_mutex_lock (&u->lock);
4012 }
4013
4014The event loop thread first acquires the mutex, and then jumps straight
4015into C<ev_loop>:
4016
4017 void *
4018 l_run (void *thr_arg)
4019 {
4020 struct ev_loop *loop = (struct ev_loop *)thr_arg;
4021
4022 l_acquire (EV_A);
4023 pthread_setcanceltype (PTHREAD_CANCEL_ASYNCHRONOUS, 0);
4024 ev_loop (EV_A_ 0);
4025 l_release (EV_A);
4026
4027 return 0;
4028 }
4029
4030Instead of invoking all pending watchers, the C<l_invoke> callback will
4031signal the main thread via some unspecified mechanism (signals? pipe
4032writes? C<Async::Interrupt>?) and then waits until all pending watchers
4033have been called:
4034
4035 static void
4036 l_invoke (EV_P)
4037 {
4038 udat *u = ev_userdata (EV_A);
4039
4040 wake_up_other_thread_in_some_magic_or_not_so_magic_way ();
4041
4042 pthread_cond_wait (&u->invoke_cv, &u->lock);
4043 }
4044
4045Now, whenever the main thread gets told to invoke pending watchers, it
4046will grab the lock, call C<ev_invoke_pending> and then signal the loop
4047thread to continue:
4048
4049 static void
4050 real_invoke_pending (EV_P)
4051 {
4052 udat *u = ev_userdata (EV_A);
4053
4054 pthread_mutex_lock (&u->lock);
4055 ev_invoke_pending (EV_A);
4056 pthread_cond_signal (&u->invoke_cv);
4057 pthread_mutex_unlock (&u->lock);
4058 }
4059
4060Whenever you want to start/stop a watcher or do other modifications to an
4061event loop, you will now have to lock:
4062
4063 ev_timer timeout_watcher;
4064 udat *u = ev_userdata (EV_A);
4065
4066 ev_timer_init (&timeout_watcher, timeout_cb, 5.5, 0.);
4067
4068 pthread_mutex_lock (&u->lock);
4069 ev_timer_start (EV_A_ &timeout_watcher);
4070 ev_async_send (EV_A_ &u->async_w);
4071 pthread_mutex_unlock (&u->lock);
4072
4073Note that sending the C<ev_async> watcher is required because otherwise
4074an event loop currently blocking in the kernel will have no knowledge
4075about the newly added timer. By waking up the loop it will pick up any new
4076watchers in the next event loop iteration.
4077
3846=head3 COROUTINES 4078=head3 COROUTINES
3847 4079
3848Libev is very accommodating to coroutines ("cooperative threads"): 4080Libev is very accommodating to coroutines ("cooperative threads"):
3849libev fully supports nesting calls to its functions from different 4081libev fully supports nesting calls to its functions from different
3850coroutines (e.g. you can call C<ev_loop> on the same loop from two 4082coroutines (e.g. you can call C<ev_loop> on the same loop from two
3935way (note also that glib is the slowest event library known to man). 4167way (note also that glib is the slowest event library known to man).
3936 4168
3937There is no supported compilation method available on windows except 4169There is no supported compilation method available on windows except
3938embedding it into other applications. 4170embedding it into other applications.
3939 4171
4172Sensible signal handling is officially unsupported by Microsoft - libev
4173tries its best, but under most conditions, signals will simply not work.
4174
3940Not a libev limitation but worth mentioning: windows apparently doesn't 4175Not a libev limitation but worth mentioning: windows apparently doesn't
3941accept large writes: instead of resulting in a partial write, windows will 4176accept large writes: instead of resulting in a partial write, windows will
3942either accept everything or return C<ENOBUFS> if the buffer is too large, 4177either accept everything or return C<ENOBUFS> if the buffer is too large,
3943so make sure you only write small amounts into your sockets (less than a 4178so make sure you only write small amounts into your sockets (less than a
3944megabyte seems safe, but this apparently depends on the amount of memory 4179megabyte seems safe, but this apparently depends on the amount of memory
3948the abysmal performance of winsockets, using a large number of sockets 4183the abysmal performance of winsockets, using a large number of sockets
3949is not recommended (and not reasonable). If your program needs to use 4184is not recommended (and not reasonable). If your program needs to use
3950more than a hundred or so sockets, then likely it needs to use a totally 4185more than a hundred or so sockets, then likely it needs to use a totally
3951different implementation for windows, as libev offers the POSIX readiness 4186different implementation for windows, as libev offers the POSIX readiness
3952notification model, which cannot be implemented efficiently on windows 4187notification model, which cannot be implemented efficiently on windows
3953(Microsoft monopoly games). 4188(due to Microsoft monopoly games).
3954 4189
3955A typical way to use libev under windows is to embed it (see the embedding 4190A typical way to use libev under windows is to embed it (see the embedding
3956section for details) and use the following F<evwrap.h> header file instead 4191section for details) and use the following F<evwrap.h> header file instead
3957of F<ev.h>: 4192of F<ev.h>:
3958 4193
3994 4229
3995Early versions of winsocket's select only supported waiting for a maximum 4230Early versions of winsocket's select only supported waiting for a maximum
3996of C<64> handles (probably owning to the fact that all windows kernels 4231of C<64> handles (probably owning to the fact that all windows kernels
3997can only wait for C<64> things at the same time internally; Microsoft 4232can only wait for C<64> things at the same time internally; Microsoft
3998recommends spawning a chain of threads and wait for 63 handles and the 4233recommends spawning a chain of threads and wait for 63 handles and the
3999previous thread in each. Great). 4234previous thread in each. Sounds great!).
4000 4235
4001Newer versions support more handles, but you need to define C<FD_SETSIZE> 4236Newer versions support more handles, but you need to define C<FD_SETSIZE>
4002to some high number (e.g. C<2048>) before compiling the winsocket select 4237to some high number (e.g. C<2048>) before compiling the winsocket select
4003call (which might be in libev or elsewhere, for example, perl does its own 4238call (which might be in libev or elsewhere, for example, perl and many
4004select emulation on windows). 4239other interpreters do their own select emulation on windows).
4005 4240
4006Another limit is the number of file descriptors in the Microsoft runtime 4241Another limit is the number of file descriptors in the Microsoft runtime
4007libraries, which by default is C<64> (there must be a hidden I<64> fetish 4242libraries, which by default is C<64> (there must be a hidden I<64>
4008or something like this inside Microsoft). You can increase this by calling 4243fetish or something like this inside Microsoft). You can increase this
4009C<_setmaxstdio>, which can increase this limit to C<2048> (another 4244by calling C<_setmaxstdio>, which can increase this limit to C<2048>
4010arbitrary limit), but is broken in many versions of the Microsoft runtime 4245(another arbitrary limit), but is broken in many versions of the Microsoft
4011libraries.
4012
4013This might get you to about C<512> or C<2048> sockets (depending on 4246runtime libraries. This might get you to about C<512> or C<2048> sockets
4014windows version and/or the phase of the moon). To get more, you need to 4247(depending on windows version and/or the phase of the moon). To get more,
4015wrap all I/O functions and provide your own fd management, but the cost of 4248you need to wrap all I/O functions and provide your own fd management, but
4016calling select (O(n²)) will likely make this unworkable. 4249the cost of calling select (O(n²)) will likely make this unworkable.
4017 4250
4018=back 4251=back
4019 4252
4020=head2 PORTABILITY REQUIREMENTS 4253=head2 PORTABILITY REQUIREMENTS
4021 4254
4064=item C<double> must hold a time value in seconds with enough accuracy 4297=item C<double> must hold a time value in seconds with enough accuracy
4065 4298
4066The type C<double> is used to represent timestamps. It is required to 4299The type C<double> is used to represent timestamps. It is required to
4067have at least 51 bits of mantissa (and 9 bits of exponent), which is good 4300have at least 51 bits of mantissa (and 9 bits of exponent), which is good
4068enough for at least into the year 4000. This requirement is fulfilled by 4301enough for at least into the year 4000. This requirement is fulfilled by
4069implementations implementing IEEE 754 (basically all existing ones). 4302implementations implementing IEEE 754, which is basically all existing
4303ones. With IEEE 754 doubles, you get microsecond accuracy until at least
43042200.
4070 4305
4071=back 4306=back
4072 4307
4073If you know of other additional requirements drop me a note. 4308If you know of other additional requirements drop me a note.
4074 4309

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines