ViewVC Help
View File | Revision Log | Show Annotations | Download File
/cvs/BDB/BDB.pm
Revision: 1.4
Committed: Mon Feb 5 23:46:15 2007 UTC (17 years, 3 months ago) by root
Branch: MAIN
Changes since 1.3: +2 -1 lines
Log Message:
*** empty log message ***

File Contents

# User Rev Content
1 root 1.1 =head1 NAME
2    
3 root 1.2 BDB - Asynchronous Berkeley DB access
4 root 1.1
5     =head1 SYNOPSIS
6    
7 root 1.2 use BDB;
8 root 1.1
9     =head1 DESCRIPTION
10    
11     =head2 EXAMPLE
12    
13     =head1 REQUEST ANATOMY AND LIFETIME
14    
15     Every request method creates a request. which is a C data structure not
16     directly visible to Perl.
17    
18     During their existance, bdb requests travel through the following states,
19     in order:
20    
21     =over 4
22    
23     =item ready
24    
25     Immediately after a request is created it is put into the ready state,
26     waiting for a thread to execute it.
27    
28     =item execute
29    
30     A thread has accepted the request for processing and is currently
31     executing it (e.g. blocking in read).
32    
33     =item pending
34    
35     The request has been executed and is waiting for result processing.
36    
37     While request submission and execution is fully asynchronous, result
38     processing is not and relies on the perl interpreter calling C<poll_cb>
39     (or another function with the same effect).
40    
41     =item result
42    
43     The request results are processed synchronously by C<poll_cb>.
44    
45     The C<poll_cb> function will process all outstanding aio requests by
46     calling their callbacks, freeing memory associated with them and managing
47     any groups they are contained in.
48    
49     =item done
50    
51     Request has reached the end of its lifetime and holds no resources anymore
52     (except possibly for the Perl object, but its connection to the actual
53     aio request is severed and calling its methods will either do nothing or
54     result in a runtime error).
55    
56     =back
57    
58     =cut
59    
60 root 1.2 package BDB;
61 root 1.1
62     no warnings;
63     use strict 'vars';
64    
65     use base 'Exporter';
66    
67     BEGIN {
68     our $VERSION = '0.1';
69    
70 root 1.3 our @BDB_REQ = qw(
71     db_env_create db_env_open db_env_close
72 root 1.4 db_create db_open db_close db_compact db_sync db_put db_get db_pget
73     db_txn_commit db_txn_abort
74 root 1.3 );
75     our @EXPORT = (@BDB_REQ, qw(dbreq_pri dbreq_nice));
76 root 1.1 our @EXPORT_OK = qw(poll_fileno poll_cb poll_wait flush
77     min_parallel max_parallel max_idle
78     nreqs nready npending nthreads
79     max_poll_time max_poll_reqs);
80    
81     require XSLoader;
82 root 1.2 XSLoader::load ("BDB", $VERSION);
83 root 1.1 }
84    
85     =head2 SUPPORT FUNCTIONS
86    
87     =head3 EVENT PROCESSING AND EVENT LOOP INTEGRATION
88    
89     =over 4
90    
91 root 1.2 =item $fileno = BDB::poll_fileno
92 root 1.1
93     Return the I<request result pipe file descriptor>. This filehandle must be
94     polled for reading by some mechanism outside this module (e.g. Event or
95     select, see below or the SYNOPSIS). If the pipe becomes readable you have
96     to call C<poll_cb> to check the results.
97    
98     See C<poll_cb> for an example.
99    
100 root 1.2 =item BDB::poll_cb
101 root 1.1
102     Process some outstanding events on the result pipe. You have to call this
103     regularly. Returns the number of events processed. Returns immediately
104     when no events are outstanding. The amount of events processed depends on
105 root 1.2 the settings of C<BDB::max_poll_req> and C<BDB::max_poll_time>.
106 root 1.1
107     If not all requests were processed for whatever reason, the filehandle
108     will still be ready when C<poll_cb> returns.
109    
110     Example: Install an Event watcher that automatically calls
111 root 1.2 BDB::poll_cb with high priority:
112 root 1.1
113 root 1.2 Event->io (fd => BDB::poll_fileno,
114 root 1.1 poll => 'r', async => 1,
115 root 1.2 cb => \&BDB::poll_cb);
116 root 1.1
117 root 1.2 =item BDB::max_poll_reqs $nreqs
118 root 1.1
119 root 1.2 =item BDB::max_poll_time $seconds
120 root 1.1
121     These set the maximum number of requests (default C<0>, meaning infinity)
122 root 1.2 that are being processed by C<BDB::poll_cb> in one call, respectively
123 root 1.1 the maximum amount of time (default C<0>, meaning infinity) spent in
124 root 1.2 C<BDB::poll_cb> to process requests (more correctly the mininum amount
125 root 1.1 of time C<poll_cb> is allowed to use).
126    
127     Setting C<max_poll_time> to a non-zero value creates an overhead of one
128     syscall per request processed, which is not normally a problem unless your
129     callbacks are really really fast or your OS is really really slow (I am
130     not mentioning Solaris here). Using C<max_poll_reqs> incurs no overhead.
131    
132     Setting these is useful if you want to ensure some level of
133     interactiveness when perl is not fast enough to process all requests in
134     time.
135    
136     For interactive programs, values such as C<0.01> to C<0.1> should be fine.
137    
138     Example: Install an Event watcher that automatically calls
139 root 1.2 BDB::poll_cb with low priority, to ensure that other parts of the
140 root 1.1 program get the CPU sometimes even under high AIO load.
141    
142     # try not to spend much more than 0.1s in poll_cb
143 root 1.2 BDB::max_poll_time 0.1;
144 root 1.1
145     # use a low priority so other tasks have priority
146 root 1.2 Event->io (fd => BDB::poll_fileno,
147 root 1.1 poll => 'r', nice => 1,
148 root 1.2 cb => &BDB::poll_cb);
149 root 1.1
150 root 1.2 =item BDB::poll_wait
151 root 1.1
152     If there are any outstanding requests and none of them in the result
153     phase, wait till the result filehandle becomes ready for reading (simply
154     does a C<select> on the filehandle. This is useful if you want to
155     synchronously wait for some requests to finish).
156    
157     See C<nreqs> for an example.
158    
159 root 1.2 =item BDB::poll
160 root 1.1
161     Waits until some requests have been handled.
162    
163     Returns the number of requests processed, but is otherwise strictly
164     equivalent to:
165    
166 root 1.2 BDB::poll_wait, BDB::poll_cb
167 root 1.1
168 root 1.2 =item BDB::flush
169 root 1.1
170     Wait till all outstanding AIO requests have been handled.
171    
172     Strictly equivalent to:
173    
174 root 1.2 BDB::poll_wait, BDB::poll_cb
175     while BDB::nreqs;
176 root 1.1
177     =head3 CONTROLLING THE NUMBER OF THREADS
178    
179 root 1.2 =item BDB::min_parallel $nthreads
180 root 1.1
181     Set the minimum number of AIO threads to C<$nthreads>. The current
182     default is C<8>, which means eight asynchronous operations can execute
183     concurrently at any one time (the number of outstanding requests,
184     however, is unlimited).
185    
186 root 1.2 BDB starts threads only on demand, when an AIO request is queued and
187 root 1.1 no free thread exists. Please note that queueing up a hundred requests can
188     create demand for a hundred threads, even if it turns out that everything
189     is in the cache and could have been processed faster by a single thread.
190    
191     It is recommended to keep the number of threads relatively low, as some
192     Linux kernel versions will scale negatively with the number of threads
193     (higher parallelity => MUCH higher latency). With current Linux 2.6
194     versions, 4-32 threads should be fine.
195    
196     Under most circumstances you don't need to call this function, as the
197     module selects a default that is suitable for low to moderate load.
198    
199 root 1.2 =item BDB::max_parallel $nthreads
200 root 1.1
201     Sets the maximum number of AIO threads to C<$nthreads>. If more than the
202     specified number of threads are currently running, this function kills
203     them. This function blocks until the limit is reached.
204    
205     While C<$nthreads> are zero, aio requests get queued but not executed
206     until the number of threads has been increased again.
207    
208     This module automatically runs C<max_parallel 0> at program end, to ensure
209     that all threads are killed and that there are no outstanding requests.
210    
211     Under normal circumstances you don't need to call this function.
212    
213 root 1.2 =item BDB::max_idle $nthreads
214 root 1.1
215     Limit the number of threads (default: 4) that are allowed to idle (i.e.,
216     threads that did not get a request to process within 10 seconds). That
217     means if a thread becomes idle while C<$nthreads> other threads are also
218     idle, it will free its resources and exit.
219    
220     This is useful when you allow a large number of threads (e.g. 100 or 1000)
221     to allow for extremely high load situations, but want to free resources
222     under normal circumstances (1000 threads can easily consume 30MB of RAM).
223    
224     The default is probably ok in most situations, especially if thread
225     creation is fast. If thread creation is very slow on your system you might
226     want to use larger values.
227    
228 root 1.2 =item $oldmaxreqs = BDB::max_outstanding $maxreqs
229 root 1.1
230     This is a very bad function to use in interactive programs because it
231     blocks, and a bad way to reduce concurrency because it is inexact: Better
232     use an C<aio_group> together with a feed callback.
233    
234     Sets the maximum number of outstanding requests to C<$nreqs>. If you
235     to queue up more than this number of requests, the next call to the
236     C<poll_cb> (and C<poll_some> and other functions calling C<poll_cb>)
237     function will block until the limit is no longer exceeded.
238    
239     The default value is very large, so there is no practical limit on the
240     number of outstanding requests.
241    
242     You can still queue as many requests as you want. Therefore,
243     C<max_oustsanding> is mainly useful in simple scripts (with low values) or
244     as a stop gap to shield against fatal memory overflow (with large values).
245    
246 root 1.3 =item BDB::set_sync_prepare $cb
247    
248     Sets a callback that is called whenever a request is created without an
249     explicit callback. It has to return two code references. The first is used
250     as the request callback, and the second is called to wait until the first
251     callback has been called. The default implementation works like this:
252    
253     sub {
254     my $status;
255     (
256     sub { $status = $! },
257     sub { BDB::poll while !defined $status; $! = $status },
258     )
259     }
260    
261     =back
262    
263 root 1.1 =head3 STATISTICAL INFORMATION
264    
265 root 1.3 =over 4
266    
267 root 1.2 =item BDB::nreqs
268 root 1.1
269     Returns the number of requests currently in the ready, execute or pending
270     states (i.e. for which their callback has not been invoked yet).
271    
272     Example: wait till there are no outstanding requests anymore:
273    
274 root 1.2 BDB::poll_wait, BDB::poll_cb
275     while BDB::nreqs;
276 root 1.1
277 root 1.2 =item BDB::nready
278 root 1.1
279     Returns the number of requests currently in the ready state (not yet
280     executed).
281    
282 root 1.2 =item BDB::npending
283 root 1.1
284     Returns the number of requests currently in the pending state (executed,
285     but not yet processed by poll_cb).
286    
287     =back
288    
289     =cut
290    
291 root 1.3 set_sync_prepare {
292     my $status;
293     (
294     sub {
295     $status = $!;
296     },
297     sub {
298     BDB::poll while !defined $status;
299     $! = $status;
300     },
301     )
302     };
303    
304 root 1.1 min_parallel 8;
305    
306     END { flush }
307    
308     1;
309    
310     =head2 FORK BEHAVIOUR
311    
312     This module should do "the right thing" when the process using it forks:
313    
314     Before the fork, IO::AIO enters a quiescent state where no requests
315     can be added in other threads and no results will be processed. After
316     the fork the parent simply leaves the quiescent state and continues
317     request/result processing, while the child frees the request/result queue
318     (so that the requests started before the fork will only be handled in the
319     parent). Threads will be started on demand until the limit set in the
320     parent process has been reached again.
321    
322     In short: the parent will, after a short pause, continue as if fork had
323     not been called, while the child will act as if IO::AIO has not been used
324     yet.
325    
326     =head2 MEMORY USAGE
327    
328     Per-request usage:
329    
330     Each aio request uses - depending on your architecture - around 100-200
331     bytes of memory. In addition, stat requests need a stat buffer (possibly
332     a few hundred bytes), readdir requires a result buffer and so on. Perl
333     scalars and other data passed into aio requests will also be locked and
334     will consume memory till the request has entered the done state.
335    
336     This is now awfully much, so queuing lots of requests is not usually a
337     problem.
338    
339     Per-thread usage:
340    
341     In the execution phase, some aio requests require more memory for
342     temporary buffers, and each thread requires a stack and other data
343     structures (usually around 16k-128k, depending on the OS).
344    
345     =head1 KNOWN BUGS
346    
347     Known bugs will be fixed in the next release.
348    
349     =head1 SEE ALSO
350    
351     L<Coro::AIO>.
352    
353     =head1 AUTHOR
354    
355     Marc Lehmann <schmorp@schmorp.de>
356     http://home.schmorp.de/
357    
358     =cut
359