ViewVC Help
View File | Revision Log | Show Annotations | Download File
/cvs/IO-AIO/AIO.xs
(Generate patch)

Comparing IO-AIO/AIO.xs (file contents):
Revision 1.53 by root, Mon Oct 23 00:17:07 2006 UTC vs.
Revision 1.76 by root, Thu Oct 26 12:38:04 2006 UTC

1/* solaris */
2#define _POSIX_PTHREAD_SEMANTICS 1
3
4#if __linux && !defined(_GNU_SOURCE)
5# define _GNU_SOURCE
6#endif
7
8/* just in case */
1#define _REENTRANT 1 9#define _REENTRANT 1
10
2#include <errno.h> 11#include <errno.h>
3 12
4#include "EXTERN.h" 13#include "EXTERN.h"
5#include "perl.h" 14#include "perl.h"
6#include "XSUB.h" 15#include "XSUB.h"
39/* used for struct dirent, AIX doesn't provide it */ 48/* used for struct dirent, AIX doesn't provide it */
40#ifndef NAME_MAX 49#ifndef NAME_MAX
41# define NAME_MAX 4096 50# define NAME_MAX 4096
42#endif 51#endif
43 52
53#ifndef PTHREAD_STACK_MIN
54/* care for broken platforms, e.g. windows */
55# define PTHREAD_STACK_MIN 16384
56#endif
57
44#if __ia64 58#if __ia64
45# define STACKSIZE 65536 59# define STACKSIZE 65536
60#elif __i386 || __x86_64 /* 16k is unreasonably high :( */
61# define STACKSIZE PTHREAD_STACK_MIN
46#else 62#else
47# define STACKSIZE 8192 63# define STACKSIZE 16384
48#endif 64#endif
65
66/* buffer size for various temporary buffers */
67#define AIO_BUFSIZE 65536
68
69#define dBUF \
70 char *aio_buf; \
71 LOCK (wrklock); \
72 self->dbuf = aio_buf = malloc (AIO_BUFSIZE); \
73 UNLOCK (wrklock); \
74 if (!aio_buf) \
75 return -1;
49 76
50enum { 77enum {
51 REQ_QUIT, 78 REQ_QUIT,
52 REQ_OPEN, REQ_CLOSE, 79 REQ_OPEN, REQ_CLOSE,
53 REQ_READ, REQ_WRITE, REQ_READAHEAD, 80 REQ_READ, REQ_WRITE, REQ_READAHEAD,
55 REQ_STAT, REQ_LSTAT, REQ_FSTAT, 82 REQ_STAT, REQ_LSTAT, REQ_FSTAT,
56 REQ_FSYNC, REQ_FDATASYNC, 83 REQ_FSYNC, REQ_FDATASYNC,
57 REQ_UNLINK, REQ_RMDIR, REQ_RENAME, 84 REQ_UNLINK, REQ_RMDIR, REQ_RENAME,
58 REQ_READDIR, 85 REQ_READDIR,
59 REQ_LINK, REQ_SYMLINK, 86 REQ_LINK, REQ_SYMLINK,
60 REQ_SLEEP, 87 REQ_GROUP, REQ_NOP,
61 REQ_GROUP, 88 REQ_BUSY,
62}; 89};
63 90
64#define AIO_REQ_KLASS "IO::AIO::REQ" 91#define AIO_REQ_KLASS "IO::AIO::REQ"
65#define AIO_GRP_KLASS "IO::AIO::GRP" 92#define AIO_GRP_KLASS "IO::AIO::GRP"
66 93
67typedef struct aio_cb 94typedef struct aio_cb
68{ 95{
69 struct aio_cb *volatile next; 96 struct aio_cb *volatile next;
70
71 struct aio_cb *grp, *grp_prev, *grp_next, *grp_first;
72
73 SV *self; /* the perl counterpart of this request, if any */
74 97
75 SV *data, *callback; 98 SV *data, *callback;
76 SV *fh, *fh2; 99 SV *fh, *fh2;
77 void *dataptr, *data2ptr; 100 void *dataptr, *data2ptr;
78 Stat_t *statdata; 101 Stat_t *statdata;
79 off_t offset; 102 off_t offset;
80 size_t length; 103 size_t length;
81 ssize_t result; 104 ssize_t result;
82 105
106 STRLEN dataoffset;
83 int type; 107 int type;
84 int fd, fd2; 108 int fd, fd2;
85 int errorno; 109 int errorno;
86 STRLEN dataoffset;
87 mode_t mode; /* open */ 110 mode_t mode; /* open */
111
88 unsigned char cancelled; 112 unsigned char flags;
113 unsigned char pri;
114
115 SV *self; /* the perl counterpart of this request, if any */
116 struct aio_cb *grp, *grp_prev, *grp_next, *grp_first;
89} aio_cb; 117} aio_cb;
118
119enum {
120 FLAG_CANCELLED = 0x01,
121};
90 122
91typedef aio_cb *aio_req; 123typedef aio_cb *aio_req;
92typedef aio_cb *aio_req_ornot; 124typedef aio_cb *aio_req_ornot;
93 125
126enum {
127 PRI_MIN = -4,
128 PRI_MAX = 4,
129
130 DEFAULT_PRI = 0,
131 PRI_BIAS = -PRI_MIN,
132 NUM_PRI = PRI_MAX + PRI_BIAS + 1,
133};
134
135static int next_pri = DEFAULT_PRI + PRI_BIAS;
136
94static int started, wanted; 137static int started, wanted;
95static volatile int nreqs; 138static volatile int nreqs;
96static int max_outstanding = 1<<30;
97static int respipe [2]; 139static int respipe [2];
98 140
141#if __linux && defined (PTHREAD_ADAPTIVE_MUTEX_INITIALIZER_NP)
142# define AIO_MUTEX_INIT PTHREAD_ADAPTIVE_MUTEX_INITIALIZER_NP
143#else
144# define AIO_MUTEX_INIT PTHREAD_MUTEX_INITIALIZER
145#endif
146
147#define LOCK(mutex) pthread_mutex_lock (&(mutex))
148#define UNLOCK(mutex) pthread_mutex_unlock (&(mutex))
149
150/* worker threads management */
151static pthread_mutex_t wrklock = AIO_MUTEX_INIT;
152
153typedef struct worker {
154 /* locked by wrklock */
155 struct worker *prev, *next;
156
157 pthread_t tid;
158
159 /* locked by reslock, reqlock or wrklock */
160 aio_req req; /* currently processed request */
161 void *dbuf;
162 DIR *dirp;
163} worker;
164
165static worker wrk_first = { &wrk_first, &wrk_first, 0 };
166
167static void worker_clear (worker *wrk)
168{
169 if (wrk->dirp)
170 {
171 closedir (wrk->dirp);
172 wrk->dirp = 0;
173 }
174
175 if (wrk->dbuf)
176 {
177 free (wrk->dbuf);
178 wrk->dbuf = 0;
179 }
180}
181
182static void worker_free (worker *wrk)
183{
184 wrk->next->prev = wrk->prev;
185 wrk->prev->next = wrk->next;
186
187 free (wrk);
188}
189
99static pthread_mutex_t reslock = PTHREAD_MUTEX_INITIALIZER; 190static pthread_mutex_t reslock = AIO_MUTEX_INIT;
100static pthread_mutex_t reqlock = PTHREAD_MUTEX_INITIALIZER; 191static pthread_mutex_t reqlock = AIO_MUTEX_INIT;
101static pthread_cond_t reqwait = PTHREAD_COND_INITIALIZER; 192static pthread_cond_t reqwait = PTHREAD_COND_INITIALIZER;
102 193
103static volatile aio_req reqs, reqe; /* queue start, queue end */ 194/*
104static volatile aio_req ress, rese; /* queue start, queue end */ 195 * a somewhat faster data structure might be nice, but
196 * with 8 priorities this actually needs <20 insns
197 * per shift, the most expensive operation.
198 */
199typedef struct {
200 aio_req qs[NUM_PRI], qe[NUM_PRI]; /* qstart, qend */
201 int size;
202} reqq;
105 203
204static reqq req_queue;
205static reqq res_queue;
206
207int reqq_push (reqq *q, aio_req req)
208{
209 int pri = req->pri;
210 req->next = 0;
211
212 if (q->qe[pri])
213 {
214 q->qe[pri]->next = req;
215 q->qe[pri] = req;
216 }
217 else
218 q->qe[pri] = q->qs[pri] = req;
219
220 return q->size++;
221}
222
223aio_req reqq_shift (reqq *q)
224{
225 int pri;
226
227 if (!q->size)
228 return 0;
229
230 --q->size;
231
232 for (pri = NUM_PRI; pri--; )
233 {
234 aio_req req = q->qs[pri];
235
236 if (req)
237 {
238 if (!(q->qs[pri] = req->next))
239 q->qe[pri] = 0;
240
241 return req;
242 }
243 }
244
245 abort ();
246}
247
248static int poll_cb (int max);
106static void req_invoke (aio_req req); 249static void req_invoke (aio_req req);
107static void req_free (aio_req req); 250static void req_free (aio_req req);
251static void req_cancel (aio_req req);
108 252
109/* must be called at most once */ 253/* must be called at most once */
110static SV *req_sv (aio_req req, const char *klass) 254static SV *req_sv (aio_req req, const char *klass)
111{ 255{
112 if (!req->self) 256 if (!req->self)
130 return mg ? (aio_req)mg->mg_ptr : 0; 274 return mg ? (aio_req)mg->mg_ptr : 0;
131} 275}
132 276
133static void aio_grp_feed (aio_req grp) 277static void aio_grp_feed (aio_req grp)
134{ 278{
135 while (grp->length < grp->fd2) 279 while (grp->length < grp->fd2 && !(grp->flags & FLAG_CANCELLED))
136 { 280 {
137 int old_len = grp->length; 281 int old_len = grp->length;
138 282
139 if (grp->fh2 && SvOK (grp->fh2)) 283 if (grp->fh2 && SvOK (grp->fh2))
140 { 284 {
143 ENTER; 287 ENTER;
144 SAVETMPS; 288 SAVETMPS;
145 PUSHMARK (SP); 289 PUSHMARK (SP);
146 XPUSHs (req_sv (grp, AIO_GRP_KLASS)); 290 XPUSHs (req_sv (grp, AIO_GRP_KLASS));
147 PUTBACK; 291 PUTBACK;
148 call_sv (grp->fh2, G_VOID | G_EVAL); 292 call_sv (grp->fh2, G_VOID | G_EVAL | G_KEEPERR);
149 SPAGAIN; 293 SPAGAIN;
150 FREETMPS; 294 FREETMPS;
151 LEAVE; 295 LEAVE;
152 } 296 }
153 297
176 } 320 }
177} 321}
178 322
179static void poll_wait () 323static void poll_wait ()
180{ 324{
181 if (nreqs && !ress)
182 {
183 fd_set rfd; 325 fd_set rfd;
326
327 while (nreqs)
328 {
329 int size;
330#if !(__i386 || __x86_64) /* safe without sempahore on these archs */
331 LOCK (reslock);
332#endif
333 size = res_queue.size;
334#if !(__i386 || __x86_64) /* safe without sempahore on these archs */
335 UNLOCK (reslock);
336#endif
337
338 if (size)
339 return;
340
184 FD_ZERO(&rfd); 341 FD_ZERO(&rfd);
185 FD_SET(respipe [0], &rfd); 342 FD_SET(respipe [0], &rfd);
186 343
187 select (respipe [0] + 1, &rfd, 0, 0, 0); 344 select (respipe [0] + 1, &rfd, 0, 0, 0);
188 } 345 }
189} 346}
190 347
191static void req_invoke (aio_req req) 348static void req_invoke (aio_req req)
192{ 349{
193 dSP; 350 dSP;
194 int errorno = errno;
195 351
196 if (req->cancelled || !SvOK (req->callback)) 352 if (!(req->flags & FLAG_CANCELLED) && SvOK (req->callback))
197 return; 353 {
198
199 errno = req->errorno; 354 errno = req->errorno;
200 355
201 ENTER; 356 ENTER;
202 SAVETMPS; 357 SAVETMPS;
203 PUSHMARK (SP); 358 PUSHMARK (SP);
204 EXTEND (SP, 1); 359 EXTEND (SP, 1);
205 360
206 switch (req->type) 361 switch (req->type)
207 {
208 case REQ_READDIR:
209 { 362 {
210 SV *rv = &PL_sv_undef; 363 case REQ_READDIR:
211
212 if (req->result >= 0)
213 { 364 {
214 char *buf = req->data2ptr; 365 SV *rv = &PL_sv_undef;
215 AV *av = newAV ();
216 366
217 while (req->result) 367 if (req->result >= 0)
218 { 368 {
369 int i;
370 char *buf = req->data2ptr;
371 AV *av = newAV ();
372
373 av_extend (av, req->result - 1);
374
375 for (i = 0; i < req->result; ++i)
376 {
219 SV *sv = newSVpv (buf, 0); 377 SV *sv = newSVpv (buf, 0);
220 378
221 av_push (av, sv); 379 av_store (av, i, sv);
222 buf += SvCUR (sv) + 1; 380 buf += SvCUR (sv) + 1;
223 req->result--; 381 }
382
383 rv = sv_2mortal (newRV_noinc ((SV *)av));
224 } 384 }
225 385
226 rv = sv_2mortal (newRV_noinc ((SV *)av)); 386 PUSHs (rv);
227 } 387 }
388 break;
228 389
229 PUSHs (rv); 390 case REQ_OPEN:
391 {
392 /* convert fd to fh */
393 SV *fh;
394
395 PUSHs (sv_2mortal (newSViv (req->result)));
396 PUTBACK;
397 call_pv ("IO::AIO::_fd2fh", G_SCALAR | G_EVAL);
398 SPAGAIN;
399
400 fh = SvREFCNT_inc (POPs);
401
402 PUSHMARK (SP);
403 XPUSHs (sv_2mortal (fh));
404 }
405 break;
406
407 case REQ_GROUP:
408 req->fd = 2; /* mark group as finished */
409
410 if (req->data)
411 {
412 int i;
413 AV *av = (AV *)req->data;
414
415 EXTEND (SP, AvFILL (av) + 1);
416 for (i = 0; i <= AvFILL (av); ++i)
417 PUSHs (*av_fetch (av, i, 0));
418 }
419 break;
420
421 case REQ_NOP:
422 case REQ_BUSY:
423 break;
424
425 default:
426 PUSHs (sv_2mortal (newSViv (req->result)));
427 break;
230 } 428 }
231 break;
232 429
233 case REQ_OPEN:
234 {
235 /* convert fd to fh */
236 SV *fh;
237 430
238 PUSHs (sv_2mortal (newSViv (req->result)));
239 PUTBACK; 431 PUTBACK;
240 call_pv ("IO::AIO::_fd2fh", G_SCALAR | G_EVAL);
241 SPAGAIN;
242
243 fh = SvREFCNT_inc (POPs);
244
245 PUSHMARK (SP);
246 XPUSHs (sv_2mortal (fh));
247 }
248 break;
249
250 case REQ_GROUP:
251 req->fd = 2; /* mark group as finished */
252
253 if (req->data)
254 {
255 int i;
256 AV *av = (AV *)req->data;
257
258 EXTEND (SP, AvFILL (av) + 1);
259 for (i = 0; i <= AvFILL (av); ++i)
260 PUSHs (*av_fetch (av, i, 0));
261 }
262 break;
263
264 case REQ_SLEEP:
265 break;
266
267 default:
268 PUSHs (sv_2mortal (newSViv (req->result)));
269 break;
270 }
271
272
273 PUTBACK;
274 call_sv (req->callback, G_VOID | G_EVAL); 432 call_sv (req->callback, G_VOID | G_EVAL);
275 SPAGAIN; 433 SPAGAIN;
276 434
277 FREETMPS; 435 FREETMPS;
278 LEAVE; 436 LEAVE;
279
280 errno = errorno;
281
282 if (SvTRUE (ERRSV))
283 { 437 }
284 req_free (req);
285 croak (0);
286 }
287}
288 438
289static void req_free (aio_req req)
290{
291 if (req->grp) 439 if (req->grp)
292 { 440 {
293 aio_req grp = req->grp; 441 aio_req grp = req->grp;
294 442
295 /* unlink request */ 443 /* unlink request */
300 grp->grp_first = req->grp_next; 448 grp->grp_first = req->grp_next;
301 449
302 aio_grp_dec (grp); 450 aio_grp_dec (grp);
303 } 451 }
304 452
453 if (SvTRUE (ERRSV))
454 {
455 req_free (req);
456 croak (0);
457 }
458}
459
460static void req_free (aio_req req)
461{
305 if (req->self) 462 if (req->self)
306 { 463 {
307 sv_unmagic (req->self, PERL_MAGIC_ext); 464 sv_unmagic (req->self, PERL_MAGIC_ext);
308 SvREFCNT_dec (req->self); 465 SvREFCNT_dec (req->self);
309 } 466 }
312 SvREFCNT_dec (req->fh); 469 SvREFCNT_dec (req->fh);
313 SvREFCNT_dec (req->fh2); 470 SvREFCNT_dec (req->fh2);
314 SvREFCNT_dec (req->callback); 471 SvREFCNT_dec (req->callback);
315 Safefree (req->statdata); 472 Safefree (req->statdata);
316 473
317 if (req->type == REQ_READDIR && req->result >= 0) 474 if (req->type == REQ_READDIR)
318 free (req->data2ptr); 475 free (req->data2ptr);
319 476
320 Safefree (req); 477 Safefree (req);
321} 478}
322 479
480static void req_cancel_subs (aio_req grp)
481{
482 aio_req sub;
483
484 if (grp->type != REQ_GROUP)
485 return;
486
487 SvREFCNT_dec (grp->fh2);
488 grp->fh2 = 0;
489
490 for (sub = grp->grp_first; sub; sub = sub->grp_next)
491 req_cancel (sub);
492}
493
323static void req_cancel (aio_req req) 494static void req_cancel (aio_req req)
324{ 495{
325 req->cancelled = 1; 496 req->flags |= FLAG_CANCELLED;
326 497
327 if (req->type == REQ_GROUP) 498 req_cancel_subs (req);
328 {
329 aio_req sub;
330
331 for (sub = req->grp_first; sub; sub = sub->grp_next)
332 req_cancel (sub);
333 }
334} 499}
335 500
336static int poll_cb () 501static int poll_cb (int max)
337{ 502{
338 dSP; 503 dSP;
339 int count = 0; 504 int count = 0;
340 int do_croak = 0; 505 int do_croak = 0;
341 aio_req req; 506 aio_req req;
342 507
343 for (;;) 508 while (max <= 0 || count < max)
344 { 509 {
345 pthread_mutex_lock (&reslock); 510 LOCK (reslock);
346 req = ress; 511 req = reqq_shift (&res_queue);
347 512
348 if (req) 513 if (req)
349 { 514 {
350 ress = req->next;
351
352 if (!ress) 515 if (!res_queue.size)
353 { 516 {
354 /* read any signals sent by the worker threads */ 517 /* read any signals sent by the worker threads */
355 char buf [32]; 518 char buf [32];
356 while (read (respipe [0], buf, 32) == 32) 519 while (read (respipe [0], buf, 32) == 32)
357 ; 520 ;
358
359 rese = 0;
360 } 521 }
361 } 522 }
362 523
363 pthread_mutex_unlock (&reslock); 524 UNLOCK (reslock);
364 525
365 if (!req) 526 if (!req)
366 break; 527 break;
367 528
368 --nreqs; 529 --nreqs;
403static void *aio_proc(void *arg); 564static void *aio_proc(void *arg);
404 565
405static void start_thread (void) 566static void start_thread (void)
406{ 567{
407 sigset_t fullsigset, oldsigset; 568 sigset_t fullsigset, oldsigset;
408 pthread_t tid;
409 pthread_attr_t attr; 569 pthread_attr_t attr;
570
571 worker *wrk = calloc (1, sizeof (worker));
572
573 if (!wrk)
574 croak ("unable to allocate worker thread data");
410 575
411 pthread_attr_init (&attr); 576 pthread_attr_init (&attr);
412 pthread_attr_setstacksize (&attr, STACKSIZE); 577 pthread_attr_setstacksize (&attr, STACKSIZE);
413 pthread_attr_setdetachstate (&attr, PTHREAD_CREATE_DETACHED); 578 pthread_attr_setdetachstate (&attr, PTHREAD_CREATE_DETACHED);
414 579
415 sigfillset (&fullsigset); 580 sigfillset (&fullsigset);
581
582 LOCK (wrklock);
416 sigprocmask (SIG_SETMASK, &fullsigset, &oldsigset); 583 sigprocmask (SIG_SETMASK, &fullsigset, &oldsigset);
417 584
418 if (pthread_create (&tid, &attr, aio_proc, 0) == 0) 585 if (pthread_create (&wrk->tid, &attr, aio_proc, (void *)wrk) == 0)
586 {
587 wrk->prev = &wrk_first;
588 wrk->next = wrk_first.next;
589 wrk_first.next->prev = wrk;
590 wrk_first.next = wrk;
419 started++; 591 started++;
592 }
593 else
594 free (wrk);
420 595
421 sigprocmask (SIG_SETMASK, &oldsigset, 0); 596 sigprocmask (SIG_SETMASK, &oldsigset, 0);
597 UNLOCK (wrklock);
422} 598}
423 599
424static void req_send (aio_req req) 600static void req_send (aio_req req)
425{ 601{
426 while (started < wanted && nreqs >= started) 602 while (started < wanted && nreqs >= started)
427 start_thread (); 603 start_thread ();
428 604
429 ++nreqs; 605 ++nreqs;
430 606
431 pthread_mutex_lock (&reqlock); 607 LOCK (reqlock);
432 608 reqq_push (&req_queue, req);
433 req->next = 0;
434
435 if (reqe)
436 {
437 reqe->next = req;
438 reqe = req;
439 }
440 else
441 reqe = reqs = req;
442
443 pthread_cond_signal (&reqwait); 609 pthread_cond_signal (&reqwait);
444 pthread_mutex_unlock (&reqlock); 610 UNLOCK (reqlock);
445
446 if (nreqs > max_outstanding)
447 for (;;)
448 {
449 poll_cb ();
450
451 if (nreqs <= max_outstanding)
452 break;
453
454 poll_wait ();
455 }
456} 611}
457 612
458static void end_thread (void) 613static void end_thread (void)
459{ 614{
460 aio_req req; 615 aio_req req;
616
461 Newz (0, req, 1, aio_cb); 617 Newz (0, req, 1, aio_cb);
618
462 req->type = REQ_QUIT; 619 req->type = REQ_QUIT;
620 req->pri = PRI_MAX + PRI_BIAS;
463 621
464 req_send (req); 622 req_send (req);
465} 623}
466 624
467static void min_parallel (int nthreads) 625static void min_parallel (int nthreads)
484 } 642 }
485 643
486 while (started > wanted) 644 while (started > wanted)
487 { 645 {
488 poll_wait (); 646 poll_wait ();
489 poll_cb (); 647 poll_cb (0);
490 } 648 }
491} 649}
492 650
493static void create_pipe () 651static void create_pipe ()
494{ 652{
519static ssize_t pread (int fd, void *buf, size_t count, off_t offset) 677static ssize_t pread (int fd, void *buf, size_t count, off_t offset)
520{ 678{
521 ssize_t res; 679 ssize_t res;
522 off_t ooffset; 680 off_t ooffset;
523 681
524 pthread_mutex_lock (&preadwritelock); 682 LOCK (preadwritelock);
525 ooffset = lseek (fd, 0, SEEK_CUR); 683 ooffset = lseek (fd, 0, SEEK_CUR);
526 lseek (fd, offset, SEEK_SET); 684 lseek (fd, offset, SEEK_SET);
527 res = read (fd, buf, count); 685 res = read (fd, buf, count);
528 lseek (fd, ooffset, SEEK_SET); 686 lseek (fd, ooffset, SEEK_SET);
529 pthread_mutex_unlock (&preadwritelock); 687 UNLOCK (preadwritelock);
530 688
531 return res; 689 return res;
532} 690}
533 691
534static ssize_t pwrite (int fd, void *buf, size_t count, off_t offset) 692static ssize_t pwrite (int fd, void *buf, size_t count, off_t offset)
535{ 693{
536 ssize_t res; 694 ssize_t res;
537 off_t ooffset; 695 off_t ooffset;
538 696
539 pthread_mutex_lock (&preadwritelock); 697 LOCK (preadwritelock);
540 ooffset = lseek (fd, 0, SEEK_CUR); 698 ooffset = lseek (fd, 0, SEEK_CUR);
541 lseek (fd, offset, SEEK_SET); 699 lseek (fd, offset, SEEK_SET);
542 res = write (fd, buf, count); 700 res = write (fd, buf, count);
543 lseek (fd, offset, SEEK_SET); 701 lseek (fd, offset, SEEK_SET);
544 pthread_mutex_unlock (&preadwritelock); 702 UNLOCK (preadwritelock);
545 703
546 return res; 704 return res;
547} 705}
548#endif 706#endif
549 707
550#if !HAVE_FDATASYNC 708#if !HAVE_FDATASYNC
551# define fdatasync fsync 709# define fdatasync fsync
552#endif 710#endif
553 711
554#if !HAVE_READAHEAD 712#if !HAVE_READAHEAD
555# define readahead aio_readahead 713# define readahead(fd,offset,count) aio_readahead (fd, offset, count, self)
556 714
557static ssize_t readahead (int fd, off_t offset, size_t count) 715static ssize_t aio_readahead (int fd, off_t offset, size_t count, worker *self)
558{ 716{
559 char readahead_buf[4096]; 717 dBUF;
560 718
561 while (count > 0) 719 while (count > 0)
562 { 720 {
563 size_t len = count < sizeof (readahead_buf) ? count : sizeof (readahead_buf); 721 size_t len = count < AIO_BUFSIZE ? count : AIO_BUFSIZE;
564 722
565 pread (fd, readahead_buf, len, offset); 723 pread (fd, aio_buf, len, offset);
566 offset += len; 724 offset += len;
567 count -= len; 725 count -= len;
568 } 726 }
569 727
570 errno = 0; 728 errno = 0;
571} 729}
730
572#endif 731#endif
573 732
574#if !HAVE_READDIR_R 733#if !HAVE_READDIR_R
575# define readdir_r aio_readdir_r 734# define readdir_r aio_readdir_r
576 735
579static int readdir_r (DIR *dirp, struct dirent *ent, struct dirent **res) 738static int readdir_r (DIR *dirp, struct dirent *ent, struct dirent **res)
580{ 739{
581 struct dirent *e; 740 struct dirent *e;
582 int errorno; 741 int errorno;
583 742
584 pthread_mutex_lock (&readdirlock); 743 LOCK (readdirlock);
585 744
586 e = readdir (dirp); 745 e = readdir (dirp);
587 errorno = errno; 746 errorno = errno;
588 747
589 if (e) 748 if (e)
592 strcpy (ent->d_name, e->d_name); 751 strcpy (ent->d_name, e->d_name);
593 } 752 }
594 else 753 else
595 *res = 0; 754 *res = 0;
596 755
597 pthread_mutex_unlock (&readdirlock); 756 UNLOCK (readdirlock);
598 757
599 errno = errorno; 758 errno = errorno;
600 return e ? 0 : -1; 759 return e ? 0 : -1;
601} 760}
602#endif 761#endif
603 762
604/* sendfile always needs emulation */ 763/* sendfile always needs emulation */
605static ssize_t sendfile_ (int ofd, int ifd, off_t offset, size_t count) 764static ssize_t sendfile_ (int ofd, int ifd, off_t offset, size_t count, worker *self)
606{ 765{
607 ssize_t res; 766 ssize_t res;
608 767
609 if (!count) 768 if (!count)
610 return 0; 769 return 0;
659#endif 818#endif
660 ) 819 )
661 ) 820 )
662 { 821 {
663 /* emulate sendfile. this is a major pain in the ass */ 822 /* emulate sendfile. this is a major pain in the ass */
664 char buf[4096]; 823 dBUF;
824
665 res = 0; 825 res = 0;
666 826
667 while (count) 827 while (count)
668 { 828 {
669 ssize_t cnt; 829 ssize_t cnt;
670 830
671 cnt = pread (ifd, buf, count > 4096 ? 4096 : count, offset); 831 cnt = pread (ifd, aio_buf, count > AIO_BUFSIZE ? AIO_BUFSIZE : count, offset);
672 832
673 if (cnt <= 0) 833 if (cnt <= 0)
674 { 834 {
675 if (cnt && !res) res = -1; 835 if (cnt && !res) res = -1;
676 break; 836 break;
677 } 837 }
678 838
679 cnt = write (ofd, buf, cnt); 839 cnt = write (ofd, aio_buf, cnt);
680 840
681 if (cnt <= 0) 841 if (cnt <= 0)
682 { 842 {
683 if (cnt && !res) res = -1; 843 if (cnt && !res) res = -1;
684 break; 844 break;
692 852
693 return res; 853 return res;
694} 854}
695 855
696/* read a full directory */ 856/* read a full directory */
697static int scandir_ (const char *path, void **namesp) 857static void scandir_ (aio_req req, worker *self)
698{ 858{
699 DIR *dirp = opendir (path); 859 DIR *dirp;
700 union 860 union
701 { 861 {
702 struct dirent d; 862 struct dirent d;
703 char b [offsetof (struct dirent, d_name) + NAME_MAX + 1]; 863 char b [offsetof (struct dirent, d_name) + NAME_MAX + 1];
704 } u; 864 } *u;
705 struct dirent *entp; 865 struct dirent *entp;
706 char *name, *names; 866 char *name, *names;
707 int memlen = 4096; 867 int memlen = 4096;
708 int memofs = 0; 868 int memofs = 0;
709 int res = 0; 869 int res = 0;
710 int errorno; 870 int errorno;
711 871
712 if (!dirp) 872 LOCK (wrklock);
713 return -1; 873 self->dirp = dirp = opendir (req->dataptr);
714 874 self->dbuf = u = malloc (sizeof (*u));
715 names = malloc (memlen); 875 req->data2ptr = names = malloc (memlen);
876 UNLOCK (wrklock);
716 877
878 if (dirp && u && names)
717 for (;;) 879 for (;;)
718 { 880 {
881 errno = 0;
719 errno = 0, readdir_r (dirp, &u.d, &entp); 882 readdir_r (dirp, &u->d, &entp);
720 883
721 if (!entp) 884 if (!entp)
722 break; 885 break;
723 886
724 name = entp->d_name; 887 name = entp->d_name;
725 888
726 if (name [0] != '.' || (name [1] && (name [1] != '.' || name [2]))) 889 if (name [0] != '.' || (name [1] && (name [1] != '.' || name [2])))
727 { 890 {
728 int len = strlen (name) + 1; 891 int len = strlen (name) + 1;
729 892
730 res++; 893 res++;
731 894
732 while (memofs + len > memlen) 895 while (memofs + len > memlen)
733 { 896 {
734 memlen *= 2; 897 memlen *= 2;
898 LOCK (wrklock);
735 names = realloc (names, memlen); 899 req->data2ptr = names = realloc (names, memlen);
900 UNLOCK (wrklock);
901
736 if (!names) 902 if (!names)
737 break; 903 break;
738 } 904 }
739 905
740 memcpy (names + memofs, name, len); 906 memcpy (names + memofs, name, len);
741 memofs += len; 907 memofs += len;
742 } 908 }
743 } 909 }
744 910
745 errorno = errno;
746 closedir (dirp);
747
748 if (errorno) 911 if (errno)
749 {
750 free (names);
751 errno = errorno;
752 res = -1; 912 res = -1;
753 } 913
754 914 req->result = res;
755 *namesp = (void *)names;
756 return res;
757} 915}
758 916
759/*****************************************************************************/ 917/*****************************************************************************/
760 918
761static void *aio_proc (void *thr_arg) 919static void *aio_proc (void *thr_arg)
762{ 920{
763 aio_req req; 921 aio_req req;
764 int type; 922 int type;
923 worker *self = (worker *)thr_arg;
765 924
766 do 925 do
767 { 926 {
768 pthread_mutex_lock (&reqlock); 927 LOCK (reqlock);
769 928
770 for (;;) 929 for (;;)
771 { 930 {
772 req = reqs; 931 self->req = req = reqq_shift (&req_queue);
773
774 if (reqs)
775 {
776 reqs = reqs->next;
777 if (!reqs) reqe = 0;
778 }
779 932
780 if (req) 933 if (req)
781 break; 934 break;
782 935
783 pthread_cond_wait (&reqwait, &reqlock); 936 pthread_cond_wait (&reqwait, &reqlock);
784 } 937 }
785 938
786 pthread_mutex_unlock (&reqlock); 939 UNLOCK (reqlock);
787 940
788 errno = 0; /* strictly unnecessary */ 941 errno = 0; /* strictly unnecessary */
789
790 if (!req->cancelled)
791 switch (type = req->type) /* remember type for QUIT check */ 942 type = req->type; /* remember type for QUIT check */
943
944 if (!(req->flags & FLAG_CANCELLED))
945 switch (type)
792 { 946 {
793 case REQ_READ: req->result = pread (req->fd, req->dataptr, req->length, req->offset); break; 947 case REQ_READ: req->result = pread (req->fd, req->dataptr, req->length, req->offset); break;
794 case REQ_WRITE: req->result = pwrite (req->fd, req->dataptr, req->length, req->offset); break; 948 case REQ_WRITE: req->result = pwrite (req->fd, req->dataptr, req->length, req->offset); break;
795 949
796 case REQ_READAHEAD: req->result = readahead (req->fd, req->offset, req->length); break; 950 case REQ_READAHEAD: req->result = readahead (req->fd, req->offset, req->length); break;
797 case REQ_SENDFILE: req->result = sendfile_ (req->fd, req->fd2, req->offset, req->length); break; 951 case REQ_SENDFILE: req->result = sendfile_ (req->fd, req->fd2, req->offset, req->length, self); break;
798 952
799 case REQ_STAT: req->result = stat (req->dataptr, req->statdata); break; 953 case REQ_STAT: req->result = stat (req->dataptr, req->statdata); break;
800 case REQ_LSTAT: req->result = lstat (req->dataptr, req->statdata); break; 954 case REQ_LSTAT: req->result = lstat (req->dataptr, req->statdata); break;
801 case REQ_FSTAT: req->result = fstat (req->fd , req->statdata); break; 955 case REQ_FSTAT: req->result = fstat (req->fd , req->statdata); break;
802 956
808 case REQ_LINK: req->result = link (req->data2ptr, req->dataptr); break; 962 case REQ_LINK: req->result = link (req->data2ptr, req->dataptr); break;
809 case REQ_SYMLINK: req->result = symlink (req->data2ptr, req->dataptr); break; 963 case REQ_SYMLINK: req->result = symlink (req->data2ptr, req->dataptr); break;
810 964
811 case REQ_FDATASYNC: req->result = fdatasync (req->fd); break; 965 case REQ_FDATASYNC: req->result = fdatasync (req->fd); break;
812 case REQ_FSYNC: req->result = fsync (req->fd); break; 966 case REQ_FSYNC: req->result = fsync (req->fd); break;
813 case REQ_READDIR: req->result = scandir_ (req->dataptr, &req->data2ptr); break; 967 case REQ_READDIR: scandir_ (req, self); break;
814 968
815 case REQ_SLEEP: 969 case REQ_BUSY:
816 { 970 {
817 struct timeval tv; 971 struct timeval tv;
818 972
819 tv.tv_sec = req->fd; 973 tv.tv_sec = req->fd;
820 tv.tv_usec = req->fd2; 974 tv.tv_usec = req->fd2;
821 975
822 req->result = select (0, 0, 0, 0, &tv); 976 req->result = select (0, 0, 0, 0, &tv);
823 } 977 }
824 978
979 case REQ_GROUP:
980 case REQ_NOP:
825 case REQ_QUIT: 981 case REQ_QUIT:
826 break; 982 break;
827 983
828 default: 984 default:
829 req->result = ENOSYS; 985 req->result = ENOSYS;
830 break; 986 break;
831 } 987 }
832 988
833 req->errorno = errno; 989 req->errorno = errno;
834 990
835 pthread_mutex_lock (&reslock); 991 LOCK (reslock);
836 992
837 req->next = 0; 993 if (!reqq_push (&res_queue, req))
838
839 if (rese)
840 {
841 rese->next = req;
842 rese = req;
843 }
844 else
845 {
846 rese = ress = req;
847
848 /* write a dummy byte to the pipe so fh becomes ready */ 994 /* write a dummy byte to the pipe so fh becomes ready */
849 write (respipe [1], &respipe, 1); 995 write (respipe [1], &respipe, 1);
850 }
851 996
852 pthread_mutex_unlock (&reslock); 997 self->req = 0;
998 worker_clear (self);
999
1000 UNLOCK (reslock);
853 } 1001 }
854 while (type != REQ_QUIT); 1002 while (type != REQ_QUIT);
855 1003
1004 LOCK (wrklock);
1005 worker_free (self);
1006 UNLOCK (wrklock);
1007
856 return 0; 1008 return 0;
857} 1009}
858 1010
859/*****************************************************************************/ 1011/*****************************************************************************/
860 1012
861static void atfork_prepare (void) 1013static void atfork_prepare (void)
862{ 1014{
863 pthread_mutex_lock (&reqlock); 1015 LOCK (wrklock);
864 pthread_mutex_lock (&reslock); 1016 LOCK (reqlock);
1017 LOCK (reslock);
865#if !HAVE_PREADWRITE 1018#if !HAVE_PREADWRITE
866 pthread_mutex_lock (&preadwritelock); 1019 LOCK (preadwritelock);
867#endif 1020#endif
868#if !HAVE_READDIR_R 1021#if !HAVE_READDIR_R
869 pthread_mutex_lock (&readdirlock); 1022 LOCK (readdirlock);
870#endif 1023#endif
871} 1024}
872 1025
873static void atfork_parent (void) 1026static void atfork_parent (void)
874{ 1027{
875#if !HAVE_READDIR_R 1028#if !HAVE_READDIR_R
876 pthread_mutex_unlock (&readdirlock); 1029 UNLOCK (readdirlock);
877#endif 1030#endif
878#if !HAVE_PREADWRITE 1031#if !HAVE_PREADWRITE
879 pthread_mutex_unlock (&preadwritelock); 1032 UNLOCK (preadwritelock);
880#endif 1033#endif
881 pthread_mutex_unlock (&reslock); 1034 UNLOCK (reslock);
882 pthread_mutex_unlock (&reqlock); 1035 UNLOCK (reqlock);
1036 UNLOCK (wrklock);
883} 1037}
884 1038
885static void atfork_child (void) 1039static void atfork_child (void)
886{ 1040{
887 aio_req prv; 1041 aio_req prv;
888 1042
1043 while (prv = reqq_shift (&req_queue))
1044 req_free (prv);
1045
1046 while (prv = reqq_shift (&res_queue))
1047 req_free (prv);
1048
1049 while (wrk_first.next != &wrk_first)
1050 {
1051 worker *wrk = wrk_first.next;
1052
1053 if (wrk->req)
1054 req_free (wrk->req);
1055
1056 worker_clear (wrk);
1057 worker_free (wrk);
1058 }
1059
889 started = 0; 1060 started = 0;
890 1061 nreqs = 0;
891 while (reqs)
892 {
893 prv = reqs;
894 reqs = prv->next;
895 req_free (prv);
896 }
897
898 reqs = reqe = 0;
899
900 while (ress)
901 {
902 prv = ress;
903 ress = prv->next;
904 req_free (prv);
905 }
906
907 ress = rese = 0;
908 1062
909 close (respipe [0]); 1063 close (respipe [0]);
910 close (respipe [1]); 1064 close (respipe [1]);
911 create_pipe (); 1065 create_pipe ();
912 1066
913 atfork_parent (); 1067 atfork_parent ();
914} 1068}
915 1069
916#define dREQ \ 1070#define dREQ \
917 aio_req req; \ 1071 aio_req req; \
1072 int req_pri = next_pri; \
1073 next_pri = DEFAULT_PRI + PRI_BIAS; \
918 \ 1074 \
919 if (SvOK (callback) && !SvROK (callback)) \ 1075 if (SvOK (callback) && !SvROK (callback)) \
920 croak ("callback must be undef or of reference type"); \ 1076 croak ("callback must be undef or of reference type"); \
921 \ 1077 \
922 Newz (0, req, 1, aio_cb); \ 1078 Newz (0, req, 1, aio_cb); \
923 if (!req) \ 1079 if (!req) \
924 croak ("out of memory during aio_req allocation"); \ 1080 croak ("out of memory during aio_req allocation"); \
925 \ 1081 \
926 req->callback = newSVsv (callback) 1082 req->callback = newSVsv (callback); \
1083 req->pri = req_pri
927 1084
928#define REQ_SEND \ 1085#define REQ_SEND \
929 req_send (req); \ 1086 req_send (req); \
930 \ 1087 \
931 if (GIMME_V != G_VOID) \ 1088 if (GIMME_V != G_VOID) \
953 1110
954void 1111void
955max_parallel (nthreads) 1112max_parallel (nthreads)
956 int nthreads 1113 int nthreads
957 PROTOTYPE: $ 1114 PROTOTYPE: $
958
959int
960max_outstanding (nreqs)
961 int nreqs
962 PROTOTYPE: $
963 CODE:
964 RETVAL = max_outstanding;
965 max_outstanding = nreqs;
966 1115
967void 1116void
968aio_open (pathname,flags,mode,callback=&PL_sv_undef) 1117aio_open (pathname,flags,mode,callback=&PL_sv_undef)
969 SV * pathname 1118 SV * pathname
970 int flags 1119 int flags
1185 1334
1186 REQ_SEND; 1335 REQ_SEND;
1187} 1336}
1188 1337
1189void 1338void
1190aio_sleep (delay,callback=&PL_sv_undef) 1339aio_busy (delay,callback=&PL_sv_undef)
1191 double delay 1340 double delay
1192 SV * callback 1341 SV * callback
1193 PPCODE: 1342 PPCODE:
1194{ 1343{
1195 dREQ; 1344 dREQ;
1196 1345
1197 req->type = REQ_SLEEP; 1346 req->type = REQ_BUSY;
1198 req->fd = delay < 0. ? 0 : delay; 1347 req->fd = delay < 0. ? 0 : delay;
1199 req->fd2 = delay < 0. ? 0 : 1000. * (delay - req->fd); 1348 req->fd2 = delay < 0. ? 0 : 1000. * (delay - req->fd);
1200 1349
1201 REQ_SEND; 1350 REQ_SEND;
1202} 1351}
1206 SV * callback 1355 SV * callback
1207 PROTOTYPE: ;$ 1356 PROTOTYPE: ;$
1208 PPCODE: 1357 PPCODE:
1209{ 1358{
1210 dREQ; 1359 dREQ;
1360
1211 req->type = REQ_GROUP; 1361 req->type = REQ_GROUP;
1212 req_send (req); 1362 req_send (req);
1363
1213 XPUSHs (req_sv (req, AIO_GRP_KLASS)); 1364 XPUSHs (req_sv (req, AIO_GRP_KLASS));
1214} 1365}
1366
1367void
1368aio_nop (callback=&PL_sv_undef)
1369 SV * callback
1370 PPCODE:
1371{
1372 dREQ;
1373
1374 req->type = REQ_NOP;
1375
1376 REQ_SEND;
1377}
1378
1379void
1380aioreq_pri (int pri = DEFAULT_PRI)
1381 CODE:
1382 if (pri < PRI_MIN) pri = PRI_MIN;
1383 if (pri > PRI_MAX) pri = PRI_MAX;
1384 next_pri = pri + PRI_BIAS;
1385
1386void
1387aioreq_nice (int nice = 0)
1388 CODE:
1389 nice = next_pri - nice;
1390 if (nice < PRI_MIN) nice = PRI_MIN;
1391 if (nice > PRI_MAX) nice = PRI_MAX;
1392 next_pri = nice + PRI_BIAS;
1215 1393
1216void 1394void
1217flush () 1395flush ()
1218 PROTOTYPE: 1396 PROTOTYPE:
1219 CODE: 1397 CODE:
1220 while (nreqs) 1398 while (nreqs)
1221 { 1399 {
1222 poll_wait (); 1400 poll_wait ();
1223 poll_cb (); 1401 poll_cb (0);
1224 } 1402 }
1225 1403
1226void 1404void
1227poll() 1405poll()
1228 PROTOTYPE: 1406 PROTOTYPE:
1229 CODE: 1407 CODE:
1230 if (nreqs) 1408 if (nreqs)
1231 { 1409 {
1232 poll_wait (); 1410 poll_wait ();
1233 poll_cb (); 1411 poll_cb (0);
1234 } 1412 }
1235 1413
1236int 1414int
1237poll_fileno() 1415poll_fileno()
1238 PROTOTYPE: 1416 PROTOTYPE:
1243 1421
1244int 1422int
1245poll_cb(...) 1423poll_cb(...)
1246 PROTOTYPE: 1424 PROTOTYPE:
1247 CODE: 1425 CODE:
1248 RETVAL = poll_cb (); 1426 RETVAL = poll_cb (0);
1427 OUTPUT:
1428 RETVAL
1429
1430int
1431poll_some(int max = 0)
1432 PROTOTYPE: $
1433 CODE:
1434 RETVAL = poll_cb (max);
1249 OUTPUT: 1435 OUTPUT:
1250 RETVAL 1436 RETVAL
1251 1437
1252void 1438void
1253poll_wait() 1439poll_wait()
1268 1454
1269MODULE = IO::AIO PACKAGE = IO::AIO::REQ 1455MODULE = IO::AIO PACKAGE = IO::AIO::REQ
1270 1456
1271void 1457void
1272cancel (aio_req_ornot req) 1458cancel (aio_req_ornot req)
1273 PROTOTYPE:
1274 CODE: 1459 CODE:
1275 req_cancel (req); 1460 req_cancel (req);
1461
1462void
1463cb (aio_req_ornot req, SV *callback=&PL_sv_undef)
1464 CODE:
1465 SvREFCNT_dec (req->callback);
1466 req->callback = newSVsv (callback);
1276 1467
1277MODULE = IO::AIO PACKAGE = IO::AIO::GRP 1468MODULE = IO::AIO PACKAGE = IO::AIO::GRP
1278 1469
1279void 1470void
1280add (aio_req grp, ...) 1471add (aio_req grp, ...)
1308 } 1499 }
1309 } 1500 }
1310} 1501}
1311 1502
1312void 1503void
1504cancel_subs (aio_req_ornot req)
1505 CODE:
1506 req_cancel_subs (req);
1507
1508void
1313result (aio_req grp, ...) 1509result (aio_req grp, ...)
1314 CODE: 1510 CODE:
1315{ 1511{
1316 int i; 1512 int i;
1317 AV *av = newAV (); 1513 AV *av = newAV ();
1322 SvREFCNT_dec (grp->data); 1518 SvREFCNT_dec (grp->data);
1323 grp->data = (SV *)av; 1519 grp->data = (SV *)av;
1324} 1520}
1325 1521
1326void 1522void
1327lock (aio_req grp)
1328 CODE:
1329 ++grp->length;
1330
1331void
1332unlock (aio_req grp)
1333 CODE:
1334 aio_grp_dec (grp);
1335
1336void
1337feeder_limit (aio_req grp, int limit) 1523limit (aio_req grp, int limit)
1338 CODE: 1524 CODE:
1339 grp->fd2 = limit; 1525 grp->fd2 = limit;
1340 aio_grp_feed (grp); 1526 aio_grp_feed (grp);
1341 1527
1342void 1528void
1343set_feeder (aio_req grp, SV *callback=&PL_sv_undef) 1529feed (aio_req grp, SV *callback=&PL_sv_undef)
1344 CODE: 1530 CODE:
1345{ 1531{
1346 SvREFCNT_dec (grp->fh2); 1532 SvREFCNT_dec (grp->fh2);
1347 grp->fh2 = newSVsv (callback); 1533 grp->fh2 = newSVsv (callback);
1348 1534

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines