ViewVC Help
View File | Revision Log | Show Annotations | Download File
/cvs/IO-AIO/AIO.xs
Revision: 1.76
Committed: Thu Oct 26 12:38:04 2006 UTC (17 years, 6 months ago) by root
Branch: MAIN
Changes since 1.75: +19 -11 lines
Log Message:
*** empty log message ***

File Contents

# Content
1 /* solaris */
2 #define _POSIX_PTHREAD_SEMANTICS 1
3
4 #if __linux && !defined(_GNU_SOURCE)
5 # define _GNU_SOURCE
6 #endif
7
8 /* just in case */
9 #define _REENTRANT 1
10
11 #include <errno.h>
12
13 #include "EXTERN.h"
14 #include "perl.h"
15 #include "XSUB.h"
16
17 #include "autoconf/config.h"
18
19 #include <pthread.h>
20
21 #include <stddef.h>
22 #include <errno.h>
23 #include <sys/time.h>
24 #include <sys/select.h>
25 #include <sys/types.h>
26 #include <sys/stat.h>
27 #include <limits.h>
28 #include <unistd.h>
29 #include <fcntl.h>
30 #include <signal.h>
31 #include <sched.h>
32
33 #if HAVE_SENDFILE
34 # if __linux
35 # include <sys/sendfile.h>
36 # elif __freebsd
37 # include <sys/socket.h>
38 # include <sys/uio.h>
39 # elif __hpux
40 # include <sys/socket.h>
41 # elif __solaris /* not yet */
42 # include <sys/sendfile.h>
43 # else
44 # error sendfile support requested but not available
45 # endif
46 #endif
47
48 /* used for struct dirent, AIX doesn't provide it */
49 #ifndef NAME_MAX
50 # define NAME_MAX 4096
51 #endif
52
53 #ifndef PTHREAD_STACK_MIN
54 /* care for broken platforms, e.g. windows */
55 # define PTHREAD_STACK_MIN 16384
56 #endif
57
58 #if __ia64
59 # define STACKSIZE 65536
60 #elif __i386 || __x86_64 /* 16k is unreasonably high :( */
61 # define STACKSIZE PTHREAD_STACK_MIN
62 #else
63 # define STACKSIZE 16384
64 #endif
65
66 /* buffer size for various temporary buffers */
67 #define AIO_BUFSIZE 65536
68
69 #define dBUF \
70 char *aio_buf; \
71 LOCK (wrklock); \
72 self->dbuf = aio_buf = malloc (AIO_BUFSIZE); \
73 UNLOCK (wrklock); \
74 if (!aio_buf) \
75 return -1;
76
77 enum {
78 REQ_QUIT,
79 REQ_OPEN, REQ_CLOSE,
80 REQ_READ, REQ_WRITE, REQ_READAHEAD,
81 REQ_SENDFILE,
82 REQ_STAT, REQ_LSTAT, REQ_FSTAT,
83 REQ_FSYNC, REQ_FDATASYNC,
84 REQ_UNLINK, REQ_RMDIR, REQ_RENAME,
85 REQ_READDIR,
86 REQ_LINK, REQ_SYMLINK,
87 REQ_GROUP, REQ_NOP,
88 REQ_BUSY,
89 };
90
91 #define AIO_REQ_KLASS "IO::AIO::REQ"
92 #define AIO_GRP_KLASS "IO::AIO::GRP"
93
94 typedef struct aio_cb
95 {
96 struct aio_cb *volatile next;
97
98 SV *data, *callback;
99 SV *fh, *fh2;
100 void *dataptr, *data2ptr;
101 Stat_t *statdata;
102 off_t offset;
103 size_t length;
104 ssize_t result;
105
106 STRLEN dataoffset;
107 int type;
108 int fd, fd2;
109 int errorno;
110 mode_t mode; /* open */
111
112 unsigned char flags;
113 unsigned char pri;
114
115 SV *self; /* the perl counterpart of this request, if any */
116 struct aio_cb *grp, *grp_prev, *grp_next, *grp_first;
117 } aio_cb;
118
119 enum {
120 FLAG_CANCELLED = 0x01,
121 };
122
123 typedef aio_cb *aio_req;
124 typedef aio_cb *aio_req_ornot;
125
126 enum {
127 PRI_MIN = -4,
128 PRI_MAX = 4,
129
130 DEFAULT_PRI = 0,
131 PRI_BIAS = -PRI_MIN,
132 NUM_PRI = PRI_MAX + PRI_BIAS + 1,
133 };
134
135 static int next_pri = DEFAULT_PRI + PRI_BIAS;
136
137 static int started, wanted;
138 static volatile int nreqs;
139 static int respipe [2];
140
141 #if __linux && defined (PTHREAD_ADAPTIVE_MUTEX_INITIALIZER_NP)
142 # define AIO_MUTEX_INIT PTHREAD_ADAPTIVE_MUTEX_INITIALIZER_NP
143 #else
144 # define AIO_MUTEX_INIT PTHREAD_MUTEX_INITIALIZER
145 #endif
146
147 #define LOCK(mutex) pthread_mutex_lock (&(mutex))
148 #define UNLOCK(mutex) pthread_mutex_unlock (&(mutex))
149
150 /* worker threads management */
151 static pthread_mutex_t wrklock = AIO_MUTEX_INIT;
152
153 typedef struct worker {
154 /* locked by wrklock */
155 struct worker *prev, *next;
156
157 pthread_t tid;
158
159 /* locked by reslock, reqlock or wrklock */
160 aio_req req; /* currently processed request */
161 void *dbuf;
162 DIR *dirp;
163 } worker;
164
165 static worker wrk_first = { &wrk_first, &wrk_first, 0 };
166
167 static void worker_clear (worker *wrk)
168 {
169 if (wrk->dirp)
170 {
171 closedir (wrk->dirp);
172 wrk->dirp = 0;
173 }
174
175 if (wrk->dbuf)
176 {
177 free (wrk->dbuf);
178 wrk->dbuf = 0;
179 }
180 }
181
182 static void worker_free (worker *wrk)
183 {
184 wrk->next->prev = wrk->prev;
185 wrk->prev->next = wrk->next;
186
187 free (wrk);
188 }
189
190 static pthread_mutex_t reslock = AIO_MUTEX_INIT;
191 static pthread_mutex_t reqlock = AIO_MUTEX_INIT;
192 static pthread_cond_t reqwait = PTHREAD_COND_INITIALIZER;
193
194 /*
195 * a somewhat faster data structure might be nice, but
196 * with 8 priorities this actually needs <20 insns
197 * per shift, the most expensive operation.
198 */
199 typedef struct {
200 aio_req qs[NUM_PRI], qe[NUM_PRI]; /* qstart, qend */
201 int size;
202 } reqq;
203
204 static reqq req_queue;
205 static reqq res_queue;
206
207 int reqq_push (reqq *q, aio_req req)
208 {
209 int pri = req->pri;
210 req->next = 0;
211
212 if (q->qe[pri])
213 {
214 q->qe[pri]->next = req;
215 q->qe[pri] = req;
216 }
217 else
218 q->qe[pri] = q->qs[pri] = req;
219
220 return q->size++;
221 }
222
223 aio_req reqq_shift (reqq *q)
224 {
225 int pri;
226
227 if (!q->size)
228 return 0;
229
230 --q->size;
231
232 for (pri = NUM_PRI; pri--; )
233 {
234 aio_req req = q->qs[pri];
235
236 if (req)
237 {
238 if (!(q->qs[pri] = req->next))
239 q->qe[pri] = 0;
240
241 return req;
242 }
243 }
244
245 abort ();
246 }
247
248 static int poll_cb (int max);
249 static void req_invoke (aio_req req);
250 static void req_free (aio_req req);
251 static void req_cancel (aio_req req);
252
253 /* must be called at most once */
254 static SV *req_sv (aio_req req, const char *klass)
255 {
256 if (!req->self)
257 {
258 req->self = (SV *)newHV ();
259 sv_magic (req->self, 0, PERL_MAGIC_ext, (char *)req, 0);
260 }
261
262 return sv_2mortal (sv_bless (newRV_inc (req->self), gv_stashpv (klass, 1)));
263 }
264
265 static aio_req SvAIO_REQ (SV *sv)
266 {
267 MAGIC *mg;
268
269 if (!sv_derived_from (sv, AIO_REQ_KLASS) || !SvROK (sv))
270 croak ("object of class " AIO_REQ_KLASS " expected");
271
272 mg = mg_find (SvRV (sv), PERL_MAGIC_ext);
273
274 return mg ? (aio_req)mg->mg_ptr : 0;
275 }
276
277 static void aio_grp_feed (aio_req grp)
278 {
279 while (grp->length < grp->fd2 && !(grp->flags & FLAG_CANCELLED))
280 {
281 int old_len = grp->length;
282
283 if (grp->fh2 && SvOK (grp->fh2))
284 {
285 dSP;
286
287 ENTER;
288 SAVETMPS;
289 PUSHMARK (SP);
290 XPUSHs (req_sv (grp, AIO_GRP_KLASS));
291 PUTBACK;
292 call_sv (grp->fh2, G_VOID | G_EVAL | G_KEEPERR);
293 SPAGAIN;
294 FREETMPS;
295 LEAVE;
296 }
297
298 /* stop if no progress has been made */
299 if (old_len == grp->length)
300 {
301 SvREFCNT_dec (grp->fh2);
302 grp->fh2 = 0;
303 break;
304 }
305 }
306 }
307
308 static void aio_grp_dec (aio_req grp)
309 {
310 --grp->length;
311
312 /* call feeder, if applicable */
313 aio_grp_feed (grp);
314
315 /* finish, if done */
316 if (!grp->length && grp->fd)
317 {
318 req_invoke (grp);
319 req_free (grp);
320 }
321 }
322
323 static void poll_wait ()
324 {
325 fd_set rfd;
326
327 while (nreqs)
328 {
329 int size;
330 #if !(__i386 || __x86_64) /* safe without sempahore on these archs */
331 LOCK (reslock);
332 #endif
333 size = res_queue.size;
334 #if !(__i386 || __x86_64) /* safe without sempahore on these archs */
335 UNLOCK (reslock);
336 #endif
337
338 if (size)
339 return;
340
341 FD_ZERO(&rfd);
342 FD_SET(respipe [0], &rfd);
343
344 select (respipe [0] + 1, &rfd, 0, 0, 0);
345 }
346 }
347
348 static void req_invoke (aio_req req)
349 {
350 dSP;
351
352 if (!(req->flags & FLAG_CANCELLED) && SvOK (req->callback))
353 {
354 errno = req->errorno;
355
356 ENTER;
357 SAVETMPS;
358 PUSHMARK (SP);
359 EXTEND (SP, 1);
360
361 switch (req->type)
362 {
363 case REQ_READDIR:
364 {
365 SV *rv = &PL_sv_undef;
366
367 if (req->result >= 0)
368 {
369 int i;
370 char *buf = req->data2ptr;
371 AV *av = newAV ();
372
373 av_extend (av, req->result - 1);
374
375 for (i = 0; i < req->result; ++i)
376 {
377 SV *sv = newSVpv (buf, 0);
378
379 av_store (av, i, sv);
380 buf += SvCUR (sv) + 1;
381 }
382
383 rv = sv_2mortal (newRV_noinc ((SV *)av));
384 }
385
386 PUSHs (rv);
387 }
388 break;
389
390 case REQ_OPEN:
391 {
392 /* convert fd to fh */
393 SV *fh;
394
395 PUSHs (sv_2mortal (newSViv (req->result)));
396 PUTBACK;
397 call_pv ("IO::AIO::_fd2fh", G_SCALAR | G_EVAL);
398 SPAGAIN;
399
400 fh = SvREFCNT_inc (POPs);
401
402 PUSHMARK (SP);
403 XPUSHs (sv_2mortal (fh));
404 }
405 break;
406
407 case REQ_GROUP:
408 req->fd = 2; /* mark group as finished */
409
410 if (req->data)
411 {
412 int i;
413 AV *av = (AV *)req->data;
414
415 EXTEND (SP, AvFILL (av) + 1);
416 for (i = 0; i <= AvFILL (av); ++i)
417 PUSHs (*av_fetch (av, i, 0));
418 }
419 break;
420
421 case REQ_NOP:
422 case REQ_BUSY:
423 break;
424
425 default:
426 PUSHs (sv_2mortal (newSViv (req->result)));
427 break;
428 }
429
430
431 PUTBACK;
432 call_sv (req->callback, G_VOID | G_EVAL);
433 SPAGAIN;
434
435 FREETMPS;
436 LEAVE;
437 }
438
439 if (req->grp)
440 {
441 aio_req grp = req->grp;
442
443 /* unlink request */
444 if (req->grp_next) req->grp_next->grp_prev = req->grp_prev;
445 if (req->grp_prev) req->grp_prev->grp_next = req->grp_next;
446
447 if (grp->grp_first == req)
448 grp->grp_first = req->grp_next;
449
450 aio_grp_dec (grp);
451 }
452
453 if (SvTRUE (ERRSV))
454 {
455 req_free (req);
456 croak (0);
457 }
458 }
459
460 static void req_free (aio_req req)
461 {
462 if (req->self)
463 {
464 sv_unmagic (req->self, PERL_MAGIC_ext);
465 SvREFCNT_dec (req->self);
466 }
467
468 SvREFCNT_dec (req->data);
469 SvREFCNT_dec (req->fh);
470 SvREFCNT_dec (req->fh2);
471 SvREFCNT_dec (req->callback);
472 Safefree (req->statdata);
473
474 if (req->type == REQ_READDIR)
475 free (req->data2ptr);
476
477 Safefree (req);
478 }
479
480 static void req_cancel_subs (aio_req grp)
481 {
482 aio_req sub;
483
484 if (grp->type != REQ_GROUP)
485 return;
486
487 SvREFCNT_dec (grp->fh2);
488 grp->fh2 = 0;
489
490 for (sub = grp->grp_first; sub; sub = sub->grp_next)
491 req_cancel (sub);
492 }
493
494 static void req_cancel (aio_req req)
495 {
496 req->flags |= FLAG_CANCELLED;
497
498 req_cancel_subs (req);
499 }
500
501 static int poll_cb (int max)
502 {
503 dSP;
504 int count = 0;
505 int do_croak = 0;
506 aio_req req;
507
508 while (max <= 0 || count < max)
509 {
510 LOCK (reslock);
511 req = reqq_shift (&res_queue);
512
513 if (req)
514 {
515 if (!res_queue.size)
516 {
517 /* read any signals sent by the worker threads */
518 char buf [32];
519 while (read (respipe [0], buf, 32) == 32)
520 ;
521 }
522 }
523
524 UNLOCK (reslock);
525
526 if (!req)
527 break;
528
529 --nreqs;
530
531 if (req->type == REQ_QUIT)
532 started--;
533 else if (req->type == REQ_GROUP && req->length)
534 {
535 req->fd = 1; /* mark request as delayed */
536 continue;
537 }
538 else
539 {
540 if (req->type == REQ_READ)
541 SvCUR_set (req->data, req->dataoffset + (req->result > 0 ? req->result : 0));
542
543 if (req->data2ptr && (req->type == REQ_READ || req->type == REQ_WRITE))
544 SvREADONLY_off (req->data);
545
546 if (req->statdata)
547 {
548 PL_laststype = req->type == REQ_LSTAT ? OP_LSTAT : OP_STAT;
549 PL_laststatval = req->result;
550 PL_statcache = *(req->statdata);
551 }
552
553 req_invoke (req);
554
555 count++;
556 }
557
558 req_free (req);
559 }
560
561 return count;
562 }
563
564 static void *aio_proc(void *arg);
565
566 static void start_thread (void)
567 {
568 sigset_t fullsigset, oldsigset;
569 pthread_attr_t attr;
570
571 worker *wrk = calloc (1, sizeof (worker));
572
573 if (!wrk)
574 croak ("unable to allocate worker thread data");
575
576 pthread_attr_init (&attr);
577 pthread_attr_setstacksize (&attr, STACKSIZE);
578 pthread_attr_setdetachstate (&attr, PTHREAD_CREATE_DETACHED);
579
580 sigfillset (&fullsigset);
581
582 LOCK (wrklock);
583 sigprocmask (SIG_SETMASK, &fullsigset, &oldsigset);
584
585 if (pthread_create (&wrk->tid, &attr, aio_proc, (void *)wrk) == 0)
586 {
587 wrk->prev = &wrk_first;
588 wrk->next = wrk_first.next;
589 wrk_first.next->prev = wrk;
590 wrk_first.next = wrk;
591 started++;
592 }
593 else
594 free (wrk);
595
596 sigprocmask (SIG_SETMASK, &oldsigset, 0);
597 UNLOCK (wrklock);
598 }
599
600 static void req_send (aio_req req)
601 {
602 while (started < wanted && nreqs >= started)
603 start_thread ();
604
605 ++nreqs;
606
607 LOCK (reqlock);
608 reqq_push (&req_queue, req);
609 pthread_cond_signal (&reqwait);
610 UNLOCK (reqlock);
611 }
612
613 static void end_thread (void)
614 {
615 aio_req req;
616
617 Newz (0, req, 1, aio_cb);
618
619 req->type = REQ_QUIT;
620 req->pri = PRI_MAX + PRI_BIAS;
621
622 req_send (req);
623 }
624
625 static void min_parallel (int nthreads)
626 {
627 if (wanted < nthreads)
628 wanted = nthreads;
629 }
630
631 static void max_parallel (int nthreads)
632 {
633 int cur = started;
634
635 if (wanted > nthreads)
636 wanted = nthreads;
637
638 while (cur > wanted)
639 {
640 end_thread ();
641 cur--;
642 }
643
644 while (started > wanted)
645 {
646 poll_wait ();
647 poll_cb (0);
648 }
649 }
650
651 static void create_pipe ()
652 {
653 if (pipe (respipe))
654 croak ("unable to initialize result pipe");
655
656 if (fcntl (respipe [0], F_SETFL, O_NONBLOCK))
657 croak ("cannot set result pipe to nonblocking mode");
658
659 if (fcntl (respipe [1], F_SETFL, O_NONBLOCK))
660 croak ("cannot set result pipe to nonblocking mode");
661 }
662
663 /*****************************************************************************/
664 /* work around various missing functions */
665
666 #if !HAVE_PREADWRITE
667 # define pread aio_pread
668 # define pwrite aio_pwrite
669
670 /*
671 * make our pread/pwrite safe against themselves, but not against
672 * normal read/write by using a mutex. slows down execution a lot,
673 * but that's your problem, not mine.
674 */
675 static pthread_mutex_t preadwritelock = PTHREAD_MUTEX_INITIALIZER;
676
677 static ssize_t pread (int fd, void *buf, size_t count, off_t offset)
678 {
679 ssize_t res;
680 off_t ooffset;
681
682 LOCK (preadwritelock);
683 ooffset = lseek (fd, 0, SEEK_CUR);
684 lseek (fd, offset, SEEK_SET);
685 res = read (fd, buf, count);
686 lseek (fd, ooffset, SEEK_SET);
687 UNLOCK (preadwritelock);
688
689 return res;
690 }
691
692 static ssize_t pwrite (int fd, void *buf, size_t count, off_t offset)
693 {
694 ssize_t res;
695 off_t ooffset;
696
697 LOCK (preadwritelock);
698 ooffset = lseek (fd, 0, SEEK_CUR);
699 lseek (fd, offset, SEEK_SET);
700 res = write (fd, buf, count);
701 lseek (fd, offset, SEEK_SET);
702 UNLOCK (preadwritelock);
703
704 return res;
705 }
706 #endif
707
708 #if !HAVE_FDATASYNC
709 # define fdatasync fsync
710 #endif
711
712 #if !HAVE_READAHEAD
713 # define readahead(fd,offset,count) aio_readahead (fd, offset, count, self)
714
715 static ssize_t aio_readahead (int fd, off_t offset, size_t count, worker *self)
716 {
717 dBUF;
718
719 while (count > 0)
720 {
721 size_t len = count < AIO_BUFSIZE ? count : AIO_BUFSIZE;
722
723 pread (fd, aio_buf, len, offset);
724 offset += len;
725 count -= len;
726 }
727
728 errno = 0;
729 }
730
731 #endif
732
733 #if !HAVE_READDIR_R
734 # define readdir_r aio_readdir_r
735
736 static pthread_mutex_t readdirlock = PTHREAD_MUTEX_INITIALIZER;
737
738 static int readdir_r (DIR *dirp, struct dirent *ent, struct dirent **res)
739 {
740 struct dirent *e;
741 int errorno;
742
743 LOCK (readdirlock);
744
745 e = readdir (dirp);
746 errorno = errno;
747
748 if (e)
749 {
750 *res = ent;
751 strcpy (ent->d_name, e->d_name);
752 }
753 else
754 *res = 0;
755
756 UNLOCK (readdirlock);
757
758 errno = errorno;
759 return e ? 0 : -1;
760 }
761 #endif
762
763 /* sendfile always needs emulation */
764 static ssize_t sendfile_ (int ofd, int ifd, off_t offset, size_t count, worker *self)
765 {
766 ssize_t res;
767
768 if (!count)
769 return 0;
770
771 #if HAVE_SENDFILE
772 # if __linux
773 res = sendfile (ofd, ifd, &offset, count);
774
775 # elif __freebsd
776 /*
777 * Of course, the freebsd sendfile is a dire hack with no thoughts
778 * wasted on making it similar to other I/O functions.
779 */
780 {
781 off_t sbytes;
782 res = sendfile (ifd, ofd, offset, count, 0, &sbytes, 0);
783
784 if (res < 0 && sbytes)
785 /* maybe only on EAGAIN only: as usual, the manpage leaves you guessing */
786 res = sbytes;
787 }
788
789 # elif __hpux
790 res = sendfile (ofd, ifd, offset, count, 0, 0);
791
792 # elif __solaris
793 {
794 struct sendfilevec vec;
795 size_t sbytes;
796
797 vec.sfv_fd = ifd;
798 vec.sfv_flag = 0;
799 vec.sfv_off = offset;
800 vec.sfv_len = count;
801
802 res = sendfilev (ofd, &vec, 1, &sbytes);
803
804 if (res < 0 && sbytes)
805 res = sbytes;
806 }
807
808 # endif
809 #else
810 res = -1;
811 errno = ENOSYS;
812 #endif
813
814 if (res < 0
815 && (errno == ENOSYS || errno == EINVAL || errno == ENOTSOCK
816 #if __solaris
817 || errno == EAFNOSUPPORT || errno == EPROTOTYPE
818 #endif
819 )
820 )
821 {
822 /* emulate sendfile. this is a major pain in the ass */
823 dBUF;
824
825 res = 0;
826
827 while (count)
828 {
829 ssize_t cnt;
830
831 cnt = pread (ifd, aio_buf, count > AIO_BUFSIZE ? AIO_BUFSIZE : count, offset);
832
833 if (cnt <= 0)
834 {
835 if (cnt && !res) res = -1;
836 break;
837 }
838
839 cnt = write (ofd, aio_buf, cnt);
840
841 if (cnt <= 0)
842 {
843 if (cnt && !res) res = -1;
844 break;
845 }
846
847 offset += cnt;
848 res += cnt;
849 count -= cnt;
850 }
851 }
852
853 return res;
854 }
855
856 /* read a full directory */
857 static void scandir_ (aio_req req, worker *self)
858 {
859 DIR *dirp;
860 union
861 {
862 struct dirent d;
863 char b [offsetof (struct dirent, d_name) + NAME_MAX + 1];
864 } *u;
865 struct dirent *entp;
866 char *name, *names;
867 int memlen = 4096;
868 int memofs = 0;
869 int res = 0;
870 int errorno;
871
872 LOCK (wrklock);
873 self->dirp = dirp = opendir (req->dataptr);
874 self->dbuf = u = malloc (sizeof (*u));
875 req->data2ptr = names = malloc (memlen);
876 UNLOCK (wrklock);
877
878 if (dirp && u && names)
879 for (;;)
880 {
881 errno = 0;
882 readdir_r (dirp, &u->d, &entp);
883
884 if (!entp)
885 break;
886
887 name = entp->d_name;
888
889 if (name [0] != '.' || (name [1] && (name [1] != '.' || name [2])))
890 {
891 int len = strlen (name) + 1;
892
893 res++;
894
895 while (memofs + len > memlen)
896 {
897 memlen *= 2;
898 LOCK (wrklock);
899 req->data2ptr = names = realloc (names, memlen);
900 UNLOCK (wrklock);
901
902 if (!names)
903 break;
904 }
905
906 memcpy (names + memofs, name, len);
907 memofs += len;
908 }
909 }
910
911 if (errno)
912 res = -1;
913
914 req->result = res;
915 }
916
917 /*****************************************************************************/
918
919 static void *aio_proc (void *thr_arg)
920 {
921 aio_req req;
922 int type;
923 worker *self = (worker *)thr_arg;
924
925 do
926 {
927 LOCK (reqlock);
928
929 for (;;)
930 {
931 self->req = req = reqq_shift (&req_queue);
932
933 if (req)
934 break;
935
936 pthread_cond_wait (&reqwait, &reqlock);
937 }
938
939 UNLOCK (reqlock);
940
941 errno = 0; /* strictly unnecessary */
942 type = req->type; /* remember type for QUIT check */
943
944 if (!(req->flags & FLAG_CANCELLED))
945 switch (type)
946 {
947 case REQ_READ: req->result = pread (req->fd, req->dataptr, req->length, req->offset); break;
948 case REQ_WRITE: req->result = pwrite (req->fd, req->dataptr, req->length, req->offset); break;
949
950 case REQ_READAHEAD: req->result = readahead (req->fd, req->offset, req->length); break;
951 case REQ_SENDFILE: req->result = sendfile_ (req->fd, req->fd2, req->offset, req->length, self); break;
952
953 case REQ_STAT: req->result = stat (req->dataptr, req->statdata); break;
954 case REQ_LSTAT: req->result = lstat (req->dataptr, req->statdata); break;
955 case REQ_FSTAT: req->result = fstat (req->fd , req->statdata); break;
956
957 case REQ_OPEN: req->result = open (req->dataptr, req->fd, req->mode); break;
958 case REQ_CLOSE: req->result = close (req->fd); break;
959 case REQ_UNLINK: req->result = unlink (req->dataptr); break;
960 case REQ_RMDIR: req->result = rmdir (req->dataptr); break;
961 case REQ_RENAME: req->result = rename (req->data2ptr, req->dataptr); break;
962 case REQ_LINK: req->result = link (req->data2ptr, req->dataptr); break;
963 case REQ_SYMLINK: req->result = symlink (req->data2ptr, req->dataptr); break;
964
965 case REQ_FDATASYNC: req->result = fdatasync (req->fd); break;
966 case REQ_FSYNC: req->result = fsync (req->fd); break;
967 case REQ_READDIR: scandir_ (req, self); break;
968
969 case REQ_BUSY:
970 {
971 struct timeval tv;
972
973 tv.tv_sec = req->fd;
974 tv.tv_usec = req->fd2;
975
976 req->result = select (0, 0, 0, 0, &tv);
977 }
978
979 case REQ_GROUP:
980 case REQ_NOP:
981 case REQ_QUIT:
982 break;
983
984 default:
985 req->result = ENOSYS;
986 break;
987 }
988
989 req->errorno = errno;
990
991 LOCK (reslock);
992
993 if (!reqq_push (&res_queue, req))
994 /* write a dummy byte to the pipe so fh becomes ready */
995 write (respipe [1], &respipe, 1);
996
997 self->req = 0;
998 worker_clear (self);
999
1000 UNLOCK (reslock);
1001 }
1002 while (type != REQ_QUIT);
1003
1004 LOCK (wrklock);
1005 worker_free (self);
1006 UNLOCK (wrklock);
1007
1008 return 0;
1009 }
1010
1011 /*****************************************************************************/
1012
1013 static void atfork_prepare (void)
1014 {
1015 LOCK (wrklock);
1016 LOCK (reqlock);
1017 LOCK (reslock);
1018 #if !HAVE_PREADWRITE
1019 LOCK (preadwritelock);
1020 #endif
1021 #if !HAVE_READDIR_R
1022 LOCK (readdirlock);
1023 #endif
1024 }
1025
1026 static void atfork_parent (void)
1027 {
1028 #if !HAVE_READDIR_R
1029 UNLOCK (readdirlock);
1030 #endif
1031 #if !HAVE_PREADWRITE
1032 UNLOCK (preadwritelock);
1033 #endif
1034 UNLOCK (reslock);
1035 UNLOCK (reqlock);
1036 UNLOCK (wrklock);
1037 }
1038
1039 static void atfork_child (void)
1040 {
1041 aio_req prv;
1042
1043 while (prv = reqq_shift (&req_queue))
1044 req_free (prv);
1045
1046 while (prv = reqq_shift (&res_queue))
1047 req_free (prv);
1048
1049 while (wrk_first.next != &wrk_first)
1050 {
1051 worker *wrk = wrk_first.next;
1052
1053 if (wrk->req)
1054 req_free (wrk->req);
1055
1056 worker_clear (wrk);
1057 worker_free (wrk);
1058 }
1059
1060 started = 0;
1061 nreqs = 0;
1062
1063 close (respipe [0]);
1064 close (respipe [1]);
1065 create_pipe ();
1066
1067 atfork_parent ();
1068 }
1069
1070 #define dREQ \
1071 aio_req req; \
1072 int req_pri = next_pri; \
1073 next_pri = DEFAULT_PRI + PRI_BIAS; \
1074 \
1075 if (SvOK (callback) && !SvROK (callback)) \
1076 croak ("callback must be undef or of reference type"); \
1077 \
1078 Newz (0, req, 1, aio_cb); \
1079 if (!req) \
1080 croak ("out of memory during aio_req allocation"); \
1081 \
1082 req->callback = newSVsv (callback); \
1083 req->pri = req_pri
1084
1085 #define REQ_SEND \
1086 req_send (req); \
1087 \
1088 if (GIMME_V != G_VOID) \
1089 XPUSHs (req_sv (req, AIO_REQ_KLASS));
1090
1091 MODULE = IO::AIO PACKAGE = IO::AIO
1092
1093 PROTOTYPES: ENABLE
1094
1095 BOOT:
1096 {
1097 HV *stash = gv_stashpv ("IO::AIO", 1);
1098 newCONSTSUB (stash, "EXDEV", newSViv (EXDEV));
1099 newCONSTSUB (stash, "O_RDONLY", newSViv (O_RDONLY));
1100 newCONSTSUB (stash, "O_WRONLY", newSViv (O_WRONLY));
1101
1102 create_pipe ();
1103 pthread_atfork (atfork_prepare, atfork_parent, atfork_child);
1104 }
1105
1106 void
1107 min_parallel (nthreads)
1108 int nthreads
1109 PROTOTYPE: $
1110
1111 void
1112 max_parallel (nthreads)
1113 int nthreads
1114 PROTOTYPE: $
1115
1116 void
1117 aio_open (pathname,flags,mode,callback=&PL_sv_undef)
1118 SV * pathname
1119 int flags
1120 int mode
1121 SV * callback
1122 PROTOTYPE: $$$;$
1123 PPCODE:
1124 {
1125 dREQ;
1126
1127 req->type = REQ_OPEN;
1128 req->data = newSVsv (pathname);
1129 req->dataptr = SvPVbyte_nolen (req->data);
1130 req->fd = flags;
1131 req->mode = mode;
1132
1133 REQ_SEND;
1134 }
1135
1136 void
1137 aio_close (fh,callback=&PL_sv_undef)
1138 SV * fh
1139 SV * callback
1140 PROTOTYPE: $;$
1141 ALIAS:
1142 aio_close = REQ_CLOSE
1143 aio_fsync = REQ_FSYNC
1144 aio_fdatasync = REQ_FDATASYNC
1145 PPCODE:
1146 {
1147 dREQ;
1148
1149 req->type = ix;
1150 req->fh = newSVsv (fh);
1151 req->fd = PerlIO_fileno (IoIFP (sv_2io (fh)));
1152
1153 REQ_SEND (req);
1154 }
1155
1156 void
1157 aio_read (fh,offset,length,data,dataoffset,callback=&PL_sv_undef)
1158 SV * fh
1159 UV offset
1160 UV length
1161 SV * data
1162 UV dataoffset
1163 SV * callback
1164 ALIAS:
1165 aio_read = REQ_READ
1166 aio_write = REQ_WRITE
1167 PROTOTYPE: $$$$$;$
1168 PPCODE:
1169 {
1170 aio_req req;
1171 STRLEN svlen;
1172 char *svptr = SvPVbyte (data, svlen);
1173
1174 SvUPGRADE (data, SVt_PV);
1175 SvPOK_on (data);
1176
1177 if (dataoffset < 0)
1178 dataoffset += svlen;
1179
1180 if (dataoffset < 0 || dataoffset > svlen)
1181 croak ("data offset outside of string");
1182
1183 if (ix == REQ_WRITE)
1184 {
1185 /* write: check length and adjust. */
1186 if (length < 0 || length + dataoffset > svlen)
1187 length = svlen - dataoffset;
1188 }
1189 else
1190 {
1191 /* read: grow scalar as necessary */
1192 svptr = SvGROW (data, length + dataoffset);
1193 }
1194
1195 if (length < 0)
1196 croak ("length must not be negative");
1197
1198 {
1199 dREQ;
1200
1201 req->type = ix;
1202 req->fh = newSVsv (fh);
1203 req->fd = PerlIO_fileno (ix == REQ_READ ? IoIFP (sv_2io (fh))
1204 : IoOFP (sv_2io (fh)));
1205 req->offset = offset;
1206 req->length = length;
1207 req->data = SvREFCNT_inc (data);
1208 req->dataptr = (char *)svptr + dataoffset;
1209
1210 if (!SvREADONLY (data))
1211 {
1212 SvREADONLY_on (data);
1213 req->data2ptr = (void *)data;
1214 }
1215
1216 REQ_SEND;
1217 }
1218 }
1219
1220 void
1221 aio_sendfile (out_fh,in_fh,in_offset,length,callback=&PL_sv_undef)
1222 SV * out_fh
1223 SV * in_fh
1224 UV in_offset
1225 UV length
1226 SV * callback
1227 PROTOTYPE: $$$$;$
1228 PPCODE:
1229 {
1230 dREQ;
1231
1232 req->type = REQ_SENDFILE;
1233 req->fh = newSVsv (out_fh);
1234 req->fd = PerlIO_fileno (IoIFP (sv_2io (out_fh)));
1235 req->fh2 = newSVsv (in_fh);
1236 req->fd2 = PerlIO_fileno (IoIFP (sv_2io (in_fh)));
1237 req->offset = in_offset;
1238 req->length = length;
1239
1240 REQ_SEND;
1241 }
1242
1243 void
1244 aio_readahead (fh,offset,length,callback=&PL_sv_undef)
1245 SV * fh
1246 UV offset
1247 IV length
1248 SV * callback
1249 PROTOTYPE: $$$;$
1250 PPCODE:
1251 {
1252 dREQ;
1253
1254 req->type = REQ_READAHEAD;
1255 req->fh = newSVsv (fh);
1256 req->fd = PerlIO_fileno (IoIFP (sv_2io (fh)));
1257 req->offset = offset;
1258 req->length = length;
1259
1260 REQ_SEND;
1261 }
1262
1263 void
1264 aio_stat (fh_or_path,callback=&PL_sv_undef)
1265 SV * fh_or_path
1266 SV * callback
1267 ALIAS:
1268 aio_stat = REQ_STAT
1269 aio_lstat = REQ_LSTAT
1270 PPCODE:
1271 {
1272 dREQ;
1273
1274 New (0, req->statdata, 1, Stat_t);
1275 if (!req->statdata)
1276 {
1277 req_free (req);
1278 croak ("out of memory during aio_req->statdata allocation");
1279 }
1280
1281 if (SvPOK (fh_or_path))
1282 {
1283 req->type = ix;
1284 req->data = newSVsv (fh_or_path);
1285 req->dataptr = SvPVbyte_nolen (req->data);
1286 }
1287 else
1288 {
1289 req->type = REQ_FSTAT;
1290 req->fh = newSVsv (fh_or_path);
1291 req->fd = PerlIO_fileno (IoIFP (sv_2io (fh_or_path)));
1292 }
1293
1294 REQ_SEND;
1295 }
1296
1297 void
1298 aio_unlink (pathname,callback=&PL_sv_undef)
1299 SV * pathname
1300 SV * callback
1301 ALIAS:
1302 aio_unlink = REQ_UNLINK
1303 aio_rmdir = REQ_RMDIR
1304 aio_readdir = REQ_READDIR
1305 PPCODE:
1306 {
1307 dREQ;
1308
1309 req->type = ix;
1310 req->data = newSVsv (pathname);
1311 req->dataptr = SvPVbyte_nolen (req->data);
1312
1313 REQ_SEND;
1314 }
1315
1316 void
1317 aio_link (oldpath,newpath,callback=&PL_sv_undef)
1318 SV * oldpath
1319 SV * newpath
1320 SV * callback
1321 ALIAS:
1322 aio_link = REQ_LINK
1323 aio_symlink = REQ_SYMLINK
1324 aio_rename = REQ_RENAME
1325 PPCODE:
1326 {
1327 dREQ;
1328
1329 req->type = ix;
1330 req->fh = newSVsv (oldpath);
1331 req->data2ptr = SvPVbyte_nolen (req->fh);
1332 req->data = newSVsv (newpath);
1333 req->dataptr = SvPVbyte_nolen (req->data);
1334
1335 REQ_SEND;
1336 }
1337
1338 void
1339 aio_busy (delay,callback=&PL_sv_undef)
1340 double delay
1341 SV * callback
1342 PPCODE:
1343 {
1344 dREQ;
1345
1346 req->type = REQ_BUSY;
1347 req->fd = delay < 0. ? 0 : delay;
1348 req->fd2 = delay < 0. ? 0 : 1000. * (delay - req->fd);
1349
1350 REQ_SEND;
1351 }
1352
1353 void
1354 aio_group (callback=&PL_sv_undef)
1355 SV * callback
1356 PROTOTYPE: ;$
1357 PPCODE:
1358 {
1359 dREQ;
1360
1361 req->type = REQ_GROUP;
1362 req_send (req);
1363
1364 XPUSHs (req_sv (req, AIO_GRP_KLASS));
1365 }
1366
1367 void
1368 aio_nop (callback=&PL_sv_undef)
1369 SV * callback
1370 PPCODE:
1371 {
1372 dREQ;
1373
1374 req->type = REQ_NOP;
1375
1376 REQ_SEND;
1377 }
1378
1379 void
1380 aioreq_pri (int pri = DEFAULT_PRI)
1381 CODE:
1382 if (pri < PRI_MIN) pri = PRI_MIN;
1383 if (pri > PRI_MAX) pri = PRI_MAX;
1384 next_pri = pri + PRI_BIAS;
1385
1386 void
1387 aioreq_nice (int nice = 0)
1388 CODE:
1389 nice = next_pri - nice;
1390 if (nice < PRI_MIN) nice = PRI_MIN;
1391 if (nice > PRI_MAX) nice = PRI_MAX;
1392 next_pri = nice + PRI_BIAS;
1393
1394 void
1395 flush ()
1396 PROTOTYPE:
1397 CODE:
1398 while (nreqs)
1399 {
1400 poll_wait ();
1401 poll_cb (0);
1402 }
1403
1404 void
1405 poll()
1406 PROTOTYPE:
1407 CODE:
1408 if (nreqs)
1409 {
1410 poll_wait ();
1411 poll_cb (0);
1412 }
1413
1414 int
1415 poll_fileno()
1416 PROTOTYPE:
1417 CODE:
1418 RETVAL = respipe [0];
1419 OUTPUT:
1420 RETVAL
1421
1422 int
1423 poll_cb(...)
1424 PROTOTYPE:
1425 CODE:
1426 RETVAL = poll_cb (0);
1427 OUTPUT:
1428 RETVAL
1429
1430 int
1431 poll_some(int max = 0)
1432 PROTOTYPE: $
1433 CODE:
1434 RETVAL = poll_cb (max);
1435 OUTPUT:
1436 RETVAL
1437
1438 void
1439 poll_wait()
1440 PROTOTYPE:
1441 CODE:
1442 if (nreqs)
1443 poll_wait ();
1444
1445 int
1446 nreqs()
1447 PROTOTYPE:
1448 CODE:
1449 RETVAL = nreqs;
1450 OUTPUT:
1451 RETVAL
1452
1453 PROTOTYPES: DISABLE
1454
1455 MODULE = IO::AIO PACKAGE = IO::AIO::REQ
1456
1457 void
1458 cancel (aio_req_ornot req)
1459 CODE:
1460 req_cancel (req);
1461
1462 void
1463 cb (aio_req_ornot req, SV *callback=&PL_sv_undef)
1464 CODE:
1465 SvREFCNT_dec (req->callback);
1466 req->callback = newSVsv (callback);
1467
1468 MODULE = IO::AIO PACKAGE = IO::AIO::GRP
1469
1470 void
1471 add (aio_req grp, ...)
1472 PPCODE:
1473 {
1474 int i;
1475 aio_req req;
1476
1477 if (grp->fd == 2)
1478 croak ("cannot add requests to IO::AIO::GRP after the group finished");
1479
1480 for (i = 1; i < items; ++i )
1481 {
1482 if (GIMME_V != G_VOID)
1483 XPUSHs (sv_2mortal (newSVsv (ST (i))));
1484
1485 req = SvAIO_REQ (ST (i));
1486
1487 if (req)
1488 {
1489 ++grp->length;
1490 req->grp = grp;
1491
1492 req->grp_prev = 0;
1493 req->grp_next = grp->grp_first;
1494
1495 if (grp->grp_first)
1496 grp->grp_first->grp_prev = req;
1497
1498 grp->grp_first = req;
1499 }
1500 }
1501 }
1502
1503 void
1504 cancel_subs (aio_req_ornot req)
1505 CODE:
1506 req_cancel_subs (req);
1507
1508 void
1509 result (aio_req grp, ...)
1510 CODE:
1511 {
1512 int i;
1513 AV *av = newAV ();
1514
1515 for (i = 1; i < items; ++i )
1516 av_push (av, newSVsv (ST (i)));
1517
1518 SvREFCNT_dec (grp->data);
1519 grp->data = (SV *)av;
1520 }
1521
1522 void
1523 limit (aio_req grp, int limit)
1524 CODE:
1525 grp->fd2 = limit;
1526 aio_grp_feed (grp);
1527
1528 void
1529 feed (aio_req grp, SV *callback=&PL_sv_undef)
1530 CODE:
1531 {
1532 SvREFCNT_dec (grp->fh2);
1533 grp->fh2 = newSVsv (callback);
1534
1535 if (grp->fd2 <= 0)
1536 grp->fd2 = 2;
1537
1538 aio_grp_feed (grp);
1539 }
1540