ViewVC Help
View File | Revision Log | Show Annotations | Download File
/cvs/IO-AIO/AIO.xs
Revision: 1.99
Committed: Fri Jun 1 05:51:21 2007 UTC (16 years, 11 months ago) by root
Branch: MAIN
Changes since 1.98: +169 -80 lines
Log Message:
*** empty log message ***

File Contents

# Content
1 #include "xthread.h"
2
3 #include <errno.h>
4
5 #include "EXTERN.h"
6 #include "perl.h"
7 #include "XSUB.h"
8
9 #include "autoconf/config.h"
10
11 #include <stddef.h>
12 #include <stdlib.h>
13 #include <errno.h>
14 #include <sys/time.h>
15 #include <sys/select.h>
16 #include <sys/types.h>
17 #include <sys/stat.h>
18 #include <limits.h>
19 #include <unistd.h>
20 #include <fcntl.h>
21 #include <signal.h>
22 #include <sched.h>
23 #include <utime.h>
24
25 #if HAVE_SENDFILE
26 # if __linux
27 # include <sys/sendfile.h>
28 # elif __freebsd
29 # include <sys/socket.h>
30 # include <sys/uio.h>
31 # elif __hpux
32 # include <sys/socket.h>
33 # elif __solaris /* not yet */
34 # include <sys/sendfile.h>
35 # else
36 # error sendfile support requested but not available
37 # endif
38 #endif
39
40 /* number of seconds after which idle threads exit */
41 #define IDLE_TIMEOUT 10
42
43 /* used for struct dirent, AIX doesn't provide it */
44 #ifndef NAME_MAX
45 # define NAME_MAX 4096
46 #endif
47
48 /* buffer size for various temporary buffers */
49 #define AIO_BUFSIZE 65536
50
51 #define dBUF \
52 char *aio_buf; \
53 LOCK (wrklock); \
54 self->dbuf = aio_buf = malloc (AIO_BUFSIZE); \
55 UNLOCK (wrklock); \
56 if (!aio_buf) \
57 return -1;
58
59 typedef SV SV8; /* byte-sv, used for argument-checking */
60
61 enum {
62 REQ_QUIT,
63 REQ_OPEN, REQ_CLOSE,
64 REQ_READ, REQ_WRITE, REQ_READAHEAD,
65 REQ_SENDFILE,
66 REQ_STAT, REQ_LSTAT, REQ_FSTAT,
67 REQ_UTIME, REQ_FUTIME, /* must be consecutive */
68 REQ_CHMOD, REQ_FCHMOD, /* must be consecutive */
69 REQ_CHOWN, REQ_FCHOWN, /* must be consecutive */
70 REQ_FSYNC, REQ_FDATASYNC,
71 REQ_UNLINK, REQ_RMDIR, REQ_MKDIR, REQ_RENAME,
72 REQ_MKNOD, REQ_READDIR,
73 REQ_LINK, REQ_SYMLINK, REQ_READLINK,
74 REQ_GROUP, REQ_NOP,
75 REQ_BUSY,
76 };
77
78 #define AIO_REQ_KLASS "IO::AIO::REQ"
79 #define AIO_GRP_KLASS "IO::AIO::GRP"
80
81 typedef struct aio_cb
82 {
83 struct aio_cb *volatile next;
84
85 SV *callback;
86 SV *sv1, *sv2;
87 void *ptr1, *ptr2;
88 off_t offs;
89 size_t size;
90 ssize_t result;
91 double nv1, nv2;
92
93 STRLEN stroffset;
94 int type;
95 int int1, int2, int3;
96 int errorno;
97 mode_t mode; /* open */
98
99 unsigned char flags;
100 unsigned char pri;
101
102 SV *self; /* the perl counterpart of this request, if any */
103 struct aio_cb *grp, *grp_prev, *grp_next, *grp_first;
104 } aio_cb;
105
106 enum {
107 FLAG_CANCELLED = 0x01, /* request was cancelled */
108 FLAG_SV1_RO_OFF = 0x40, /* data was set readonly */
109 FLAG_PTR2_FREE = 0x80, /* need to free(ptr2) */
110 };
111
112 typedef aio_cb *aio_req;
113 typedef aio_cb *aio_req_ornot;
114
115 enum {
116 PRI_MIN = -4,
117 PRI_MAX = 4,
118
119 DEFAULT_PRI = 0,
120 PRI_BIAS = -PRI_MIN,
121 NUM_PRI = PRI_MAX + PRI_BIAS + 1,
122 };
123
124 #define AIO_TICKS ((1000000 + 1023) >> 10)
125
126 static unsigned int max_poll_time = 0;
127 static unsigned int max_poll_reqs = 0;
128
129 /* calculcate time difference in ~1/AIO_TICKS of a second */
130 static int tvdiff (struct timeval *tv1, struct timeval *tv2)
131 {
132 return (tv2->tv_sec - tv1->tv_sec ) * AIO_TICKS
133 + ((tv2->tv_usec - tv1->tv_usec) >> 10);
134 }
135
136 static thread_t main_tid;
137 static int main_sig;
138 static int block_sig_level;
139
140 void block_sig ()
141 {
142 sigset_t ss;
143
144 if (block_sig_level++)
145 return;
146
147 if (!main_sig)
148 return;
149
150 sigemptyset (&ss);
151 sigaddset (&ss, main_sig);
152 pthread_sigmask (SIG_BLOCK, &ss, 0);
153 }
154
155 void unblock_sig ()
156 {
157 sigset_t ss;
158
159 if (--block_sig_level)
160 return;
161
162 if (!main_sig)
163 return;
164
165 sigemptyset (&ss);
166 sigaddset (&ss, main_sig);
167 pthread_sigmask (SIG_UNBLOCK, &ss, 0);
168 }
169
170 static int next_pri = DEFAULT_PRI + PRI_BIAS;
171
172 static unsigned int started, idle, wanted;
173
174 /* worker threads management */
175 static mutex_t wrklock = MUTEX_INIT;
176
177 typedef struct worker {
178 /* locked by wrklock */
179 struct worker *prev, *next;
180
181 thread_t tid;
182
183 /* locked by reslock, reqlock or wrklock */
184 aio_req req; /* currently processed request */
185 void *dbuf;
186 DIR *dirp;
187 } worker;
188
189 static worker wrk_first = { &wrk_first, &wrk_first, 0 };
190
191 static void worker_clear (worker *wrk)
192 {
193 if (wrk->dirp)
194 {
195 closedir (wrk->dirp);
196 wrk->dirp = 0;
197 }
198
199 if (wrk->dbuf)
200 {
201 free (wrk->dbuf);
202 wrk->dbuf = 0;
203 }
204 }
205
206 static void worker_free (worker *wrk)
207 {
208 wrk->next->prev = wrk->prev;
209 wrk->prev->next = wrk->next;
210
211 free (wrk);
212 }
213
214 static volatile unsigned int nreqs, nready, npending;
215 static volatile unsigned int max_idle = 4;
216 static volatile unsigned int max_outstanding = 0xffffffff;
217 static int respipe [2];
218
219 static mutex_t reslock = MUTEX_INIT;
220 static mutex_t reqlock = MUTEX_INIT;
221 static cond_t reqwait = COND_INIT;
222
223 #if WORDACCESS_UNSAFE
224
225 static unsigned int get_nready ()
226 {
227 unsigned int retval;
228
229 LOCK (reqlock);
230 retval = nready;
231 UNLOCK (reqlock);
232
233 return retval;
234 }
235
236 static unsigned int get_npending ()
237 {
238 unsigned int retval;
239
240 LOCK (reslock);
241 retval = npending;
242 UNLOCK (reslock);
243
244 return retval;
245 }
246
247 static unsigned int get_nthreads ()
248 {
249 unsigned int retval;
250
251 LOCK (wrklock);
252 retval = started;
253 UNLOCK (wrklock);
254
255 return retval;
256 }
257
258 #else
259
260 # define get_nready() nready
261 # define get_npending() npending
262 # define get_nthreads() started
263
264 #endif
265
266 /*
267 * a somewhat faster data structure might be nice, but
268 * with 8 priorities this actually needs <20 insns
269 * per shift, the most expensive operation.
270 */
271 typedef struct {
272 aio_req qs[NUM_PRI], qe[NUM_PRI]; /* qstart, qend */
273 int size;
274 } reqq;
275
276 static reqq req_queue;
277 static reqq res_queue;
278
279 int reqq_push (reqq *q, aio_req req)
280 {
281 int pri = req->pri;
282 req->next = 0;
283
284 if (q->qe[pri])
285 {
286 q->qe[pri]->next = req;
287 q->qe[pri] = req;
288 }
289 else
290 q->qe[pri] = q->qs[pri] = req;
291
292 return q->size++;
293 }
294
295 aio_req reqq_shift (reqq *q)
296 {
297 int pri;
298
299 if (!q->size)
300 return 0;
301
302 --q->size;
303
304 for (pri = NUM_PRI; pri--; )
305 {
306 aio_req req = q->qs[pri];
307
308 if (req)
309 {
310 if (!(q->qs[pri] = req->next))
311 q->qe[pri] = 0;
312
313 return req;
314 }
315 }
316
317 abort ();
318 }
319
320 static int poll_cb ();
321 static int req_invoke (aio_req req);
322 static void req_free (aio_req req);
323 static void req_cancel (aio_req req);
324
325 /* must be called at most once */
326 static SV *req_sv (aio_req req, const char *klass)
327 {
328 if (!req->self)
329 {
330 req->self = (SV *)newHV ();
331 sv_magic (req->self, 0, PERL_MAGIC_ext, (char *)req, 0);
332 }
333
334 return sv_2mortal (sv_bless (newRV_inc (req->self), gv_stashpv (klass, 1)));
335 }
336
337 static aio_req SvAIO_REQ (SV *sv)
338 {
339 MAGIC *mg;
340
341 if (!sv_derived_from (sv, AIO_REQ_KLASS) || !SvROK (sv))
342 croak ("object of class " AIO_REQ_KLASS " expected");
343
344 mg = mg_find (SvRV (sv), PERL_MAGIC_ext);
345
346 return mg ? (aio_req)mg->mg_ptr : 0;
347 }
348
349 static void aio_grp_feed (aio_req grp)
350 {
351 block_sig ();
352
353 while (grp->size < grp->int2 && !(grp->flags & FLAG_CANCELLED))
354 {
355 int old_len = grp->size;
356
357 if (grp->sv2 && SvOK (grp->sv2))
358 {
359 dSP;
360
361 ENTER;
362 SAVETMPS;
363 PUSHMARK (SP);
364 XPUSHs (req_sv (grp, AIO_GRP_KLASS));
365 PUTBACK;
366 call_sv (grp->sv2, G_VOID | G_EVAL | G_KEEPERR);
367 SPAGAIN;
368 FREETMPS;
369 LEAVE;
370 }
371
372 /* stop if no progress has been made */
373 if (old_len == grp->size)
374 {
375 SvREFCNT_dec (grp->sv2);
376 grp->sv2 = 0;
377 break;
378 }
379 }
380
381 unblock_sig ();
382 }
383
384 static void aio_grp_dec (aio_req grp)
385 {
386 --grp->size;
387
388 /* call feeder, if applicable */
389 aio_grp_feed (grp);
390
391 /* finish, if done */
392 if (!grp->size && grp->int1)
393 {
394 block_sig ();
395
396 if (!req_invoke (grp))
397 {
398 req_free (grp);
399 unblock_sig ();
400 croak (0);
401 }
402
403 req_free (grp);
404 unblock_sig ();
405 }
406 }
407
408 static int req_invoke (aio_req req)
409 {
410 dSP;
411
412 if (req->flags & FLAG_SV1_RO_OFF)
413 SvREADONLY_off (req->sv1);
414
415 if (!(req->flags & FLAG_CANCELLED) && SvOK (req->callback))
416 {
417 ENTER;
418 SAVETMPS;
419 PUSHMARK (SP);
420 EXTEND (SP, 1);
421
422 switch (req->type)
423 {
424 case REQ_READDIR:
425 {
426 SV *rv = &PL_sv_undef;
427
428 if (req->result >= 0)
429 {
430 int i;
431 char *buf = req->ptr2;
432 AV *av = newAV ();
433
434 av_extend (av, req->result - 1);
435
436 for (i = 0; i < req->result; ++i)
437 {
438 SV *sv = newSVpv (buf, 0);
439
440 av_store (av, i, sv);
441 buf += SvCUR (sv) + 1;
442 }
443
444 rv = sv_2mortal (newRV_noinc ((SV *)av));
445 }
446
447 PUSHs (rv);
448 }
449 break;
450
451 case REQ_OPEN:
452 {
453 /* convert fd to fh */
454 SV *fh;
455
456 PUSHs (sv_2mortal (newSViv (req->result)));
457 PUTBACK;
458 call_pv ("IO::AIO::_fd2fh", G_SCALAR | G_EVAL);
459 SPAGAIN;
460
461 fh = POPs;
462 PUSHMARK (SP);
463 XPUSHs (fh);
464 }
465 break;
466
467 case REQ_GROUP:
468 req->int1 = 2; /* mark group as finished */
469
470 if (req->sv1)
471 {
472 int i;
473 AV *av = (AV *)req->sv1;
474
475 EXTEND (SP, AvFILL (av) + 1);
476 for (i = 0; i <= AvFILL (av); ++i)
477 PUSHs (*av_fetch (av, i, 0));
478 }
479 break;
480
481 case REQ_NOP:
482 case REQ_BUSY:
483 break;
484
485 case REQ_READLINK:
486 if (req->result > 0)
487 {
488 SvCUR_set (req->sv2, req->result);
489 *SvEND (req->sv2) = 0;
490 PUSHs (req->sv2);
491 }
492 break;
493
494 case REQ_STAT:
495 case REQ_LSTAT:
496 case REQ_FSTAT:
497 PL_laststype = req->type == REQ_LSTAT ? OP_LSTAT : OP_STAT;
498 PL_laststatval = req->result;
499 PL_statcache = *(Stat_t *)(req->ptr2);
500 PUSHs (sv_2mortal (newSViv (req->result)));
501 break;
502
503 case REQ_READ:
504 SvCUR_set (req->sv2, req->stroffset + (req->result > 0 ? req->result : 0));
505 *SvEND (req->sv2) = 0;
506 PUSHs (sv_2mortal (newSViv (req->result)));
507 break;
508
509 default:
510 PUSHs (sv_2mortal (newSViv (req->result)));
511 break;
512 }
513
514 errno = req->errorno;
515
516 PUTBACK;
517 call_sv (req->callback, G_VOID | G_EVAL);
518 SPAGAIN;
519
520 FREETMPS;
521 LEAVE;
522 }
523
524 if (req->grp)
525 {
526 aio_req grp = req->grp;
527
528 /* unlink request */
529 if (req->grp_next) req->grp_next->grp_prev = req->grp_prev;
530 if (req->grp_prev) req->grp_prev->grp_next = req->grp_next;
531
532 if (grp->grp_first == req)
533 grp->grp_first = req->grp_next;
534
535 aio_grp_dec (grp);
536 }
537
538 return !SvTRUE (ERRSV);
539 }
540
541 static void req_free (aio_req req)
542 {
543 if (req->self)
544 {
545 sv_unmagic (req->self, PERL_MAGIC_ext);
546 SvREFCNT_dec (req->self);
547 }
548
549 SvREFCNT_dec (req->sv1);
550 SvREFCNT_dec (req->sv2);
551 SvREFCNT_dec (req->callback);
552
553 if (req->flags & FLAG_PTR2_FREE)
554 free (req->ptr2);
555
556 Safefree (req);
557 }
558
559 static void req_cancel_subs (aio_req grp)
560 {
561 aio_req sub;
562
563 if (grp->type != REQ_GROUP)
564 return;
565
566 SvREFCNT_dec (grp->sv2);
567 grp->sv2 = 0;
568
569 for (sub = grp->grp_first; sub; sub = sub->grp_next)
570 req_cancel (sub);
571 }
572
573 static void req_cancel (aio_req req)
574 {
575 req->flags |= FLAG_CANCELLED;
576
577 req_cancel_subs (req);
578 }
579
580 static void *aio_proc(void *arg);
581
582 static void start_thread (void)
583 {
584 worker *wrk = calloc (1, sizeof (worker));
585
586 if (!wrk)
587 croak ("unable to allocate worker thread data");
588
589 LOCK (wrklock);
590
591 if (thread_create (&wrk->tid, aio_proc, (void *)wrk))
592 {
593 wrk->prev = &wrk_first;
594 wrk->next = wrk_first.next;
595 wrk_first.next->prev = wrk;
596 wrk_first.next = wrk;
597 ++started;
598 }
599 else
600 free (wrk);
601
602 UNLOCK (wrklock);
603 }
604
605 static void maybe_start_thread ()
606 {
607 if (get_nthreads () >= wanted)
608 return;
609
610 /* todo: maybe use idle here, but might be less exact */
611 if (0 <= (int)get_nthreads () + (int)get_npending () - (int)nreqs)
612 return;
613
614 start_thread ();
615 }
616
617 static void req_send (aio_req req)
618 {
619 block_sig ();
620
621 ++nreqs;
622
623 LOCK (reqlock);
624 ++nready;
625 reqq_push (&req_queue, req);
626 COND_SIGNAL (reqwait);
627 UNLOCK (reqlock);
628
629 unblock_sig ();
630
631 maybe_start_thread ();
632 }
633
634 static void end_thread (void)
635 {
636 aio_req req;
637
638 Newz (0, req, 1, aio_cb);
639
640 req->type = REQ_QUIT;
641 req->pri = PRI_MAX + PRI_BIAS;
642
643 LOCK (reqlock);
644 reqq_push (&req_queue, req);
645 COND_SIGNAL (reqwait);
646 UNLOCK (reqlock);
647
648 LOCK (wrklock);
649 --started;
650 UNLOCK (wrklock);
651 }
652
653 static void set_max_idle (int nthreads)
654 {
655 if (WORDACCESS_UNSAFE) LOCK (reqlock);
656 max_idle = nthreads <= 0 ? 1 : nthreads;
657 if (WORDACCESS_UNSAFE) UNLOCK (reqlock);
658 }
659
660 static void min_parallel (int nthreads)
661 {
662 if (wanted < nthreads)
663 wanted = nthreads;
664 }
665
666 static void max_parallel (int nthreads)
667 {
668 if (wanted > nthreads)
669 wanted = nthreads;
670
671 while (started > wanted)
672 end_thread ();
673 }
674
675 static void poll_wait ()
676 {
677 fd_set rfd;
678
679 while (nreqs)
680 {
681 int size;
682 if (WORDACCESS_UNSAFE) LOCK (reslock);
683 size = res_queue.size;
684 if (WORDACCESS_UNSAFE) UNLOCK (reslock);
685
686 if (size)
687 return;
688
689 maybe_start_thread ();
690
691 FD_ZERO(&rfd);
692 FD_SET(respipe [0], &rfd);
693
694 select (respipe [0] + 1, &rfd, 0, 0, 0);
695 }
696 }
697
698 static int poll_cb ()
699 {
700 dSP;
701 int count = 0;
702 int maxreqs = max_poll_reqs;
703 int do_croak = 0;
704 struct timeval tv_start, tv_now;
705 aio_req req;
706
707 if (max_poll_time)
708 gettimeofday (&tv_start, 0);
709
710 block_sig ();
711
712 for (;;)
713 {
714 for (;;)
715 {
716 maybe_start_thread ();
717
718 LOCK (reslock);
719 req = reqq_shift (&res_queue);
720
721 if (req)
722 {
723 --npending;
724
725 if (!res_queue.size)
726 {
727 /* read any signals sent by the worker threads */
728 char buf [4];
729 while (read (respipe [0], buf, 4) == 4)
730 ;
731 }
732 }
733
734 UNLOCK (reslock);
735
736 if (!req)
737 break;
738
739 --nreqs;
740
741 if (req->type == REQ_GROUP && req->size)
742 {
743 req->int1 = 1; /* mark request as delayed */
744 continue;
745 }
746 else
747 {
748 if (!req_invoke (req))
749 {
750 req_free (req);
751 unblock_sig ();
752 croak (0);
753 }
754
755 count++;
756 }
757
758 req_free (req);
759
760 if (maxreqs && !--maxreqs)
761 break;
762
763 if (max_poll_time)
764 {
765 gettimeofday (&tv_now, 0);
766
767 if (tvdiff (&tv_start, &tv_now) >= max_poll_time)
768 break;
769 }
770 }
771
772 if (nreqs <= max_outstanding)
773 break;
774
775 poll_wait ();
776
777 ++maxreqs;
778 }
779
780 unblock_sig ();
781 return count;
782 }
783
784 static void create_pipe ()
785 {
786 if (pipe (respipe))
787 croak ("unable to initialize result pipe");
788
789 if (fcntl (respipe [0], F_SETFL, O_NONBLOCK))
790 croak ("cannot set result pipe to nonblocking mode");
791
792 if (fcntl (respipe [1], F_SETFL, O_NONBLOCK))
793 croak ("cannot set result pipe to nonblocking mode");
794 }
795
796 /*****************************************************************************/
797 /* work around various missing functions */
798
799 #if !HAVE_PREADWRITE
800 # define pread aio_pread
801 # define pwrite aio_pwrite
802
803 /*
804 * make our pread/pwrite safe against themselves, but not against
805 * normal read/write by using a mutex. slows down execution a lot,
806 * but that's your problem, not mine.
807 */
808 static mutex_t preadwritelock = MUTEX_INIT;
809
810 static ssize_t pread (int fd, void *buf, size_t count, off_t offset)
811 {
812 ssize_t res;
813 off_t ooffset;
814
815 LOCK (preadwritelock);
816 ooffset = lseek (fd, 0, SEEK_CUR);
817 lseek (fd, offset, SEEK_SET);
818 res = read (fd, buf, count);
819 lseek (fd, ooffset, SEEK_SET);
820 UNLOCK (preadwritelock);
821
822 return res;
823 }
824
825 static ssize_t pwrite (int fd, void *buf, size_t count, off_t offset)
826 {
827 ssize_t res;
828 off_t ooffset;
829
830 LOCK (preadwritelock);
831 ooffset = lseek (fd, 0, SEEK_CUR);
832 lseek (fd, offset, SEEK_SET);
833 res = write (fd, buf, count);
834 lseek (fd, offset, SEEK_SET);
835 UNLOCK (preadwritelock);
836
837 return res;
838 }
839 #endif
840
841 #ifndef HAVE_FUTIMES
842
843 # define utimes(path,times) aio_utimes (path, times)
844 # define futimes(fd,times) aio_futimes (fd, times)
845
846 int aio_utimes (const char *filename, const struct timeval times[2])
847 {
848 if (times)
849 {
850 struct utimbuf buf;
851
852 buf.actime = times[0].tv_sec;
853 buf.modtime = times[1].tv_sec;
854
855 return utime (filename, &buf);
856 }
857 else
858 return utime (filename, 0);
859 }
860
861 int aio_futimes (int fd, const struct timeval tv[2])
862 {
863 errno = ENOSYS;
864 return -1;
865 }
866
867 #endif
868
869 #if !HAVE_FDATASYNC
870 # define fdatasync fsync
871 #endif
872
873 #if !HAVE_READAHEAD
874 # define readahead(fd,offset,count) aio_readahead (fd, offset, count, self)
875
876 static ssize_t aio_readahead (int fd, off_t offset, size_t count, worker *self)
877 {
878 dBUF;
879
880 while (count > 0)
881 {
882 size_t len = count < AIO_BUFSIZE ? count : AIO_BUFSIZE;
883
884 pread (fd, aio_buf, len, offset);
885 offset += len;
886 count -= len;
887 }
888
889 errno = 0;
890 }
891
892 #endif
893
894 #if !HAVE_READDIR_R
895 # define readdir_r aio_readdir_r
896
897 static mutex_t readdirlock = MUTEX_INIT;
898
899 static int readdir_r (DIR *dirp, struct dirent *ent, struct dirent **res)
900 {
901 struct dirent *e;
902 int errorno;
903
904 LOCK (readdirlock);
905
906 e = readdir (dirp);
907 errorno = errno;
908
909 if (e)
910 {
911 *res = ent;
912 strcpy (ent->d_name, e->d_name);
913 }
914 else
915 *res = 0;
916
917 UNLOCK (readdirlock);
918
919 errno = errorno;
920 return e ? 0 : -1;
921 }
922 #endif
923
924 /* sendfile always needs emulation */
925 static ssize_t sendfile_ (int ofd, int ifd, off_t offset, size_t count, worker *self)
926 {
927 ssize_t res;
928
929 if (!count)
930 return 0;
931
932 #if HAVE_SENDFILE
933 # if __linux
934 res = sendfile (ofd, ifd, &offset, count);
935
936 # elif __freebsd
937 /*
938 * Of course, the freebsd sendfile is a dire hack with no thoughts
939 * wasted on making it similar to other I/O functions.
940 */
941 {
942 off_t sbytes;
943 res = sendfile (ifd, ofd, offset, count, 0, &sbytes, 0);
944
945 if (res < 0 && sbytes)
946 /* maybe only on EAGAIN: as usual, the manpage leaves you guessing */
947 res = sbytes;
948 }
949
950 # elif __hpux
951 res = sendfile (ofd, ifd, offset, count, 0, 0);
952
953 # elif __solaris
954 {
955 struct sendfilevec vec;
956 size_t sbytes;
957
958 vec.sfv_fd = ifd;
959 vec.sfv_flag = 0;
960 vec.sfv_off = offset;
961 vec.sfv_len = count;
962
963 res = sendfilev (ofd, &vec, 1, &sbytes);
964
965 if (res < 0 && sbytes)
966 res = sbytes;
967 }
968
969 # endif
970 #else
971 res = -1;
972 errno = ENOSYS;
973 #endif
974
975 if (res < 0
976 && (errno == ENOSYS || errno == EINVAL || errno == ENOTSOCK
977 #if __solaris
978 || errno == EAFNOSUPPORT || errno == EPROTOTYPE
979 #endif
980 )
981 )
982 {
983 /* emulate sendfile. this is a major pain in the ass */
984 dBUF;
985
986 res = 0;
987
988 while (count)
989 {
990 ssize_t cnt;
991
992 cnt = pread (ifd, aio_buf, count > AIO_BUFSIZE ? AIO_BUFSIZE : count, offset);
993
994 if (cnt <= 0)
995 {
996 if (cnt && !res) res = -1;
997 break;
998 }
999
1000 cnt = write (ofd, aio_buf, cnt);
1001
1002 if (cnt <= 0)
1003 {
1004 if (cnt && !res) res = -1;
1005 break;
1006 }
1007
1008 offset += cnt;
1009 res += cnt;
1010 count -= cnt;
1011 }
1012 }
1013
1014 return res;
1015 }
1016
1017 /* read a full directory */
1018 static void scandir_ (aio_req req, worker *self)
1019 {
1020 DIR *dirp;
1021 union
1022 {
1023 struct dirent d;
1024 char b [offsetof (struct dirent, d_name) + NAME_MAX + 1];
1025 } *u;
1026 struct dirent *entp;
1027 char *name, *names;
1028 int memlen = 4096;
1029 int memofs = 0;
1030 int res = 0;
1031 int errorno;
1032
1033 LOCK (wrklock);
1034 self->dirp = dirp = opendir (req->ptr1);
1035 self->dbuf = u = malloc (sizeof (*u));
1036 req->flags |= FLAG_PTR2_FREE;
1037 req->ptr2 = names = malloc (memlen);
1038 UNLOCK (wrklock);
1039
1040 if (dirp && u && names)
1041 for (;;)
1042 {
1043 errno = 0;
1044 readdir_r (dirp, &u->d, &entp);
1045
1046 if (!entp)
1047 break;
1048
1049 name = entp->d_name;
1050
1051 if (name [0] != '.' || (name [1] && (name [1] != '.' || name [2])))
1052 {
1053 int len = strlen (name) + 1;
1054
1055 res++;
1056
1057 while (memofs + len > memlen)
1058 {
1059 memlen *= 2;
1060 LOCK (wrklock);
1061 req->ptr2 = names = realloc (names, memlen);
1062 UNLOCK (wrklock);
1063
1064 if (!names)
1065 break;
1066 }
1067
1068 memcpy (names + memofs, name, len);
1069 memofs += len;
1070 }
1071 }
1072
1073 if (errno)
1074 res = -1;
1075
1076 req->result = res;
1077 }
1078
1079 /*****************************************************************************/
1080
1081 static void *aio_proc (void *thr_arg)
1082 {
1083 aio_req req;
1084 struct timespec ts;
1085 worker *self = (worker *)thr_arg;
1086
1087 /* try to distribute timeouts somewhat evenly */
1088 ts.tv_nsec = (((unsigned long)self + (unsigned long)ts.tv_sec) & 1023UL)
1089 * (1000000000UL / 1024UL);
1090
1091 for (;;)
1092 {
1093 ts.tv_sec = time (0) + IDLE_TIMEOUT;
1094
1095 LOCK (reqlock);
1096
1097 for (;;)
1098 {
1099 self->req = req = reqq_shift (&req_queue);
1100
1101 if (req)
1102 break;
1103
1104 ++idle;
1105
1106 if (COND_TIMEDWAIT (reqwait, reqlock, ts)
1107 == ETIMEDOUT)
1108 {
1109 if (idle > max_idle)
1110 {
1111 --idle;
1112 UNLOCK (reqlock);
1113 LOCK (wrklock);
1114 --started;
1115 UNLOCK (wrklock);
1116 goto quit;
1117 }
1118
1119 /* we are allowed to idle, so do so without any timeout */
1120 COND_WAIT (reqwait, reqlock);
1121 ts.tv_sec = time (0) + IDLE_TIMEOUT;
1122 }
1123
1124 --idle;
1125 }
1126
1127 --nready;
1128
1129 UNLOCK (reqlock);
1130
1131 errno = 0; /* strictly unnecessary */
1132
1133 if (!(req->flags & FLAG_CANCELLED))
1134 switch (req->type)
1135 {
1136 case REQ_READ: req->result = pread (req->int1, req->ptr1, req->size, req->offs); break;
1137 case REQ_WRITE: req->result = pwrite (req->int1, req->ptr1, req->size, req->offs); break;
1138
1139 case REQ_READAHEAD: req->result = readahead (req->int1, req->offs, req->size); break;
1140 case REQ_SENDFILE: req->result = sendfile_ (req->int1, req->int2, req->offs, req->size, self); break;
1141
1142 case REQ_STAT: req->result = stat (req->ptr1, (Stat_t *)req->ptr2); break;
1143 case REQ_LSTAT: req->result = lstat (req->ptr1, (Stat_t *)req->ptr2); break;
1144 case REQ_FSTAT: req->result = fstat (req->int1, (Stat_t *)req->ptr2); break;
1145
1146 case REQ_CHOWN: req->result = chown (req->ptr1, req->int2, req->int3); break;
1147 case REQ_FCHOWN: req->result = fchown (req->int1, req->int2, req->int3); break;
1148 case REQ_CHMOD: req->result = chmod (req->ptr1, req->mode); break;
1149 case REQ_FCHMOD: req->result = fchmod (req->int1, req->mode); break;
1150
1151 case REQ_OPEN: req->result = open (req->ptr1, req->int1, req->mode); break;
1152 case REQ_CLOSE: req->result = close (req->int1); break;
1153 case REQ_UNLINK: req->result = unlink (req->ptr1); break;
1154 case REQ_RMDIR: req->result = rmdir (req->ptr1); break;
1155 case REQ_MKDIR: req->result = mkdir (req->ptr1, req->mode); break;
1156 case REQ_RENAME: req->result = rename (req->ptr2, req->ptr1); break;
1157 case REQ_LINK: req->result = link (req->ptr2, req->ptr1); break;
1158 case REQ_SYMLINK: req->result = symlink (req->ptr2, req->ptr1); break;
1159 case REQ_MKNOD: req->result = mknod (req->ptr2, req->mode, (dev_t)req->offs); break;
1160 case REQ_READLINK: req->result = readlink (req->ptr2, req->ptr1, NAME_MAX); break;
1161
1162 case REQ_FDATASYNC: req->result = fdatasync (req->int1); break;
1163 case REQ_FSYNC: req->result = fsync (req->int1); break;
1164 case REQ_READDIR: scandir_ (req, self); break;
1165
1166 case REQ_BUSY:
1167 {
1168 struct timeval tv;
1169
1170 tv.tv_sec = req->nv1;
1171 tv.tv_usec = (req->nv1 - tv.tv_usec) * 1000000.;
1172
1173 req->result = select (0, 0, 0, 0, &tv);
1174 }
1175
1176 case REQ_UTIME:
1177 case REQ_FUTIME:
1178 {
1179 struct timeval tv[2];
1180 struct timeval *times;
1181
1182 if (req->nv1 != -1. || req->nv2 != -1.)
1183 {
1184 tv[0].tv_sec = req->nv1;
1185 tv[0].tv_usec = (req->nv1 - tv[0].tv_sec) * 1000000.;
1186 tv[1].tv_sec = req->nv2;
1187 tv[1].tv_usec = (req->nv2 - tv[1].tv_sec) * 1000000.;
1188
1189 times = tv;
1190 }
1191 else
1192 times = 0;
1193
1194
1195 req->result = req->type == REQ_FUTIME
1196 ? futimes (req->int1, times)
1197 : utimes (req->ptr1, times);
1198 }
1199
1200 case REQ_GROUP:
1201 case REQ_NOP:
1202 break;
1203
1204 case REQ_QUIT:
1205 goto quit;
1206
1207 default:
1208 req->result = -1;
1209 break;
1210 }
1211
1212 req->errorno = errno;
1213
1214 LOCK (reslock);
1215
1216 ++npending;
1217
1218 if (!reqq_push (&res_queue, req))
1219 {
1220 /* write a dummy byte to the pipe so fh becomes ready */
1221 write (respipe [1], &respipe, 1);
1222
1223 /* optionally signal the main thread asynchronously */
1224 if (main_sig)
1225 pthread_kill (main_tid, main_sig);
1226 }
1227
1228 self->req = 0;
1229 worker_clear (self);
1230
1231 UNLOCK (reslock);
1232 }
1233
1234 quit:
1235 LOCK (wrklock);
1236 worker_free (self);
1237 UNLOCK (wrklock);
1238
1239 return 0;
1240 }
1241
1242 /*****************************************************************************/
1243
1244 static void atfork_prepare (void)
1245 {
1246 LOCK (wrklock);
1247 LOCK (reqlock);
1248 LOCK (reslock);
1249 #if !HAVE_PREADWRITE
1250 LOCK (preadwritelock);
1251 #endif
1252 #if !HAVE_READDIR_R
1253 LOCK (readdirlock);
1254 #endif
1255 }
1256
1257 static void atfork_parent (void)
1258 {
1259 #if !HAVE_READDIR_R
1260 UNLOCK (readdirlock);
1261 #endif
1262 #if !HAVE_PREADWRITE
1263 UNLOCK (preadwritelock);
1264 #endif
1265 UNLOCK (reslock);
1266 UNLOCK (reqlock);
1267 UNLOCK (wrklock);
1268 }
1269
1270 static void atfork_child (void)
1271 {
1272 aio_req prv;
1273
1274 while (prv = reqq_shift (&req_queue))
1275 req_free (prv);
1276
1277 while (prv = reqq_shift (&res_queue))
1278 req_free (prv);
1279
1280 while (wrk_first.next != &wrk_first)
1281 {
1282 worker *wrk = wrk_first.next;
1283
1284 if (wrk->req)
1285 req_free (wrk->req);
1286
1287 worker_clear (wrk);
1288 worker_free (wrk);
1289 }
1290
1291 started = 0;
1292 idle = 0;
1293 nreqs = 0;
1294 nready = 0;
1295 npending = 0;
1296
1297 close (respipe [0]);
1298 close (respipe [1]);
1299 create_pipe ();
1300
1301 atfork_parent ();
1302 }
1303
1304 #define dREQ \
1305 aio_req req; \
1306 int req_pri = next_pri; \
1307 next_pri = DEFAULT_PRI + PRI_BIAS; \
1308 \
1309 if (SvOK (callback) && !SvROK (callback)) \
1310 croak ("callback must be undef or of reference type"); \
1311 \
1312 Newz (0, req, 1, aio_cb); \
1313 if (!req) \
1314 croak ("out of memory during aio_req allocation"); \
1315 \
1316 req->callback = newSVsv (callback); \
1317 req->pri = req_pri
1318
1319 #define REQ_SEND \
1320 req_send (req); \
1321 \
1322 if (GIMME_V != G_VOID) \
1323 XPUSHs (req_sv (req, AIO_REQ_KLASS));
1324
1325 MODULE = IO::AIO PACKAGE = IO::AIO
1326
1327 PROTOTYPES: ENABLE
1328
1329 BOOT:
1330 {
1331 HV *stash = gv_stashpv ("IO::AIO", 1);
1332
1333 newCONSTSUB (stash, "EXDEV", newSViv (EXDEV));
1334 newCONSTSUB (stash, "O_RDONLY", newSViv (O_RDONLY));
1335 newCONSTSUB (stash, "O_WRONLY", newSViv (O_WRONLY));
1336 newCONSTSUB (stash, "O_CREAT", newSViv (O_CREAT));
1337 newCONSTSUB (stash, "O_TRUNC", newSViv (O_TRUNC));
1338 newCONSTSUB (stash, "S_IFIFO", newSViv (S_IFIFO));
1339 newCONSTSUB (stash, "SIGIO", newSViv (SIGIO));
1340
1341 create_pipe ();
1342 ATFORK (atfork_prepare, atfork_parent, atfork_child);
1343 }
1344
1345 void
1346 max_poll_reqs (int nreqs)
1347 PROTOTYPE: $
1348 CODE:
1349 max_poll_reqs = nreqs;
1350
1351 void
1352 max_poll_time (double nseconds)
1353 PROTOTYPE: $
1354 CODE:
1355 max_poll_time = nseconds * AIO_TICKS;
1356
1357 void
1358 min_parallel (int nthreads)
1359 PROTOTYPE: $
1360
1361 void
1362 max_parallel (int nthreads)
1363 PROTOTYPE: $
1364
1365 void
1366 max_idle (int nthreads)
1367 PROTOTYPE: $
1368 CODE:
1369 set_max_idle (nthreads);
1370
1371 int
1372 max_outstanding (int maxreqs)
1373 PROTOTYPE: $
1374 CODE:
1375 RETVAL = max_outstanding;
1376 max_outstanding = maxreqs;
1377 OUTPUT:
1378 RETVAL
1379
1380 void
1381 aio_open (SV8 *pathname, int flags, int mode, SV *callback=&PL_sv_undef)
1382 PROTOTYPE: $$$;$
1383 PPCODE:
1384 {
1385 dREQ;
1386
1387 req->type = REQ_OPEN;
1388 req->sv1 = newSVsv (pathname);
1389 req->ptr1 = SvPVbyte_nolen (req->sv1);
1390 req->int1 = flags;
1391 req->mode = mode;
1392
1393 REQ_SEND;
1394 }
1395
1396 void
1397 aio_close (SV *fh, SV *callback=&PL_sv_undef)
1398 PROTOTYPE: $;$
1399 ALIAS:
1400 aio_close = REQ_CLOSE
1401 aio_fsync = REQ_FSYNC
1402 aio_fdatasync = REQ_FDATASYNC
1403 PPCODE:
1404 {
1405 dREQ;
1406
1407 req->type = ix;
1408 req->sv1 = newSVsv (fh);
1409 req->int1 = PerlIO_fileno (IoIFP (sv_2io (fh)));
1410
1411 REQ_SEND (req);
1412 }
1413
1414 void
1415 aio_read (SV *fh, UV offset, UV length, SV8 *data, UV dataoffset, SV *callback=&PL_sv_undef)
1416 ALIAS:
1417 aio_read = REQ_READ
1418 aio_write = REQ_WRITE
1419 PROTOTYPE: $$$$$;$
1420 PPCODE:
1421 {
1422 STRLEN svlen;
1423 char *svptr = SvPVbyte (data, svlen);
1424
1425 SvUPGRADE (data, SVt_PV);
1426 SvPOK_on (data);
1427
1428 if (dataoffset < 0)
1429 dataoffset += svlen;
1430
1431 if (dataoffset < 0 || dataoffset > svlen)
1432 croak ("data offset outside of string");
1433
1434 if (ix == REQ_WRITE)
1435 {
1436 /* write: check length and adjust. */
1437 if (length < 0 || length + dataoffset > svlen)
1438 length = svlen - dataoffset;
1439 }
1440 else
1441 {
1442 /* read: grow scalar as necessary */
1443 svptr = SvGROW (data, length + dataoffset + 1);
1444 }
1445
1446 if (length < 0)
1447 croak ("length must not be negative");
1448
1449 {
1450 dREQ;
1451
1452 req->type = ix;
1453 req->sv1 = newSVsv (fh);
1454 req->int1 = PerlIO_fileno (ix == REQ_READ ? IoIFP (sv_2io (fh))
1455 : IoOFP (sv_2io (fh)));
1456 req->offs = offset;
1457 req->size = length;
1458 req->sv2 = SvREFCNT_inc (data);
1459 req->ptr1 = (char *)svptr + dataoffset;
1460 req->stroffset = dataoffset;
1461
1462 if (!SvREADONLY (data))
1463 {
1464 SvREADONLY_on (data);
1465 req->flags |= FLAG_SV1_RO_OFF;
1466 }
1467
1468 REQ_SEND;
1469 }
1470 }
1471
1472 void
1473 aio_readlink (SV8 *path, SV *callback=&PL_sv_undef)
1474 PROTOTYPE: $$;$
1475 PPCODE:
1476 {
1477 SV *data;
1478 dREQ;
1479
1480 data = newSV (NAME_MAX);
1481 SvPOK_on (data);
1482
1483 req->type = REQ_READLINK;
1484 req->sv1 = newSVsv (path);
1485 req->ptr2 = SvPVbyte_nolen (req->sv1);
1486 req->sv2 = data;
1487 req->ptr1 = SvPVbyte_nolen (data);
1488
1489 REQ_SEND;
1490 }
1491
1492 void
1493 aio_sendfile (SV *out_fh, SV *in_fh, UV in_offset, UV length, SV *callback=&PL_sv_undef)
1494 PROTOTYPE: $$$$;$
1495 PPCODE:
1496 {
1497 dREQ;
1498
1499 req->type = REQ_SENDFILE;
1500 req->sv1 = newSVsv (out_fh);
1501 req->int1 = PerlIO_fileno (IoIFP (sv_2io (out_fh)));
1502 req->sv2 = newSVsv (in_fh);
1503 req->int2 = PerlIO_fileno (IoIFP (sv_2io (in_fh)));
1504 req->offs = in_offset;
1505 req->size = length;
1506
1507 REQ_SEND;
1508 }
1509
1510 void
1511 aio_readahead (SV *fh, UV offset, IV length, SV *callback=&PL_sv_undef)
1512 PROTOTYPE: $$$;$
1513 PPCODE:
1514 {
1515 dREQ;
1516
1517 req->type = REQ_READAHEAD;
1518 req->sv1 = newSVsv (fh);
1519 req->int1 = PerlIO_fileno (IoIFP (sv_2io (fh)));
1520 req->offs = offset;
1521 req->size = length;
1522
1523 REQ_SEND;
1524 }
1525
1526 void
1527 aio_stat (SV8 *fh_or_path, SV *callback=&PL_sv_undef)
1528 ALIAS:
1529 aio_stat = REQ_STAT
1530 aio_lstat = REQ_LSTAT
1531 PPCODE:
1532 {
1533 dREQ;
1534
1535 req->ptr2 = malloc (sizeof (Stat_t));
1536 if (!req->ptr2)
1537 {
1538 req_free (req);
1539 croak ("out of memory during aio_stat statdata allocation");
1540 }
1541
1542 req->flags |= FLAG_PTR2_FREE;
1543 req->sv1 = newSVsv (fh_or_path);
1544
1545 if (SvPOK (fh_or_path))
1546 {
1547 req->type = ix;
1548 req->ptr1 = SvPVbyte_nolen (req->sv1);
1549 }
1550 else
1551 {
1552 req->type = REQ_FSTAT;
1553 req->int1 = PerlIO_fileno (IoIFP (sv_2io (fh_or_path)));
1554 }
1555
1556 REQ_SEND;
1557 }
1558
1559 void
1560 aio_utime (SV8 *fh_or_path, SV *atime, SV *mtime, SV *callback=&PL_sv_undef)
1561 PPCODE:
1562 {
1563 dREQ;
1564
1565 req->nv1 = SvOK (atime) ? SvNV (atime) : -1.;
1566 req->nv2 = SvOK (mtime) ? SvNV (mtime) : -1.;
1567 req->sv1 = newSVsv (fh_or_path);
1568
1569 if (SvPOK (fh_or_path))
1570 {
1571 req->type = REQ_UTIME;
1572 req->ptr1 = SvPVbyte_nolen (req->sv1);
1573 }
1574 else
1575 {
1576 req->type = REQ_FUTIME;
1577 req->int1 = PerlIO_fileno (IoIFP (sv_2io (fh_or_path)));
1578 }
1579
1580 REQ_SEND;
1581 }
1582
1583 void
1584 aio_chmod (SV8 *fh_or_path, int mode, SV *callback=&PL_sv_undef)
1585 PPCODE:
1586 {
1587 dREQ;
1588
1589 req->mode = mode;
1590 req->sv1 = newSVsv (fh_or_path);
1591
1592 if (SvPOK (fh_or_path))
1593 {
1594 req->type = REQ_CHMOD;
1595 req->ptr1 = SvPVbyte_nolen (req->sv1);
1596 }
1597 else
1598 {
1599 req->type = REQ_FCHMOD;
1600 req->int1 = PerlIO_fileno (IoIFP (sv_2io (fh_or_path)));
1601 }
1602
1603 REQ_SEND;
1604 }
1605
1606 void
1607 aio_chown (SV8 *fh_or_path, SV *uid, SV *gid, SV *callback=&PL_sv_undef)
1608 PPCODE:
1609 {
1610 dREQ;
1611
1612 req->int2 = SvOK (uid) ? SvIV (uid) : -1;
1613 req->int3 = SvOK (gid) ? SvIV (gid) : -1;
1614 req->sv1 = newSVsv (fh_or_path);
1615
1616 if (SvPOK (fh_or_path))
1617 {
1618 req->type = REQ_CHOWN;
1619 req->ptr1 = SvPVbyte_nolen (req->sv1);
1620 }
1621 else
1622 {
1623 req->type = REQ_FCHOWN;
1624 req->int1 = PerlIO_fileno (IoIFP (sv_2io (fh_or_path)));
1625 }
1626
1627 REQ_SEND;
1628 }
1629
1630 void
1631 aio_unlink (SV8 *pathname, SV *callback=&PL_sv_undef)
1632 ALIAS:
1633 aio_unlink = REQ_UNLINK
1634 aio_rmdir = REQ_RMDIR
1635 aio_readdir = REQ_READDIR
1636 PPCODE:
1637 {
1638 dREQ;
1639
1640 req->type = ix;
1641 req->sv1 = newSVsv (pathname);
1642 req->ptr1 = SvPVbyte_nolen (req->sv1);
1643
1644 REQ_SEND;
1645 }
1646
1647 void
1648 aio_mkdir (SV8 *pathname, int mode, SV *callback=&PL_sv_undef)
1649 PPCODE:
1650 {
1651 dREQ;
1652
1653 req->type = REQ_MKDIR;
1654 req->sv1 = newSVsv (pathname);
1655 req->ptr1 = SvPVbyte_nolen (req->sv1);
1656 req->mode = mode;
1657
1658 REQ_SEND;
1659 }
1660
1661 void
1662 aio_link (SV8 *oldpath, SV8 *newpath, SV *callback=&PL_sv_undef)
1663 ALIAS:
1664 aio_link = REQ_LINK
1665 aio_symlink = REQ_SYMLINK
1666 aio_rename = REQ_RENAME
1667 PPCODE:
1668 {
1669 dREQ;
1670
1671 req->type = ix;
1672 req->sv2 = newSVsv (oldpath);
1673 req->ptr2 = SvPVbyte_nolen (req->sv2);
1674 req->sv1 = newSVsv (newpath);
1675 req->ptr1 = SvPVbyte_nolen (req->sv1);
1676
1677 REQ_SEND;
1678 }
1679
1680 void
1681 aio_mknod (SV8 *pathname, int mode, UV dev, SV *callback=&PL_sv_undef)
1682 PPCODE:
1683 {
1684 dREQ;
1685
1686 req->type = REQ_MKNOD;
1687 req->sv1 = newSVsv (pathname);
1688 req->ptr1 = SvPVbyte_nolen (req->sv1);
1689 req->mode = (mode_t)mode;
1690 req->offs = dev;
1691
1692 REQ_SEND;
1693 }
1694
1695 void
1696 aio_busy (double delay, SV *callback=&PL_sv_undef)
1697 PPCODE:
1698 {
1699 dREQ;
1700
1701 req->type = REQ_BUSY;
1702 req->nv1 = delay < 0. ? 0. : delay;
1703
1704 REQ_SEND;
1705 }
1706
1707 void
1708 aio_group (SV *callback=&PL_sv_undef)
1709 PROTOTYPE: ;$
1710 PPCODE:
1711 {
1712 dREQ;
1713
1714 req->type = REQ_GROUP;
1715
1716 req_send (req);
1717 XPUSHs (req_sv (req, AIO_GRP_KLASS));
1718 }
1719
1720 void
1721 aio_nop (SV *callback=&PL_sv_undef)
1722 PPCODE:
1723 {
1724 dREQ;
1725
1726 req->type = REQ_NOP;
1727
1728 REQ_SEND;
1729 }
1730
1731 int
1732 aioreq_pri (int pri = 0)
1733 PROTOTYPE: ;$
1734 CODE:
1735 RETVAL = next_pri - PRI_BIAS;
1736 if (items > 0)
1737 {
1738 if (pri < PRI_MIN) pri = PRI_MIN;
1739 if (pri > PRI_MAX) pri = PRI_MAX;
1740 next_pri = pri + PRI_BIAS;
1741 }
1742 OUTPUT:
1743 RETVAL
1744
1745 void
1746 aioreq_nice (int nice = 0)
1747 CODE:
1748 nice = next_pri - nice;
1749 if (nice < PRI_MIN) nice = PRI_MIN;
1750 if (nice > PRI_MAX) nice = PRI_MAX;
1751 next_pri = nice + PRI_BIAS;
1752
1753 void
1754 flush ()
1755 PROTOTYPE:
1756 CODE:
1757 while (nreqs)
1758 {
1759 poll_wait ();
1760 poll_cb ();
1761 }
1762
1763 int
1764 poll()
1765 PROTOTYPE:
1766 CODE:
1767 poll_wait ();
1768 RETVAL = poll_cb ();
1769 OUTPUT:
1770 RETVAL
1771
1772 int
1773 poll_fileno()
1774 PROTOTYPE:
1775 CODE:
1776 RETVAL = respipe [0];
1777 OUTPUT:
1778 RETVAL
1779
1780 int
1781 poll_cb(...)
1782 PROTOTYPE:
1783 CODE:
1784 RETVAL = poll_cb ();
1785 OUTPUT:
1786 RETVAL
1787
1788 void
1789 poll_wait()
1790 PROTOTYPE:
1791 CODE:
1792 poll_wait ();
1793
1794 void
1795 setsig (int signum = SIGIO)
1796 PROTOTYPE: ;$
1797 CODE:
1798 {
1799 if (block_sig_level)
1800 croak ("cannot call IO::AIO::setsig from within aio_block/callback");
1801
1802 LOCK (reslock);
1803 main_tid = pthread_self ();
1804 main_sig = signum;
1805 UNLOCK (reslock);
1806
1807 if (main_sig && npending)
1808 pthread_kill (main_tid, main_sig);
1809 }
1810
1811 void
1812 aio_block (SV *cb)
1813 PROTOTYPE: &
1814 PPCODE:
1815 {
1816 int count;
1817
1818 block_sig ();
1819 PUSHMARK (SP);
1820 PUTBACK;
1821 count = call_sv (cb, GIMME_V | G_NOARGS | G_EVAL);
1822 SPAGAIN;
1823 unblock_sig ();
1824
1825 if (SvTRUE (ERRSV))
1826 croak (0);
1827
1828 XSRETURN (count);
1829 }
1830
1831 int
1832 nreqs()
1833 PROTOTYPE:
1834 CODE:
1835 RETVAL = nreqs;
1836 OUTPUT:
1837 RETVAL
1838
1839 int
1840 nready()
1841 PROTOTYPE:
1842 CODE:
1843 RETVAL = get_nready ();
1844 OUTPUT:
1845 RETVAL
1846
1847 int
1848 npending()
1849 PROTOTYPE:
1850 CODE:
1851 RETVAL = get_npending ();
1852 OUTPUT:
1853 RETVAL
1854
1855 int
1856 nthreads()
1857 PROTOTYPE:
1858 CODE:
1859 if (WORDACCESS_UNSAFE) LOCK (wrklock);
1860 RETVAL = started;
1861 if (WORDACCESS_UNSAFE) UNLOCK (wrklock);
1862 OUTPUT:
1863 RETVAL
1864
1865 PROTOTYPES: DISABLE
1866
1867 MODULE = IO::AIO PACKAGE = IO::AIO::REQ
1868
1869 void
1870 cancel (aio_req_ornot req)
1871 CODE:
1872 req_cancel (req);
1873
1874 void
1875 cb (aio_req_ornot req, SV *callback=&PL_sv_undef)
1876 CODE:
1877 SvREFCNT_dec (req->callback);
1878 req->callback = newSVsv (callback);
1879
1880 MODULE = IO::AIO PACKAGE = IO::AIO::GRP
1881
1882 void
1883 add (aio_req grp, ...)
1884 PPCODE:
1885 {
1886 int i;
1887 aio_req req;
1888
1889 if (main_sig && !block_sig_level)
1890 croak ("aio_group->add called outside aio_block/callback context while IO::AIO::setsig is in use");
1891
1892 if (grp->int1 == 2)
1893 croak ("cannot add requests to IO::AIO::GRP after the group finished");
1894
1895 for (i = 1; i < items; ++i )
1896 {
1897 if (GIMME_V != G_VOID)
1898 XPUSHs (sv_2mortal (newSVsv (ST (i))));
1899
1900 req = SvAIO_REQ (ST (i));
1901
1902 if (req)
1903 {
1904 ++grp->size;
1905 req->grp = grp;
1906
1907 req->grp_prev = 0;
1908 req->grp_next = grp->grp_first;
1909
1910 if (grp->grp_first)
1911 grp->grp_first->grp_prev = req;
1912
1913 grp->grp_first = req;
1914 }
1915 }
1916 }
1917
1918 void
1919 cancel_subs (aio_req_ornot req)
1920 CODE:
1921 req_cancel_subs (req);
1922
1923 void
1924 result (aio_req grp, ...)
1925 CODE:
1926 {
1927 int i;
1928 AV *av;
1929
1930 grp->errorno = errno;
1931
1932 av = newAV ();
1933
1934 for (i = 1; i < items; ++i )
1935 av_push (av, newSVsv (ST (i)));
1936
1937 SvREFCNT_dec (grp->sv1);
1938 grp->sv1 = (SV *)av;
1939 }
1940
1941 void
1942 errno (aio_req grp, int errorno = errno)
1943 CODE:
1944 grp->errorno = errorno;
1945
1946 void
1947 limit (aio_req grp, int limit)
1948 CODE:
1949 grp->int2 = limit;
1950 aio_grp_feed (grp);
1951
1952 void
1953 feed (aio_req grp, SV *callback=&PL_sv_undef)
1954 CODE:
1955 {
1956 SvREFCNT_dec (grp->sv2);
1957 grp->sv2 = newSVsv (callback);
1958
1959 if (grp->int2 <= 0)
1960 grp->int2 = 2;
1961
1962 aio_grp_feed (grp);
1963 }
1964