… | |
… | |
46 | # define BOOT_PAGESIZE (void)0 |
46 | # define BOOT_PAGESIZE (void)0 |
47 | #endif |
47 | #endif |
48 | |
48 | |
49 | #if CORO_USE_VALGRIND |
49 | #if CORO_USE_VALGRIND |
50 | # include <valgrind/valgrind.h> |
50 | # include <valgrind/valgrind.h> |
51 | # define REGISTER_STACK(cctx,start,end) (cctx)->valgrind_id = VALGRIND_STACK_REGISTER ((start), (end)) |
|
|
52 | #else |
|
|
53 | # define REGISTER_STACK(cctx,start,end) |
|
|
54 | #endif |
51 | #endif |
55 | |
52 | |
56 | /* the maximum number of idle cctx that will be pooled */ |
53 | /* the maximum number of idle cctx that will be pooled */ |
57 | static int cctx_max_idle = 4; |
54 | static int cctx_max_idle = 4; |
58 | |
55 | |
… | |
… | |
119 | # define CORO_PREFER_PERL_FUNCTIONS 0 |
116 | # define CORO_PREFER_PERL_FUNCTIONS 0 |
120 | #endif |
117 | #endif |
121 | |
118 | |
122 | /* The next macros try to return the current stack pointer, in an as |
119 | /* The next macros try to return the current stack pointer, in an as |
123 | * portable way as possible. */ |
120 | * portable way as possible. */ |
124 | #define dSTACKLEVEL volatile char stacklevel |
121 | #if __GNUC__ >= 4 |
125 | #define STACKLEVEL ((void *)&stacklevel) |
122 | # define dSTACKLEVEL void *stacklevel = __builtin_frame_address (0) |
|
|
123 | #else |
|
|
124 | # define dSTACKLEVEL volatile void *stacklevel = (volatile void *)&stacklevel |
|
|
125 | #endif |
126 | |
126 | |
127 | #define IN_DESTRUCT (PL_main_cv == Nullcv) |
127 | #define IN_DESTRUCT (PL_main_cv == Nullcv) |
128 | |
128 | |
129 | #if __GNUC__ >= 3 |
129 | #if __GNUC__ >= 3 |
130 | # define attribute(x) __attribute__(x) |
130 | # define attribute(x) __attribute__(x) |
131 | # define BARRIER __asm__ __volatile__ ("" : : : "memory") |
|
|
132 | # define expect(expr,value) __builtin_expect ((expr),(value)) |
131 | # define expect(expr,value) __builtin_expect ((expr),(value)) |
|
|
132 | # define INLINE static inline |
133 | #else |
133 | #else |
134 | # define attribute(x) |
134 | # define attribute(x) |
135 | # define BARRIER |
|
|
136 | # define expect(expr,value) (expr) |
135 | # define expect(expr,value) (expr) |
|
|
136 | # define INLINE static |
137 | #endif |
137 | #endif |
138 | |
138 | |
139 | #define expect_false(expr) expect ((expr) != 0, 0) |
139 | #define expect_false(expr) expect ((expr) != 0, 0) |
140 | #define expect_true(expr) expect ((expr) != 0, 1) |
140 | #define expect_true(expr) expect ((expr) != 0, 1) |
141 | |
141 | |
… | |
… | |
181 | static struct CoroAPI coroapi; |
181 | static struct CoroAPI coroapi; |
182 | static AV *main_mainstack; /* used to differentiate between $main and others */ |
182 | static AV *main_mainstack; /* used to differentiate between $main and others */ |
183 | static JMPENV *main_top_env; |
183 | static JMPENV *main_top_env; |
184 | static HV *coro_state_stash, *coro_stash; |
184 | static HV *coro_state_stash, *coro_stash; |
185 | static volatile SV *coro_mortal; /* will be freed/thrown after next transfer */ |
185 | static volatile SV *coro_mortal; /* will be freed/thrown after next transfer */ |
|
|
186 | static volatile struct coro *transfer_next; |
|
|
187 | |
|
|
188 | struct transfer_args |
|
|
189 | { |
|
|
190 | struct coro *prev, *next; |
|
|
191 | }; |
186 | |
192 | |
187 | static GV *irsgv; /* $/ */ |
193 | static GV *irsgv; /* $/ */ |
188 | static GV *stdoutgv; /* *STDOUT */ |
194 | static GV *stdoutgv; /* *STDOUT */ |
189 | static SV *rv_diehook; |
195 | static SV *rv_diehook; |
190 | static SV *rv_warnhook; |
196 | static SV *rv_warnhook; |
… | |
… | |
407 | : 0 |
413 | : 0 |
408 | |
414 | |
409 | #define CORO_MAGIC_cv(cv) CORO_MAGIC (((SV *)(cv)), CORO_MAGIC_type_cv) |
415 | #define CORO_MAGIC_cv(cv) CORO_MAGIC (((SV *)(cv)), CORO_MAGIC_type_cv) |
410 | #define CORO_MAGIC_state(sv) CORO_MAGIC (((SV *)(sv)), CORO_MAGIC_type_state) |
416 | #define CORO_MAGIC_state(sv) CORO_MAGIC (((SV *)(sv)), CORO_MAGIC_type_state) |
411 | |
417 | |
412 | static struct coro * |
418 | INLINE struct coro * |
413 | SvSTATE_ (pTHX_ SV *coro) |
419 | SvSTATE_ (pTHX_ SV *coro) |
414 | { |
420 | { |
415 | HV *stash; |
421 | HV *stash; |
416 | MAGIC *mg; |
422 | MAGIC *mg; |
417 | |
423 | |
… | |
… | |
708 | } |
714 | } |
709 | |
715 | |
710 | return rss; |
716 | return rss; |
711 | } |
717 | } |
712 | |
718 | |
|
|
719 | /** set stacklevel support **************************************************/ |
|
|
720 | |
|
|
721 | /* we sometimes need to create the effect of pp_slf calling us */ |
|
|
722 | #define SLF_HEAD (void)0 |
|
|
723 | /* we sometimes need to create the effect of leaving via pp_slf */ |
|
|
724 | #define SLF_TAIL slf_tail (aTHX) |
|
|
725 | |
|
|
726 | INLINE void |
|
|
727 | slf_tail (pTHX) |
|
|
728 | { |
|
|
729 | dSP; |
|
|
730 | SV **bot = SP; |
|
|
731 | |
|
|
732 | int gimme = GIMME_V; |
|
|
733 | |
|
|
734 | /* make sure we put something on the stack in scalar context */ |
|
|
735 | if (gimme == G_SCALAR) |
|
|
736 | { |
|
|
737 | if (sp == bot) |
|
|
738 | XPUSHs (&PL_sv_undef); |
|
|
739 | |
|
|
740 | SP = bot + 1; |
|
|
741 | } |
|
|
742 | |
|
|
743 | PUTBACK; |
|
|
744 | } |
|
|
745 | |
713 | /** coroutine stack handling ************************************************/ |
746 | /** coroutine stack handling ************************************************/ |
714 | |
747 | |
715 | static int (*orig_sigelem_get) (pTHX_ SV *sv, MAGIC *mg); |
748 | static int (*orig_sigelem_get) (pTHX_ SV *sv, MAGIC *mg); |
716 | static int (*orig_sigelem_set) (pTHX_ SV *sv, MAGIC *mg); |
749 | static int (*orig_sigelem_set) (pTHX_ SV *sv, MAGIC *mg); |
717 | static int (*orig_sigelem_clr) (pTHX_ SV *sv, MAGIC *mg); |
750 | static int (*orig_sigelem_clr) (pTHX_ SV *sv, MAGIC *mg); |
… | |
… | |
834 | PL_rs = newSVsv (GvSV (irsgv)); |
867 | PL_rs = newSVsv (GvSV (irsgv)); |
835 | PL_defoutgv = (GV *)SvREFCNT_inc_NN (stdoutgv); |
868 | PL_defoutgv = (GV *)SvREFCNT_inc_NN (stdoutgv); |
836 | |
869 | |
837 | { |
870 | { |
838 | dSP; |
871 | dSP; |
839 | LOGOP myop; |
872 | UNOP myop; |
840 | |
873 | |
841 | Zero (&myop, 1, LOGOP); |
874 | Zero (&myop, 1, UNOP); |
842 | myop.op_next = Nullop; |
875 | myop.op_next = Nullop; |
843 | myop.op_flags = OPf_WANT_VOID; |
876 | myop.op_flags = OPf_WANT_VOID; |
844 | |
877 | |
845 | PUSHMARK (SP); |
878 | PUSHMARK (SP); |
846 | XPUSHs (sv_2mortal (av_shift (GvAV (PL_defgv)))); |
879 | XPUSHs (sv_2mortal (av_shift (GvAV (PL_defgv)))); |
… | |
… | |
849 | PL_op = PL_ppaddr[OP_ENTERSUB](aTHX); |
882 | PL_op = PL_ppaddr[OP_ENTERSUB](aTHX); |
850 | SPAGAIN; |
883 | SPAGAIN; |
851 | } |
884 | } |
852 | |
885 | |
853 | /* this newly created coroutine might be run on an existing cctx which most |
886 | /* this newly created coroutine might be run on an existing cctx which most |
854 | * likely was suspended in set_stacklevel, called from entersub. |
887 | * likely was suspended in set_stacklevel, called from pp_set_stacklevel, |
855 | * set_stacklevl doesn't do anything on return, but entersub does LEAVE, |
888 | * so we have to emulate entering pp_set_stacklevel here. |
856 | * so we ENTER here for symmetry |
|
|
857 | */ |
889 | */ |
858 | ENTER; |
890 | SLF_HEAD; |
859 | } |
891 | } |
860 | |
892 | |
861 | static void |
893 | static void |
862 | coro_destruct (pTHX_ struct coro *coro) |
894 | coro_destruct (pTHX_ struct coro *coro) |
863 | { |
895 | { |
… | |
… | |
892 | SvREFCNT_dec (coro->throw); |
924 | SvREFCNT_dec (coro->throw); |
893 | |
925 | |
894 | coro_destruct_stacks (aTHX); |
926 | coro_destruct_stacks (aTHX); |
895 | } |
927 | } |
896 | |
928 | |
897 | static void |
929 | INLINE void |
898 | free_coro_mortal (pTHX) |
930 | free_coro_mortal (pTHX) |
899 | { |
931 | { |
900 | if (expect_true (coro_mortal)) |
932 | if (expect_true (coro_mortal)) |
901 | { |
933 | { |
902 | SvREFCNT_dec (coro_mortal); |
934 | SvREFCNT_dec (coro_mortal); |
… | |
… | |
1026 | |
1058 | |
1027 | TAINT_NOT; |
1059 | TAINT_NOT; |
1028 | return 0; |
1060 | return 0; |
1029 | } |
1061 | } |
1030 | |
1062 | |
|
|
1063 | static void |
|
|
1064 | prepare_set_stacklevel (struct transfer_args *ta, struct coro_cctx *cctx) |
|
|
1065 | { |
|
|
1066 | ta->prev = (struct coro *)cctx; |
|
|
1067 | ta->next = 0; |
|
|
1068 | } |
|
|
1069 | |
1031 | /* inject a fake call to Coro::State::_cctx_init into the execution */ |
1070 | /* inject a fake call to Coro::State::_cctx_init into the execution */ |
1032 | /* _cctx_init should be careful, as it could be called at almost any time */ |
1071 | /* _cctx_init should be careful, as it could be called at almost any time */ |
1033 | /* during execution of a perl program */ |
1072 | /* during execution of a perl program */ |
|
|
1073 | /* also initialises PL_top_env */ |
1034 | static void NOINLINE |
1074 | static void NOINLINE |
1035 | cctx_prepare (pTHX_ coro_cctx *cctx) |
1075 | cctx_prepare (pTHX_ coro_cctx *cctx) |
1036 | { |
1076 | { |
1037 | dSP; |
1077 | dSP; |
1038 | LOGOP myop; |
1078 | UNOP myop; |
1039 | |
1079 | |
1040 | PL_top_env = &PL_start_env; |
1080 | PL_top_env = &PL_start_env; |
1041 | |
1081 | |
1042 | if (cctx->flags & CC_TRACE) |
1082 | if (cctx->flags & CC_TRACE) |
1043 | PL_runops = runops_trace; |
1083 | PL_runops = runops_trace; |
1044 | |
1084 | |
1045 | Zero (&myop, 1, LOGOP); |
1085 | Zero (&myop, 1, UNOP); |
1046 | myop.op_next = PL_op; |
1086 | myop.op_next = PL_op; |
1047 | myop.op_flags = OPf_WANT_VOID | OPf_STACKED; |
1087 | myop.op_flags = OPf_WANT_VOID | OPf_STACKED; |
1048 | |
1088 | |
1049 | PUSHMARK (SP); |
1089 | PUSHMARK (SP); |
1050 | EXTEND (SP, 2); |
1090 | EXTEND (SP, 2); |
1051 | PUSHs (sv_2mortal (newSViv (PTR2IV (cctx)))); |
1091 | PUSHs (sv_2mortal (newSViv ((IV)cctx))); |
1052 | PUSHs ((SV *)get_cv ("Coro::State::_cctx_init", FALSE)); |
1092 | PUSHs ((SV *)get_cv ("Coro::State::_cctx_init", FALSE)); |
1053 | PUTBACK; |
1093 | PUTBACK; |
1054 | PL_op = (OP *)&myop; |
1094 | PL_op = (OP *)&myop; |
1055 | PL_op = PL_ppaddr[OP_ENTERSUB](aTHX); |
1095 | PL_op = PL_ppaddr[OP_ENTERSUB](aTHX); |
1056 | SPAGAIN; |
1096 | SPAGAIN; |
1057 | } |
1097 | } |
1058 | |
1098 | |
|
|
1099 | /* the tail of transfer: execute stuff we can only do after a transfer */ |
|
|
1100 | INLINE void |
|
|
1101 | transfer_tail (pTHX) |
|
|
1102 | { |
|
|
1103 | struct coro *next = (struct coro *)transfer_next; |
|
|
1104 | assert (!(transfer_next = 0)); /* just used for the side effect when asserts are enabled */ |
|
|
1105 | assert (("FATAL: next coroutine was zero in transfer_tail (please report)", next)); |
|
|
1106 | |
|
|
1107 | free_coro_mortal (aTHX); |
|
|
1108 | UNLOCK; |
|
|
1109 | |
|
|
1110 | if (expect_false (next->throw)) |
|
|
1111 | { |
|
|
1112 | SV *exception = sv_2mortal (next->throw); |
|
|
1113 | |
|
|
1114 | next->throw = 0; |
|
|
1115 | sv_setsv (ERRSV, exception); |
|
|
1116 | croak (0); |
|
|
1117 | } |
|
|
1118 | } |
|
|
1119 | |
1059 | /* |
1120 | /* |
1060 | * this is a _very_ stripped down perl interpreter ;) |
1121 | * this is a _very_ stripped down perl interpreter ;) |
1061 | */ |
1122 | */ |
1062 | static void |
1123 | static void |
1063 | cctx_run (void *arg) |
1124 | cctx_run (void *arg) |
… | |
… | |
1068 | # endif |
1129 | # endif |
1069 | #endif |
1130 | #endif |
1070 | { |
1131 | { |
1071 | dTHX; |
1132 | dTHX; |
1072 | |
1133 | |
1073 | /* cctx_run is the alternative tail of transfer(), so unlock here. */ |
1134 | /* we are the alternative tail to pp_set_stacklevel */ |
1074 | UNLOCK; |
1135 | /* so do the same things here */ |
|
|
1136 | SLF_TAIL; |
1075 | |
1137 | |
1076 | /* we now skip the entersub that lead to transfer() */ |
1138 | /* we now skip the op that did lead to transfer() */ |
1077 | PL_op = PL_op->op_next; |
1139 | PL_op = PL_op->op_next; |
1078 | |
1140 | |
1079 | /* inject a fake subroutine call to cctx_init */ |
1141 | /* inject a fake subroutine call to cctx_init */ |
1080 | cctx_prepare (aTHX_ (coro_cctx *)arg); |
1142 | cctx_prepare (aTHX_ (coro_cctx *)arg); |
|
|
1143 | |
|
|
1144 | /* cctx_run is the alternative tail of transfer() */ |
|
|
1145 | transfer_tail (aTHX); |
1081 | |
1146 | |
1082 | /* somebody or something will hit me for both perl_run and PL_restartop */ |
1147 | /* somebody or something will hit me for both perl_run and PL_restartop */ |
1083 | PL_restartop = PL_op; |
1148 | PL_restartop = PL_op; |
1084 | perl_run (PL_curinterp); |
1149 | perl_run (PL_curinterp); |
1085 | |
1150 | |
… | |
… | |
1097 | |
1162 | |
1098 | static coro_cctx * |
1163 | static coro_cctx * |
1099 | cctx_new () |
1164 | cctx_new () |
1100 | { |
1165 | { |
1101 | coro_cctx *cctx; |
1166 | coro_cctx *cctx; |
|
|
1167 | |
|
|
1168 | ++cctx_count; |
|
|
1169 | New (0, cctx, 1, coro_cctx); |
|
|
1170 | |
|
|
1171 | cctx->gen = cctx_gen; |
|
|
1172 | cctx->flags = 0; |
|
|
1173 | cctx->idle_sp = 0; /* can be accessed by transfer between cctx_run and set_stacklevel, on throw */ |
|
|
1174 | |
|
|
1175 | return cctx; |
|
|
1176 | } |
|
|
1177 | |
|
|
1178 | /* create a new cctx only suitable as source */ |
|
|
1179 | static coro_cctx * |
|
|
1180 | cctx_new_empty () |
|
|
1181 | { |
|
|
1182 | coro_cctx *cctx = cctx_new (); |
|
|
1183 | |
|
|
1184 | cctx->sptr = 0; |
|
|
1185 | coro_create (&cctx->cctx, 0, 0, 0, 0); |
|
|
1186 | |
|
|
1187 | return cctx; |
|
|
1188 | } |
|
|
1189 | |
|
|
1190 | /* create a new cctx suitable as destination/running a perl interpreter */ |
|
|
1191 | static coro_cctx * |
|
|
1192 | cctx_new_run () |
|
|
1193 | { |
|
|
1194 | coro_cctx *cctx = cctx_new (); |
1102 | void *stack_start; |
1195 | void *stack_start; |
1103 | size_t stack_size; |
1196 | size_t stack_size; |
1104 | |
|
|
1105 | ++cctx_count; |
|
|
1106 | Newz (0, cctx, 1, coro_cctx); |
|
|
1107 | |
|
|
1108 | cctx->gen = cctx_gen; |
|
|
1109 | |
1197 | |
1110 | #if HAVE_MMAP |
1198 | #if HAVE_MMAP |
1111 | cctx->ssize = ((cctx_stacksize * sizeof (long) + PAGESIZE - 1) / PAGESIZE + CORO_STACKGUARD) * PAGESIZE; |
1199 | cctx->ssize = ((cctx_stacksize * sizeof (long) + PAGESIZE - 1) / PAGESIZE + CORO_STACKGUARD) * PAGESIZE; |
1112 | /* mmap supposedly does allocate-on-write for us */ |
1200 | /* mmap supposedly does allocate-on-write for us */ |
1113 | cctx->sptr = mmap (0, cctx->ssize, PROT_EXEC|PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, 0, 0); |
1201 | cctx->sptr = mmap (0, cctx->ssize, PROT_EXEC|PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, 0, 0); |
1114 | |
1202 | |
1115 | if (cctx->sptr != (void *)-1) |
1203 | if (cctx->sptr != (void *)-1) |
1116 | { |
1204 | { |
1117 | # if CORO_STACKGUARD |
1205 | #if CORO_STACKGUARD |
1118 | mprotect (cctx->sptr, CORO_STACKGUARD * PAGESIZE, PROT_NONE); |
1206 | mprotect (cctx->sptr, CORO_STACKGUARD * PAGESIZE, PROT_NONE); |
1119 | # endif |
1207 | #endif |
1120 | stack_start = CORO_STACKGUARD * PAGESIZE + (char *)cctx->sptr; |
1208 | stack_start = (char *)cctx->sptr + CORO_STACKGUARD * PAGESIZE; |
1121 | stack_size = cctx->ssize - CORO_STACKGUARD * PAGESIZE; |
1209 | stack_size = cctx->ssize - CORO_STACKGUARD * PAGESIZE; |
1122 | cctx->flags |= CC_MAPPED; |
1210 | cctx->flags |= CC_MAPPED; |
1123 | } |
1211 | } |
1124 | else |
1212 | else |
1125 | #endif |
1213 | #endif |
1126 | { |
1214 | { |
1127 | cctx->ssize = cctx_stacksize * (long)sizeof (long); |
1215 | cctx->ssize = cctx_stacksize * (long)sizeof (long); |
1128 | New (0, cctx->sptr, cctx_stacksize, long); |
1216 | New (0, cctx->sptr, cctx_stacksize, long); |
1129 | |
1217 | |
1130 | if (!cctx->sptr) |
1218 | if (!cctx->sptr) |
1131 | { |
1219 | { |
1132 | perror ("FATAL: unable to allocate stack for coroutine"); |
1220 | perror ("FATAL: unable to allocate stack for coroutine, exiting."); |
1133 | _exit (EXIT_FAILURE); |
1221 | _exit (EXIT_FAILURE); |
1134 | } |
1222 | } |
1135 | |
1223 | |
1136 | stack_start = cctx->sptr; |
1224 | stack_start = cctx->sptr; |
1137 | stack_size = cctx->ssize; |
1225 | stack_size = cctx->ssize; |
1138 | } |
1226 | } |
1139 | |
1227 | |
1140 | REGISTER_STACK (cctx, (char *)stack_start, (char *)stack_start + stack_size); |
1228 | #if CORO_USE_VALGRIND |
|
|
1229 | cctx->valgrind_id = VALGRIND_STACK_REGISTER ((char *)stack_start, (char *)stack_start + stack_size); |
|
|
1230 | #endif |
|
|
1231 | |
1141 | coro_create (&cctx->cctx, cctx_run, (void *)cctx, stack_start, stack_size); |
1232 | coro_create (&cctx->cctx, cctx_run, (void *)cctx, stack_start, stack_size); |
1142 | |
1233 | |
1143 | return cctx; |
1234 | return cctx; |
1144 | } |
1235 | } |
1145 | |
1236 | |
… | |
… | |
1153 | coro_destroy (&cctx->cctx); |
1244 | coro_destroy (&cctx->cctx); |
1154 | |
1245 | |
1155 | /* coro_transfer creates new, empty cctx's */ |
1246 | /* coro_transfer creates new, empty cctx's */ |
1156 | if (cctx->sptr) |
1247 | if (cctx->sptr) |
1157 | { |
1248 | { |
1158 | #if CORO_USE_VALGRIND |
1249 | #if CORO_USE_VALGRIND |
1159 | VALGRIND_STACK_DEREGISTER (cctx->valgrind_id); |
1250 | VALGRIND_STACK_DEREGISTER (cctx->valgrind_id); |
1160 | #endif |
1251 | #endif |
1161 | |
1252 | |
1162 | #if HAVE_MMAP |
1253 | #if HAVE_MMAP |
1163 | if (cctx->flags & CC_MAPPED) |
1254 | if (cctx->flags & CC_MAPPED) |
1164 | munmap (cctx->sptr, cctx->ssize); |
1255 | munmap (cctx->sptr, cctx->ssize); |
1165 | else |
1256 | else |
… | |
… | |
1186 | return cctx; |
1277 | return cctx; |
1187 | |
1278 | |
1188 | cctx_destroy (cctx); |
1279 | cctx_destroy (cctx); |
1189 | } |
1280 | } |
1190 | |
1281 | |
1191 | return cctx_new (); |
1282 | return cctx_new_run (); |
1192 | } |
1283 | } |
1193 | |
1284 | |
1194 | static void |
1285 | static void |
1195 | cctx_put (coro_cctx *cctx) |
1286 | cctx_put (coro_cctx *cctx) |
1196 | { |
1287 | { |
1197 | assert (("cctx_put called on non-initialised cctx", cctx->sptr)); |
1288 | assert (("FATAL: cctx_put called on non-initialised cctx in Coro (please report)", cctx->sptr)); |
1198 | |
1289 | |
1199 | /* free another cctx if overlimit */ |
1290 | /* free another cctx if overlimit */ |
1200 | if (expect_false (cctx_idle >= cctx_max_idle)) |
1291 | if (expect_false (cctx_idle >= cctx_max_idle)) |
1201 | { |
1292 | { |
1202 | coro_cctx *first = cctx_first; |
1293 | coro_cctx *first = cctx_first; |
… | |
… | |
1241 | dSTACKLEVEL; |
1332 | dSTACKLEVEL; |
1242 | |
1333 | |
1243 | /* sometimes transfer is only called to set idle_sp */ |
1334 | /* sometimes transfer is only called to set idle_sp */ |
1244 | if (expect_false (!next)) |
1335 | if (expect_false (!next)) |
1245 | { |
1336 | { |
1246 | ((coro_cctx *)prev)->idle_sp = STACKLEVEL; |
1337 | ((coro_cctx *)prev)->idle_sp = stacklevel; |
1247 | assert (((coro_cctx *)prev)->idle_te = PL_top_env); /* just for the side-effect when asserts are enabled */ |
1338 | assert (((coro_cctx *)prev)->idle_te = PL_top_env); /* just for the side-effect when asserts are enabled */ |
1248 | } |
1339 | } |
1249 | else if (expect_true (prev != next)) |
1340 | else if (expect_true (prev != next)) |
1250 | { |
1341 | { |
1251 | static volatile int has_throw; |
|
|
1252 | coro_cctx *prev__cctx; |
1342 | coro_cctx *prev__cctx; |
1253 | |
1343 | |
1254 | if (expect_false (prev->flags & CF_NEW)) |
1344 | if (expect_false (prev->flags & CF_NEW)) |
1255 | { |
1345 | { |
1256 | /* create a new empty/source context */ |
1346 | /* create a new empty/source context */ |
1257 | ++cctx_count; |
1347 | prev->cctx = cctx_new_empty (); |
1258 | New (0, prev->cctx, 1, coro_cctx); |
|
|
1259 | prev->cctx->sptr = 0; |
|
|
1260 | coro_create (&prev->cctx->cctx, 0, 0, 0, 0); |
|
|
1261 | |
|
|
1262 | prev->flags &= ~CF_NEW; |
1348 | prev->flags &= ~CF_NEW; |
1263 | prev->flags |= CF_RUNNING; |
1349 | prev->flags |= CF_RUNNING; |
1264 | } |
1350 | } |
1265 | |
1351 | |
1266 | prev->flags &= ~CF_RUNNING; |
1352 | prev->flags &= ~CF_RUNNING; |
… | |
… | |
1281 | else |
1367 | else |
1282 | load_perl (aTHX_ next); |
1368 | load_perl (aTHX_ next); |
1283 | |
1369 | |
1284 | prev__cctx = prev->cctx; |
1370 | prev__cctx = prev->cctx; |
1285 | |
1371 | |
1286 | /* possibly "free" the cctx */ |
1372 | /* possibly untie and reuse the cctx */ |
1287 | if (expect_true ( |
1373 | if (expect_true ( |
1288 | prev__cctx->idle_sp == STACKLEVEL |
1374 | prev__cctx->idle_sp == stacklevel |
1289 | && !(prev__cctx->flags & CC_TRACE) |
1375 | && !(prev__cctx->flags & CC_TRACE) |
1290 | && !force_cctx |
1376 | && !force_cctx |
1291 | )) |
1377 | )) |
1292 | { |
1378 | { |
1293 | /* I assume that STACKLEVEL is a stronger indicator than PL_top_env changes */ |
1379 | /* I assume that stacklevel is a stronger indicator than PL_top_env changes */ |
1294 | assert (("ERROR: current top_env must equal previous top_env", PL_top_env == prev__cctx->idle_te)); |
1380 | assert (("FATAL: current top_env must equal previous top_env in Coro (please report)", PL_top_env == prev__cctx->idle_te)); |
1295 | |
1381 | |
1296 | prev->cctx = 0; |
1382 | prev->cctx = 0; |
1297 | |
1383 | |
1298 | /* if the cctx is about to be destroyed we need to make sure we won't see it in cctx_get */ |
1384 | /* if the cctx is about to be destroyed we need to make sure we won't see it in cctx_get */ |
1299 | /* without this the next cctx_get might destroy the prev__cctx while still in use */ |
1385 | /* without this the next cctx_get might destroy the prev__cctx while still in use */ |
… | |
… | |
1307 | ++next->usecount; |
1393 | ++next->usecount; |
1308 | |
1394 | |
1309 | if (expect_true (!next->cctx)) |
1395 | if (expect_true (!next->cctx)) |
1310 | next->cctx = cctx_get (aTHX); |
1396 | next->cctx = cctx_get (aTHX); |
1311 | |
1397 | |
1312 | has_throw = !!next->throw; |
1398 | assert (("FATAL: transfer_next already nonzero in Coro (please report)", !transfer_next)); |
|
|
1399 | transfer_next = next; |
1313 | |
1400 | |
1314 | if (expect_false (prev__cctx != next->cctx)) |
1401 | if (expect_false (prev__cctx != next->cctx)) |
1315 | { |
1402 | { |
1316 | prev__cctx->top_env = PL_top_env; |
1403 | prev__cctx->top_env = PL_top_env; |
1317 | PL_top_env = next->cctx->top_env; |
1404 | PL_top_env = next->cctx->top_env; |
1318 | coro_transfer (&prev__cctx->cctx, &next->cctx->cctx); |
1405 | coro_transfer (&prev__cctx->cctx, &next->cctx->cctx); |
1319 | } |
1406 | } |
1320 | |
1407 | |
1321 | free_coro_mortal (aTHX); |
1408 | transfer_tail (aTHX); |
1322 | UNLOCK; |
|
|
1323 | |
|
|
1324 | if (expect_false (has_throw)) |
|
|
1325 | { |
|
|
1326 | struct coro *coro = SvSTATE (coro_current); |
|
|
1327 | |
|
|
1328 | if (coro->throw) |
|
|
1329 | { |
|
|
1330 | SV *exception = coro->throw; |
|
|
1331 | coro->throw = 0; |
|
|
1332 | sv_setsv (ERRSV, exception); |
|
|
1333 | croak (0); |
|
|
1334 | } |
|
|
1335 | } |
|
|
1336 | } |
1409 | } |
1337 | } |
1410 | } |
1338 | |
|
|
1339 | struct transfer_args |
|
|
1340 | { |
|
|
1341 | struct coro *prev, *next; |
|
|
1342 | }; |
|
|
1343 | |
1411 | |
1344 | #define TRANSFER(ta, force_cctx) transfer (aTHX_ (ta).prev, (ta).next, (force_cctx)) |
1412 | #define TRANSFER(ta, force_cctx) transfer (aTHX_ (ta).prev, (ta).next, (force_cctx)) |
1345 | #define TRANSFER_CHECK(ta) transfer_check (aTHX_ (ta).prev, (ta).next) |
1413 | #define TRANSFER_CHECK(ta) transfer_check (aTHX_ (ta).prev, (ta).next) |
1346 | |
1414 | |
1347 | /** high level stuff ********************************************************/ |
1415 | /** high level stuff ********************************************************/ |
… | |
… | |
1520 | |
1588 | |
1521 | static int |
1589 | static int |
1522 | api_is_ready (SV *coro_sv) |
1590 | api_is_ready (SV *coro_sv) |
1523 | { |
1591 | { |
1524 | dTHX; |
1592 | dTHX; |
|
|
1593 | |
1525 | return !!(SvSTATE (coro_sv)->flags & CF_READY); |
1594 | return !!(SvSTATE (coro_sv)->flags & CF_READY); |
1526 | } |
1595 | } |
1527 | |
1596 | |
1528 | static void |
1597 | INLINE void |
1529 | prepare_schedule (pTHX_ struct transfer_args *ta) |
1598 | prepare_schedule (pTHX_ struct transfer_args *ta) |
1530 | { |
1599 | { |
1531 | SV *prev_sv, *next_sv; |
1600 | SV *prev_sv, *next_sv; |
1532 | |
1601 | |
1533 | for (;;) |
1602 | for (;;) |
… | |
… | |
1559 | /* cannot transfer to destroyed coros, skip and look for next */ |
1628 | /* cannot transfer to destroyed coros, skip and look for next */ |
1560 | if (expect_false (ta->next->flags & CF_DESTROYED)) |
1629 | if (expect_false (ta->next->flags & CF_DESTROYED)) |
1561 | { |
1630 | { |
1562 | UNLOCK; |
1631 | UNLOCK; |
1563 | SvREFCNT_dec (next_sv); |
1632 | SvREFCNT_dec (next_sv); |
1564 | /* coro_nready is already taken care of by destroy */ |
1633 | /* coro_nready has already been taken care of by destroy */ |
1565 | continue; |
1634 | continue; |
1566 | } |
1635 | } |
1567 | |
1636 | |
1568 | --coro_nready; |
1637 | --coro_nready; |
1569 | UNLOCK; |
1638 | UNLOCK; |
… | |
… | |
1572 | |
1641 | |
1573 | /* free this only after the transfer */ |
1642 | /* free this only after the transfer */ |
1574 | prev_sv = SvRV (coro_current); |
1643 | prev_sv = SvRV (coro_current); |
1575 | ta->prev = SvSTATE (prev_sv); |
1644 | ta->prev = SvSTATE (prev_sv); |
1576 | TRANSFER_CHECK (*ta); |
1645 | TRANSFER_CHECK (*ta); |
1577 | assert (ta->next->flags & CF_READY); |
1646 | assert (("FATAL: next coroutine isn't marked as ready in Coro (please report)", ta->next->flags & CF_READY)); |
1578 | ta->next->flags &= ~CF_READY; |
1647 | ta->next->flags &= ~CF_READY; |
1579 | SvRV_set (coro_current, next_sv); |
1648 | SvRV_set (coro_current, next_sv); |
1580 | |
1649 | |
1581 | LOCK; |
1650 | LOCK; |
1582 | free_coro_mortal (aTHX); |
1651 | free_coro_mortal (aTHX); |
1583 | coro_mortal = prev_sv; |
1652 | coro_mortal = prev_sv; |
1584 | UNLOCK; |
1653 | UNLOCK; |
1585 | } |
1654 | } |
1586 | |
1655 | |
1587 | static void |
1656 | INLINE void |
1588 | prepare_cede (pTHX_ struct transfer_args *ta) |
1657 | prepare_cede (pTHX_ struct transfer_args *ta) |
1589 | { |
1658 | { |
1590 | api_ready (coro_current); |
1659 | api_ready (coro_current); |
1591 | prepare_schedule (aTHX_ ta); |
1660 | prepare_schedule (aTHX_ ta); |
1592 | } |
1661 | } |
1593 | |
1662 | |
1594 | static int |
1663 | static void |
1595 | prepare_cede_notself (pTHX_ struct transfer_args *ta) |
1664 | prepare_cede_notself (pTHX_ struct transfer_args *ta) |
1596 | { |
1665 | { |
|
|
1666 | SV *prev = SvRV (coro_current); |
|
|
1667 | |
1597 | if (coro_nready) |
1668 | if (coro_nready) |
1598 | { |
1669 | { |
1599 | SV *prev = SvRV (coro_current); |
|
|
1600 | prepare_schedule (aTHX_ ta); |
1670 | prepare_schedule (aTHX_ ta); |
1601 | api_ready (prev); |
1671 | api_ready (prev); |
|
|
1672 | } |
|
|
1673 | else |
|
|
1674 | ta->prev = ta->next = SvSTATE (prev); |
|
|
1675 | } |
|
|
1676 | |
|
|
1677 | static void |
|
|
1678 | api_schedule (void) |
|
|
1679 | { |
|
|
1680 | dTHX; |
|
|
1681 | struct transfer_args ta; |
|
|
1682 | |
|
|
1683 | prepare_schedule (aTHX_ &ta); |
|
|
1684 | TRANSFER (ta, 1); |
|
|
1685 | } |
|
|
1686 | |
|
|
1687 | static int |
|
|
1688 | api_cede (void) |
|
|
1689 | { |
|
|
1690 | dTHX; |
|
|
1691 | struct transfer_args ta; |
|
|
1692 | |
|
|
1693 | prepare_cede (aTHX_ &ta); |
|
|
1694 | |
|
|
1695 | if (expect_true (ta.prev != ta.next)) |
|
|
1696 | { |
|
|
1697 | TRANSFER (ta, 1); |
1602 | return 1; |
1698 | return 1; |
1603 | } |
1699 | } |
1604 | else |
1700 | else |
1605 | return 0; |
1701 | return 0; |
1606 | } |
1702 | } |
1607 | |
1703 | |
1608 | static void |
|
|
1609 | api_schedule (void) |
|
|
1610 | { |
|
|
1611 | dTHX; |
|
|
1612 | struct transfer_args ta; |
|
|
1613 | |
|
|
1614 | prepare_schedule (aTHX_ &ta); |
|
|
1615 | TRANSFER (ta, 1); |
|
|
1616 | } |
|
|
1617 | |
|
|
1618 | static int |
1704 | static int |
1619 | api_cede (void) |
1705 | api_cede_notself (void) |
1620 | { |
1706 | { |
|
|
1707 | if (coro_nready) |
|
|
1708 | { |
1621 | dTHX; |
1709 | dTHX; |
1622 | struct transfer_args ta; |
1710 | struct transfer_args ta; |
1623 | |
1711 | |
1624 | prepare_cede (aTHX_ &ta); |
1712 | prepare_cede_notself (aTHX_ &ta); |
1625 | |
|
|
1626 | if (expect_true (ta.prev != ta.next)) |
|
|
1627 | { |
|
|
1628 | TRANSFER (ta, 1); |
1713 | TRANSFER (ta, 1); |
1629 | return 1; |
1714 | return 1; |
1630 | } |
1715 | } |
1631 | else |
1716 | else |
1632 | return 0; |
1717 | return 0; |
1633 | } |
1718 | } |
1634 | |
1719 | |
1635 | static int |
|
|
1636 | api_cede_notself (void) |
|
|
1637 | { |
|
|
1638 | dTHX; |
|
|
1639 | struct transfer_args ta; |
|
|
1640 | |
|
|
1641 | if (prepare_cede_notself (aTHX_ &ta)) |
|
|
1642 | { |
|
|
1643 | TRANSFER (ta, 1); |
|
|
1644 | return 1; |
|
|
1645 | } |
|
|
1646 | else |
|
|
1647 | return 0; |
|
|
1648 | } |
|
|
1649 | |
|
|
1650 | static void |
1720 | static void |
1651 | api_trace (SV *coro_sv, int flags) |
1721 | api_trace (SV *coro_sv, int flags) |
1652 | { |
1722 | { |
1653 | dTHX; |
1723 | dTHX; |
1654 | struct coro *coro = SvSTATE (coro_sv); |
1724 | struct coro *coro = SvSTATE (coro_sv); |
1655 | |
1725 | |
1656 | if (flags & CC_TRACE) |
1726 | if (flags & CC_TRACE) |
1657 | { |
1727 | { |
1658 | if (!coro->cctx) |
1728 | if (!coro->cctx) |
1659 | coro->cctx = cctx_new (); |
1729 | coro->cctx = cctx_new_run (); |
1660 | else if (!(coro->cctx->flags & CC_TRACE)) |
1730 | else if (!(coro->cctx->flags & CC_TRACE)) |
1661 | croak ("cannot enable tracing on coroutine with custom stack"); |
1731 | croak ("cannot enable tracing on coroutine with custom stack"); |
1662 | |
1732 | |
1663 | coro->cctx->flags |= CC_NOREUSE | (flags & (CC_TRACE | CC_TRACE_ALL)); |
1733 | coro->cctx->flags |= CC_NOREUSE | (flags & (CC_TRACE | CC_TRACE_ALL)); |
1664 | } |
1734 | } |
… | |
… | |
1764 | PerlIOBuf_get_ptr, |
1834 | PerlIOBuf_get_ptr, |
1765 | PerlIOBuf_get_cnt, |
1835 | PerlIOBuf_get_cnt, |
1766 | PerlIOBuf_set_ptrcnt, |
1836 | PerlIOBuf_set_ptrcnt, |
1767 | }; |
1837 | }; |
1768 | |
1838 | |
|
|
1839 | /*****************************************************************************/ |
|
|
1840 | |
|
|
1841 | static const CV *slf_cv; /* for quick consistency check */ |
|
|
1842 | |
|
|
1843 | static UNOP slf_restore; /* restore stack as entersub did, for first-re-run */ |
|
|
1844 | static SV *slf_arg0; |
|
|
1845 | static SV *slf_arg1; |
|
|
1846 | |
|
|
1847 | /* this restores the stack in the case we patched the entersub, to */ |
|
|
1848 | /* recreate the stack frame as perl will on following calls */ |
|
|
1849 | /* since entersub cleared the stack */ |
|
|
1850 | static OP * |
|
|
1851 | pp_restore (pTHX) |
|
|
1852 | { |
|
|
1853 | dSP; |
|
|
1854 | |
|
|
1855 | PUSHMARK (SP); |
|
|
1856 | |
|
|
1857 | EXTEND (SP, 3); |
|
|
1858 | if (slf_arg0) PUSHs (sv_2mortal (slf_arg0)); |
|
|
1859 | if (slf_arg1) PUSHs (sv_2mortal (slf_arg1)); |
|
|
1860 | PUSHs ((SV *)CvGV (slf_cv)); |
|
|
1861 | |
|
|
1862 | RETURNOP (slf_restore.op_first); |
|
|
1863 | } |
|
|
1864 | |
|
|
1865 | #define OPpENTERSUB_SLF 15 /* the part of op_private entersub hopefully doesn't use */ |
|
|
1866 | |
|
|
1867 | /* declare prototype */ |
|
|
1868 | XS(XS_Coro__State__set_stacklevel); |
|
|
1869 | |
|
|
1870 | /* |
|
|
1871 | * these not obviously related functions are all rolled into one |
|
|
1872 | * function to increase chances that they all will call transfer with the same |
|
|
1873 | * stack offset |
|
|
1874 | * SLF stands for "schedule-like-function". |
|
|
1875 | */ |
|
|
1876 | static OP * |
|
|
1877 | pp_slf (pTHX) |
|
|
1878 | { |
|
|
1879 | dSP; |
|
|
1880 | struct transfer_args ta; |
|
|
1881 | SV **arg = PL_stack_base + TOPMARK + 1; |
|
|
1882 | int items = SP - arg; /* args without function object */ |
|
|
1883 | |
|
|
1884 | /* do a quick consistency check on the "function" object, and if it isn't */ |
|
|
1885 | /* for us, divert to the real entersub */ |
|
|
1886 | if (SvTYPE (*sp) != SVt_PVGV || CvXSUB (GvCV (*sp)) != XS_Coro__State__set_stacklevel) |
|
|
1887 | return PL_ppaddr[OP_ENTERSUB](aTHX); |
|
|
1888 | |
|
|
1889 | /* pop args */ |
|
|
1890 | SP = PL_stack_base + POPMARK; |
|
|
1891 | |
|
|
1892 | if (!(PL_op->op_flags & OPf_STACKED)) |
|
|
1893 | { |
|
|
1894 | /* ampersand-form of call, use @_ instead of stack */ |
|
|
1895 | AV *av = GvAV (PL_defgv); |
|
|
1896 | arg = AvARRAY (av); |
|
|
1897 | items = AvFILLp (av) + 1; |
|
|
1898 | } |
|
|
1899 | |
|
|
1900 | PUTBACK; |
|
|
1901 | switch (PL_op->op_private & OPpENTERSUB_SLF) |
|
|
1902 | { |
|
|
1903 | case 0: |
|
|
1904 | prepare_set_stacklevel (&ta, (struct coro_cctx *)SvIV (arg [0])); |
|
|
1905 | break; |
|
|
1906 | |
|
|
1907 | case 1: |
|
|
1908 | if (items != 2) |
|
|
1909 | croak ("Coro::State::transfer (prev, next) expects two arguments, not %d.", items); |
|
|
1910 | |
|
|
1911 | prepare_transfer (aTHX_ &ta, arg [0], arg [1]); |
|
|
1912 | break; |
|
|
1913 | |
|
|
1914 | case 2: |
|
|
1915 | prepare_schedule (aTHX_ &ta); |
|
|
1916 | break; |
|
|
1917 | |
|
|
1918 | case 3: |
|
|
1919 | prepare_cede (aTHX_ &ta); |
|
|
1920 | break; |
|
|
1921 | |
|
|
1922 | case 4: |
|
|
1923 | prepare_cede_notself (aTHX_ &ta); |
|
|
1924 | break; |
|
|
1925 | |
|
|
1926 | case 5: |
|
|
1927 | abort (); |
|
|
1928 | |
|
|
1929 | default: |
|
|
1930 | abort (); |
|
|
1931 | } |
|
|
1932 | |
|
|
1933 | TRANSFER (ta, 0); |
|
|
1934 | SPAGAIN; |
|
|
1935 | |
|
|
1936 | PUTBACK; |
|
|
1937 | SLF_TAIL; |
|
|
1938 | SPAGAIN; |
|
|
1939 | RETURN; |
|
|
1940 | } |
|
|
1941 | |
|
|
1942 | static void |
|
|
1943 | coro_slf_patch (pTHX_ CV *cv, int ix, SV **args, int items) |
|
|
1944 | { |
|
|
1945 | assert (("FATAL: SLF call recursion in Coro module (please report)", PL_op->op_ppaddr != pp_slf)); |
|
|
1946 | |
|
|
1947 | assert (("FATAL: SLF call with illegal CV value", CvGV (cv))); |
|
|
1948 | slf_cv = cv; |
|
|
1949 | |
|
|
1950 | /* we patch the op, and then re-run the whole call */ |
|
|
1951 | /* we have to put the same argument on the stack for this to work */ |
|
|
1952 | /* and this will be done by pp_restore */ |
|
|
1953 | slf_restore.op_next = (OP *)&slf_restore; |
|
|
1954 | slf_restore.op_type = OP_NULL; |
|
|
1955 | slf_restore.op_ppaddr = pp_restore; |
|
|
1956 | slf_restore.op_first = PL_op; |
|
|
1957 | |
|
|
1958 | slf_arg0 = items > 0 ? SvREFCNT_inc (args [0]) : 0; |
|
|
1959 | slf_arg1 = items > 1 ? SvREFCNT_inc (args [1]) : 0; |
|
|
1960 | |
|
|
1961 | PL_op->op_ppaddr = pp_slf; |
|
|
1962 | PL_op->op_private = PL_op->op_private & ~OPpENTERSUB_SLF | ix; /* we potentially share our private flags with entersub */ |
|
|
1963 | |
|
|
1964 | PL_op = (OP *)&slf_restore; |
|
|
1965 | } |
1769 | |
1966 | |
1770 | MODULE = Coro::State PACKAGE = Coro::State PREFIX = api_ |
1967 | MODULE = Coro::State PACKAGE = Coro::State PREFIX = api_ |
1771 | |
1968 | |
1772 | PROTOTYPES: DISABLE |
1969 | PROTOTYPES: DISABLE |
1773 | |
1970 | |
… | |
… | |
1848 | av_push (coro->args, newSVsv (ST (i))); |
2045 | av_push (coro->args, newSVsv (ST (i))); |
1849 | } |
2046 | } |
1850 | OUTPUT: |
2047 | OUTPUT: |
1851 | RETVAL |
2048 | RETVAL |
1852 | |
2049 | |
1853 | # these not obviously related functions are all rolled into the same xs |
|
|
1854 | # function to increase chances that they all will call transfer with the same |
|
|
1855 | # stack offset |
|
|
1856 | void |
2050 | void |
1857 | _set_stacklevel (...) |
2051 | _set_stacklevel (...) |
1858 | ALIAS: |
2052 | ALIAS: |
1859 | Coro::State::transfer = 1 |
2053 | Coro::State::transfer = 1 |
1860 | Coro::schedule = 2 |
2054 | Coro::schedule = 2 |
1861 | Coro::cede = 3 |
2055 | Coro::cede = 3 |
1862 | Coro::cede_notself = 4 |
2056 | Coro::cede_notself = 4 |
1863 | CODE: |
2057 | CODE: |
1864 | { |
2058 | coro_slf_patch (aTHX_ cv, ix, &ST (0), items); |
1865 | struct transfer_args ta; |
|
|
1866 | |
|
|
1867 | PUTBACK; |
|
|
1868 | switch (ix) |
|
|
1869 | { |
|
|
1870 | case 0: |
|
|
1871 | ta.prev = (struct coro *)INT2PTR (coro_cctx *, SvIV (ST (0))); |
|
|
1872 | ta.next = 0; |
|
|
1873 | break; |
|
|
1874 | |
|
|
1875 | case 1: |
|
|
1876 | if (items != 2) |
|
|
1877 | croak ("Coro::State::transfer (prev, next) expects two arguments, not %d", items); |
|
|
1878 | |
|
|
1879 | prepare_transfer (aTHX_ &ta, ST (0), ST (1)); |
|
|
1880 | break; |
|
|
1881 | |
|
|
1882 | case 2: |
|
|
1883 | prepare_schedule (aTHX_ &ta); |
|
|
1884 | break; |
|
|
1885 | |
|
|
1886 | case 3: |
|
|
1887 | prepare_cede (aTHX_ &ta); |
|
|
1888 | break; |
|
|
1889 | |
|
|
1890 | case 4: |
|
|
1891 | if (!prepare_cede_notself (aTHX_ &ta)) |
|
|
1892 | XSRETURN_EMPTY; |
|
|
1893 | |
|
|
1894 | break; |
|
|
1895 | } |
|
|
1896 | SPAGAIN; |
|
|
1897 | |
|
|
1898 | BARRIER; |
|
|
1899 | PUTBACK; |
|
|
1900 | TRANSFER (ta, 0); |
|
|
1901 | SPAGAIN; /* might be the sp of a different coroutine now */ |
|
|
1902 | /* be extra careful not to ever do anything after TRANSFER */ |
|
|
1903 | } |
|
|
1904 | |
2059 | |
1905 | bool |
2060 | bool |
1906 | _destroy (SV *coro_sv) |
2061 | _destroy (SV *coro_sv) |
1907 | CODE: |
2062 | CODE: |
1908 | RETVAL = coro_state_destroy (aTHX_ SvSTATE (coro_sv)); |
2063 | RETVAL = coro_state_destroy (aTHX_ SvSTATE (coro_sv)); |
… | |
… | |
2016 | is_destroyed = CF_DESTROYED |
2171 | is_destroyed = CF_DESTROYED |
2017 | CODE: |
2172 | CODE: |
2018 | RETVAL = boolSV (coro->flags & ix); |
2173 | RETVAL = boolSV (coro->flags & ix); |
2019 | OUTPUT: |
2174 | OUTPUT: |
2020 | RETVAL |
2175 | RETVAL |
|
|
2176 | |
|
|
2177 | void |
|
|
2178 | throw (Coro::State self, SV *throw = &PL_sv_undef) |
|
|
2179 | PROTOTYPE: $;$ |
|
|
2180 | CODE: |
|
|
2181 | SvREFCNT_dec (self->throw); |
|
|
2182 | self->throw = SvOK (throw) ? newSVsv (throw) : 0; |
2021 | |
2183 | |
2022 | void |
2184 | void |
2023 | api_trace (SV *coro, int flags = CC_TRACE | CC_TRACE_SUB) |
2185 | api_trace (SV *coro, int flags = CC_TRACE | CC_TRACE_SUB) |
2024 | |
2186 | |
2025 | SV * |
2187 | SV * |
… | |
… | |
2098 | |
2260 | |
2099 | for (i = PRIO_MAX - PRIO_MIN + 1; i--; ) |
2261 | for (i = PRIO_MAX - PRIO_MIN + 1; i--; ) |
2100 | coro_ready[i] = newAV (); |
2262 | coro_ready[i] = newAV (); |
2101 | |
2263 | |
2102 | { |
2264 | { |
2103 | SV *sv = perl_get_sv ("Coro::API", TRUE); |
2265 | SV *sv = coro_get_sv (aTHX_ "Coro::API", TRUE); |
2104 | perl_get_sv ("Coro::API", TRUE); /* silence 5.10 warning */ |
|
|
2105 | |
2266 | |
2106 | coroapi.schedule = api_schedule; |
2267 | coroapi.schedule = api_schedule; |
2107 | coroapi.cede = api_cede; |
2268 | coroapi.cede = api_cede; |
2108 | coroapi.cede_notself = api_cede_notself; |
2269 | coroapi.cede_notself = api_cede_notself; |
2109 | coroapi.ready = api_ready; |
2270 | coroapi.ready = api_ready; |
… | |
… | |
2169 | CODE: |
2330 | CODE: |
2170 | RETVAL = coro_nready; |
2331 | RETVAL = coro_nready; |
2171 | OUTPUT: |
2332 | OUTPUT: |
2172 | RETVAL |
2333 | RETVAL |
2173 | |
2334 | |
2174 | void |
|
|
2175 | throw (Coro::State self, SV *throw = &PL_sv_undef) |
|
|
2176 | PROTOTYPE: $;$ |
|
|
2177 | CODE: |
|
|
2178 | SvREFCNT_dec (self->throw); |
|
|
2179 | self->throw = SvOK (throw) ? newSVsv (throw) : 0; |
|
|
2180 | |
|
|
2181 | # for async_pool speedup |
2335 | # for async_pool speedup |
2182 | void |
2336 | void |
2183 | _pool_1 (SV *cb) |
2337 | _pool_1 (SV *cb) |
2184 | CODE: |
2338 | CODE: |
2185 | { |
2339 | { |
… | |
… | |
2384 | |
2538 | |
2385 | MODULE = Coro::State PACKAGE = PerlIO::cede |
2539 | MODULE = Coro::State PACKAGE = PerlIO::cede |
2386 | |
2540 | |
2387 | BOOT: |
2541 | BOOT: |
2388 | PerlIO_define_layer (aTHX_ &PerlIO_cede); |
2542 | PerlIO_define_layer (aTHX_ &PerlIO_cede); |
|
|
2543 | |