ViewVC Help
View File | Revision Log | Show Annotations | Download File
/cvs/libcoro/coro.c
(Generate patch)

Comparing libcoro/coro.c (file contents):
Revision 1.36 by root, Wed Nov 5 01:54:34 2008 UTC vs.
Revision 1.64 by root, Wed Dec 5 13:01:18 2012 UTC

1/* 1/*
2 * Copyright (c) 2001-2008 Marc Alexander Lehmann <schmorp@schmorp.de> 2 * Copyright (c) 2001-2011 Marc Alexander Lehmann <schmorp@schmorp.de>
3 * 3 *
4 * Redistribution and use in source and binary forms, with or without modifica- 4 * Redistribution and use in source and binary forms, with or without modifica-
5 * tion, are permitted provided that the following conditions are met: 5 * tion, are permitted provided that the following conditions are met:
6 * 6 *
7 * 1. Redistributions of source code must retain the above copyright notice, 7 * 1. Redistributions of source code must retain the above copyright notice,
8 * this list of conditions and the following disclaimer. 8 * this list of conditions and the following disclaimer.
9 * 9 *
10 * 2. Redistributions in binary form must reproduce the above copyright 10 * 2. Redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the 11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution. 12 * documentation and/or other materials provided with the distribution.
13 * 13 *
14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED 14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED
15 * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MER- 15 * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MER-
16 * CHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO 16 * CHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO
17 * EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPE- 17 * EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPE-
18 * CIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, 18 * CIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
38 * go to Ralf S. Engelschall <rse@engelschall.com>. 38 * go to Ralf S. Engelschall <rse@engelschall.com>.
39 */ 39 */
40 40
41#include "coro.h" 41#include "coro.h"
42 42
43#include <string.h>
44
45/*****************************************************************************/
46/* ucontext/setjmp/asm backends */
47/*****************************************************************************/
48#if CORO_UCONTEXT || CORO_SJLJ || CORO_LOSER || CORO_LINUX || CORO_IRIX || CORO_ASM
49
50# if CORO_UCONTEXT
51# include <stddef.h>
52# endif
53
43#if !defined(STACK_ADJUST_PTR) 54# if !defined(STACK_ADJUST_PTR)
55# if __sgi
44/* IRIX is decidedly NON-unix */ 56/* IRIX is decidedly NON-unix */
45# if __sgi
46# define STACK_ADJUST_PTR(sp,ss) ((char *)(sp) + (ss) - 8) 57# define STACK_ADJUST_PTR(sp,ss) ((char *)(sp) + (ss) - 8)
47# define STACK_ADJUST_SIZE(sp,ss) ((ss) - 8) 58# define STACK_ADJUST_SIZE(sp,ss) ((ss) - 8)
48# elif (__i386__ && CORO_LINUX) || (_M_IX86 && CORO_LOSER) 59# elif (__i386__ && CORO_LINUX) || (_M_IX86 && CORO_LOSER)
49# define STACK_ADJUST_PTR(sp,ss) ((char *)(sp) + (ss)) 60# define STACK_ADJUST_PTR(sp,ss) ((char *)(sp) + (ss))
50# define STACK_ADJUST_SIZE(sp,ss) (ss) 61# define STACK_ADJUST_SIZE(sp,ss) (ss)
51# elif (__amd64__ && CORO_LINUX) || ((_M_AMD64 || _M_IA64) && CORO_LOSER) 62# elif (__amd64__ && CORO_LINUX) || ((_M_AMD64 || _M_IA64) && CORO_LOSER)
52# define STACK_ADJUST_PTR(sp,ss) ((char *)(sp) + (ss) - 8) 63# define STACK_ADJUST_PTR(sp,ss) ((char *)(sp) + (ss) - 8)
53# define STACK_ADJUST_SIZE(sp,ss) (ss) 64# define STACK_ADJUST_SIZE(sp,ss) (ss)
54# else 65# else
55# define STACK_ADJUST_PTR(sp,ss) (sp) 66# define STACK_ADJUST_PTR(sp,ss) (sp)
56# define STACK_ADJUST_SIZE(sp,ss) (ss) 67# define STACK_ADJUST_SIZE(sp,ss) (ss)
57# endif 68# endif
58#endif 69# endif
59 70
60#if CORO_UCONTEXT
61# include <stddef.h>
62#endif
63
64#if CORO_SJLJ || CORO_LOSER || CORO_LINUX || CORO_IRIX || CORO_ASM
65
66#include <stdlib.h> 71# include <stdlib.h>
67 72
68#if CORO_SJLJ 73# if CORO_SJLJ
69# include <stdio.h> 74# include <stdio.h>
70# include <signal.h> 75# include <signal.h>
71# include <unistd.h> 76# include <unistd.h>
72#endif 77# endif
73 78
74static volatile coro_func coro_init_func; 79static coro_func coro_init_func;
75static volatile void *coro_init_arg; 80static void *coro_init_arg;
76static volatile coro_context *new_coro, *create_coro; 81static coro_context *new_coro, *create_coro;
77
78/* what we really want to detect here is wether we use a new-enough version of GAS */
79/* instead, check for gcc 3, ELF and GNU/Linux and hope for the best */
80#if __GNUC__ >= 3 && __ELF__ && __linux__
81# define HAVE_CFI 1
82#endif
83 82
84static void 83static void
85coro_init (void) 84coro_init (void)
86{ 85{
87 volatile coro_func func = coro_init_func; 86 volatile coro_func func = coro_init_func;
88 volatile void *arg = coro_init_arg; 87 volatile void *arg = coro_init_arg;
89 88
90 coro_transfer ((coro_context *)new_coro, (coro_context *)create_coro); 89 coro_transfer (new_coro, create_coro);
90
91#if __GCC_HAVE_DWARF2_CFI_ASM && __amd64
92 asm (".cfi_undefined rip");
93#endif
91 94
92 func ((void *)arg); 95 func ((void *)arg);
93 96
94 /* the new coro returned. bad. just abort() for now */ 97 /* the new coro returned. bad. just abort() for now */
95 abort (); 98 abort ();
96} 99}
97 100
98# if CORO_SJLJ 101# if CORO_SJLJ
99 102
100static volatile int trampoline_count; 103static volatile int trampoline_done;
101 104
102/* trampoline signal handler */ 105/* trampoline signal handler */
103static void 106static void
104trampoline (int sig) 107trampoline (int sig)
105{ 108{
106 if (setjmp (((coro_context *)new_coro)->env)) 109 if (coro_setjmp (new_coro->env))
107 {
108#if HAVE_CFI
109 asm (".cfi_startproc");
110#endif
111 coro_init (); /* start it */ 110 coro_init (); /* start it */
112#if HAVE_CFI
113 asm (".cfi_endproc");
114#endif
115 }
116 else 111 else
117 trampoline_count++; 112 trampoline_done = 1;
118} 113}
119 114
120# endif
121
122#endif 115# endif
123 116
124#if CORO_ASM 117# if CORO_ASM
118
119 #if _WIN32 || __CYGWIN__
120 #define CORO_WIN_TIB 1
121 #endif
122
125asm ( 123 asm (
126 ".text\n" 124 "\t.text\n"
125 #if _WIN32 || __CYGWIN__
126 "\t.globl _coro_transfer\n"
127 "_coro_transfer:\n"
128 #else
127 ".globl coro_transfer\n" 129 "\t.globl coro_transfer\n"
128 ".type coro_transfer, @function\n"
129 "coro_transfer:\n" 130 "coro_transfer:\n"
131 #endif
132 /* windows, of course, gives a shit on the amd64 ABI and uses different registers */
133 /* http://blogs.msdn.com/freik/archive/2005/03/17/398200.aspx */
130#if __amd64 134 #if __amd64
131# define NUM_ALIGN 1 135
132# define NUM_SAVED 5 136 #if _WIN32 || __CYGWIN__
137 #define NUM_SAVED 29
138 "\tsubq $168, %rsp\t" /* one dummy qword to improve alignment */
139 "\tmovaps %xmm6, (%rsp)\n"
140 "\tmovaps %xmm7, 16(%rsp)\n"
141 "\tmovaps %xmm8, 32(%rsp)\n"
142 "\tmovaps %xmm9, 48(%rsp)\n"
143 "\tmovaps %xmm10, 64(%rsp)\n"
144 "\tmovaps %xmm11, 80(%rsp)\n"
145 "\tmovaps %xmm12, 96(%rsp)\n"
146 "\tmovaps %xmm13, 112(%rsp)\n"
147 "\tmovaps %xmm14, 128(%rsp)\n"
148 "\tmovaps %xmm15, 144(%rsp)\n"
149 "\tpushq %rsi\n"
150 "\tpushq %rdi\n"
151 "\tpushq %rbp\n"
133 "\tpush %rbx\n" 152 "\tpushq %rbx\n"
134 "\tpush %r12\n" 153 "\tpushq %r12\n"
135 "\tpush %r13\n" 154 "\tpushq %r13\n"
136 "\tpush %r14\n" 155 "\tpushq %r14\n"
137 "\tpush %r15\n" 156 "\tpushq %r15\n"
157 #if CORO_WIN_TIB
158 "\tpushq %fs:0x0\n"
159 "\tpushq %fs:0x8\n"
160 "\tpushq %fs:0xc\n"
161 #endif
162 "\tmovq %rsp, (%rcx)\n"
163 "\tmovq (%rdx), %rsp\n"
164 #if CORO_WIN_TIB
165 "\tpopq %fs:0xc\n"
166 "\tpopq %fs:0x8\n"
167 "\tpopq %fs:0x0\n"
168 #endif
169 "\tpopq %r15\n"
170 "\tpopq %r14\n"
171 "\tpopq %r13\n"
172 "\tpopq %r12\n"
173 "\tpopq %rbx\n"
174 "\tpopq %rbp\n"
175 "\tpopq %rdi\n"
176 "\tpopq %rsi\n"
177 "\tmovaps (%rsp), %xmm6\n"
178 "\tmovaps 16(%rsp), %xmm7\n"
179 "\tmovaps 32(%rsp), %xmm8\n"
180 "\tmovaps 48(%rsp), %xmm9\n"
181 "\tmovaps 64(%rsp), %xmm10\n"
182 "\tmovaps 80(%rsp), %xmm11\n"
183 "\tmovaps 96(%rsp), %xmm12\n"
184 "\tmovaps 112(%rsp), %xmm13\n"
185 "\tmovaps 128(%rsp), %xmm14\n"
186 "\tmovaps 144(%rsp), %xmm15\n"
187 "\taddq $168, %rsp\n"
188 #else
189 #define NUM_SAVED 6
190 "\tpushq %rbp\n"
191 "\tpushq %rbx\n"
192 "\tpushq %r12\n"
193 "\tpushq %r13\n"
194 "\tpushq %r14\n"
195 "\tpushq %r15\n"
138 "\tmov %rsp, (%rdi)\n" 196 "\tmovq %rsp, (%rdi)\n"
139 "\tmov (%rsi), %rsp\n" 197 "\tmovq (%rsi), %rsp\n"
140 "\tpop %r15\n" 198 "\tpopq %r15\n"
141 "\tpop %r14\n" 199 "\tpopq %r14\n"
142 "\tpop %r13\n" 200 "\tpopq %r13\n"
143 "\tpop %r12\n" 201 "\tpopq %r12\n"
144 "\tpop %rbx\n" 202 "\tpopq %rbx\n"
203 "\tpopq %rbp\n"
204 #endif
205 "\tpopq %rcx\n"
206 "\tjmpq *%rcx\n"
207
145#elif __i386 208 #elif __i386
146# define NUM_ALIGN 1 209
147# define NUM_SAVED 4 210 #define NUM_SAVED 4
148 "\tpush %ebx\n"
149 "\tpush %esi\n"
150 "\tpush %edi\n"
151 "\tpush %ebp\n" 211 "\tpushl %ebp\n"
212 "\tpushl %ebx\n"
213 "\tpushl %esi\n"
214 "\tpushl %edi\n"
215 #if CORO_WIN_TIB
216 #undef NUM_SAVED
217 #define NUM_SAVED 7
218 "\tpushl %fs:0\n"
219 "\tpushl %fs:4\n"
220 "\tpushl %fs:8\n"
221 #endif
152 "\tmov %esp, (%eax)\n" 222 "\tmovl %esp, (%eax)\n"
153 "\tmov (%edx), %esp\n" 223 "\tmovl (%edx), %esp\n"
154 "\tpop %ebp\n" 224 #if CORO_WIN_TIB
225 "\tpopl %fs:8\n"
226 "\tpopl %fs:4\n"
227 "\tpopl %fs:0\n"
228 #endif
155 "\tpop %edi\n" 229 "\tpopl %edi\n"
156 "\tpop %esi\n" 230 "\tpopl %esi\n"
157 "\tpop %ebx\n" 231 "\tpopl %ebx\n"
158#else 232 "\tpopl %ebp\n"
233 "\tpopl %ecx\n"
234 "\tjmpl *%ecx\n"
235
236 #else
159# error unsupported architecture 237 #error unsupported architecture
238 #endif
239 );
240
160#endif 241# endif
161 "\tret\n"
162);
163#endif
164 242
165#if CORO_PTHREAD 243void
166 244coro_create (coro_context *ctx, coro_func coro, void *arg, void *sptr, long ssize)
167struct coro_init_args {
168 coro_func func;
169 void *arg;
170 coro_context *self, *main;
171};
172
173pthread_mutex_t coro_mutex = PTHREAD_MUTEX_INITIALIZER;
174
175static void *
176trampoline (void *args_)
177{ 245{
178 struct coro_init_args *args = (struct coro_init_args *)args_; 246 coro_context nctx;
179 coro_func func = args->func;
180 void *arg = args->arg;
181
182 pthread_mutex_lock (&coro_mutex);
183 pthread_cond_destroy (&args->self->c);
184 coro_transfer (args->self, args->main);
185 func (arg);
186 pthread_mutex_unlock (&coro_mutex);
187
188 return 0;
189}
190
191asm("");
192
193void coro_transfer(coro_context *prev, coro_context *next)
194{
195 pthread_cond_init (&prev->c, 0);
196 pthread_cond_signal (&next->c);
197 pthread_cond_wait (&prev->c, &coro_mutex);
198 pthread_cond_destroy (&prev->c);
199}
200
201#endif
202
203/* initialize a machine state */
204void coro_create (coro_context *ctx,
205 coro_func coro, void *arg,
206 void *sptr, long ssize)
207{
208#if CORO_UCONTEXT
209
210 getcontext (&(ctx->uc));
211
212 ctx->uc.uc_link = 0;
213 ctx->uc.uc_stack.ss_sp = STACK_ADJUST_PTR (sptr,ssize);
214 ctx->uc.uc_stack.ss_size = (size_t)STACK_ADJUST_SIZE (sptr,ssize);
215 ctx->uc.uc_stack.ss_flags = 0;
216
217 makecontext (&(ctx->uc), (void (*)()) coro, 1, arg);
218
219#elif CORO_SJLJ || CORO_LOSER || CORO_LINUX || CORO_IRIX || CORO_ASM
220
221# if CORO_SJLJ 247# if CORO_SJLJ
222 stack_t ostk, nstk; 248 stack_t ostk, nstk;
223 struct sigaction osa, nsa; 249 struct sigaction osa, nsa;
224 sigset_t nsig, osig; 250 sigset_t nsig, osig;
225# endif 251# endif
226 coro_context nctx; 252
253 if (!coro)
254 return;
227 255
228 coro_init_func = coro; 256 coro_init_func = coro;
229 coro_init_arg = arg; 257 coro_init_arg = arg;
230 258
231 new_coro = ctx; 259 new_coro = ctx;
247 perror ("sigaction"); 275 perror ("sigaction");
248 abort (); 276 abort ();
249 } 277 }
250 278
251 /* set the new stack */ 279 /* set the new stack */
252 nstk.ss_sp = STACK_ADJUST_PTR (sptr,ssize); /* yes, some platforms (IRIX) get this wrong. */ 280 nstk.ss_sp = STACK_ADJUST_PTR (sptr, ssize); /* yes, some platforms (IRIX) get this wrong. */
253 nstk.ss_size = STACK_ADJUST_SIZE (sptr,ssize); 281 nstk.ss_size = STACK_ADJUST_SIZE (sptr, ssize);
254 nstk.ss_flags = 0; 282 nstk.ss_flags = 0;
255 283
256 if (sigaltstack (&nstk, &ostk) < 0) 284 if (sigaltstack (&nstk, &ostk) < 0)
257 { 285 {
258 perror ("sigaltstack"); 286 perror ("sigaltstack");
259 abort (); 287 abort ();
260 } 288 }
261 289
262 trampoline_count = 0; 290 trampoline_done = 0;
263 kill (getpid (), SIGUSR2); 291 kill (getpid (), SIGUSR2);
264 sigfillset (&nsig); sigdelset (&nsig, SIGUSR2); 292 sigfillset (&nsig); sigdelset (&nsig, SIGUSR2);
265 293
266 while (!trampoline_count) 294 while (!trampoline_done)
267 sigsuspend (&nsig); 295 sigsuspend (&nsig);
268 296
269 sigaltstack (0, &nstk); 297 sigaltstack (0, &nstk);
270 nstk.ss_flags = SS_DISABLE; 298 nstk.ss_flags = SS_DISABLE;
271 if (sigaltstack (&nstk, 0) < 0) 299 if (sigaltstack (&nstk, 0) < 0)
277 305
278 if (~ostk.ss_flags & SS_DISABLE) 306 if (~ostk.ss_flags & SS_DISABLE)
279 sigaltstack (&ostk, 0); 307 sigaltstack (&ostk, 0);
280 308
281 sigaction (SIGUSR2, &osa, 0); 309 sigaction (SIGUSR2, &osa, 0);
282
283 sigprocmask (SIG_SETMASK, &osig, 0); 310 sigprocmask (SIG_SETMASK, &osig, 0);
284 311
285# elif CORO_LOSER 312# elif CORO_LOSER
286 313
287 setjmp (ctx->env); 314 coro_setjmp (ctx->env);
288#if __CYGWIN__ 315 #if __CYGWIN__ && __i386
316 ctx->env[8] = (long) coro_init;
289 ctx->env[7] = (long)((char *)sptr + ssize) - sizeof (long); 317 ctx->env[7] = (long) ((char *)sptr + ssize) - sizeof (long);
290 ctx->env[8] = (long)coro_init; 318 #elif __CYGWIN__ && __x86_64
319 ctx->env[7] = (long) coro_init;
320 ctx->env[6] = (long) ((char *)sptr + ssize) - sizeof (long);
321 #elif defined __MINGW32__
322 ctx->env[5] = (long) coro_init;
323 ctx->env[4] = (long) ((char *)sptr + ssize) - sizeof (long);
291#elif defined(_M_IX86) 324 #elif defined _M_IX86
292 ((_JUMP_BUFFER *)&ctx->env)->Eip = (long)coro_init; 325 ((_JUMP_BUFFER *)&ctx->env)->Eip = (long) coro_init;
293 ((_JUMP_BUFFER *)&ctx->env)->Esp = (long)STACK_ADJUST_PTR (sptr, ssize) - sizeof (long); 326 ((_JUMP_BUFFER *)&ctx->env)->Esp = (long) STACK_ADJUST_PTR (sptr, ssize) - sizeof (long);
294#elif defined(_M_AMD64) 327 #elif defined _M_AMD64
295 ((_JUMP_BUFFER *)&ctx->env)->Rip = (__int64)coro_init; 328 ((_JUMP_BUFFER *)&ctx->env)->Rip = (__int64) coro_init;
296 ((_JUMP_BUFFER *)&ctx->env)->Rsp = (__int64)STACK_ADJUST_PTR (sptr, ssize) - sizeof (long); 329 ((_JUMP_BUFFER *)&ctx->env)->Rsp = (__int64) STACK_ADJUST_PTR (sptr, ssize) - sizeof (__int64);
297#elif defined(_M_IA64) 330 #elif defined _M_IA64
298 ((_JUMP_BUFFER *)&ctx->env)->StIIP = (__int64)coro_init; 331 ((_JUMP_BUFFER *)&ctx->env)->StIIP = (__int64) coro_init;
299 ((_JUMP_BUFFER *)&ctx->env)->IntSp = (__int64)STACK_ADJUST_PTR (sptr, ssize) - sizeof (long); 332 ((_JUMP_BUFFER *)&ctx->env)->IntSp = (__int64) STACK_ADJUST_PTR (sptr, ssize) - sizeof (__int64);
300#else 333 #else
301# error "microsoft libc or architecture not supported" 334 #error "microsoft libc or architecture not supported"
335 #endif
336
337# elif CORO_LINUX
338
339 coro_setjmp (ctx->env);
340 #if __GLIBC__ >= 2 && __GLIBC_MINOR__ >= 0 && defined (JB_PC) && defined (JB_SP)
341 ctx->env[0].__jmpbuf[JB_PC] = (long) coro_init;
342 ctx->env[0].__jmpbuf[JB_SP] = (long) STACK_ADJUST_PTR (sptr, ssize) - sizeof (long);
343 #elif __GLIBC__ >= 2 && __GLIBC_MINOR__ >= 0 && defined (__mc68000__)
344 ctx->env[0].__jmpbuf[0].__aregs[0] = (long int)coro_init;
345 ctx->env[0].__jmpbuf[0].__sp = (int *) ((char *)sptr + ssize) - sizeof (long);
346 #elif defined (__GNU_LIBRARY__) && defined (__i386__)
347 ctx->env[0].__jmpbuf[0].__pc = (char *) coro_init;
348 ctx->env[0].__jmpbuf[0].__sp = (void *) ((char *)sptr + ssize) - sizeof (long);
349 #elif defined (__GNU_LIBRARY__) && defined (__amd64__)
350 ctx->env[0].__jmpbuf[JB_PC] = (long) coro_init;
351 ctx->env[0].__jmpbuf[0].__sp = (void *) ((char *)sptr + ssize) - sizeof (long);
352 #else
353 #error "linux libc or architecture not supported"
354 #endif
355
356# elif CORO_IRIX
357
358 coro_setjmp (ctx->env, 0);
359 ctx->env[JB_PC] = (__uint64_t)coro_init;
360 ctx->env[JB_SP] = (__uint64_t)STACK_ADJUST_PTR (sptr, ssize) - sizeof (long);
361
362# elif CORO_ASM
363
364 ctx->sp = (void **)(ssize + (char *)sptr);
365 *--ctx->sp = (void *)abort; /* needed for alignment only */
366 *--ctx->sp = (void *)coro_init;
367
368 #if CORO_WIN_TIB
369 *--ctx->sp = 0; /* ExceptionList */
370 *--ctx->sp = (char *)sptr + ssize; /* StackBase */
371 *--ctx->sp = sptr; /* StackLimit */
372 #endif
373
374 ctx->sp -= NUM_SAVED;
375 memset (ctx->sp, 0, sizeof (*ctx->sp) * NUM_SAVED);
376
377# elif CORO_UCONTEXT
378
379 getcontext (&(ctx->uc));
380
381 ctx->uc.uc_link = 0;
382 ctx->uc.uc_stack.ss_sp = sptr;
383 ctx->uc.uc_stack.ss_size = (size_t)ssize;
384 ctx->uc.uc_stack.ss_flags = 0;
385
386 makecontext (&(ctx->uc), (void (*)())coro_init, 0);
387
388# endif
389
390 coro_transfer (create_coro, new_coro);
391}
392
393/*****************************************************************************/
394/* pthread backend */
395/*****************************************************************************/
396#elif CORO_PTHREAD
397
398/* this mutex will be locked by the running coroutine */
399pthread_mutex_t coro_mutex = PTHREAD_MUTEX_INITIALIZER;
400
401struct coro_init_args
402{
403 coro_func func;
404 void *arg;
405 coro_context *self, *main;
406};
407
408static pthread_t null_tid;
409
410/* I'd so love to cast pthread_mutex_unlock to void (*)(void *)... */
411static void
412mutex_unlock_wrapper (void *arg)
413{
414 pthread_mutex_unlock ((pthread_mutex_t *)arg);
415}
416
417static void *
418coro_init (void *args_)
419{
420 struct coro_init_args *args = (struct coro_init_args *)args_;
421 coro_func func = args->func;
422 void *arg = args->arg;
423
424 pthread_mutex_lock (&coro_mutex);
425
426 /* we try to be good citizens and use deferred cancellation and cleanup handlers */
427 pthread_cleanup_push (mutex_unlock_wrapper, &coro_mutex);
428 coro_transfer (args->self, args->main);
429 func (arg);
430 pthread_cleanup_pop (1);
431
432 return 0;
433}
434
435void
436coro_transfer (coro_context *prev, coro_context *next)
437{
438 pthread_cond_signal (&next->cv);
439 pthread_cond_wait (&prev->cv, &coro_mutex);
440#if __FreeBSD__ /* freebsd is of course broken and needs manual testcancel calls... yay... */
441 pthread_testcancel ();
302#endif 442#endif
443}
303 444
304# elif CORO_LINUX 445void
305 446coro_create (coro_context *ctx, coro_func coro, void *arg, void *sptr, long ssize)
306 _setjmp (ctx->env); 447{
307#if __GLIBC__ >= 2 && __GLIBC_MINOR__ >= 0 && defined (JB_PC) && defined (JB_SP)
308 ctx->env[0].__jmpbuf[JB_PC] = (long)coro_init;
309 ctx->env[0].__jmpbuf[JB_SP] = (long)STACK_ADJUST_PTR (sptr, ssize) - sizeof (long);
310#elif __GLIBC__ >= 2 && __GLIBC_MINOR__ >= 0 && defined (__mc68000__)
311 ctx->env[0].__jmpbuf[0].__aregs[0] = (long int)coro_init;
312 ctx->env[0].__jmpbuf[0].__sp = (int *)((char *)sptr + ssize) - sizeof (long);
313#elif defined (__GNU_LIBRARY__) && defined (__i386__)
314 ctx->env[0].__jmpbuf[0].__pc = (char *)coro_init;
315 ctx->env[0].__jmpbuf[0].__sp = (void *)((char *)sptr + ssize) - sizeof (long);
316#elif defined (__GNU_LIBRARY__) && defined (__amd64__)
317 ctx->env[0].__jmpbuf[JB_PC] = (long)coro_init;
318 ctx->env[0].__jmpbuf[0].__sp = (void *)((char *)sptr + ssize) - sizeof (long);
319#else
320# error "linux libc or architecture not supported"
321#endif
322
323# elif CORO_IRIX
324
325 setjmp (ctx->env);
326 ctx->env[JB_PC] = (__uint64_t)coro_init;
327 ctx->env[JB_SP] = (__uint64_t)STACK_ADJUST_PTR (sptr, ssize) - sizeof (long);
328
329# elif CORO_ASM
330
331 ctx->sp = (volatile void **)(ssize + (char *)sptr);
332 ctx->sp -= NUM_ALIGN;
333 *--ctx->sp = (void *)coro_init;
334 ctx->sp -= NUM_SAVED;
335
336# endif
337
338 coro_transfer ((coro_context *)create_coro, (coro_context *)new_coro);
339
340# elif CORO_PTHREAD
341
342 pthread_t id;
343 pthread_attr_t attr;
344 coro_context nctx; 448 static coro_context nctx;
345 struct coro_init_args args;
346 static int once; 449 static int once;
347 450
348 if (!once) 451 if (!once)
349 { 452 {
453 once = 1;
454
350 pthread_mutex_lock (&coro_mutex); 455 pthread_mutex_lock (&coro_mutex);
351 once = 1; 456 pthread_cond_init (&nctx.cv, 0);
457 null_tid = pthread_self ();
352 } 458 }
353 459
460 pthread_cond_init (&ctx->cv, 0);
461
462 if (coro)
463 {
464 pthread_attr_t attr;
465 struct coro_init_args args;
466
354 args.func = coro; 467 args.func = coro;
355 args.arg = arg; 468 args.arg = arg;
356 args.self = ctx; 469 args.self = ctx;
357 args.main = &nctx; 470 args.main = &nctx;
358 471
359 pthread_attr_init (&attr); 472 pthread_attr_init (&attr);
360 pthread_attr_setdetachstate (&attr, PTHREAD_CREATE_DETACHED); 473#if __UCLIBC__
474 /* exists, but is borked */
475 /*pthread_attr_setstacksize (&attr, (size_t)ssize);*/
476#elif __CYGWIN__
477 /* POSIX, not here */
478 pthread_attr_setstacksize (&attr, (size_t)ssize);
479#else
361 pthread_attr_setstack (&attr, sptr, (size_t)ssize); 480 pthread_attr_setstack (&attr, sptr, (size_t)ssize);
481#endif
482 pthread_attr_setscope (&attr, PTHREAD_SCOPE_PROCESS);
362 pthread_create (&id, &attr, trampoline, &args); 483 pthread_create (&ctx->id, &attr, coro_init, &args);
363 484
364 pthread_cond_init (&args.self->c, 0);
365 coro_transfer (args.main, args.self); 485 coro_transfer (args.main, args.self);
486 }
487 else
488 ctx->id = null_tid;
489}
490
491void
492coro_destroy (coro_context *ctx)
493{
494 if (!pthread_equal (ctx->id, null_tid))
495 {
496 pthread_cancel (ctx->id);
497 pthread_mutex_unlock (&coro_mutex);
498 pthread_join (ctx->id, 0);
499 pthread_mutex_lock (&coro_mutex);
500 }
501
502 pthread_cond_destroy (&ctx->cv);
503}
504
505/*****************************************************************************/
506/* fiber backend */
507/*****************************************************************************/
508#elif CORO_FIBER
509
510#define WIN32_LEAN_AND_MEAN
511#include <windows.h>
512
513VOID CALLBACK
514coro_init (PVOID arg)
515{
516 coro_context *ctx = (coro_context *)arg;
517
518 ctx->coro (ctx->arg);
519}
520
521void
522coro_transfer (coro_context *prev, coro_context *next)
523{
524 if (!prev->fiber)
525 {
526 prev->fiber = GetCurrentFiber ();
527
528 if (prev->fiber == 0 || prev->fiber == (void *)0x1e00)
529 prev->fiber = ConvertThreadToFiber (0);
530 }
531
532 SwitchToFiber (next->fiber);
533}
534
535void
536coro_create (coro_context *ctx, coro_func coro, void *arg, void *sptr, long ssize)
537{
538 ctx->fiber = 0;
539 ctx->coro = coro;
540 ctx->arg = arg;
541
542 if (!coro)
543 return;
544
545 ctx->fiber = CreateFiber (ssize, coro_init, ctx);
546}
547
548void
549coro_destroy (coro_context *ctx)
550{
551 DeleteFiber (ctx->fiber);
552}
366 553
367#else 554#else
368# error unsupported backend 555# error unsupported backend
369#endif 556#endif
370}
371 557

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines