1 | /* |
1 | /* |
2 | * libeio implementation |
2 | * libeio implementation |
3 | * |
3 | * |
4 | * Copyright (c) 2007,2008,2009,2010,2011,2012 Marc Alexander Lehmann <libeio@schmorp.de> |
4 | * Copyright (c) 2007,2008,2009,2010,2011,2012,2013 Marc Alexander Lehmann <libeio@schmorp.de> |
5 | * All rights reserved. |
5 | * All rights reserved. |
6 | * |
6 | * |
7 | * Redistribution and use in source and binary forms, with or without modifica- |
7 | * Redistribution and use in source and binary forms, with or without modifica- |
8 | * tion, are permitted provided that the following conditions are met: |
8 | * tion, are permitted provided that the following conditions are met: |
9 | * |
9 | * |
10 | * 1. Redistributions of source code must retain the above copyright notice, |
10 | * 1. Redistributions of source code must retain the above copyright notice, |
11 | * this list of conditions and the following disclaimer. |
11 | * this list of conditions and the following disclaimer. |
12 | * |
12 | * |
13 | * 2. Redistributions in binary form must reproduce the above copyright |
13 | * 2. Redistributions in binary form must reproduce the above copyright |
14 | * notice, this list of conditions and the following disclaimer in the |
14 | * notice, this list of conditions and the following disclaimer in the |
15 | * documentation and/or other materials provided with the distribution. |
15 | * documentation and/or other materials provided with the distribution. |
16 | * |
16 | * |
17 | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED |
17 | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED |
18 | * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MER- |
18 | * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MER- |
19 | * CHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO |
19 | * CHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO |
20 | * EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPE- |
20 | * EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPE- |
21 | * CIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
21 | * CIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
… | |
… | |
120 | #define link(old,neu) (CreateHardLink (neu, old, 0) ? 0 : EIO_ERRNO (ENOENT, -1)) |
120 | #define link(old,neu) (CreateHardLink (neu, old, 0) ? 0 : EIO_ERRNO (ENOENT, -1)) |
121 | |
121 | |
122 | #define chmod(path,mode) _chmod (path, mode) |
122 | #define chmod(path,mode) _chmod (path, mode) |
123 | #define dup(fd) _dup (fd) |
123 | #define dup(fd) _dup (fd) |
124 | #define dup2(fd1,fd2) _dup2 (fd1, fd2) |
124 | #define dup2(fd1,fd2) _dup2 (fd1, fd2) |
|
|
125 | #define pipe(fds) _pipe (fds, 4096, O_BINARY) |
125 | |
126 | |
126 | #define fchmod(fd,mode) EIO_ENOSYS () |
127 | #define fchmod(fd,mode) EIO_ENOSYS () |
127 | #define chown(path,uid,gid) EIO_ENOSYS () |
128 | #define chown(path,uid,gid) EIO_ENOSYS () |
128 | #define fchown(fd,uid,gid) EIO_ENOSYS () |
129 | #define fchown(fd,uid,gid) EIO_ENOSYS () |
129 | #define truncate(path,offs) EIO_ENOSYS () /* far-miss: SetEndOfFile */ |
130 | #define truncate(path,offs) EIO_ENOSYS () /* far-miss: SetEndOfFile */ |
… | |
… | |
131 | #define mknod(path,mode,dev) EIO_ENOSYS () |
132 | #define mknod(path,mode,dev) EIO_ENOSYS () |
132 | #define sync() EIO_ENOSYS () |
133 | #define sync() EIO_ENOSYS () |
133 | #define readlink(path,buf,s) EIO_ENOSYS () |
134 | #define readlink(path,buf,s) EIO_ENOSYS () |
134 | #define statvfs(path,buf) EIO_ENOSYS () |
135 | #define statvfs(path,buf) EIO_ENOSYS () |
135 | #define fstatvfs(fd,buf) EIO_ENOSYS () |
136 | #define fstatvfs(fd,buf) EIO_ENOSYS () |
|
|
137 | |
|
|
138 | #define pread(fd,buf,count,offset) eio__pread (fd, buf, count, offset) |
|
|
139 | #define pwrite(fd,buf,count,offset) eio__pwrite (fd, buf, count, offset) |
|
|
140 | |
|
|
141 | #if __GNUC__ |
|
|
142 | typedef long long eio_off_t; /* signed for compatibility to msvc */ |
|
|
143 | #else |
|
|
144 | typedef __int64 eio_off_t; /* unsigned not supported by msvc */ |
|
|
145 | #endif |
|
|
146 | |
|
|
147 | static eio_ssize_t |
|
|
148 | eio__pread (int fd, void *buf, eio_ssize_t count, eio_off_t offset) |
|
|
149 | { |
|
|
150 | OVERLAPPED o = { 0 }; |
|
|
151 | DWORD got; |
|
|
152 | |
|
|
153 | o.Offset = offset; |
|
|
154 | o.OffsetHigh = offset >> 32; |
|
|
155 | |
|
|
156 | return ReadFile ((HANDLE)EIO_FD_TO_WIN32_HANDLE (fd), buf, count, &got, &o) |
|
|
157 | ? got : -1; |
|
|
158 | } |
|
|
159 | |
|
|
160 | static eio_ssize_t |
|
|
161 | eio__pwrite (int fd, void *buf, eio_ssize_t count, eio_off_t offset) |
|
|
162 | { |
|
|
163 | OVERLAPPED o = { 0 }; |
|
|
164 | DWORD got; |
|
|
165 | |
|
|
166 | o.Offset = offset; |
|
|
167 | o.OffsetHigh = offset >> 32; |
|
|
168 | |
|
|
169 | return WriteFile ((HANDLE)EIO_FD_TO_WIN32_HANDLE (fd), buf, count, &got, &o) |
|
|
170 | ? got : -1; |
|
|
171 | } |
136 | |
172 | |
137 | /* rename() uses MoveFile, which fails to overwrite */ |
173 | /* rename() uses MoveFile, which fails to overwrite */ |
138 | #define rename(old,neu) eio__rename (old, neu) |
174 | #define rename(old,neu) eio__rename (old, neu) |
139 | |
175 | |
140 | static int |
176 | static int |
… | |
… | |
178 | #endif |
214 | #endif |
179 | |
215 | |
180 | return EIO_ERRNO (ENOENT, -1); |
216 | return EIO_ERRNO (ENOENT, -1); |
181 | } |
217 | } |
182 | |
218 | |
183 | /* POSIX API only */ |
219 | /* POSIX API only, causing trouble for win32 apps */ |
184 | #define CreateHardLink(neu,old,flags) 0 |
220 | #define CreateHardLink(neu,old,flags) 0 /* not really creating hardlink, still using relative paths? */ |
185 | #define CreateSymbolicLink(neu,old,flags) 0 |
221 | #define CreateSymbolicLink(neu,old,flags) 0 /* vista+ only */ |
186 | |
222 | |
187 | struct statvfs |
223 | struct statvfs |
188 | { |
224 | { |
189 | int dummy; |
225 | int dummy; |
190 | }; |
226 | }; |
… | |
… | |
196 | |
232 | |
197 | #else |
233 | #else |
198 | |
234 | |
199 | #include <sys/time.h> |
235 | #include <sys/time.h> |
200 | #include <sys/select.h> |
236 | #include <sys/select.h> |
201 | #include <sys/statvfs.h> |
|
|
202 | #include <unistd.h> |
237 | #include <unistd.h> |
203 | #include <signal.h> |
238 | #include <signal.h> |
204 | #include <dirent.h> |
239 | #include <dirent.h> |
|
|
240 | |
|
|
241 | #ifdef ANDROID |
|
|
242 | #include <sys/vfs.h> |
|
|
243 | #define statvfs statfs |
|
|
244 | #define fstatvfs fstatfs |
|
|
245 | #include <asm/page.h> /* supposedly limits.h does #define PAGESIZE PAGESIZE */ |
|
|
246 | #else |
|
|
247 | #include <sys/statvfs.h> |
|
|
248 | #endif |
205 | |
249 | |
206 | #if _POSIX_MEMLOCK || _POSIX_MEMLOCK_RANGE || _POSIX_MAPPED_FILES |
250 | #if _POSIX_MEMLOCK || _POSIX_MEMLOCK_RANGE || _POSIX_MAPPED_FILES |
207 | #include <sys/mman.h> |
251 | #include <sys/mman.h> |
208 | #endif |
252 | #endif |
209 | |
253 | |
… | |
… | |
281 | #endif |
325 | #endif |
282 | |
326 | |
283 | /* buffer size for various temporary buffers */ |
327 | /* buffer size for various temporary buffers */ |
284 | #define EIO_BUFSIZE 65536 |
328 | #define EIO_BUFSIZE 65536 |
285 | |
329 | |
286 | #define dBUF \ |
330 | #define dBUF \ |
287 | char *eio_buf = malloc (EIO_BUFSIZE); \ |
331 | char *eio_buf = malloc (EIO_BUFSIZE); \ |
288 | errno = ENOMEM; \ |
332 | errno = ENOMEM; \ |
289 | if (!eio_buf) \ |
333 | if (!eio_buf) \ |
290 | return -1 |
334 | return -1 |
291 | |
335 | |
292 | #define FUBd \ |
336 | #define FUBd \ |
293 | free (eio_buf) |
337 | free (eio_buf) |
294 | |
338 | |
295 | #define EIO_TICKS ((1000000 + 1023) >> 10) |
|
|
296 | |
|
|
297 | /*****************************************************************************/ |
339 | /*****************************************************************************/ |
298 | |
340 | |
299 | struct tmpbuf |
|
|
300 | { |
|
|
301 | void *ptr; |
|
|
302 | int len; |
|
|
303 | }; |
|
|
304 | |
|
|
305 | static void * |
|
|
306 | tmpbuf_get (struct tmpbuf *buf, int len) |
|
|
307 | { |
|
|
308 | if (buf->len < len) |
|
|
309 | { |
|
|
310 | free (buf->ptr); |
|
|
311 | buf->ptr = malloc (buf->len = len); |
|
|
312 | } |
|
|
313 | |
|
|
314 | return buf->ptr; |
|
|
315 | } |
|
|
316 | |
|
|
317 | struct tmpbuf; |
341 | struct etp_tmpbuf; |
318 | |
342 | |
319 | #if _POSIX_VERSION >= 200809L |
343 | #if _POSIX_VERSION >= 200809L |
320 | #define HAVE_AT 1 |
344 | #define HAVE_AT 1 |
321 | #define WD2FD(wd) ((wd) ? (wd)->fd : AT_FDCWD) |
345 | #define WD2FD(wd) ((wd) ? (wd)->fd : AT_FDCWD) |
322 | #ifndef O_SEARCH |
346 | #ifndef O_SEARCH |
323 | #define O_SEARCH O_RDONLY |
347 | #define O_SEARCH O_RDONLY |
324 | #endif |
348 | #endif |
325 | #else |
349 | #else |
326 | #define HAVE_AT 0 |
350 | #define HAVE_AT 0 |
327 | static const char *wd_expand (struct tmpbuf *tmpbuf, eio_wd wd, const char *path); |
351 | static const char *wd_expand (struct etp_tmpbuf *tmpbuf, eio_wd wd, const char *path); |
328 | #endif |
352 | #endif |
329 | |
353 | |
330 | struct eio_pwd |
354 | struct eio_pwd |
331 | { |
355 | { |
332 | #if HAVE_AT |
356 | #if HAVE_AT |
… | |
… | |
339 | /*****************************************************************************/ |
363 | /*****************************************************************************/ |
340 | |
364 | |
341 | #define ETP_PRI_MIN EIO_PRI_MIN |
365 | #define ETP_PRI_MIN EIO_PRI_MIN |
342 | #define ETP_PRI_MAX EIO_PRI_MAX |
366 | #define ETP_PRI_MAX EIO_PRI_MAX |
343 | |
367 | |
|
|
368 | #define ETP_TYPE_QUIT -1 |
|
|
369 | #define ETP_TYPE_GROUP EIO_GROUP |
|
|
370 | |
|
|
371 | static void eio_nop_callback (void) { } |
|
|
372 | static void (*eio_want_poll_cb)(void) = eio_nop_callback; |
|
|
373 | static void (*eio_done_poll_cb)(void) = eio_nop_callback; |
|
|
374 | |
|
|
375 | #define ETP_WANT_POLL(pool) eio_want_poll_cb () |
|
|
376 | #define ETP_DONE_POLL(pool) eio_done_poll_cb () |
|
|
377 | |
344 | struct etp_worker; |
378 | struct etp_worker; |
345 | |
|
|
346 | #define ETP_REQ eio_req |
379 | #define ETP_REQ eio_req |
347 | #define ETP_DESTROY(req) eio_destroy (req) |
380 | #define ETP_DESTROY(req) eio_destroy (req) |
348 | static int eio_finish (eio_req *req); |
381 | static int eio_finish (eio_req *req); |
349 | #define ETP_FINISH(req) eio_finish (req) |
382 | #define ETP_FINISH(req) eio_finish (req) |
350 | static void eio_execute (struct etp_worker *self, eio_req *req); |
383 | static void eio_execute (struct etp_worker *self, eio_req *req); |
351 | #define ETP_EXECUTE(wrk,req) eio_execute (wrk,req) |
384 | #define ETP_EXECUTE(wrk,req) eio_execute (wrk, req) |
352 | |
385 | |
353 | /*****************************************************************************/ |
386 | #include "etp.c" |
354 | |
387 | |
355 | #define ETP_NUM_PRI (ETP_PRI_MAX - ETP_PRI_MIN + 1) |
388 | static struct etp_pool eio_pool; |
356 | |
389 | #define EIO_POOL (&eio_pool) |
357 | /* calculate time difference in ~1/EIO_TICKS of a second */ |
|
|
358 | ecb_inline int |
|
|
359 | tvdiff (struct timeval *tv1, struct timeval *tv2) |
|
|
360 | { |
|
|
361 | return (tv2->tv_sec - tv1->tv_sec ) * EIO_TICKS |
|
|
362 | + ((tv2->tv_usec - tv1->tv_usec) >> 10); |
|
|
363 | } |
|
|
364 | |
|
|
365 | static unsigned int started, idle, wanted = 4; |
|
|
366 | |
|
|
367 | static void (*want_poll_cb) (void); |
|
|
368 | static void (*done_poll_cb) (void); |
|
|
369 | |
|
|
370 | static unsigned int max_poll_time; /* reslock */ |
|
|
371 | static unsigned int max_poll_reqs; /* reslock */ |
|
|
372 | |
|
|
373 | static unsigned int nreqs; /* reqlock */ |
|
|
374 | static unsigned int nready; /* reqlock */ |
|
|
375 | static unsigned int npending; /* reqlock */ |
|
|
376 | static unsigned int max_idle = 4; /* maximum number of threads that can idle indefinitely */ |
|
|
377 | static unsigned int idle_timeout = 10; /* number of seconds after which an idle threads exit */ |
|
|
378 | |
|
|
379 | static xmutex_t wrklock; |
|
|
380 | static xmutex_t reslock; |
|
|
381 | static xmutex_t reqlock; |
|
|
382 | static xcond_t reqwait; |
|
|
383 | |
|
|
384 | #if !HAVE_PREADWRITE |
|
|
385 | /* |
|
|
386 | * make our pread/pwrite emulation safe against themselves, but not against |
|
|
387 | * normal read/write by using a mutex. slows down execution a lot, |
|
|
388 | * but that's your problem, not mine. |
|
|
389 | */ |
|
|
390 | static xmutex_t preadwritelock; |
|
|
391 | #endif |
|
|
392 | |
|
|
393 | typedef struct etp_worker |
|
|
394 | { |
|
|
395 | struct tmpbuf tmpbuf; |
|
|
396 | |
|
|
397 | /* locked by wrklock */ |
|
|
398 | struct etp_worker *prev, *next; |
|
|
399 | |
|
|
400 | xthread_t tid; |
|
|
401 | |
|
|
402 | #ifdef ETP_WORKER_COMMON |
|
|
403 | ETP_WORKER_COMMON |
|
|
404 | #endif |
|
|
405 | } etp_worker; |
|
|
406 | |
|
|
407 | static etp_worker wrk_first; /* NOT etp */ |
|
|
408 | |
|
|
409 | #define ETP_WORKER_LOCK(wrk) X_LOCK (wrklock) |
|
|
410 | #define ETP_WORKER_UNLOCK(wrk) X_UNLOCK (wrklock) |
|
|
411 | |
|
|
412 | /* worker threads management */ |
|
|
413 | |
|
|
414 | static void |
|
|
415 | etp_worker_clear (etp_worker *wrk) |
|
|
416 | { |
|
|
417 | } |
|
|
418 | |
|
|
419 | static void ecb_cold |
|
|
420 | etp_worker_free (etp_worker *wrk) |
|
|
421 | { |
|
|
422 | free (wrk->tmpbuf.ptr); |
|
|
423 | |
|
|
424 | wrk->next->prev = wrk->prev; |
|
|
425 | wrk->prev->next = wrk->next; |
|
|
426 | |
|
|
427 | free (wrk); |
|
|
428 | } |
|
|
429 | |
|
|
430 | static unsigned int |
|
|
431 | etp_nreqs (void) |
|
|
432 | { |
|
|
433 | int retval; |
|
|
434 | if (WORDACCESS_UNSAFE) X_LOCK (reqlock); |
|
|
435 | retval = nreqs; |
|
|
436 | if (WORDACCESS_UNSAFE) X_UNLOCK (reqlock); |
|
|
437 | return retval; |
|
|
438 | } |
|
|
439 | |
|
|
440 | static unsigned int |
|
|
441 | etp_nready (void) |
|
|
442 | { |
|
|
443 | unsigned int retval; |
|
|
444 | |
|
|
445 | if (WORDACCESS_UNSAFE) X_LOCK (reqlock); |
|
|
446 | retval = nready; |
|
|
447 | if (WORDACCESS_UNSAFE) X_UNLOCK (reqlock); |
|
|
448 | |
|
|
449 | return retval; |
|
|
450 | } |
|
|
451 | |
|
|
452 | static unsigned int |
|
|
453 | etp_npending (void) |
|
|
454 | { |
|
|
455 | unsigned int retval; |
|
|
456 | |
|
|
457 | if (WORDACCESS_UNSAFE) X_LOCK (reqlock); |
|
|
458 | retval = npending; |
|
|
459 | if (WORDACCESS_UNSAFE) X_UNLOCK (reqlock); |
|
|
460 | |
|
|
461 | return retval; |
|
|
462 | } |
|
|
463 | |
|
|
464 | static unsigned int |
|
|
465 | etp_nthreads (void) |
|
|
466 | { |
|
|
467 | unsigned int retval; |
|
|
468 | |
|
|
469 | if (WORDACCESS_UNSAFE) X_LOCK (reqlock); |
|
|
470 | retval = started; |
|
|
471 | if (WORDACCESS_UNSAFE) X_UNLOCK (reqlock); |
|
|
472 | |
|
|
473 | return retval; |
|
|
474 | } |
|
|
475 | |
|
|
476 | /* |
|
|
477 | * a somewhat faster data structure might be nice, but |
|
|
478 | * with 8 priorities this actually needs <20 insns |
|
|
479 | * per shift, the most expensive operation. |
|
|
480 | */ |
|
|
481 | typedef struct { |
|
|
482 | ETP_REQ *qs[ETP_NUM_PRI], *qe[ETP_NUM_PRI]; /* qstart, qend */ |
|
|
483 | int size; |
|
|
484 | } etp_reqq; |
|
|
485 | |
|
|
486 | static etp_reqq req_queue; |
|
|
487 | static etp_reqq res_queue; |
|
|
488 | |
|
|
489 | static void ecb_noinline ecb_cold |
|
|
490 | reqq_init (etp_reqq *q) |
|
|
491 | { |
|
|
492 | int pri; |
|
|
493 | |
|
|
494 | for (pri = 0; pri < ETP_NUM_PRI; ++pri) |
|
|
495 | q->qs[pri] = q->qe[pri] = 0; |
|
|
496 | |
|
|
497 | q->size = 0; |
|
|
498 | } |
|
|
499 | |
|
|
500 | static int ecb_noinline |
|
|
501 | reqq_push (etp_reqq *q, ETP_REQ *req) |
|
|
502 | { |
|
|
503 | int pri = req->pri; |
|
|
504 | req->next = 0; |
|
|
505 | |
|
|
506 | if (q->qe[pri]) |
|
|
507 | { |
|
|
508 | q->qe[pri]->next = req; |
|
|
509 | q->qe[pri] = req; |
|
|
510 | } |
|
|
511 | else |
|
|
512 | q->qe[pri] = q->qs[pri] = req; |
|
|
513 | |
|
|
514 | return q->size++; |
|
|
515 | } |
|
|
516 | |
|
|
517 | static ETP_REQ * ecb_noinline |
|
|
518 | reqq_shift (etp_reqq *q) |
|
|
519 | { |
|
|
520 | int pri; |
|
|
521 | |
|
|
522 | if (!q->size) |
|
|
523 | return 0; |
|
|
524 | |
|
|
525 | --q->size; |
|
|
526 | |
|
|
527 | for (pri = ETP_NUM_PRI; pri--; ) |
|
|
528 | { |
|
|
529 | eio_req *req = q->qs[pri]; |
|
|
530 | |
|
|
531 | if (req) |
|
|
532 | { |
|
|
533 | if (!(q->qs[pri] = (eio_req *)req->next)) |
|
|
534 | q->qe[pri] = 0; |
|
|
535 | |
|
|
536 | return req; |
|
|
537 | } |
|
|
538 | } |
|
|
539 | |
|
|
540 | abort (); |
|
|
541 | } |
|
|
542 | |
|
|
543 | static int ecb_cold |
|
|
544 | etp_init (void (*want_poll)(void), void (*done_poll)(void)) |
|
|
545 | { |
|
|
546 | X_MUTEX_CREATE (wrklock); |
|
|
547 | X_MUTEX_CREATE (reslock); |
|
|
548 | X_MUTEX_CREATE (reqlock); |
|
|
549 | X_COND_CREATE (reqwait); |
|
|
550 | |
|
|
551 | reqq_init (&req_queue); |
|
|
552 | reqq_init (&res_queue); |
|
|
553 | |
|
|
554 | wrk_first.next = |
|
|
555 | wrk_first.prev = &wrk_first; |
|
|
556 | |
|
|
557 | started = 0; |
|
|
558 | idle = 0; |
|
|
559 | nreqs = 0; |
|
|
560 | nready = 0; |
|
|
561 | npending = 0; |
|
|
562 | |
|
|
563 | want_poll_cb = want_poll; |
|
|
564 | done_poll_cb = done_poll; |
|
|
565 | |
|
|
566 | return 0; |
|
|
567 | } |
|
|
568 | |
|
|
569 | X_THREAD_PROC (etp_proc); |
|
|
570 | |
|
|
571 | static void ecb_cold |
|
|
572 | etp_start_thread (void) |
|
|
573 | { |
|
|
574 | etp_worker *wrk = calloc (1, sizeof (etp_worker)); |
|
|
575 | |
|
|
576 | /*TODO*/ |
|
|
577 | assert (("unable to allocate worker thread data", wrk)); |
|
|
578 | |
|
|
579 | X_LOCK (wrklock); |
|
|
580 | |
|
|
581 | if (thread_create (&wrk->tid, etp_proc, (void *)wrk)) |
|
|
582 | { |
|
|
583 | wrk->prev = &wrk_first; |
|
|
584 | wrk->next = wrk_first.next; |
|
|
585 | wrk_first.next->prev = wrk; |
|
|
586 | wrk_first.next = wrk; |
|
|
587 | ++started; |
|
|
588 | } |
|
|
589 | else |
|
|
590 | free (wrk); |
|
|
591 | |
|
|
592 | X_UNLOCK (wrklock); |
|
|
593 | } |
|
|
594 | |
|
|
595 | static void |
|
|
596 | etp_maybe_start_thread (void) |
|
|
597 | { |
|
|
598 | if (ecb_expect_true (etp_nthreads () >= wanted)) |
|
|
599 | return; |
|
|
600 | |
|
|
601 | /* todo: maybe use idle here, but might be less exact */ |
|
|
602 | if (ecb_expect_true (0 <= (int)etp_nthreads () + (int)etp_npending () - (int)etp_nreqs ())) |
|
|
603 | return; |
|
|
604 | |
|
|
605 | etp_start_thread (); |
|
|
606 | } |
|
|
607 | |
|
|
608 | static void ecb_cold |
|
|
609 | etp_end_thread (void) |
|
|
610 | { |
|
|
611 | eio_req *req = calloc (1, sizeof (eio_req)); /* will be freed by worker */ |
|
|
612 | |
|
|
613 | req->type = -1; |
|
|
614 | req->pri = ETP_PRI_MAX - ETP_PRI_MIN; |
|
|
615 | |
|
|
616 | X_LOCK (reqlock); |
|
|
617 | reqq_push (&req_queue, req); |
|
|
618 | X_COND_SIGNAL (reqwait); |
|
|
619 | X_UNLOCK (reqlock); |
|
|
620 | |
|
|
621 | X_LOCK (wrklock); |
|
|
622 | --started; |
|
|
623 | X_UNLOCK (wrklock); |
|
|
624 | } |
|
|
625 | |
|
|
626 | static int |
|
|
627 | etp_poll (void) |
|
|
628 | { |
|
|
629 | unsigned int maxreqs; |
|
|
630 | unsigned int maxtime; |
|
|
631 | struct timeval tv_start, tv_now; |
|
|
632 | |
|
|
633 | X_LOCK (reslock); |
|
|
634 | maxreqs = max_poll_reqs; |
|
|
635 | maxtime = max_poll_time; |
|
|
636 | X_UNLOCK (reslock); |
|
|
637 | |
|
|
638 | if (maxtime) |
|
|
639 | gettimeofday (&tv_start, 0); |
|
|
640 | |
|
|
641 | for (;;) |
|
|
642 | { |
|
|
643 | ETP_REQ *req; |
|
|
644 | |
|
|
645 | etp_maybe_start_thread (); |
|
|
646 | |
|
|
647 | X_LOCK (reslock); |
|
|
648 | req = reqq_shift (&res_queue); |
|
|
649 | |
|
|
650 | if (req) |
|
|
651 | { |
|
|
652 | --npending; |
|
|
653 | |
|
|
654 | if (!res_queue.size && done_poll_cb) |
|
|
655 | done_poll_cb (); |
|
|
656 | } |
|
|
657 | |
|
|
658 | X_UNLOCK (reslock); |
|
|
659 | |
|
|
660 | if (!req) |
|
|
661 | return 0; |
|
|
662 | |
|
|
663 | X_LOCK (reqlock); |
|
|
664 | --nreqs; |
|
|
665 | X_UNLOCK (reqlock); |
|
|
666 | |
|
|
667 | if (ecb_expect_false (req->type == EIO_GROUP && req->size)) |
|
|
668 | { |
|
|
669 | req->int1 = 1; /* mark request as delayed */ |
|
|
670 | continue; |
|
|
671 | } |
|
|
672 | else |
|
|
673 | { |
|
|
674 | int res = ETP_FINISH (req); |
|
|
675 | if (ecb_expect_false (res)) |
|
|
676 | return res; |
|
|
677 | } |
|
|
678 | |
|
|
679 | if (ecb_expect_false (maxreqs && !--maxreqs)) |
|
|
680 | break; |
|
|
681 | |
|
|
682 | if (maxtime) |
|
|
683 | { |
|
|
684 | gettimeofday (&tv_now, 0); |
|
|
685 | |
|
|
686 | if (tvdiff (&tv_start, &tv_now) >= maxtime) |
|
|
687 | break; |
|
|
688 | } |
|
|
689 | } |
|
|
690 | |
|
|
691 | errno = EAGAIN; |
|
|
692 | return -1; |
|
|
693 | } |
|
|
694 | |
|
|
695 | static void |
|
|
696 | etp_cancel (ETP_REQ *req) |
|
|
697 | { |
|
|
698 | req->cancelled = 1; |
|
|
699 | |
|
|
700 | eio_grp_cancel (req); |
|
|
701 | } |
|
|
702 | |
|
|
703 | static void |
|
|
704 | etp_submit (ETP_REQ *req) |
|
|
705 | { |
|
|
706 | req->pri -= ETP_PRI_MIN; |
|
|
707 | |
|
|
708 | if (ecb_expect_false (req->pri < ETP_PRI_MIN - ETP_PRI_MIN)) req->pri = ETP_PRI_MIN - ETP_PRI_MIN; |
|
|
709 | if (ecb_expect_false (req->pri > ETP_PRI_MAX - ETP_PRI_MIN)) req->pri = ETP_PRI_MAX - ETP_PRI_MIN; |
|
|
710 | |
|
|
711 | if (ecb_expect_false (req->type == EIO_GROUP)) |
|
|
712 | { |
|
|
713 | /* I hope this is worth it :/ */ |
|
|
714 | X_LOCK (reqlock); |
|
|
715 | ++nreqs; |
|
|
716 | X_UNLOCK (reqlock); |
|
|
717 | |
|
|
718 | X_LOCK (reslock); |
|
|
719 | |
|
|
720 | ++npending; |
|
|
721 | |
|
|
722 | if (!reqq_push (&res_queue, req) && want_poll_cb) |
|
|
723 | want_poll_cb (); |
|
|
724 | |
|
|
725 | X_UNLOCK (reslock); |
|
|
726 | } |
|
|
727 | else |
|
|
728 | { |
|
|
729 | X_LOCK (reqlock); |
|
|
730 | ++nreqs; |
|
|
731 | ++nready; |
|
|
732 | reqq_push (&req_queue, req); |
|
|
733 | X_COND_SIGNAL (reqwait); |
|
|
734 | X_UNLOCK (reqlock); |
|
|
735 | |
|
|
736 | etp_maybe_start_thread (); |
|
|
737 | } |
|
|
738 | } |
|
|
739 | |
|
|
740 | static void ecb_cold |
|
|
741 | etp_set_max_poll_time (double nseconds) |
|
|
742 | { |
|
|
743 | if (WORDACCESS_UNSAFE) X_LOCK (reslock); |
|
|
744 | max_poll_time = nseconds * EIO_TICKS; |
|
|
745 | if (WORDACCESS_UNSAFE) X_UNLOCK (reslock); |
|
|
746 | } |
|
|
747 | |
|
|
748 | static void ecb_cold |
|
|
749 | etp_set_max_poll_reqs (unsigned int maxreqs) |
|
|
750 | { |
|
|
751 | if (WORDACCESS_UNSAFE) X_LOCK (reslock); |
|
|
752 | max_poll_reqs = maxreqs; |
|
|
753 | if (WORDACCESS_UNSAFE) X_UNLOCK (reslock); |
|
|
754 | } |
|
|
755 | |
|
|
756 | static void ecb_cold |
|
|
757 | etp_set_max_idle (unsigned int nthreads) |
|
|
758 | { |
|
|
759 | if (WORDACCESS_UNSAFE) X_LOCK (reqlock); |
|
|
760 | max_idle = nthreads; |
|
|
761 | if (WORDACCESS_UNSAFE) X_UNLOCK (reqlock); |
|
|
762 | } |
|
|
763 | |
|
|
764 | static void ecb_cold |
|
|
765 | etp_set_idle_timeout (unsigned int seconds) |
|
|
766 | { |
|
|
767 | if (WORDACCESS_UNSAFE) X_LOCK (reqlock); |
|
|
768 | idle_timeout = seconds; |
|
|
769 | if (WORDACCESS_UNSAFE) X_UNLOCK (reqlock); |
|
|
770 | } |
|
|
771 | |
|
|
772 | static void ecb_cold |
|
|
773 | etp_set_min_parallel (unsigned int nthreads) |
|
|
774 | { |
|
|
775 | if (wanted < nthreads) |
|
|
776 | wanted = nthreads; |
|
|
777 | } |
|
|
778 | |
|
|
779 | static void ecb_cold |
|
|
780 | etp_set_max_parallel (unsigned int nthreads) |
|
|
781 | { |
|
|
782 | if (wanted > nthreads) |
|
|
783 | wanted = nthreads; |
|
|
784 | |
|
|
785 | while (started > wanted) |
|
|
786 | etp_end_thread (); |
|
|
787 | } |
|
|
788 | |
390 | |
789 | /*****************************************************************************/ |
391 | /*****************************************************************************/ |
790 | |
392 | |
791 | static void |
393 | static void |
792 | grp_try_feed (eio_req *grp) |
394 | grp_try_feed (eio_req *grp) |
793 | { |
395 | { |
794 | while (grp->size < grp->int2 && !EIO_CANCELLED (grp)) |
396 | while (grp->size < grp->int2 && !EIO_CANCELLED (grp)) |
795 | { |
397 | { |
796 | grp->flags &= ~EIO_FLAG_GROUPADD; |
398 | grp->flags &= ~ETP_FLAG_GROUPADD; |
797 | |
399 | |
798 | EIO_FEED (grp); |
400 | EIO_FEED (grp); |
799 | |
401 | |
800 | /* stop if no progress has been made */ |
402 | /* stop if no progress has been made */ |
801 | if (!(grp->flags & EIO_FLAG_GROUPADD)) |
403 | if (!(grp->flags & ETP_FLAG_GROUPADD)) |
802 | { |
404 | { |
803 | grp->feed = 0; |
405 | grp->feed = 0; |
804 | break; |
406 | break; |
805 | } |
407 | } |
806 | } |
408 | } |
… | |
… | |
813 | |
415 | |
814 | /* call feeder, if applicable */ |
416 | /* call feeder, if applicable */ |
815 | grp_try_feed (grp); |
417 | grp_try_feed (grp); |
816 | |
418 | |
817 | /* finish, if done */ |
419 | /* finish, if done */ |
818 | if (!grp->size && grp->int1) |
420 | if (!grp->size && grp->flags & ETP_FLAG_DELAYED) |
819 | return eio_finish (grp); |
421 | return eio_finish (grp); |
820 | else |
422 | else |
821 | return 0; |
423 | return 0; |
822 | } |
424 | } |
823 | |
425 | |
… | |
… | |
859 | } |
461 | } |
860 | |
462 | |
861 | void |
463 | void |
862 | eio_grp_cancel (eio_req *grp) |
464 | eio_grp_cancel (eio_req *grp) |
863 | { |
465 | { |
864 | for (grp = grp->grp_first; grp; grp = grp->grp_next) |
466 | etp_grp_cancel (EIO_POOL, grp); |
865 | eio_cancel (grp); |
|
|
866 | } |
467 | } |
867 | |
468 | |
868 | void |
469 | void |
869 | eio_cancel (eio_req *req) |
470 | eio_cancel (eio_req *req) |
870 | { |
471 | { |
871 | etp_cancel (req); |
472 | etp_cancel (EIO_POOL, req); |
872 | } |
473 | } |
873 | |
474 | |
874 | void |
475 | void |
875 | eio_submit (eio_req *req) |
476 | eio_submit (eio_req *req) |
876 | { |
477 | { |
877 | etp_submit (req); |
478 | etp_submit (EIO_POOL, req); |
878 | } |
479 | } |
879 | |
480 | |
880 | unsigned int |
481 | unsigned int |
881 | eio_nreqs (void) |
482 | eio_nreqs (void) |
882 | { |
483 | { |
883 | return etp_nreqs (); |
484 | return etp_nreqs (EIO_POOL); |
884 | } |
485 | } |
885 | |
486 | |
886 | unsigned int |
487 | unsigned int |
887 | eio_nready (void) |
488 | eio_nready (void) |
888 | { |
489 | { |
889 | return etp_nready (); |
490 | return etp_nready (EIO_POOL); |
890 | } |
491 | } |
891 | |
492 | |
892 | unsigned int |
493 | unsigned int |
893 | eio_npending (void) |
494 | eio_npending (void) |
894 | { |
495 | { |
895 | return etp_npending (); |
496 | return etp_npending (EIO_POOL); |
896 | } |
497 | } |
897 | |
498 | |
898 | unsigned int ecb_cold |
499 | unsigned int ecb_cold |
899 | eio_nthreads (void) |
500 | eio_nthreads (void) |
900 | { |
501 | { |
901 | return etp_nthreads (); |
502 | return etp_nthreads (EIO_POOL); |
902 | } |
503 | } |
903 | |
504 | |
904 | void ecb_cold |
505 | void ecb_cold |
905 | eio_set_max_poll_time (double nseconds) |
506 | eio_set_max_poll_time (double nseconds) |
906 | { |
507 | { |
907 | etp_set_max_poll_time (nseconds); |
508 | etp_set_max_poll_time (EIO_POOL, nseconds); |
908 | } |
509 | } |
909 | |
510 | |
910 | void ecb_cold |
511 | void ecb_cold |
911 | eio_set_max_poll_reqs (unsigned int maxreqs) |
512 | eio_set_max_poll_reqs (unsigned int maxreqs) |
912 | { |
513 | { |
913 | etp_set_max_poll_reqs (maxreqs); |
514 | etp_set_max_poll_reqs (EIO_POOL, maxreqs); |
914 | } |
515 | } |
915 | |
516 | |
916 | void ecb_cold |
517 | void ecb_cold |
917 | eio_set_max_idle (unsigned int nthreads) |
518 | eio_set_max_idle (unsigned int nthreads) |
918 | { |
519 | { |
919 | etp_set_max_idle (nthreads); |
520 | etp_set_max_idle (EIO_POOL, nthreads); |
920 | } |
521 | } |
921 | |
522 | |
922 | void ecb_cold |
523 | void ecb_cold |
923 | eio_set_idle_timeout (unsigned int seconds) |
524 | eio_set_idle_timeout (unsigned int seconds) |
924 | { |
525 | { |
925 | etp_set_idle_timeout (seconds); |
526 | etp_set_idle_timeout (EIO_POOL, seconds); |
926 | } |
527 | } |
927 | |
528 | |
928 | void ecb_cold |
529 | void ecb_cold |
929 | eio_set_min_parallel (unsigned int nthreads) |
530 | eio_set_min_parallel (unsigned int nthreads) |
930 | { |
531 | { |
931 | etp_set_min_parallel (nthreads); |
532 | etp_set_min_parallel (EIO_POOL, nthreads); |
932 | } |
533 | } |
933 | |
534 | |
934 | void ecb_cold |
535 | void ecb_cold |
935 | eio_set_max_parallel (unsigned int nthreads) |
536 | eio_set_max_parallel (unsigned int nthreads) |
936 | { |
537 | { |
937 | etp_set_max_parallel (nthreads); |
538 | etp_set_max_parallel (EIO_POOL, nthreads); |
938 | } |
539 | } |
939 | |
540 | |
940 | int eio_poll (void) |
541 | int eio_poll (void) |
941 | { |
542 | { |
942 | return etp_poll (); |
543 | return etp_poll (EIO_POOL); |
943 | } |
544 | } |
944 | |
545 | |
945 | /*****************************************************************************/ |
546 | /*****************************************************************************/ |
946 | /* work around various missing functions */ |
547 | /* work around various missing functions */ |
947 | |
|
|
948 | #if !HAVE_PREADWRITE |
|
|
949 | # undef pread |
|
|
950 | # undef pwrite |
|
|
951 | # define pread eio__pread |
|
|
952 | # define pwrite eio__pwrite |
|
|
953 | |
|
|
954 | static eio_ssize_t |
|
|
955 | eio__pread (int fd, void *buf, size_t count, off_t offset) |
|
|
956 | { |
|
|
957 | eio_ssize_t res; |
|
|
958 | off_t ooffset; |
|
|
959 | |
|
|
960 | X_LOCK (preadwritelock); |
|
|
961 | ooffset = lseek (fd, 0, SEEK_CUR); |
|
|
962 | lseek (fd, offset, SEEK_SET); |
|
|
963 | res = read (fd, buf, count); |
|
|
964 | lseek (fd, ooffset, SEEK_SET); |
|
|
965 | X_UNLOCK (preadwritelock); |
|
|
966 | |
|
|
967 | return res; |
|
|
968 | } |
|
|
969 | |
|
|
970 | static eio_ssize_t |
|
|
971 | eio__pwrite (int fd, void *buf, size_t count, off_t offset) |
|
|
972 | { |
|
|
973 | eio_ssize_t res; |
|
|
974 | off_t ooffset; |
|
|
975 | |
|
|
976 | X_LOCK (preadwritelock); |
|
|
977 | ooffset = lseek (fd, 0, SEEK_CUR); |
|
|
978 | lseek (fd, offset, SEEK_SET); |
|
|
979 | res = write (fd, buf, count); |
|
|
980 | lseek (fd, ooffset, SEEK_SET); |
|
|
981 | X_UNLOCK (preadwritelock); |
|
|
982 | |
|
|
983 | return res; |
|
|
984 | } |
|
|
985 | #endif |
|
|
986 | |
548 | |
987 | #ifndef HAVE_UTIMES |
549 | #ifndef HAVE_UTIMES |
988 | |
550 | |
989 | # undef utimes |
551 | # undef utimes |
990 | # define utimes(path,times) eio__utimes (path, times) |
552 | # define utimes(path,times) eio__utimes (path, times) |
… | |
… | |
1396 | req->result = req->offs == (off_t)-1 ? -1 : 0; |
958 | req->result = req->offs == (off_t)-1 ? -1 : 0; |
1397 | } |
959 | } |
1398 | |
960 | |
1399 | /* result will always end up in tmpbuf, there is always space for adding a 0-byte */ |
961 | /* result will always end up in tmpbuf, there is always space for adding a 0-byte */ |
1400 | static int |
962 | static int |
1401 | eio__realpath (struct tmpbuf *tmpbuf, eio_wd wd, const char *path) |
963 | eio__realpath (struct etp_tmpbuf *tmpbuf, eio_wd wd, const char *path) |
1402 | { |
964 | { |
|
|
965 | char *res; |
1403 | const char *rel = path; |
966 | const char *rel = path; |
1404 | char *res; |
|
|
1405 | char *tmp1, *tmp2; |
967 | char *tmp1, *tmp2; |
1406 | #if SYMLOOP_MAX > 32 |
968 | #if SYMLOOP_MAX > 32 |
1407 | int symlinks = SYMLOOP_MAX; |
969 | int symlinks = SYMLOOP_MAX; |
1408 | #else |
970 | #else |
1409 | int symlinks = 32; |
971 | int symlinks = 32; |
… | |
… | |
1415 | |
977 | |
1416 | errno = ENOENT; |
978 | errno = ENOENT; |
1417 | if (!*rel) |
979 | if (!*rel) |
1418 | return -1; |
980 | return -1; |
1419 | |
981 | |
1420 | res = tmpbuf_get (tmpbuf, PATH_MAX * 3); |
982 | res = etp_tmpbuf_get (tmpbuf, PATH_MAX * 3); |
|
|
983 | #ifdef _WIN32 |
|
|
984 | if (_access (rel, 4) != 0) |
|
|
985 | return -1; |
|
|
986 | |
|
|
987 | symlinks = GetFullPathName (rel, PATH_MAX * 3, res, 0); |
|
|
988 | |
|
|
989 | errno = ENAMETOOLONG; |
|
|
990 | if (symlinks >= PATH_MAX * 3) |
|
|
991 | return -1; |
|
|
992 | |
|
|
993 | errno = EIO; |
|
|
994 | if (symlinks <= 0) |
|
|
995 | return -1; |
|
|
996 | |
|
|
997 | return symlinks; |
|
|
998 | |
|
|
999 | #else |
1421 | tmp1 = res + PATH_MAX; |
1000 | tmp1 = res + PATH_MAX; |
1422 | tmp2 = tmp1 + PATH_MAX; |
1001 | tmp2 = tmp1 + PATH_MAX; |
1423 | |
1002 | |
1424 | #if 0 /* disabled, the musl way to do things is just too racy */ |
1003 | #if 0 /* disabled, the musl way to do things is just too racy */ |
1425 | #if __linux && defined(O_NONBLOCK) && defined(O_NOATIME) |
1004 | #if __linux && defined(O_NONBLOCK) && defined(O_NOATIME) |
… | |
… | |
1429 | |
1008 | |
1430 | if (fd >= 0) |
1009 | if (fd >= 0) |
1431 | { |
1010 | { |
1432 | sprintf (tmp1, "/proc/self/fd/%d", fd); |
1011 | sprintf (tmp1, "/proc/self/fd/%d", fd); |
1433 | req->result = readlink (tmp1, res, PATH_MAX); |
1012 | req->result = readlink (tmp1, res, PATH_MAX); |
|
|
1013 | /* here we should probably stat the open file and the disk file, to make sure they still match */ |
1434 | close (fd); |
1014 | close (fd); |
1435 | |
|
|
1436 | /* here we should probably stat the open file and the disk file, to make sure they still match */ |
|
|
1437 | |
1015 | |
1438 | if (req->result > 0) |
1016 | if (req->result > 0) |
1439 | goto done; |
1017 | goto done; |
1440 | } |
1018 | } |
1441 | else if (errno == ELOOP || errno == ENAMETOOLONG || errno == ENOENT || errno == ENOTDIR || errno == EIO) |
1019 | else if (errno == ELOOP || errno == ENAMETOOLONG || errno == ENOENT || errno == ENOTDIR || errno == EIO) |
1442 | return; |
1020 | return -1; |
1443 | } |
1021 | } |
1444 | #endif |
1022 | #endif |
1445 | #endif |
1023 | #endif |
1446 | |
1024 | |
1447 | if (*rel != '/') |
1025 | if (*rel != '/') |
… | |
… | |
1549 | /* special case for the lone root path */ |
1127 | /* special case for the lone root path */ |
1550 | if (res == tmpbuf->ptr) |
1128 | if (res == tmpbuf->ptr) |
1551 | *res++ = '/'; |
1129 | *res++ = '/'; |
1552 | |
1130 | |
1553 | return res - (char *)tmpbuf->ptr; |
1131 | return res - (char *)tmpbuf->ptr; |
|
|
1132 | #endif |
1554 | } |
1133 | } |
1555 | |
1134 | |
1556 | static signed char |
1135 | static signed char |
1557 | eio_dent_cmp (const eio_dirent *a, const eio_dirent *b) |
1136 | eio_dent_cmp (const eio_dirent *a, const eio_dirent *b) |
1558 | { |
1137 | { |
… | |
… | |
1931 | #ifdef DT_FIFO |
1510 | #ifdef DT_FIFO |
1932 | case DT_FIFO: ent->type = EIO_DT_FIFO; break; |
1511 | case DT_FIFO: ent->type = EIO_DT_FIFO; break; |
1933 | #endif |
1512 | #endif |
1934 | #ifdef DT_CHR |
1513 | #ifdef DT_CHR |
1935 | case DT_CHR: ent->type = EIO_DT_CHR; break; |
1514 | case DT_CHR: ent->type = EIO_DT_CHR; break; |
1936 | #endif |
1515 | #endif |
1937 | #ifdef DT_MPC |
1516 | #ifdef DT_MPC |
1938 | case DT_MPC: ent->type = EIO_DT_MPC; break; |
1517 | case DT_MPC: ent->type = EIO_DT_MPC; break; |
1939 | #endif |
1518 | #endif |
1940 | #ifdef DT_DIR |
1519 | #ifdef DT_DIR |
1941 | case DT_DIR: ent->type = EIO_DT_DIR; break; |
1520 | case DT_DIR: ent->type = EIO_DT_DIR; break; |
1942 | #endif |
1521 | #endif |
1943 | #ifdef DT_NAM |
1522 | #ifdef DT_NAM |
1944 | case DT_NAM: ent->type = EIO_DT_NAM; break; |
1523 | case DT_NAM: ent->type = EIO_DT_NAM; break; |
1945 | #endif |
1524 | #endif |
1946 | #ifdef DT_BLK |
1525 | #ifdef DT_BLK |
1947 | case DT_BLK: ent->type = EIO_DT_BLK; break; |
1526 | case DT_BLK: ent->type = EIO_DT_BLK; break; |
1948 | #endif |
1527 | #endif |
1949 | #ifdef DT_MPB |
1528 | #ifdef DT_MPB |
1950 | case DT_MPB: ent->type = EIO_DT_MPB; break; |
1529 | case DT_MPB: ent->type = EIO_DT_MPB; break; |
1951 | #endif |
1530 | #endif |
1952 | #ifdef DT_REG |
1531 | #ifdef DT_REG |
1953 | case DT_REG: ent->type = EIO_DT_REG; break; |
1532 | case DT_REG: ent->type = EIO_DT_REG; break; |
1954 | #endif |
1533 | #endif |
1955 | #ifdef DT_NWK |
1534 | #ifdef DT_NWK |
1956 | case DT_NWK: ent->type = EIO_DT_NWK; break; |
1535 | case DT_NWK: ent->type = EIO_DT_NWK; break; |
1957 | #endif |
1536 | #endif |
1958 | #ifdef DT_CMP |
1537 | #ifdef DT_CMP |
1959 | case DT_CMP: ent->type = EIO_DT_CMP; break; |
1538 | case DT_CMP: ent->type = EIO_DT_CMP; break; |
1960 | #endif |
1539 | #endif |
1961 | #ifdef DT_LNK |
1540 | #ifdef DT_LNK |
1962 | case DT_LNK: ent->type = EIO_DT_LNK; break; |
1541 | case DT_LNK: ent->type = EIO_DT_LNK; break; |
1963 | #endif |
1542 | #endif |
1964 | #ifdef DT_SOCK |
1543 | #ifdef DT_SOCK |
1965 | case DT_SOCK: ent->type = EIO_DT_SOCK; break; |
1544 | case DT_SOCK: ent->type = EIO_DT_SOCK; break; |
… | |
… | |
2017 | #if !HAVE_AT |
1596 | #if !HAVE_AT |
2018 | |
1597 | |
2019 | /* a bit like realpath, but usually faster because it doesn'T have to return */ |
1598 | /* a bit like realpath, but usually faster because it doesn'T have to return */ |
2020 | /* an absolute or canonical path */ |
1599 | /* an absolute or canonical path */ |
2021 | static const char * |
1600 | static const char * |
2022 | wd_expand (struct tmpbuf *tmpbuf, eio_wd wd, const char *path) |
1601 | wd_expand (struct etp_tmpbuf *tmpbuf, eio_wd wd, const char *path) |
2023 | { |
1602 | { |
2024 | if (!wd || *path == '/') |
1603 | if (!wd || *path == '/') |
2025 | return path; |
1604 | return path; |
2026 | |
1605 | |
2027 | if (path [0] == '.' && !path [1]) |
1606 | if (path [0] == '.' && !path [1]) |
… | |
… | |
2029 | |
1608 | |
2030 | { |
1609 | { |
2031 | int l1 = wd->len; |
1610 | int l1 = wd->len; |
2032 | int l2 = strlen (path); |
1611 | int l2 = strlen (path); |
2033 | |
1612 | |
2034 | char *res = tmpbuf_get (tmpbuf, l1 + l2 + 2); |
1613 | char *res = etp_tmpbuf_get (tmpbuf, l1 + l2 + 2); |
2035 | |
1614 | |
2036 | memcpy (res, wd->str, l1); |
1615 | memcpy (res, wd->str, l1); |
2037 | res [l1] = '/'; |
1616 | res [l1] = '/'; |
2038 | memcpy (res + l1 + 1, path, l2 + 1); |
1617 | memcpy (res + l1 + 1, path, l2 + 1); |
2039 | |
1618 | |
… | |
… | |
2042 | } |
1621 | } |
2043 | |
1622 | |
2044 | #endif |
1623 | #endif |
2045 | |
1624 | |
2046 | static eio_wd |
1625 | static eio_wd |
2047 | eio__wd_open_sync (struct tmpbuf *tmpbuf, eio_wd wd, const char *path) |
1626 | eio__wd_open_sync (struct etp_tmpbuf *tmpbuf, eio_wd wd, const char *path) |
2048 | { |
1627 | { |
2049 | int fd; |
1628 | int fd; |
2050 | eio_wd res; |
1629 | eio_wd res; |
2051 | int len = eio__realpath (tmpbuf, wd, path); |
1630 | int len = eio__realpath (tmpbuf, wd, path); |
2052 | |
1631 | |
… | |
… | |
2074 | } |
1653 | } |
2075 | |
1654 | |
2076 | eio_wd |
1655 | eio_wd |
2077 | eio_wd_open_sync (eio_wd wd, const char *path) |
1656 | eio_wd_open_sync (eio_wd wd, const char *path) |
2078 | { |
1657 | { |
2079 | struct tmpbuf tmpbuf = { 0 }; |
1658 | struct etp_tmpbuf tmpbuf = { }; |
2080 | wd = eio__wd_open_sync (&tmpbuf, wd, path); |
1659 | wd = eio__wd_open_sync (&tmpbuf, wd, path); |
2081 | free (tmpbuf.ptr); |
1660 | free (tmpbuf.ptr); |
2082 | |
1661 | |
2083 | return wd; |
1662 | return wd; |
2084 | } |
1663 | } |
… | |
… | |
2133 | /*****************************************************************************/ |
1712 | /*****************************************************************************/ |
2134 | |
1713 | |
2135 | #define ALLOC(len) \ |
1714 | #define ALLOC(len) \ |
2136 | if (!req->ptr2) \ |
1715 | if (!req->ptr2) \ |
2137 | { \ |
1716 | { \ |
2138 | X_LOCK (wrklock); \ |
1717 | X_LOCK (EIO_POOL->wrklock); \ |
2139 | req->flags |= EIO_FLAG_PTR2_FREE; \ |
1718 | req->flags |= EIO_FLAG_PTR2_FREE; \ |
2140 | X_UNLOCK (wrklock); \ |
1719 | X_UNLOCK (EIO_POOL->wrklock); \ |
2141 | req->ptr2 = malloc (len); \ |
1720 | req->ptr2 = malloc (len); \ |
2142 | if (!req->ptr2) \ |
1721 | if (!req->ptr2) \ |
2143 | { \ |
1722 | { \ |
2144 | errno = ENOMEM; \ |
1723 | errno = ENOMEM; \ |
2145 | req->result = -1; \ |
1724 | req->result = -1; \ |
2146 | break; \ |
1725 | break; \ |
2147 | } \ |
1726 | } \ |
2148 | } |
1727 | } |
2149 | |
1728 | |
2150 | static void ecb_noinline ecb_cold |
|
|
2151 | etp_proc_init (void) |
|
|
2152 | { |
|
|
2153 | #if HAVE_PRCTL_SET_NAME |
|
|
2154 | /* provide a more sensible "thread name" */ |
|
|
2155 | char name[16 + 1]; |
|
|
2156 | const int namelen = sizeof (name) - 1; |
|
|
2157 | int len; |
|
|
2158 | |
|
|
2159 | prctl (PR_GET_NAME, (unsigned long)name, 0, 0, 0); |
|
|
2160 | name [namelen] = 0; |
|
|
2161 | len = strlen (name); |
|
|
2162 | strcpy (name + (len <= namelen - 4 ? len : namelen - 4), "/eio"); |
|
|
2163 | prctl (PR_SET_NAME, (unsigned long)name, 0, 0, 0); |
|
|
2164 | #endif |
|
|
2165 | } |
|
|
2166 | |
|
|
2167 | X_THREAD_PROC (etp_proc) |
|
|
2168 | { |
|
|
2169 | ETP_REQ *req; |
|
|
2170 | struct timespec ts; |
|
|
2171 | etp_worker *self = (etp_worker *)thr_arg; |
|
|
2172 | |
|
|
2173 | etp_proc_init (); |
|
|
2174 | |
|
|
2175 | /* try to distribute timeouts somewhat evenly */ |
|
|
2176 | ts.tv_nsec = ((unsigned long)self & 1023UL) * (1000000000UL / 1024UL); |
|
|
2177 | |
|
|
2178 | for (;;) |
|
|
2179 | { |
|
|
2180 | ts.tv_sec = 0; |
|
|
2181 | |
|
|
2182 | X_LOCK (reqlock); |
|
|
2183 | |
|
|
2184 | for (;;) |
|
|
2185 | { |
|
|
2186 | req = reqq_shift (&req_queue); |
|
|
2187 | |
|
|
2188 | if (req) |
|
|
2189 | break; |
|
|
2190 | |
|
|
2191 | if (ts.tv_sec == 1) /* no request, but timeout detected, let's quit */ |
|
|
2192 | { |
|
|
2193 | X_UNLOCK (reqlock); |
|
|
2194 | X_LOCK (wrklock); |
|
|
2195 | --started; |
|
|
2196 | X_UNLOCK (wrklock); |
|
|
2197 | goto quit; |
|
|
2198 | } |
|
|
2199 | |
|
|
2200 | ++idle; |
|
|
2201 | |
|
|
2202 | if (idle <= max_idle) |
|
|
2203 | /* we are allowed to idle, so do so without any timeout */ |
|
|
2204 | X_COND_WAIT (reqwait, reqlock); |
|
|
2205 | else |
|
|
2206 | { |
|
|
2207 | /* initialise timeout once */ |
|
|
2208 | if (!ts.tv_sec) |
|
|
2209 | ts.tv_sec = time (0) + idle_timeout; |
|
|
2210 | |
|
|
2211 | if (X_COND_TIMEDWAIT (reqwait, reqlock, ts) == ETIMEDOUT) |
|
|
2212 | ts.tv_sec = 1; /* assuming this is not a value computed above.,.. */ |
|
|
2213 | } |
|
|
2214 | |
|
|
2215 | --idle; |
|
|
2216 | } |
|
|
2217 | |
|
|
2218 | --nready; |
|
|
2219 | |
|
|
2220 | X_UNLOCK (reqlock); |
|
|
2221 | |
|
|
2222 | if (req->type < 0) |
|
|
2223 | goto quit; |
|
|
2224 | |
|
|
2225 | ETP_EXECUTE (self, req); |
|
|
2226 | |
|
|
2227 | X_LOCK (reslock); |
|
|
2228 | |
|
|
2229 | ++npending; |
|
|
2230 | |
|
|
2231 | if (!reqq_push (&res_queue, req) && want_poll_cb) |
|
|
2232 | want_poll_cb (); |
|
|
2233 | |
|
|
2234 | etp_worker_clear (self); |
|
|
2235 | |
|
|
2236 | X_UNLOCK (reslock); |
|
|
2237 | } |
|
|
2238 | |
|
|
2239 | quit: |
|
|
2240 | free (req); |
|
|
2241 | |
|
|
2242 | X_LOCK (wrklock); |
|
|
2243 | etp_worker_free (self); |
|
|
2244 | X_UNLOCK (wrklock); |
|
|
2245 | |
|
|
2246 | return 0; |
|
|
2247 | } |
|
|
2248 | |
|
|
2249 | /*****************************************************************************/ |
1729 | /*****************************************************************************/ |
2250 | |
1730 | |
2251 | int ecb_cold |
1731 | int ecb_cold |
2252 | eio_init (void (*want_poll)(void), void (*done_poll)(void)) |
1732 | eio_init (void (*want_poll)(void), void (*done_poll)(void)) |
2253 | { |
1733 | { |
2254 | #if !HAVE_PREADWRITE |
1734 | eio_want_poll_cb = want_poll; |
2255 | X_MUTEX_CREATE (preadwritelock); |
1735 | eio_done_poll_cb = done_poll; |
2256 | #endif |
|
|
2257 | |
1736 | |
2258 | return etp_init (want_poll, done_poll); |
1737 | return etp_init (EIO_POOL, 0, 0, 0); |
2259 | } |
1738 | } |
2260 | |
1739 | |
2261 | ecb_inline void |
1740 | ecb_inline void |
2262 | eio_api_destroy (eio_req *req) |
1741 | eio_api_destroy (eio_req *req) |
2263 | { |
1742 | { |
2264 | free (req); |
1743 | free (req); |
2265 | } |
1744 | } |
2266 | |
1745 | |
2267 | #define REQ(rtype) \ |
1746 | #define REQ(rtype) \ |
2268 | eio_req *req; \ |
1747 | eio_req *req; \ |
2269 | \ |
1748 | \ |
2270 | req = (eio_req *)calloc (1, sizeof *req); \ |
1749 | req = (eio_req *)calloc (1, sizeof *req); \ |
2271 | if (!req) \ |
1750 | if (!req) \ |
2272 | return 0; \ |
1751 | return 0; \ |
… | |
… | |
2286 | { \ |
1765 | { \ |
2287 | eio_api_destroy (req); \ |
1766 | eio_api_destroy (req); \ |
2288 | return 0; \ |
1767 | return 0; \ |
2289 | } |
1768 | } |
2290 | |
1769 | |
|
|
1770 | #define SINGLEDOT(ptr) (0[(char *)(ptr)] == '.' && !1[(char *)(ptr)]) |
|
|
1771 | |
2291 | static void |
1772 | static void |
2292 | eio_execute (etp_worker *self, eio_req *req) |
1773 | eio_execute (etp_worker *self, eio_req *req) |
2293 | { |
1774 | { |
2294 | #if HAVE_AT |
1775 | #if HAVE_AT |
2295 | int dirfd; |
1776 | int dirfd; |
… | |
… | |
2350 | case EIO_CHMOD: req->result = fchmodat (dirfd, req->ptr1, (mode_t)req->int2, 0); break; |
1831 | case EIO_CHMOD: req->result = fchmodat (dirfd, req->ptr1, (mode_t)req->int2, 0); break; |
2351 | case EIO_TRUNCATE: req->result = eio__truncateat (dirfd, req->ptr1, req->offs); break; |
1832 | case EIO_TRUNCATE: req->result = eio__truncateat (dirfd, req->ptr1, req->offs); break; |
2352 | case EIO_OPEN: req->result = openat (dirfd, req->ptr1, req->int1, (mode_t)req->int2); break; |
1833 | case EIO_OPEN: req->result = openat (dirfd, req->ptr1, req->int1, (mode_t)req->int2); break; |
2353 | |
1834 | |
2354 | case EIO_UNLINK: req->result = unlinkat (dirfd, req->ptr1, 0); break; |
1835 | case EIO_UNLINK: req->result = unlinkat (dirfd, req->ptr1, 0); break; |
2355 | case EIO_RMDIR: req->result = unlinkat (dirfd, req->ptr1, AT_REMOVEDIR); break; |
1836 | case EIO_RMDIR: /* complications arise because "." cannot be removed, so we might have to expand */ |
|
|
1837 | req->result = req->wd && SINGLEDOT (req->ptr1) |
|
|
1838 | ? rmdir (req->wd->str) |
|
|
1839 | : unlinkat (dirfd, req->ptr1, AT_REMOVEDIR); break; |
2356 | case EIO_MKDIR: req->result = mkdirat (dirfd, req->ptr1, (mode_t)req->int2); break; |
1840 | case EIO_MKDIR: req->result = mkdirat (dirfd, req->ptr1, (mode_t)req->int2); break; |
2357 | case EIO_RENAME: req->result = renameat (dirfd, req->ptr1, WD2FD ((eio_wd)req->int3), req->ptr2); break; |
1841 | case EIO_RENAME: /* complications arise because "." cannot be renamed, so we might have to expand */ |
|
|
1842 | req->result = req->wd && SINGLEDOT (req->ptr1) |
|
|
1843 | ? rename (req->wd->str, req->ptr2) |
|
|
1844 | : renameat (dirfd, req->ptr1, WD2FD ((eio_wd)req->int3), req->ptr2); break; |
2358 | case EIO_LINK: req->result = linkat (dirfd, req->ptr1, WD2FD ((eio_wd)req->int3), req->ptr2, 0); break; |
1845 | case EIO_LINK: req->result = linkat (dirfd, req->ptr1, WD2FD ((eio_wd)req->int3), req->ptr2, 0); break; |
2359 | case EIO_SYMLINK: req->result = symlinkat (req->ptr1, dirfd, req->ptr2); break; |
1846 | case EIO_SYMLINK: req->result = symlinkat (req->ptr1, dirfd, req->ptr2); break; |
2360 | case EIO_MKNOD: req->result = mknodat (dirfd, req->ptr1, (mode_t)req->int2, (dev_t)req->offs); break; |
1847 | case EIO_MKNOD: req->result = mknodat (dirfd, req->ptr1, (mode_t)req->int2, (dev_t)req->offs); break; |
2361 | case EIO_READLINK: ALLOC (PATH_MAX); |
1848 | case EIO_READLINK: ALLOC (PATH_MAX); |
2362 | req->result = readlinkat (dirfd, req->ptr1, req->ptr2, PATH_MAX); break; |
1849 | req->result = readlinkat (dirfd, req->ptr1, req->ptr2, PATH_MAX); break; |
… | |
… | |
2480 | req->result = select (0, 0, 0, 0, &tv); |
1967 | req->result = select (0, 0, 0, 0, &tv); |
2481 | } |
1968 | } |
2482 | #endif |
1969 | #endif |
2483 | break; |
1970 | break; |
2484 | |
1971 | |
|
|
1972 | #if 0 |
2485 | case EIO_GROUP: |
1973 | case EIO_GROUP: |
2486 | abort (); /* handled in eio_request */ |
1974 | abort (); /* handled in eio_request */ |
|
|
1975 | #endif |
2487 | |
1976 | |
2488 | case EIO_NOP: |
1977 | case EIO_NOP: |
2489 | req->result = 0; |
1978 | req->result = 0; |
2490 | break; |
1979 | break; |
2491 | |
1980 | |
… | |
… | |
2791 | void |
2280 | void |
2792 | eio_grp_add (eio_req *grp, eio_req *req) |
2281 | eio_grp_add (eio_req *grp, eio_req *req) |
2793 | { |
2282 | { |
2794 | assert (("cannot add requests to IO::AIO::GRP after the group finished", grp->int1 != 2)); |
2283 | assert (("cannot add requests to IO::AIO::GRP after the group finished", grp->int1 != 2)); |
2795 | |
2284 | |
2796 | grp->flags |= EIO_FLAG_GROUPADD; |
2285 | grp->flags |= ETP_FLAG_GROUPADD; |
2797 | |
2286 | |
2798 | ++grp->size; |
2287 | ++grp->size; |
2799 | req->grp = grp; |
2288 | req->grp = grp; |
2800 | |
2289 | |
2801 | req->grp_prev = 0; |
2290 | req->grp_prev = 0; |