1 | /* |
1 | /* |
2 | * libev kqueue backend |
2 | * libev kqueue backend |
3 | * |
3 | * |
4 | * Copyright (c) 2007,2008,2009 Marc Alexander Lehmann <libev@schmorp.de> |
4 | * Copyright (c) 2007,2008,2009,2010,2011,2012,2013,2016,2019 Marc Alexander Lehmann <libev@schmorp.de> |
5 | * All rights reserved. |
5 | * All rights reserved. |
6 | * |
6 | * |
7 | * Redistribution and use in source and binary forms, with or without modifica- |
7 | * Redistribution and use in source and binary forms, with or without modifica- |
8 | * tion, are permitted provided that the following conditions are met: |
8 | * tion, are permitted provided that the following conditions are met: |
9 | * |
9 | * |
10 | * 1. Redistributions of source code must retain the above copyright notice, |
10 | * 1. Redistributions of source code must retain the above copyright notice, |
11 | * this list of conditions and the following disclaimer. |
11 | * this list of conditions and the following disclaimer. |
12 | * |
12 | * |
13 | * 2. Redistributions in binary form must reproduce the above copyright |
13 | * 2. Redistributions in binary form must reproduce the above copyright |
14 | * notice, this list of conditions and the following disclaimer in the |
14 | * notice, this list of conditions and the following disclaimer in the |
15 | * documentation and/or other materials provided with the distribution. |
15 | * documentation and/or other materials provided with the distribution. |
16 | * |
16 | * |
17 | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED |
17 | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED |
18 | * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MER- |
18 | * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MER- |
19 | * CHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO |
19 | * CHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO |
20 | * EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPE- |
20 | * EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPE- |
21 | * CIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
21 | * CIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
… | |
… | |
41 | #include <sys/time.h> |
41 | #include <sys/time.h> |
42 | #include <sys/event.h> |
42 | #include <sys/event.h> |
43 | #include <string.h> |
43 | #include <string.h> |
44 | #include <errno.h> |
44 | #include <errno.h> |
45 | |
45 | |
46 | void inline_speed |
46 | inline_speed |
|
|
47 | void |
47 | kqueue_change (EV_P_ int fd, int filter, int flags, int fflags) |
48 | kqueue_change (EV_P_ int fd, int filter, int flags, int fflags) |
48 | { |
49 | { |
49 | ++kqueue_changecnt; |
50 | ++kqueue_changecnt; |
50 | array_needsize (struct kevent, kqueue_changes, kqueue_changemax, kqueue_changecnt, EMPTY2); |
51 | array_needsize (struct kevent, kqueue_changes, kqueue_changemax, kqueue_changecnt, array_needsize_noinit); |
51 | |
52 | |
52 | EV_SET (&kqueue_changes [kqueue_changecnt - 1], fd, filter, flags, fflags, 0, 0); |
53 | EV_SET (&kqueue_changes [kqueue_changecnt - 1], fd, filter, flags, fflags, 0, 0); |
53 | } |
54 | } |
54 | |
55 | |
|
|
56 | /* OS X at least needs this */ |
|
|
57 | #ifndef EV_ENABLE |
|
|
58 | # define EV_ENABLE 0 |
|
|
59 | #endif |
55 | #ifndef NOTE_EOF |
60 | #ifndef NOTE_EOF |
56 | # define NOTE_EOF 0 |
61 | # define NOTE_EOF 0 |
57 | #endif |
62 | #endif |
58 | |
63 | |
59 | static void |
64 | static void |
… | |
… | |
70 | |
75 | |
71 | /* to detect close/reopen reliably, we have to re-add */ |
76 | /* to detect close/reopen reliably, we have to re-add */ |
72 | /* event requests even when oev == nev */ |
77 | /* event requests even when oev == nev */ |
73 | |
78 | |
74 | if (nev & EV_READ) |
79 | if (nev & EV_READ) |
75 | kqueue_change (EV_A_ fd, EVFILT_READ , EV_ADD, NOTE_EOF); |
80 | kqueue_change (EV_A_ fd, EVFILT_READ , EV_ADD | EV_ENABLE, NOTE_EOF); |
76 | |
81 | |
77 | if (nev & EV_WRITE) |
82 | if (nev & EV_WRITE) |
78 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_ADD, NOTE_EOF); |
83 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_ADD | EV_ENABLE, NOTE_EOF); |
79 | } |
84 | } |
80 | |
85 | |
81 | static void |
86 | static void |
82 | kqueue_poll (EV_P_ ev_tstamp timeout) |
87 | kqueue_poll (EV_P_ ev_tstamp timeout) |
83 | { |
88 | { |
… | |
… | |
91 | kqueue_eventmax = array_nextsize (sizeof (struct kevent), kqueue_eventmax, kqueue_changecnt); |
96 | kqueue_eventmax = array_nextsize (sizeof (struct kevent), kqueue_eventmax, kqueue_changecnt); |
92 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
97 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
93 | } |
98 | } |
94 | |
99 | |
95 | EV_RELEASE_CB; |
100 | EV_RELEASE_CB; |
96 | ts.tv_sec = (time_t)timeout; |
101 | EV_TS_SET (ts, timeout); |
97 | ts.tv_nsec = (long)((timeout - (ev_tstamp)ts.tv_sec) * 1e9); |
|
|
98 | res = kevent (backend_fd, kqueue_changes, kqueue_changecnt, kqueue_events, kqueue_eventmax, &ts); |
102 | res = kevent (backend_fd, kqueue_changes, kqueue_changecnt, kqueue_events, kqueue_eventmax, &ts); |
99 | EV_ACQUIRE_CB; |
103 | EV_ACQUIRE_CB; |
100 | kqueue_changecnt = 0; |
104 | kqueue_changecnt = 0; |
101 | |
105 | |
102 | if (expect_false (res < 0)) |
106 | if (expect_false (res < 0)) |
103 | { |
107 | { |
104 | if (errno != EINTR) |
108 | if (errno != EINTR) |
105 | ev_syserr ("(libev) kevent"); |
109 | ev_syserr ("(libev) kqueue kevent"); |
106 | |
110 | |
107 | return; |
111 | return; |
108 | } |
112 | } |
109 | |
113 | |
110 | for (i = 0; i < res; ++i) |
114 | for (i = 0; i < res; ++i) |
111 | { |
115 | { |
112 | int fd = kqueue_events [i].ident; |
116 | int fd = kqueue_events [i].ident; |
113 | |
117 | |
114 | if (expect_false (kqueue_events [i].flags & EV_ERROR)) |
118 | if (expect_false (kqueue_events [i].flags & EV_ERROR)) |
115 | { |
119 | { |
116 | int err = kqueue_events [i].data; |
120 | int err = kqueue_events [i].data; |
117 | |
121 | |
118 | /* we are only interested in errors for fds that we are interested in :) */ |
122 | /* we are only interested in errors for fds that we are interested in :) */ |
119 | if (anfds [fd].events) |
123 | if (anfds [fd].events) |
120 | { |
124 | { |
121 | if (err == ENOENT) /* resubmit changes on ENOENT */ |
125 | if (err == ENOENT) /* resubmit changes on ENOENT */ |
122 | kqueue_modify (EV_A_ fd, 0, anfds [fd].events); |
126 | kqueue_modify (EV_A_ fd, 0, anfds [fd].events); |
123 | else if (err == EBADF) /* on EBADF, we re-check the fd */ |
127 | else if (err == EBADF) /* on EBADF, we re-check the fd */ |
124 | { |
128 | { |
125 | if (fd_valid (fd)) |
129 | if (fd_valid (fd)) |
126 | kqueue_modify (EV_A_ fd, 0, anfds [fd].events); |
130 | kqueue_modify (EV_A_ fd, 0, anfds [fd].events); |
127 | else |
131 | else |
|
|
132 | { |
|
|
133 | assert (("libev: kqueue found invalid fd", 0)); |
128 | fd_kill (EV_A_ fd); |
134 | fd_kill (EV_A_ fd); |
|
|
135 | } |
129 | } |
136 | } |
130 | else /* on all other errors, we error out on the fd */ |
137 | else /* on all other errors, we error out on the fd */ |
|
|
138 | { |
|
|
139 | assert (("libev: kqueue found invalid fd", 0)); |
131 | fd_kill (EV_A_ fd); |
140 | fd_kill (EV_A_ fd); |
132 | } |
141 | } |
|
|
142 | } |
133 | } |
143 | } |
134 | else |
144 | else |
135 | fd_event ( |
145 | fd_event ( |
136 | EV_A_ |
146 | EV_A_ |
137 | fd, |
147 | fd, |
… | |
… | |
147 | kqueue_eventmax = array_nextsize (sizeof (struct kevent), kqueue_eventmax, kqueue_eventmax + 1); |
157 | kqueue_eventmax = array_nextsize (sizeof (struct kevent), kqueue_eventmax, kqueue_eventmax + 1); |
148 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
158 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
149 | } |
159 | } |
150 | } |
160 | } |
151 | |
161 | |
152 | int inline_size |
162 | inline_size |
|
|
163 | int |
153 | kqueue_init (EV_P_ int flags) |
164 | kqueue_init (EV_P_ int flags) |
154 | { |
165 | { |
155 | /* Initalize the kernel queue */ |
166 | /* initialize the kernel queue */ |
|
|
167 | kqueue_fd_pid = getpid (); |
156 | if ((backend_fd = kqueue ()) < 0) |
168 | if ((backend_fd = kqueue ()) < 0) |
157 | return 0; |
169 | return 0; |
158 | |
170 | |
159 | fcntl (backend_fd, F_SETFD, FD_CLOEXEC); /* not sure if necessary, hopefully doesn't hurt */ |
171 | fcntl (backend_fd, F_SETFD, FD_CLOEXEC); /* not sure if necessary, hopefully doesn't hurt */ |
160 | |
172 | |
161 | backend_fudge = 0.; |
173 | backend_mintime = 1e-9; /* apparently, they did the right thing in freebsd */ |
162 | backend_modify = kqueue_modify; |
174 | backend_modify = kqueue_modify; |
163 | backend_poll = kqueue_poll; |
175 | backend_poll = kqueue_poll; |
164 | |
176 | |
165 | kqueue_eventmax = 64; /* initial number of events receivable per poll */ |
177 | kqueue_eventmax = 64; /* initial number of events receivable per poll */ |
166 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
178 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
167 | |
179 | |
168 | kqueue_changes = 0; |
180 | kqueue_changes = 0; |
… | |
… | |
170 | kqueue_changecnt = 0; |
182 | kqueue_changecnt = 0; |
171 | |
183 | |
172 | return EVBACKEND_KQUEUE; |
184 | return EVBACKEND_KQUEUE; |
173 | } |
185 | } |
174 | |
186 | |
175 | void inline_size |
187 | inline_size |
|
|
188 | void |
176 | kqueue_destroy (EV_P) |
189 | kqueue_destroy (EV_P) |
177 | { |
190 | { |
178 | ev_free (kqueue_events); |
191 | ev_free (kqueue_events); |
179 | ev_free (kqueue_changes); |
192 | ev_free (kqueue_changes); |
180 | } |
193 | } |
181 | |
194 | |
182 | void inline_size |
195 | inline_size |
|
|
196 | void |
183 | kqueue_fork (EV_P) |
197 | kqueue_fork (EV_P) |
184 | { |
198 | { |
|
|
199 | /* some BSD kernels don't just destroy the kqueue itself, |
|
|
200 | * but also close the fd, which isn't documented, and |
|
|
201 | * impossible to support properly. |
|
|
202 | * we remember the pid of the kqueue call and only close |
|
|
203 | * the fd if the pid is still the same. |
|
|
204 | * this leaks fds on sane kernels, but BSD interfaces are |
|
|
205 | * notoriously buggy and rarely get fixed. |
|
|
206 | */ |
|
|
207 | pid_t newpid = getpid (); |
|
|
208 | |
|
|
209 | if (newpid == kqueue_fd_pid) |
185 | close (backend_fd); |
210 | close (backend_fd); |
186 | |
211 | |
|
|
212 | kqueue_fd_pid = newpid; |
187 | while ((backend_fd = kqueue ()) < 0) |
213 | while ((backend_fd = kqueue ()) < 0) |
188 | ev_syserr ("(libev) kqueue"); |
214 | ev_syserr ("(libev) kqueue"); |
189 | |
215 | |
190 | fcntl (backend_fd, F_SETFD, FD_CLOEXEC); |
216 | fcntl (backend_fd, F_SETFD, FD_CLOEXEC); |
191 | |
217 | |
192 | /* re-register interest in fds */ |
218 | /* re-register interest in fds */ |
193 | fd_rearm_all (EV_A); |
219 | fd_rearm_all (EV_A); |
194 | } |
220 | } |
195 | |
221 | |
|
|
222 | /* sys/event.h defines EV_ERROR */ |
|
|
223 | #undef EV_ERROR |
|
|
224 | |