… | |
… | |
40 | struct kevent *ke; |
40 | struct kevent *ke; |
41 | |
41 | |
42 | ++kqueue_changecnt; |
42 | ++kqueue_changecnt; |
43 | array_needsize (struct kevent, kqueue_changes, kqueue_changemax, kqueue_changecnt, EMPTY2); |
43 | array_needsize (struct kevent, kqueue_changes, kqueue_changemax, kqueue_changecnt, EMPTY2); |
44 | |
44 | |
45 | ke = &kqueue_changes [kqueue_changecnt - 1]; |
45 | EV_SET (&kqueue_changes [kqueue_changecnt - 1], fd, filter, flags, fflags, 0, 0); |
46 | memset (ke, 0, sizeof (struct kevent)); |
|
|
47 | ke->ident = fd; |
|
|
48 | ke->filter = filter; |
|
|
49 | ke->flags = flags; |
|
|
50 | ke->fflags = fflags; |
|
|
51 | } |
46 | } |
52 | |
47 | |
53 | #ifndef NOTE_EOF |
48 | #ifndef NOTE_EOF |
54 | # define NOTE_EOF 0 |
49 | # define NOTE_EOF 0 |
55 | #endif |
50 | #endif |
56 | |
51 | |
57 | static void |
52 | static void |
58 | kqueue_modify (EV_P_ int fd, int oev, int nev) |
53 | kqueue_modify (EV_P_ int fd, int oev, int nev) |
59 | { |
54 | { |
|
|
55 | if (oev != nev) |
|
|
56 | { |
|
|
57 | if (oev & EV_READ) |
|
|
58 | kqueue_change (EV_A_ fd, EVFILT_READ , EV_DELETE, 0); |
|
|
59 | |
|
|
60 | if (oev & EV_WRITE) |
|
|
61 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_DELETE, 0); |
|
|
62 | } |
|
|
63 | |
60 | /* to detect close/reopen reliably, we have to remove and re-add */ |
64 | /* to detect close/reopen reliably, we have to re-add */ |
61 | /* event requests even when oev == nev */ |
65 | /* event requests even when oev == nev */ |
62 | |
66 | |
63 | if (oev & EV_READ) |
|
|
64 | kqueue_change (EV_A_ fd, EVFILT_READ, EV_DELETE, 0); |
|
|
65 | |
|
|
66 | if (oev & EV_WRITE) |
|
|
67 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_DELETE, 0); |
|
|
68 | |
|
|
69 | if (nev & EV_READ) |
67 | if (nev & EV_READ) |
70 | kqueue_change (EV_A_ fd, EVFILT_READ, EV_ADD, NOTE_EOF); |
68 | kqueue_change (EV_A_ fd, EVFILT_READ , EV_ADD, NOTE_EOF); |
71 | |
69 | |
72 | if (nev & EV_WRITE) |
70 | if (nev & EV_WRITE) |
73 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_ADD, NOTE_EOF); |
71 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_ADD, NOTE_EOF); |
74 | } |
72 | } |
75 | |
73 | |
… | |
… | |
86 | kqueue_eventmax = array_roundsize (struct kevent, kqueue_changecnt); |
84 | kqueue_eventmax = array_roundsize (struct kevent, kqueue_changecnt); |
87 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
85 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
88 | } |
86 | } |
89 | |
87 | |
90 | ts.tv_sec = (time_t)timeout; |
88 | ts.tv_sec = (time_t)timeout; |
91 | ts.tv_nsec = (long)(timeout - (ev_tstamp)ts.tv_sec) * 1e9; |
89 | ts.tv_nsec = (long)((timeout - (ev_tstamp)ts.tv_sec) * 1e9); |
92 | res = kevent (kqueue_fd, kqueue_changes, kqueue_changecnt, kqueue_events, kqueue_eventmax, &ts); |
90 | res = kevent (backend_fd, kqueue_changes, kqueue_changecnt, kqueue_events, kqueue_eventmax, &ts); |
93 | kqueue_changecnt = 0; |
91 | kqueue_changecnt = 0; |
94 | |
92 | |
95 | if (res < 0) |
93 | if (res < 0) |
96 | { |
94 | { |
97 | if (errno != EINTR) |
95 | if (errno != EINTR) |
… | |
… | |
102 | |
100 | |
103 | for (i = 0; i < res; ++i) |
101 | for (i = 0; i < res; ++i) |
104 | { |
102 | { |
105 | int fd = kqueue_events [i].ident; |
103 | int fd = kqueue_events [i].ident; |
106 | |
104 | |
107 | if (kqueue_events [i].flags & EV_ERROR) |
105 | if (expect_false (kqueue_events [i].flags & EV_ERROR)) |
108 | { |
106 | { |
109 | int err = kqueue_events [i].data; |
107 | int err = kqueue_events [i].data; |
110 | |
108 | |
111 | /* |
109 | /* |
112 | * errors that may happen |
110 | * errors that may happen |
… | |
… | |
156 | kqueue_init (EV_P_ int flags) |
154 | kqueue_init (EV_P_ int flags) |
157 | { |
155 | { |
158 | struct kevent ch, ev; |
156 | struct kevent ch, ev; |
159 | |
157 | |
160 | /* Initalize the kernel queue */ |
158 | /* Initalize the kernel queue */ |
161 | if ((kqueue_fd = kqueue ()) < 0) |
159 | if ((backend_fd = kqueue ()) < 0) |
162 | return 0; |
160 | return 0; |
163 | |
161 | |
164 | fcntl (kqueue_fd, F_SETFD, FD_CLOEXEC); /* not sure if necessary, hopefully doesn't hurt */ |
162 | fcntl (backend_fd, F_SETFD, FD_CLOEXEC); /* not sure if necessary, hopefully doesn't hurt */ |
165 | |
163 | |
166 | /* Check for Mac OS X kqueue bug. */ |
164 | /* Check for Mac OS X kqueue bug. */ |
167 | ch.ident = -1; |
165 | ch.ident = -1; |
168 | ch.filter = EVFILT_READ; |
166 | ch.filter = EVFILT_READ; |
169 | ch.flags = EV_ADD; |
167 | ch.flags = EV_ADD; |
… | |
… | |
171 | /* |
169 | /* |
172 | * If kqueue works, then kevent will succeed, and it will |
170 | * If kqueue works, then kevent will succeed, and it will |
173 | * stick an error in ev. If kqueue is broken, then |
171 | * stick an error in ev. If kqueue is broken, then |
174 | * kevent will fail. |
172 | * kevent will fail. |
175 | */ |
173 | */ |
176 | if (kevent (kqueue_fd, &ch, 1, &ev, 1, 0) != 1 |
174 | if (kevent (backend_fd, &ch, 1, &ev, 1, 0) != 1 |
177 | || ev.ident != -1 |
175 | || ev.ident != -1 |
178 | || ev.flags != EV_ERROR) |
176 | || ev.flags != EV_ERROR) |
179 | { |
177 | { |
180 | /* detected broken kqueue */ |
178 | /* detected broken kqueue */ |
181 | close (kqueue_fd); |
179 | close (backend_fd); |
182 | return 0; |
180 | return 0; |
183 | } |
181 | } |
184 | |
182 | |
185 | method_fudge = 1e-3; /* needed to compensate for kevent returning early */ |
183 | backend_fudge = 1e-3; /* needed to compensate for kevent returning early */ |
186 | method_modify = kqueue_modify; |
184 | backend_modify = kqueue_modify; |
187 | method_poll = kqueue_poll; |
185 | backend_poll = kqueue_poll; |
188 | |
186 | |
189 | kqueue_eventmax = 64; /* intiial number of events receivable per poll */ |
187 | kqueue_eventmax = 64; /* intiial number of events receivable per poll */ |
190 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
188 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
191 | |
189 | |
192 | kqueue_changes = 0; |
190 | kqueue_changes = 0; |
193 | kqueue_changemax = 0; |
191 | kqueue_changemax = 0; |
194 | kqueue_changecnt = 0; |
192 | kqueue_changecnt = 0; |
195 | |
193 | |
196 | return EVMETHOD_KQUEUE; |
194 | return EVBACKEND_KQUEUE; |
197 | } |
195 | } |
198 | |
196 | |
199 | static void |
197 | static void |
200 | kqueue_destroy (EV_P) |
198 | kqueue_destroy (EV_P) |
201 | { |
199 | { |
202 | close (kqueue_fd); |
200 | close (backend_fd); |
203 | |
201 | |
204 | ev_free (kqueue_events); |
202 | ev_free (kqueue_events); |
205 | ev_free (kqueue_changes); |
203 | ev_free (kqueue_changes); |
206 | } |
204 | } |
207 | |
205 | |
208 | static void |
206 | static void |
209 | kqueue_fork (EV_P) |
207 | kqueue_fork (EV_P) |
210 | { |
208 | { |
211 | close (kqueue_fd); |
209 | close (backend_fd); |
212 | |
210 | |
213 | while ((kqueue_fd = kqueue ()) < 0) |
211 | while ((backend_fd = kqueue ()) < 0) |
214 | syserr ("(libev) kqueue"); |
212 | syserr ("(libev) kqueue"); |
215 | |
213 | |
216 | fcntl (kqueue_fd, F_SETFD, FD_CLOEXEC); |
214 | fcntl (backend_fd, F_SETFD, FD_CLOEXEC); |
217 | |
215 | |
218 | /* re-register interest in fds */ |
216 | /* re-register interest in fds */ |
219 | fd_rearm_all (EV_A); |
217 | fd_rearm_all (EV_A); |
220 | } |
218 | } |
221 | |
219 | |