… | |
… | |
37 | static void |
37 | static void |
38 | kqueue_change (EV_P_ int fd, int filter, int flags, int fflags) |
38 | kqueue_change (EV_P_ int fd, int filter, int flags, int fflags) |
39 | { |
39 | { |
40 | struct kevent *ke; |
40 | struct kevent *ke; |
41 | |
41 | |
|
|
42 | ++kqueue_changecnt; |
42 | array_needsize (kqueue_changes, kqueue_changemax, ++kqueue_changecnt, ); |
43 | array_needsize (struct kevent, kqueue_changes, kqueue_changemax, kqueue_changecnt, EMPTY2); |
43 | |
44 | |
44 | ke = &kqueue_changes [kqueue_changecnt - 1]; |
45 | ke = &kqueue_changes [kqueue_changecnt - 1]; |
45 | memset (ke, 0, sizeof (struct kevent)); |
46 | memset (ke, 0, sizeof (struct kevent)); |
46 | ke->ident = fd; |
47 | ke->ident = fd; |
47 | ke->filter = filter; |
48 | ke->filter = filter; |
… | |
… | |
54 | #endif |
55 | #endif |
55 | |
56 | |
56 | static void |
57 | static void |
57 | kqueue_modify (EV_P_ int fd, int oev, int nev) |
58 | kqueue_modify (EV_P_ int fd, int oev, int nev) |
58 | { |
59 | { |
59 | if ((oev ^ nev) & EV_READ) |
60 | if (oev != nev) |
60 | { |
61 | { |
61 | if (nev & EV_READ) |
62 | if (oev & EV_READ) |
62 | kqueue_change (fd, EVFILT_READ, EV_ADD, NOTE_EOF); |
|
|
63 | else |
|
|
64 | kqueue_change (fd, EVFILT_READ, EV_DELETE, 0); |
63 | kqueue_change (EV_A_ fd, EVFILT_READ , EV_DELETE, 0); |
65 | } |
|
|
66 | |
64 | |
67 | if ((oev ^ nev) & EV_WRITE) |
|
|
68 | { |
|
|
69 | if (nev & EV_WRITE) |
65 | if (oev & EV_WRITE) |
70 | kqueue_change (fd, EVFILT_WRITE, EV_ADD, NOTE_EOF); |
|
|
71 | else |
|
|
72 | kqueue_change (fd, EVFILT_WRITE, EV_DELETE, 0); |
66 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_DELETE, 0); |
73 | } |
67 | } |
|
|
68 | |
|
|
69 | /* to detect close/reopen reliably, we have to re-add */ |
|
|
70 | /* event requests even when oev == nev */ |
|
|
71 | |
|
|
72 | if (nev & EV_READ) |
|
|
73 | kqueue_change (EV_A_ fd, EVFILT_READ , EV_ADD, NOTE_EOF); |
|
|
74 | |
|
|
75 | if (nev & EV_WRITE) |
|
|
76 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_ADD, NOTE_EOF); |
74 | } |
77 | } |
75 | |
78 | |
76 | static void |
79 | static void |
77 | kqueue_poll (EV_P_ ev_tstamp timeout) |
80 | kqueue_poll (EV_P_ ev_tstamp timeout) |
78 | { |
81 | { |
79 | int res, i; |
82 | int res, i; |
80 | struct timespec ts; |
83 | struct timespec ts; |
81 | |
84 | |
|
|
85 | /* need to resize so there is enough space for errors */ |
|
|
86 | if (kqueue_changecnt > kqueue_eventmax) |
|
|
87 | { |
|
|
88 | ev_free (kqueue_events); |
|
|
89 | kqueue_eventmax = array_roundsize (struct kevent, kqueue_changecnt); |
|
|
90 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
|
|
91 | } |
|
|
92 | |
82 | ts.tv_sec = (time_t)timeout; |
93 | ts.tv_sec = (time_t)timeout; |
83 | ts.tv_nsec = (long)(timeout - (ev_tstamp)ts.tv_sec) * 1e9; |
94 | ts.tv_nsec = (long)((timeout - (ev_tstamp)ts.tv_sec) * 1e9); |
84 | res = kevent (kqueue_fd, kqueue_changes, kqueue_changecnt, kqueue_events, kqueue_eventmax, &ts); |
95 | res = kevent (kqueue_fd, kqueue_changes, kqueue_changecnt, kqueue_events, kqueue_eventmax, &ts); |
85 | kqueue_changecnt = 0; |
96 | kqueue_changecnt = 0; |
86 | |
97 | |
87 | if (res < 0) |
98 | if (res < 0) |
88 | { |
99 | { |
… | |
… | |
92 | return; |
103 | return; |
93 | } |
104 | } |
94 | |
105 | |
95 | for (i = 0; i < res; ++i) |
106 | for (i = 0; i < res; ++i) |
96 | { |
107 | { |
|
|
108 | int fd = kqueue_events [i].ident; |
|
|
109 | |
97 | if (kqueue_events [i].flags & EV_ERROR) |
110 | if (kqueue_events [i].flags & EV_ERROR) |
98 | { |
111 | { |
|
|
112 | int err = kqueue_events [i].data; |
|
|
113 | |
99 | /* |
114 | /* |
100 | * Error messages that can happen, when a delete fails. |
115 | * errors that may happen |
101 | * EBADF happens when the file discriptor has been |
116 | * EBADF happens when the file discriptor has been |
102 | * closed, |
117 | * closed, |
103 | * ENOENT when the file discriptor was closed and |
118 | * ENOENT when the file descriptor was closed and |
104 | * then reopened. |
119 | * then reopened. |
105 | * EINVAL for some reasons not understood; EINVAL |
120 | * EINVAL for some reasons not understood; EINVAL |
106 | * should not be returned ever; but FreeBSD does :-\ |
121 | * should not be returned ever; but FreeBSD does :-\ |
107 | * An error is also indicated when a callback deletes |
|
|
108 | * an event we are still processing. In that case |
|
|
109 | * the data field is set to ENOENT. |
|
|
110 | */ |
122 | */ |
111 | if (kqueue_events [i].data == EBADF) |
123 | |
112 | fd_kill (EV_A_ kqueue_events [i].ident); |
124 | /* we are only interested in errors for fds that we are interested in :) */ |
|
|
125 | if (anfds [fd].events) |
|
|
126 | { |
|
|
127 | if (err == ENOENT) /* resubmit changes on ENOENT */ |
|
|
128 | kqueue_modify (EV_A_ fd, 0, anfds [fd].events); |
|
|
129 | else if (err == EBADF) /* on EBADF, we re-check the fd */ |
|
|
130 | { |
|
|
131 | if (fd_valid (fd)) |
|
|
132 | kqueue_modify (EV_A_ fd, 0, anfds [fd].events); |
|
|
133 | else |
|
|
134 | fd_kill (EV_A_ fd); |
|
|
135 | } |
|
|
136 | else /* on all other errors, we error out on the fd */ |
|
|
137 | fd_kill (EV_A_ fd); |
|
|
138 | } |
113 | } |
139 | } |
114 | else |
140 | else |
115 | fd_event ( |
141 | fd_event ( |
116 | EV_A_ |
142 | EV_A_ |
117 | kqueue_events [i].ident, |
143 | fd, |
118 | kqueue_events [i].filter == EVFILT_READ ? EV_READ |
144 | kqueue_events [i].filter == EVFILT_READ ? EV_READ |
119 | : kqueue_events [i].filter == EVFILT_WRITE ? EV_WRITE |
145 | : kqueue_events [i].filter == EVFILT_WRITE ? EV_WRITE |
120 | : 0 |
146 | : 0 |
121 | ); |
147 | ); |
122 | } |
148 | } |
123 | |
149 | |
124 | if (expect_false (res == kqueue_eventmax)) |
150 | if (expect_false (res == kqueue_eventmax)) |
125 | { |
151 | { |
126 | ev_free (kqueue_events); |
152 | ev_free (kqueue_events); |
127 | kqueue_eventmax = array_roundsize (kqueue_events, kqueue_eventmax << 1); |
153 | kqueue_eventmax = array_roundsize (struct kevent, kqueue_eventmax << 1); |
128 | kqueue_events = ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
154 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
129 | } |
155 | } |
130 | } |
156 | } |
131 | |
157 | |
132 | static int |
158 | static int |
133 | kqueue_init (EV_P_ int flags) |
159 | kqueue_init (EV_P_ int flags) |
… | |
… | |
157 | /* detected broken kqueue */ |
183 | /* detected broken kqueue */ |
158 | close (kqueue_fd); |
184 | close (kqueue_fd); |
159 | return 0; |
185 | return 0; |
160 | } |
186 | } |
161 | |
187 | |
162 | method_fudge = 1e-3; /* needed to compensate for kevent returning early */ |
188 | backend_fudge = 1e-3; /* needed to compensate for kevent returning early */ |
163 | method_modify = kqueue_modify; |
189 | backend_modify = kqueue_modify; |
164 | method_poll = kqueue_poll; |
190 | backend_poll = kqueue_poll; |
165 | |
191 | |
166 | kqueue_eventmax = 64; /* intiial number of events receivable per poll */ |
192 | kqueue_eventmax = 64; /* intiial number of events receivable per poll */ |
167 | kqueue_events = ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
193 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
168 | |
194 | |
169 | kqueue_changes = 0; |
195 | kqueue_changes = 0; |
170 | kqueue_changemax = 0; |
196 | kqueue_changemax = 0; |
171 | kqueue_changecnt = 0; |
197 | kqueue_changecnt = 0; |
172 | |
198 | |
173 | return EVMETHOD_KQUEUE; |
199 | return EVBACKEND_KQUEUE; |
174 | } |
200 | } |
175 | |
201 | |
176 | static void |
202 | static void |
177 | kqueue_destroy (EV_P) |
203 | kqueue_destroy (EV_P) |
178 | { |
204 | { |
… | |
… | |
183 | } |
209 | } |
184 | |
210 | |
185 | static void |
211 | static void |
186 | kqueue_fork (EV_P) |
212 | kqueue_fork (EV_P) |
187 | { |
213 | { |
188 | for (;;) |
214 | close (kqueue_fd); |
189 | { |
|
|
190 | kqueue_fd = kqueue (); |
|
|
191 | |
215 | |
192 | if (kqueue_fd >= 0) |
216 | while ((kqueue_fd = kqueue ()) < 0) |
193 | break; |
|
|
194 | |
|
|
195 | syserr ("(libev) kqueue"); |
217 | syserr ("(libev) kqueue"); |
196 | } |
|
|
197 | |
218 | |
198 | fcntl (kqueue_fd, F_SETFD, FD_CLOEXEC); |
219 | fcntl (kqueue_fd, F_SETFD, FD_CLOEXEC); |
199 | |
220 | |
200 | /* re-register interest in fds */ |
221 | /* re-register interest in fds */ |
201 | fd_rearm_all (EV_A); |
222 | fd_rearm_all (EV_A); |