… | |
… | |
55 | #endif |
55 | #endif |
56 | |
56 | |
57 | static void |
57 | static void |
58 | kqueue_modify (EV_P_ int fd, int oev, int nev) |
58 | kqueue_modify (EV_P_ int fd, int oev, int nev) |
59 | { |
59 | { |
|
|
60 | if (oev != nev) |
|
|
61 | { |
|
|
62 | if (oev & EV_READ) |
|
|
63 | kqueue_change (EV_A_ fd, EVFILT_READ , EV_DELETE, 0); |
|
|
64 | |
|
|
65 | if (oev & EV_WRITE) |
|
|
66 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_DELETE, 0); |
|
|
67 | } |
|
|
68 | |
60 | /* to detect close/reopen reliably, we have to remove and re-add */ |
69 | /* to detect close/reopen reliably, we have to re-add */ |
61 | /* event requests even when oev == nev */ |
70 | /* event requests even when oev == nev */ |
62 | |
71 | |
63 | if (oev & EV_READ) |
|
|
64 | kqueue_change (EV_A_ fd, EVFILT_READ, EV_DELETE, 0); |
|
|
65 | |
|
|
66 | if (oev & EV_WRITE) |
|
|
67 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_DELETE, 0); |
|
|
68 | |
|
|
69 | if (nev & EV_READ) |
72 | if (nev & EV_READ) |
70 | kqueue_change (EV_A_ fd, EVFILT_READ, EV_ADD, NOTE_EOF); |
73 | kqueue_change (EV_A_ fd, EVFILT_READ , EV_ADD, NOTE_EOF); |
71 | |
74 | |
72 | if (nev & EV_WRITE) |
75 | if (nev & EV_WRITE) |
73 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_ADD, NOTE_EOF); |
76 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_ADD, NOTE_EOF); |
74 | } |
77 | } |
75 | |
78 | |
… | |
… | |
180 | /* detected broken kqueue */ |
183 | /* detected broken kqueue */ |
181 | close (kqueue_fd); |
184 | close (kqueue_fd); |
182 | return 0; |
185 | return 0; |
183 | } |
186 | } |
184 | |
187 | |
185 | method_fudge = 1e-3; /* needed to compensate for kevent returning early */ |
188 | backend_fudge = 1e-3; /* needed to compensate for kevent returning early */ |
186 | method_modify = kqueue_modify; |
189 | backend_modify = kqueue_modify; |
187 | method_poll = kqueue_poll; |
190 | backend_poll = kqueue_poll; |
188 | |
191 | |
189 | kqueue_eventmax = 64; /* intiial number of events receivable per poll */ |
192 | kqueue_eventmax = 64; /* intiial number of events receivable per poll */ |
190 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
193 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
191 | |
194 | |
192 | kqueue_changes = 0; |
195 | kqueue_changes = 0; |