… | |
… | |
55 | #endif |
55 | #endif |
56 | |
56 | |
57 | static void |
57 | static void |
58 | kqueue_modify (EV_P_ int fd, int oev, int nev) |
58 | kqueue_modify (EV_P_ int fd, int oev, int nev) |
59 | { |
59 | { |
|
|
60 | if (oev != nev) |
|
|
61 | { |
|
|
62 | if (oev & EV_READ) |
|
|
63 | kqueue_change (EV_A_ fd, EVFILT_READ , EV_DELETE, 0); |
|
|
64 | |
|
|
65 | if (oev & EV_WRITE) |
|
|
66 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_DELETE, 0); |
|
|
67 | } |
|
|
68 | |
60 | /* to detect close/reopen reliably, we have to remove and re-add */ |
69 | /* to detect close/reopen reliably, we have to re-add */ |
61 | /* event requests even when oev == nev */ |
70 | /* event requests even when oev == nev */ |
62 | |
71 | |
63 | if (oev & EV_READ) |
|
|
64 | kqueue_change (EV_A_ fd, EVFILT_READ, EV_DELETE, 0); |
|
|
65 | |
|
|
66 | if (oev & EV_WRITE) |
|
|
67 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_DELETE, 0); |
|
|
68 | |
|
|
69 | if (nev & EV_READ) |
72 | if (nev & EV_READ) |
70 | kqueue_change (EV_A_ fd, EVFILT_READ, EV_ADD, NOTE_EOF); |
73 | kqueue_change (EV_A_ fd, EVFILT_READ , EV_ADD, NOTE_EOF); |
71 | |
74 | |
72 | if (nev & EV_WRITE) |
75 | if (nev & EV_WRITE) |
73 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_ADD, NOTE_EOF); |
76 | kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_ADD, NOTE_EOF); |
74 | } |
77 | } |
75 | |
78 | |
… | |
… | |
87 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
90 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
88 | } |
91 | } |
89 | |
92 | |
90 | ts.tv_sec = (time_t)timeout; |
93 | ts.tv_sec = (time_t)timeout; |
91 | ts.tv_nsec = (long)((timeout - (ev_tstamp)ts.tv_sec) * 1e9); |
94 | ts.tv_nsec = (long)((timeout - (ev_tstamp)ts.tv_sec) * 1e9); |
92 | fprintf (stderr, "to %ld:%09ld %f\n", ts.tv_sec, ts.tv_nsec, res);//D |
|
|
93 | res = kevent (kqueue_fd, kqueue_changes, kqueue_changecnt, kqueue_events, kqueue_eventmax, &ts); |
95 | res = kevent (kqueue_fd, kqueue_changes, kqueue_changecnt, kqueue_events, kqueue_eventmax, &ts); |
94 | kqueue_changecnt = 0; |
96 | kqueue_changecnt = 0; |
95 | |
97 | |
96 | if (res < 0) |
98 | if (res < 0) |
97 | { |
99 | { |
… | |
… | |
181 | /* detected broken kqueue */ |
183 | /* detected broken kqueue */ |
182 | close (kqueue_fd); |
184 | close (kqueue_fd); |
183 | return 0; |
185 | return 0; |
184 | } |
186 | } |
185 | |
187 | |
186 | method_fudge = 1e-3; /* needed to compensate for kevent returning early */ |
188 | backend_fudge = 1e-3; /* needed to compensate for kevent returning early */ |
187 | method_modify = kqueue_modify; |
189 | backend_modify = kqueue_modify; |
188 | method_poll = kqueue_poll; |
190 | backend_poll = kqueue_poll; |
189 | |
191 | |
190 | kqueue_eventmax = 64; /* intiial number of events receivable per poll */ |
192 | kqueue_eventmax = 64; /* intiial number of events receivable per poll */ |
191 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
193 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
192 | |
194 | |
193 | kqueue_changes = 0; |
195 | kqueue_changes = 0; |
194 | kqueue_changemax = 0; |
196 | kqueue_changemax = 0; |
195 | kqueue_changecnt = 0; |
197 | kqueue_changecnt = 0; |
196 | |
198 | |
197 | return EVMETHOD_KQUEUE; |
199 | return EVBACKEND_KQUEUE; |
198 | } |
200 | } |
199 | |
201 | |
200 | static void |
202 | static void |
201 | kqueue_destroy (EV_P) |
203 | kqueue_destroy (EV_P) |
202 | { |
204 | { |