… | |
… | |
180 | /* detected broken kqueue */ |
180 | /* detected broken kqueue */ |
181 | close (kqueue_fd); |
181 | close (kqueue_fd); |
182 | return 0; |
182 | return 0; |
183 | } |
183 | } |
184 | |
184 | |
185 | method_fudge = 1e-3; /* needed to compensate for kevent returning early */ |
185 | backend_fudge = 1e-3; /* needed to compensate for kevent returning early */ |
186 | method_modify = kqueue_modify; |
186 | backend_modify = kqueue_modify; |
187 | method_poll = kqueue_poll; |
187 | backend_poll = kqueue_poll; |
188 | |
188 | |
189 | kqueue_eventmax = 64; /* intiial number of events receivable per poll */ |
189 | kqueue_eventmax = 64; /* intiial number of events receivable per poll */ |
190 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
190 | kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); |
191 | |
191 | |
192 | kqueue_changes = 0; |
192 | kqueue_changes = 0; |