… | |
… | |
115 | if (!epoll_ctl (backend_fd, EPOLL_CTL_MOD, fd, &ev)) |
115 | if (!epoll_ctl (backend_fd, EPOLL_CTL_MOD, fd, &ev)) |
116 | return; |
116 | return; |
117 | } |
117 | } |
118 | else if (expect_true (errno == EPERM)) |
118 | else if (expect_true (errno == EPERM)) |
119 | { |
119 | { |
|
|
120 | /* EPERM means the fd is always ready, but epoll is too snobbish */ |
|
|
121 | /* to handle it, unlike select or poll. */ |
120 | anfds [fd].emask = EV_EMASK_EPERM; |
122 | anfds [fd].emask = EV_EMASK_EPERM; |
121 | |
123 | |
122 | /* add fd to epoll_eperms, if not already inside */ |
124 | /* add fd to epoll_eperms, if not already inside */ |
123 | if (!(oldmask & EV_EMASK_EPERM)) |
125 | if (!(oldmask & EV_EMASK_EPERM)) |
124 | { |
126 | { |
… | |
… | |
139 | static void |
141 | static void |
140 | epoll_poll (EV_P_ ev_tstamp timeout) |
142 | epoll_poll (EV_P_ ev_tstamp timeout) |
141 | { |
143 | { |
142 | int i; |
144 | int i; |
143 | int eventcnt; |
145 | int eventcnt; |
144 | |
146 | |
145 | /* epoll wait times cannot be larger than (LONG_MAX - 999UL) / HZ msecs, which is below */ |
147 | /* epoll wait times cannot be larger than (LONG_MAX - 999UL) / HZ msecs, which is below */ |
146 | /* the default libev max wait time, however. */ |
148 | /* the default libev max wait time, however. */ |
147 | EV_RELEASE_CB; |
149 | EV_RELEASE_CB; |
148 | eventcnt = epoll_wait (backend_fd, epoll_events, epoll_eventmax, (int)ceil (timeout * 1000.)); |
150 | eventcnt = epoll_wait (backend_fd, epoll_events, epoll_eventmax, |
|
|
151 | epoll_epermcnt ? 0 : (int)ceil (timeout * 1000.)); |
149 | EV_ACQUIRE_CB; |
152 | EV_ACQUIRE_CB; |
150 | |
153 | |
151 | if (expect_false (eventcnt < 0)) |
154 | if (expect_false (eventcnt < 0)) |
152 | { |
155 | { |
153 | if (errno != EINTR) |
156 | if (errno != EINTR) |
… | |
… | |
201 | ev_free (epoll_events); |
204 | ev_free (epoll_events); |
202 | epoll_eventmax = array_nextsize (sizeof (struct epoll_event), epoll_eventmax, epoll_eventmax + 1); |
205 | epoll_eventmax = array_nextsize (sizeof (struct epoll_event), epoll_eventmax, epoll_eventmax + 1); |
203 | epoll_events = (struct epoll_event *)ev_malloc (sizeof (struct epoll_event) * epoll_eventmax); |
206 | epoll_events = (struct epoll_event *)ev_malloc (sizeof (struct epoll_event) * epoll_eventmax); |
204 | } |
207 | } |
205 | |
208 | |
206 | /* now add events for all fds where epoll fails, while select works... */ |
209 | /* now synthesize events for all fds where epoll fails, while select works... */ |
207 | for (i = epoll_epermcnt; i--; ) |
210 | for (i = epoll_epermcnt; i--; ) |
208 | { |
211 | { |
209 | int fd = epoll_eperms [i]; |
212 | int fd = epoll_eperms [i]; |
210 | unsigned char events = anfds [fd].events & (EV_READ | EV_WRITE); |
213 | unsigned char events = anfds [fd].events & (EV_READ | EV_WRITE); |
211 | |
214 | |