… | |
… | |
115 | if (!epoll_ctl (backend_fd, EPOLL_CTL_MOD, fd, &ev)) |
115 | if (!epoll_ctl (backend_fd, EPOLL_CTL_MOD, fd, &ev)) |
116 | return; |
116 | return; |
117 | } |
117 | } |
118 | else if (expect_true (errno == EPERM)) |
118 | else if (expect_true (errno == EPERM)) |
119 | { |
119 | { |
|
|
120 | /* EPERM means the fd is always ready, but epoll is too snobbish */ |
|
|
121 | /* to handle it, unlike select or poll. */ |
120 | anfds [fd].emask = EV_EMASK_EPERM; |
122 | anfds [fd].emask = EV_EMASK_EPERM; |
121 | |
123 | |
122 | /* add fd to epoll_eperms, if not already inside */ |
124 | /* add fd to epoll_eperms, if not already inside */ |
123 | if (!(oldmask & EV_EMASK_EPERM)) |
125 | if (!(oldmask & EV_EMASK_EPERM)) |
124 | { |
126 | { |
… | |
… | |
201 | ev_free (epoll_events); |
203 | ev_free (epoll_events); |
202 | epoll_eventmax = array_nextsize (sizeof (struct epoll_event), epoll_eventmax, epoll_eventmax + 1); |
204 | epoll_eventmax = array_nextsize (sizeof (struct epoll_event), epoll_eventmax, epoll_eventmax + 1); |
203 | epoll_events = (struct epoll_event *)ev_malloc (sizeof (struct epoll_event) * epoll_eventmax); |
205 | epoll_events = (struct epoll_event *)ev_malloc (sizeof (struct epoll_event) * epoll_eventmax); |
204 | } |
206 | } |
205 | |
207 | |
206 | /* now add events for all fds where epoll fails, while select works... */ |
208 | /* now synthesize events for all fds where epoll fails, while select works... */ |
207 | for (i = epoll_epermcnt; i--; ) |
209 | for (i = epoll_epermcnt; i--; ) |
208 | { |
210 | { |
209 | int fd = epoll_eperms [i]; |
211 | int fd = epoll_eperms [i]; |
210 | unsigned char events = anfds [fd].events & (EV_READ | EV_WRITE); |
212 | unsigned char events = anfds [fd].events & (EV_READ | EV_WRITE); |
211 | |
213 | |