… | |
… | |
91 | ev.data.u64 = (uint64_t)(uint32_t)fd |
91 | ev.data.u64 = (uint64_t)(uint32_t)fd |
92 | | ((uint64_t)(uint32_t)++anfds [fd].egen << 32); |
92 | | ((uint64_t)(uint32_t)++anfds [fd].egen << 32); |
93 | ev.events = (nev & EV_READ ? EPOLLIN : 0) |
93 | ev.events = (nev & EV_READ ? EPOLLIN : 0) |
94 | | (nev & EV_WRITE ? EPOLLOUT : 0); |
94 | | (nev & EV_WRITE ? EPOLLOUT : 0); |
95 | |
95 | |
96 | if (expect_true (!epoll_ctl (backend_fd, oev && oldmask != nev ? EPOLL_CTL_MOD : EPOLL_CTL_ADD, fd, &ev))) |
96 | if (ecb_expect_true (!epoll_ctl (backend_fd, oev && oldmask != nev ? EPOLL_CTL_MOD : EPOLL_CTL_ADD, fd, &ev))) |
97 | return; |
97 | return; |
98 | |
98 | |
99 | if (expect_true (errno == ENOENT)) |
99 | if (ecb_expect_true (errno == ENOENT)) |
100 | { |
100 | { |
101 | /* if ENOENT then the fd went away, so try to do the right thing */ |
101 | /* if ENOENT then the fd went away, so try to do the right thing */ |
102 | if (!nev) |
102 | if (!nev) |
103 | goto dec_egen; |
103 | goto dec_egen; |
104 | |
104 | |
105 | if (!epoll_ctl (backend_fd, EPOLL_CTL_ADD, fd, &ev)) |
105 | if (!epoll_ctl (backend_fd, EPOLL_CTL_ADD, fd, &ev)) |
106 | return; |
106 | return; |
107 | } |
107 | } |
108 | else if (expect_true (errno == EEXIST)) |
108 | else if (ecb_expect_true (errno == EEXIST)) |
109 | { |
109 | { |
110 | /* EEXIST means we ignored a previous DEL, but the fd is still active */ |
110 | /* EEXIST means we ignored a previous DEL, but the fd is still active */ |
111 | /* if the kernel mask is the same as the new mask, we assume it hasn't changed */ |
111 | /* if the kernel mask is the same as the new mask, we assume it hasn't changed */ |
112 | if (oldmask == nev) |
112 | if (oldmask == nev) |
113 | goto dec_egen; |
113 | goto dec_egen; |
114 | |
114 | |
115 | if (!epoll_ctl (backend_fd, EPOLL_CTL_MOD, fd, &ev)) |
115 | if (!epoll_ctl (backend_fd, EPOLL_CTL_MOD, fd, &ev)) |
116 | return; |
116 | return; |
117 | } |
117 | } |
118 | else if (expect_true (errno == EPERM)) |
118 | else if (ecb_expect_true (errno == EPERM)) |
119 | { |
119 | { |
120 | /* EPERM means the fd is always ready, but epoll is too snobbish */ |
120 | /* EPERM means the fd is always ready, but epoll is too snobbish */ |
121 | /* to handle it, unlike select or poll. */ |
121 | /* to handle it, unlike select or poll. */ |
122 | anfds [fd].emask = EV_EMASK_EPERM; |
122 | anfds [fd].emask = EV_EMASK_EPERM; |
123 | |
123 | |
… | |
… | |
144 | epoll_poll (EV_P_ ev_tstamp timeout) |
144 | epoll_poll (EV_P_ ev_tstamp timeout) |
145 | { |
145 | { |
146 | int i; |
146 | int i; |
147 | int eventcnt; |
147 | int eventcnt; |
148 | |
148 | |
149 | if (expect_false (epoll_epermcnt)) |
149 | if (ecb_expect_false (epoll_epermcnt)) |
150 | timeout = 0.; |
150 | timeout = 0.; |
151 | |
151 | |
152 | /* epoll wait times cannot be larger than (LONG_MAX - 999UL) / HZ msecs, which is below */ |
152 | /* epoll wait times cannot be larger than (LONG_MAX - 999UL) / HZ msecs, which is below */ |
153 | /* the default libev max wait time, however. */ |
153 | /* the default libev max wait time, however. */ |
154 | EV_RELEASE_CB; |
154 | EV_RELEASE_CB; |
155 | eventcnt = epoll_wait (backend_fd, epoll_events, epoll_eventmax, timeout * 1e3); |
155 | eventcnt = epoll_wait (backend_fd, epoll_events, epoll_eventmax, timeout * 1e3); |
156 | EV_ACQUIRE_CB; |
156 | EV_ACQUIRE_CB; |
157 | |
157 | |
158 | if (expect_false (eventcnt < 0)) |
158 | if (ecb_expect_false (eventcnt < 0)) |
159 | { |
159 | { |
160 | if (errno != EINTR) |
160 | if (errno != EINTR) |
161 | ev_syserr ("(libev) epoll_wait"); |
161 | ev_syserr ("(libev) epoll_wait"); |
162 | |
162 | |
163 | return; |
163 | return; |
… | |
… | |
176 | * check for spurious notification. |
176 | * check for spurious notification. |
177 | * this only finds spurious notifications on egen updates |
177 | * this only finds spurious notifications on egen updates |
178 | * other spurious notifications will be found by epoll_ctl, below |
178 | * other spurious notifications will be found by epoll_ctl, below |
179 | * we assume that fd is always in range, as we never shrink the anfds array |
179 | * we assume that fd is always in range, as we never shrink the anfds array |
180 | */ |
180 | */ |
181 | if (expect_false ((uint32_t)anfds [fd].egen != (uint32_t)(ev->data.u64 >> 32))) |
181 | if (ecb_expect_false ((uint32_t)anfds [fd].egen != (uint32_t)(ev->data.u64 >> 32))) |
182 | { |
182 | { |
183 | /* recreate kernel state */ |
183 | /* recreate kernel state */ |
184 | postfork |= 2; |
184 | postfork |= 2; |
185 | continue; |
185 | continue; |
186 | } |
186 | } |
187 | |
187 | |
188 | if (expect_false (got & ~want)) |
188 | if (ecb_expect_false (got & ~want)) |
189 | { |
189 | { |
190 | anfds [fd].emask = want; |
190 | anfds [fd].emask = want; |
191 | |
191 | |
192 | /* |
192 | /* |
193 | * we received an event but are not interested in it, try mod or del |
193 | * we received an event but are not interested in it, try mod or del |
… | |
… | |
212 | |
212 | |
213 | fd_event (EV_A_ fd, got); |
213 | fd_event (EV_A_ fd, got); |
214 | } |
214 | } |
215 | |
215 | |
216 | /* if the receive array was full, increase its size */ |
216 | /* if the receive array was full, increase its size */ |
217 | if (expect_false (eventcnt == epoll_eventmax)) |
217 | if (ecb_expect_false (eventcnt == epoll_eventmax)) |
218 | { |
218 | { |
219 | ev_free (epoll_events); |
219 | ev_free (epoll_events); |
220 | epoll_eventmax = array_nextsize (sizeof (struct epoll_event), epoll_eventmax, epoll_eventmax + 1); |
220 | epoll_eventmax = array_nextsize (sizeof (struct epoll_event), epoll_eventmax, epoll_eventmax + 1); |
221 | epoll_events = (struct epoll_event *)ev_malloc (sizeof (struct epoll_event) * epoll_eventmax); |
221 | epoll_events = (struct epoll_event *)ev_malloc (sizeof (struct epoll_event) * epoll_eventmax); |
222 | } |
222 | } |