… | |
… | |
320 | |
320 | |
321 | return EV_SQES + (tail & EV_SQ_VAR (ring_mask)); |
321 | return EV_SQES + (tail & EV_SQ_VAR (ring_mask)); |
322 | } |
322 | } |
323 | |
323 | |
324 | inline_size |
324 | inline_size |
325 | struct io_uring_sqe * |
325 | void |
326 | iouring_sqe_submit (EV_P_ struct io_uring_sqe *sqe) |
326 | iouring_sqe_submit (EV_P_ struct io_uring_sqe *sqe) |
327 | { |
327 | { |
328 | unsigned idx = sqe - EV_SQES; |
328 | unsigned idx = sqe - EV_SQES; |
329 | |
329 | |
330 | EV_SQ_ARRAY [idx] = idx; |
330 | EV_SQ_ARRAY [idx] = idx; |
… | |
… | |
346 | iouring_tfd_to = EV_TSTAMP_HUGE; |
346 | iouring_tfd_to = EV_TSTAMP_HUGE; |
347 | } |
347 | } |
348 | |
348 | |
349 | /* called for full and partial cleanup */ |
349 | /* called for full and partial cleanup */ |
350 | ecb_cold |
350 | ecb_cold |
351 | static int |
351 | static void |
352 | iouring_internal_destroy (EV_P) |
352 | iouring_internal_destroy (EV_P) |
353 | { |
353 | { |
354 | close (iouring_tfd); |
354 | close (iouring_tfd); |
355 | close (iouring_fd); |
355 | close (iouring_fd); |
356 | |
356 | |