1 /* Licensed under LGPLv2.1+ - see LICENSE file for details */
8 #include <sys/socket.h>
11 #include <ccan/time/time.h>
12 #include <ccan/timer/timer.h>
14 static size_t num_fds = 0, max_fds = 0, num_waiting = 0, num_always = 0, max_always = 0, num_exclusive = 0;
15 static struct pollfd *pollfds = NULL;
16 static struct fd **fds = NULL;
17 static struct io_plan **always = NULL;
18 static struct timemono (*nowfn)(void) = time_mono;
19 static int (*pollfn)(struct pollfd *fds, nfds_t nfds, int timeout) = poll;
21 struct timemono (*io_time_override(struct timemono (*now)(void)))(void)
23 struct timemono (*old)(void) = nowfn;
28 int (*io_poll_override(int (*poll)(struct pollfd *fds, nfds_t nfds, int timeout)))(struct pollfd *, nfds_t, int)
30 int (*old)(struct pollfd *fds, nfds_t nfds, int timeout) = pollfn;
35 static bool add_fd(struct fd *fd, short events)
39 pollfds = tal_arr(NULL, struct pollfd, 8);
42 fds = tal_arr(pollfds, struct fd *, 8);
48 if (num_fds + 1 > max_fds) {
49 size_t num = max_fds * 2;
51 if (!tal_resize(&pollfds, num))
53 if (!tal_resize(&fds, num))
58 pollfds[num_fds].events = events;
59 /* In case it's idle. */
61 pollfds[num_fds].fd = -fd->fd - 1;
63 pollfds[num_fds].fd = fd->fd;
64 pollfds[num_fds].revents = 0; /* In case we're iterating now */
66 fd->backend_info = num_fds;
67 fd->exclusive[0] = fd->exclusive[1] = false;
75 static void del_fd(struct fd *fd)
77 size_t n = fd->backend_info;
81 if (pollfds[n].events)
83 if (n != num_fds - 1) {
84 /* Move last one over us. */
85 pollfds[n] = pollfds[num_fds-1];
86 fds[n] = fds[num_fds-1];
87 assert(fds[n]->backend_info == num_fds-1);
88 fds[n]->backend_info = n;
89 } else if (num_fds == 1) {
90 /* Free everything when no more fds. */
91 pollfds = tal_free(pollfds);
96 fd->backend_info = -1;
98 if (fd->exclusive[IO_IN])
100 if (fd->exclusive[IO_OUT])
104 static void destroy_listener(struct io_listener *l)
110 bool add_listener(struct io_listener *l)
112 if (!add_fd(&l->fd, POLLIN))
114 tal_add_destructor(l, destroy_listener);
118 static int find_always(const struct io_plan *plan)
120 for (size_t i = 0; i < num_always; i++)
121 if (always[i] == plan)
126 static void remove_from_always(const struct io_plan *plan)
130 if (plan->status != IO_ALWAYS)
133 pos = find_always(plan);
136 /* Move last one down if we made a hole */
137 if (pos != num_always-1)
138 always[pos] = always[num_always-1];
141 /* Only free if no fds left either. */
142 if (num_always == 0 && max_fds == 0) {
148 bool backend_new_always(struct io_plan *plan)
150 assert(find_always(plan) == -1);
153 assert(num_always == 0);
154 always = tal_arr(NULL, struct io_plan *, 8);
160 if (num_always + 1 > max_always) {
161 size_t num = max_always * 2;
163 if (!tal_resize(&always, num))
168 always[num_always++] = plan;
172 static void setup_pfd(struct io_conn *conn, struct pollfd *pfd)
174 assert(pfd == &pollfds[conn->fd.backend_info]);
177 if (conn->plan[IO_IN].status == IO_POLLING_NOTSTARTED
178 || conn->plan[IO_IN].status == IO_POLLING_STARTED)
179 pfd->events |= POLLIN;
180 if (conn->plan[IO_OUT].status == IO_POLLING_NOTSTARTED
181 || conn->plan[IO_OUT].status == IO_POLLING_STARTED)
182 pfd->events |= POLLOUT;
185 pfd->fd = conn->fd.fd;
187 pfd->fd = -conn->fd.fd - 1;
191 void backend_new_plan(struct io_conn *conn)
193 struct pollfd *pfd = &pollfds[conn->fd.backend_info];
198 setup_pfd(conn, pfd);
204 void backend_wake(const void *wait)
208 for (i = 0; i < num_fds; i++) {
211 /* Ignore listeners */
212 if (fds[i]->listener)
216 if (c->plan[IO_IN].status == IO_WAITING
217 && c->plan[IO_IN].arg.u1.const_vp == wait)
218 io_do_wakeup(c, IO_IN);
220 if (c->plan[IO_OUT].status == IO_WAITING
221 && c->plan[IO_OUT].arg.u1.const_vp == wait)
222 io_do_wakeup(c, IO_OUT);
226 static void destroy_conn(struct io_conn *conn, bool close_fd)
228 int saved_errno = errno;
234 remove_from_always(&conn->plan[IO_IN]);
235 remove_from_always(&conn->plan[IO_OUT]);
237 /* errno saved/restored by tal_free itself. */
240 conn->finish(conn, conn->finish_arg);
244 static void destroy_conn_close_fd(struct io_conn *conn)
246 destroy_conn(conn, true);
249 bool add_conn(struct io_conn *c)
251 if (!add_fd(&c->fd, 0))
253 tal_add_destructor(c, destroy_conn_close_fd);
257 void cleanup_conn_without_close(struct io_conn *conn)
259 tal_del_destructor(conn, destroy_conn_close_fd);
260 destroy_conn(conn, false);
263 static void accept_conn(struct io_listener *l)
265 int fd = accept(l->fd.fd, NULL, NULL);
267 /* FIXME: What to do here? */
271 io_new_conn(l->ctx, fd, l->init, l->arg);
274 /* Return pointer to exclusive flag for this plan. */
275 static bool *exclusive(struct io_plan *plan)
277 struct io_conn *conn;
279 conn = container_of(plan, struct io_conn, plan[plan->dir]);
280 return &conn->fd.exclusive[plan->dir];
283 /* For simplicity, we do one always at a time */
284 static bool handle_always(void)
286 /* Backwards is simple easier to remove entries */
287 for (int i = num_always - 1; i >= 0; i--) {
288 struct io_plan *plan = always[i];
290 if (num_exclusive && !*exclusive(plan))
292 /* Remove first: it might re-add */
293 if (i != num_always-1)
294 always[i] = always[num_always-1];
303 bool backend_set_exclusive(struct io_plan *plan, bool excl)
305 bool *excl_ptr = exclusive(plan);
307 if (excl != *excl_ptr) {
315 return num_exclusive != 0;
318 /* FIXME: We could do this once at set_exclusive time, and catch everywhere
319 * else that we manipulate events. */
320 static void exclude_pollfds(void)
322 if (num_exclusive == 0)
325 for (size_t i = 0; i < num_fds; i++) {
326 struct pollfd *pfd = &pollfds[fds[i]->backend_info];
328 if (!fds[i]->exclusive[IO_IN])
329 pfd->events &= ~POLLIN;
330 if (!fds[i]->exclusive[IO_OUT])
331 pfd->events &= ~POLLOUT;
333 /* If we're not listening, we don't want error events
336 pfd->fd = -fds[i]->fd - 1;
340 static void restore_pollfds(void)
342 if (num_exclusive == 0)
345 for (size_t i = 0; i < num_fds; i++) {
346 struct pollfd *pfd = &pollfds[fds[i]->backend_info];
348 if (fds[i]->listener) {
349 pfd->events = POLLIN;
350 pfd->fd = fds[i]->fd;
352 struct io_conn *conn = (void *)fds[i];
353 setup_pfd(conn, pfd);
358 /* This is the main loop. */
359 void *io_loop(struct timers *timers, struct timer **expired)
363 /* if timers is NULL, expired must be. If not, not. */
364 assert(!timers == !expired);
366 /* Make sure this is NULL if we exit for some other reason. */
370 while (!io_loop_return) {
371 int i, r, ms_timeout = -1;
373 if (handle_always()) {
374 /* Could have started/finished more. */
378 /* Everything closed? */
382 /* You can't tell them all to go to sleep! */
386 struct timemono now, first;
390 /* Call functions for expired timers. */
391 *expired = timers_expire(timers, now);
395 /* Now figure out how long to wait for the next one. */
396 if (timer_earliest(timers, &first)) {
398 next = time_to_msec(timemono_between(first, now));
402 ms_timeout = INT_MAX;
406 /* We do this temporarily, assuming exclusive is unusual */
408 r = pollfn(pollfds, num_fds, ms_timeout);
412 /* Signals shouldn't break us, unless they set
419 for (i = 0; i < num_fds && !io_loop_return; i++) {
420 struct io_conn *c = (void *)fds[i];
421 int events = pollfds[i].revents;
423 /* Clear so we don't get confused if exclusive next time */
424 pollfds[i].revents = 0;
429 if (fds[i]->listener) {
430 struct io_listener *l = (void *)fds[i];
431 if (events & POLLIN) {
434 } else if (events & (POLLHUP|POLLNVAL|POLLERR)) {
437 io_close_listener(l);
439 } else if (events & (POLLIN|POLLOUT)) {
442 } else if (events & (POLLHUP|POLLNVAL|POLLERR)) {
450 ret = io_loop_return;
451 io_loop_return = NULL;