X-Git-Url: https://git.ozlabs.org/?p=ccan;a=blobdiff_plain;f=ccan%2Fio%2Fio.c;h=faf8b87bfd649b7244e8ee01087651d1039dbbf5;hp=352afe33e87638ed638729f7e56f91f37ea1db71;hb=95b59482c1bb18b7904ea60149eff4809dd28d80;hpb=3a7b8a8a8081ebbb6457527de376dec6264bc381 diff --git a/ccan/io/io.c b/ccan/io/io.c index 352afe33..faf8b87b 100644 --- a/ccan/io/io.c +++ b/ccan/io/io.c @@ -9,53 +9,131 @@ #include #include #include +#include +#include void *io_loop_return; +struct io_alloc io_alloc = { + malloc, realloc, free +}; + #ifdef DEBUG -bool io_plan_for_other; +/* Set to skip the next plan. */ +bool io_plan_nodebug; +/* The current connection to apply plan to. */ struct io_conn *current; -bool (*io_debug)(struct io_conn *conn); +/* User-defined function to select which connection(s) to debug. */ +bool (*io_debug_conn)(struct io_conn *conn); +/* Set when we wake up an connection we are debugging. */ bool io_debug_wakeup; -void io_plan_debug(struct io_plan *plan) +struct io_plan io_debug(struct io_plan plan) { - if (io_plan_for_other) { - io_plan_for_other = false; - return; - } + struct io_conn *ready = NULL; - if (!io_debug || !current) - return; + if (io_plan_nodebug) { + io_plan_nodebug = false; + return plan; + } - if (!io_debug(current) && !io_debug_wakeup) - return; + if (!current || !doing_debug_on(current)) { + if (!io_debug_wakeup) + return plan; + } io_debug_wakeup = false; - current->plan = *plan; + current->plan = plan; backend_plan_changed(current); /* Call back into the loop immediately. */ - io_loop_return = io_loop(); + io_loop_return = do_io_loop(&ready); + + if (ready) { + set_current(ready); + if (!ready->plan.next) { + /* Call finish function immediately. */ + if (ready->finish) { + errno = ready->plan.u1.s; + ready->finish(ready, ready->finish_arg); + ready->finish = NULL; + } + backend_del_conn(ready); + } else { + /* Calls back in itself, via io_debug_io(). */ + if (ready->plan.io(ready->fd.fd, &ready->plan) != 2) + abort(); + } + set_current(NULL); + } + + /* Return a do-nothing plan, so backend_plan_changed in + * io_ready doesn't do anything (it's already been called). */ + return io_idle_(); +} + +int io_debug_io(int ret) +{ + /* Cache it for debugging; current changes. */ + struct io_conn *conn = current; + int saved_errno = errno; + + if (!doing_debug_on(conn)) + return ret; + + /* These will all go linearly through the io_debug() path above. */ + switch (ret) { + case -1: + /* This will call io_debug above. */ + errno = saved_errno; + io_close(); + break; + case 0: /* Keep going with plan. */ + io_debug(conn->plan); + break; + case 1: /* Done: get next plan. */ + if (timeout_active(conn)) + backend_del_timeout(conn); + /* In case they call io_duplex, clear our poll flags so + * both sides don't seem to be both doing read or write + * (See assert(!mask || pfd->events != mask) in poll.c) */ + conn->plan.pollflag = 0; + conn->plan.next(conn, conn->plan.next_arg); + break; + default: + abort(); + } + + /* Normally-invalid value, used for sanity check. */ + return 2; } static void debug_io_wake(struct io_conn *conn) { /* We want linear if we wake a debugged connection, too. */ - if (io_debug && io_debug(conn)) + if (io_debug_conn && io_debug_conn(conn)) io_debug_wakeup = true; } + +/* Counterpart to io_plan_no_debug(), called in macros in io.h */ +static void io_plan_debug_again(void) +{ + io_plan_nodebug = false; +} #else static void debug_io_wake(struct io_conn *conn) { } +static void io_plan_debug_again(void) +{ +} #endif struct io_listener *io_new_listener_(int fd, void (*init)(int fd, void *arg), void *arg) { - struct io_listener *l = malloc(sizeof(*l)); + struct io_listener *l = io_alloc.alloc(sizeof(*l)); if (!l) return NULL; @@ -65,7 +143,7 @@ struct io_listener *io_new_listener_(int fd, l->init = init; l->arg = arg; if (!add_listener(l)) { - free(l); + io_alloc.free(l); return NULL; } return l; @@ -75,12 +153,14 @@ void io_close_listener(struct io_listener *l) { close(l->fd.fd); del_listener(l); - free(l); + io_alloc.free(l); } struct io_conn *io_new_conn_(int fd, struct io_plan plan) { - struct io_conn *conn = malloc(sizeof(*conn)); + struct io_conn *conn = io_alloc.alloc(sizeof(*conn)); + + io_plan_debug_again(); if (!conn) return NULL; @@ -93,7 +173,7 @@ struct io_conn *io_new_conn_(int fd, struct io_plan plan) conn->duplex = NULL; conn->timeout = NULL; if (!add_conn(conn)) { - free(conn); + io_alloc.free(conn); return NULL; } return conn; @@ -111,9 +191,11 @@ struct io_conn *io_duplex_(struct io_conn *old, struct io_plan plan) { struct io_conn *conn; + io_plan_debug_again(); + assert(!old->duplex); - conn = malloc(sizeof(*conn)); + conn = io_alloc.alloc(sizeof(*conn)); if (!conn) return NULL; @@ -125,7 +207,7 @@ struct io_conn *io_duplex_(struct io_conn *old, struct io_plan plan) conn->finish_arg = NULL; conn->timeout = NULL; if (!add_duplex(conn)) { - free(conn); + io_alloc.free(conn); return NULL; } old->duplex = conn; @@ -138,7 +220,7 @@ bool io_timeout_(struct io_conn *conn, struct timespec ts, assert(cb); if (!conn->timeout) { - conn->timeout = malloc(sizeof(*conn->timeout)); + conn->timeout = io_alloc.alloc(sizeof(*conn->timeout)); if (!conn->timeout) return false; } else @@ -153,13 +235,13 @@ bool io_timeout_(struct io_conn *conn, struct timespec ts, /* Returns true if we're finished. */ static int do_write(int fd, struct io_plan *plan) { - ssize_t ret = write(fd, plan->u.write.buf, plan->u.write.len); + ssize_t ret = write(fd, plan->u1.cp, plan->u2.s); if (ret < 0) - return -1; + return io_debug_io(-1); - plan->u.write.buf += ret; - plan->u.write.len -= ret; - return (plan->u.write.len == 0); + plan->u1.cp += ret; + plan->u2.s -= ret; + return io_debug_io(plan->u2.s == 0); } /* Queue some data to be written. */ @@ -170,26 +252,25 @@ struct io_plan io_write_(const void *data, size_t len, struct io_plan plan; assert(cb); - plan.u.write.buf = data; - plan.u.write.len = len; + plan.u1.const_vp = data; + plan.u2.s = len; plan.io = do_write; plan.next = cb; plan.next_arg = arg; plan.pollflag = POLLOUT; - io_plan_debug(&plan); return plan; } static int do_read(int fd, struct io_plan *plan) { - ssize_t ret = read(fd, plan->u.read.buf, plan->u.read.len); + ssize_t ret = read(fd, plan->u1.cp, plan->u2.s); if (ret <= 0) - return -1; + return io_debug_io(-1); - plan->u.read.buf += ret; - plan->u.read.len -= ret; - return (plan->u.read.len == 0); + plan->u1.cp += ret; + plan->u2.s -= ret; + return io_debug_io(plan->u2.s == 0); } /* Queue a request to read into a buffer. */ @@ -200,25 +281,24 @@ struct io_plan io_read_(void *data, size_t len, struct io_plan plan; assert(cb); - plan.u.read.buf = data; - plan.u.read.len = len; + plan.u1.cp = data; + plan.u2.s = len; plan.io = do_read; plan.next = cb; plan.next_arg = arg; plan.pollflag = POLLIN; - io_plan_debug(&plan); return plan; } static int do_read_partial(int fd, struct io_plan *plan) { - ssize_t ret = read(fd, plan->u.readpart.buf, *plan->u.readpart.lenp); + ssize_t ret = read(fd, plan->u1.cp, *(size_t *)plan->u2.vp); if (ret <= 0) - return -1; + return io_debug_io(-1); - *plan->u.readpart.lenp = ret; - return 1; + *(size_t *)plan->u2.vp = ret; + return io_debug_io(1); } /* Queue a partial request to read into a buffer. */ @@ -229,25 +309,24 @@ struct io_plan io_read_partial_(void *data, size_t *len, struct io_plan plan; assert(cb); - plan.u.readpart.buf = data; - plan.u.readpart.lenp = len; + plan.u1.cp = data; + plan.u2.vp = len; plan.io = do_read_partial; plan.next = cb; plan.next_arg = arg; plan.pollflag = POLLIN; - io_plan_debug(&plan); return plan; } static int do_write_partial(int fd, struct io_plan *plan) { - ssize_t ret = write(fd, plan->u.writepart.buf, *plan->u.writepart.lenp); + ssize_t ret = write(fd, plan->u1.cp, *(size_t *)plan->u2.vp); if (ret < 0) - return -1; + return io_debug_io(-1); - *plan->u.writepart.lenp = ret; - return 1; + *(size_t *)plan->u2.vp = ret; + return io_debug_io(1); } /* Queue a partial write request. */ @@ -258,33 +337,91 @@ struct io_plan io_write_partial_(const void *data, size_t *len, struct io_plan plan; assert(cb); - plan.u.writepart.buf = data; - plan.u.writepart.lenp = len; + plan.u1.const_vp = data; + plan.u2.vp = len; plan.io = do_write_partial; plan.next = cb; plan.next_arg = arg; plan.pollflag = POLLOUT; - io_plan_debug(&plan); return plan; } -struct io_plan io_idle(void) +static int already_connected(int fd, struct io_plan *plan) +{ + return io_debug_io(1); +} + +static int do_connect(int fd, struct io_plan *plan) +{ + int err, ret; + socklen_t len = sizeof(err); + + /* Has async connect finished? */ + ret = getsockopt(fd, SOL_SOCKET, SO_ERROR, &err, &len); + if (ret < 0) + return -1; + + if (err == 0) { + /* Restore blocking if it was initially. */ + fcntl(fd, F_SETFD, plan->u1.s); + return 1; + } + return 0; +} + +struct io_plan io_connect_(int fd, const struct addrinfo *addr, + struct io_plan (*cb)(struct io_conn*, void *), + void *arg) +{ + struct io_plan plan; + + assert(cb); + + plan.next = cb; + plan.next_arg = arg; + + /* Save old flags, set nonblock if not already. */ + plan.u1.s = fcntl(fd, F_GETFD); + fcntl(fd, F_SETFD, plan.u1.s | O_NONBLOCK); + + /* Immediate connect can happen. */ + if (connect(fd, addr->ai_addr, addr->ai_addrlen) == 0) { + /* Dummy will be called immediately. */ + plan.pollflag = POLLOUT; + plan.io = already_connected; + } else { + if (errno != EINPROGRESS) + return io_close_(); + + plan.pollflag = POLLIN; + plan.io = do_connect; + } + return plan; +} + +struct io_plan io_idle_(void) { struct io_plan plan; plan.pollflag = 0; plan.io = NULL; /* Never called (overridden by io_wake), but NULL means closing */ - plan.next = (void *)io_idle; + plan.next = (void *)io_idle_; - io_plan_debug(&plan); return plan; } +bool io_is_idle(const struct io_conn *conn) +{ + return conn->plan.io == NULL; +} + void io_wake_(struct io_conn *conn, struct io_plan plan) { + io_plan_debug_again(); + /* It might be closing, but we haven't called its finish() yet. */ if (!conn->plan.next) return; @@ -298,35 +435,41 @@ void io_wake_(struct io_conn *conn, struct io_plan plan) void io_ready(struct io_conn *conn) { + /* Beware io_close_other! */ + if (!conn->plan.next) + return; + + set_current(conn); switch (conn->plan.io(conn->fd.fd, &conn->plan)) { case -1: /* Failure means a new plan: close up. */ - set_current(conn); conn->plan = io_close(); backend_plan_changed(conn); - set_current(NULL); break; case 0: /* Keep going with plan. */ break; case 1: /* Done: get next plan. */ - set_current(conn); if (timeout_active(conn)) backend_del_timeout(conn); + /* In case they call io_duplex, clear our poll flags so + * both sides don't seem to be both doing read or write + * (See assert(!mask || pfd->events != mask) in poll.c) */ + conn->plan.pollflag = 0; conn->plan = conn->plan.next(conn, conn->plan.next_arg); backend_plan_changed(conn); - set_current(NULL); } + set_current(NULL); } /* Close the connection, we're done. */ -struct io_plan io_close(void) +struct io_plan io_close_(void) { struct io_plan plan; plan.pollflag = 0; /* This means we're closing. */ plan.next = NULL; + plan.u1.s = errno; - io_plan_debug(&plan); return plan; } @@ -335,11 +478,33 @@ struct io_plan io_close_cb(struct io_conn *conn, void *arg) return io_close(); } +void io_close_other(struct io_conn *conn) +{ + conn->plan = io_close_(); + backend_plan_changed(conn); +} + /* Exit the loop, returning this (non-NULL) arg. */ struct io_plan io_break_(void *ret, struct io_plan plan) { + io_plan_debug_again(); + assert(ret); io_loop_return = ret; return plan; } + +int io_conn_fd(const struct io_conn *conn) +{ + return conn->fd.fd; +} + +void io_set_alloc(void *(*allocfn)(size_t size), + void *(*reallocfn)(void *ptr, size_t size), + void (*freefn)(void *ptr)) +{ + io_alloc.alloc = allocfn; + io_alloc.realloc = reallocfn; + io_alloc.free = freefn; +}