/* * Copyright (C) Igor Sysoev * Copyright (C) NGINX, Inc. */ #include static void nxt_port_write_handler(nxt_task_t *task, void *obj, void *data); static nxt_buf_t *nxt_port_buf_completion(nxt_task_t *task, nxt_work_queue_t *wq, nxt_buf_t *b, size_t sent, nxt_bool_t mmap_mode); static void nxt_port_read_handler(nxt_task_t *task, void *obj, void *data); static void nxt_port_read_msg_process(nxt_task_t *task, nxt_port_t *port, nxt_port_recv_msg_t *msg); static nxt_buf_t *nxt_port_buf_alloc(nxt_port_t *port); static void nxt_port_buf_free(nxt_port_t *port, nxt_buf_t *b); static void nxt_port_error_handler(nxt_task_t *task, void *obj, void *data); nxt_int_t nxt_port_socket_init(nxt_task_t *task, nxt_port_t *port, size_t max_size) { nxt_int_t sndbuf, rcvbuf, size; nxt_socket_t snd, rcv; port->socket.task = task; port->pair[0] = -1; port->pair[1] = -1; if (nxt_slow_path(nxt_socketpair_create(task, port->pair) != NXT_OK)) { goto socketpair_fail; } snd = port->pair[1]; sndbuf = nxt_socket_getsockopt(task, snd, SOL_SOCKET, SO_SNDBUF); if (nxt_slow_path(sndbuf < 0)) { goto getsockopt_fail; } rcv = port->pair[0]; rcvbuf = nxt_socket_getsockopt(task, rcv, SOL_SOCKET, SO_RCVBUF); if (nxt_slow_path(rcvbuf < 0)) { goto getsockopt_fail; } if (max_size == 0) { max_size = 16 * 1024; } if ((size_t) sndbuf < max_size) { /* * On Unix domain sockets * Linux uses 224K on both send and receive directions; * FreeBSD, MacOSX, NetBSD, and OpenBSD use 2K buffer size * on send direction and 4K buffer size on receive direction; * Solaris uses 16K on send direction and 5K on receive direction. */ (void) nxt_socket_setsockopt(task, snd, SOL_SOCKET, SO_SNDBUF, max_size); sndbuf = nxt_socket_getsockopt(task, snd, SOL_SOCKET, SO_SNDBUF); if (nxt_slow_path(sndbuf < 0)) { goto getsockopt_fail; } size = sndbuf * 4; if (rcvbuf < size) { (void) nxt_socket_setsockopt(task, rcv, SOL_SOCKET, SO_RCVBUF, size); rcvbuf = nxt_socket_getsockopt(task, rcv, SOL_SOCKET, SO_RCVBUF); if (nxt_slow_path(rcvbuf < 0)) { goto getsockopt_fail; } } } port->max_size = nxt_min(max_size, (size_t) sndbuf); port->max_share = (64 * 1024); return NXT_OK; getsockopt_fail: nxt_socket_close(task, port->pair[0]); nxt_socket_close(task, port->pair[1]); socketpair_fail: return NXT_ERROR; } void nxt_port_destroy(nxt_port_t *port) { nxt_socket_close(port->socket.task, port->socket.fd); nxt_mp_destroy(port->mem_pool); } void nxt_port_write_enable(nxt_task_t *task, nxt_port_t *port) { port->socket.fd = port->pair[1]; port->socket.log = &nxt_main_log; port->socket.write_ready = 1; port->engine = task->thread->engine; port->socket.write_work_queue = &port->engine->fast_work_queue; port->socket.write_handler = nxt_port_write_handler; port->socket.error_handler = nxt_port_error_handler; if (port->iov == NULL) { port->iov = nxt_mp_get(port->mem_pool, sizeof(struct iovec) * NXT_IOBUF_MAX * 10); port->mmsg_buf = nxt_mp_get(port->mem_pool, sizeof(uint32_t) * 3 * NXT_IOBUF_MAX * 10); } } void nxt_port_write_close(nxt_port_t *port) { nxt_socket_close(port->socket.task, port->pair[1]); port->pair[1] = -1; } static void nxt_port_release_send_msg(nxt_task_t *task, void *obj, void *data) { nxt_event_engine_t *engine; nxt_port_send_msg_t *msg; msg = obj; engine = data; nxt_assert(data == msg->work.data); if (engine != task->thread->engine) { nxt_debug(task, "current thread is %PT, expected %PT", task->thread->tid, engine->task.thread->tid); nxt_event_engine_post(engine, &msg->work); return; } nxt_mp_free(engine->mem_pool, obj); nxt_mp_release(engine->mem_pool); } static nxt_port_send_msg_t * nxt_port_msg_create(nxt_task_t *task, nxt_port_send_msg_t *m) { nxt_mp_t *mp; nxt_port_send_msg_t *msg; mp = task->thread->engine->mem_pool; msg = nxt_mp_alloc(mp, sizeof(nxt_port_send_msg_t)); if (nxt_slow_path(msg == NULL)) { return NULL; } nxt_mp_retain(mp); msg->link.next = NULL; msg->link.prev = NULL; msg->buf = m->buf; msg->share = m->share; msg->fd = m->fd; msg->close_fd = m->close_fd; msg->port_msg = m->port_msg; msg->work.next = NULL; msg->work.handler = nxt_port_release_send_msg; msg->work.task = task; msg->work.obj = msg; msg->work.data = task->thread->engine; return msg; } static nxt_port_send_msg_t * nxt_port_msg_push(nxt_task_t *task, nxt_port_t *port, nxt_port_send_msg_t *msg) { if (msg->work.data == NULL) { msg = nxt_port_msg_create(task, msg); } if (msg != NULL) { nxt_queue_insert_tail(&port->messages, &msg->link); } return msg; } static nxt_port_send_msg_t * nxt_port_msg_first(nxt_task_t *task, nxt_port_t *port, nxt_port_send_msg_t *msg) { nxt_queue_link_t *lnk; lnk = nxt_queue_first(&port->messages); if (lnk == nxt_queue_tail(&port->messages)) { return msg; } return nxt_queue_link_data(lnk, nxt_port_send_msg_t, link); } nxt_int_t nxt_port_socket_twrite(nxt_task_t *task, nxt_port_t *port, nxt_uint_t type, nxt_fd_t fd, uint32_t stream, nxt_port_id_t reply_port, nxt_buf_t *b, void *tracking) { nxt_port_send_msg_t msg, *res; msg.link.next = NULL; msg.link.prev = NULL; msg.buf = b; msg.fd = fd; msg.close_fd = (type & NXT_PORT_MSG_CLOSE_FD) != 0; msg.share = 0; if (tracking != NULL) { nxt_port_mmap_tracking_write(msg.tracking_msg, tracking); } msg.port_msg.stream = stream; msg.port_msg.pid = nxt_pid; msg.port_msg.reply_port = reply_port; msg.port_msg.type = type & NXT_PORT_MSG_MASK; msg.port_msg.last = (type & NXT_PORT_MSG_LAST) != 0; msg.port_msg.mmap = 0; msg.port_msg.nf = 0; msg.port_msg.mf = 0; msg.port_msg.tracking = tracking != NULL; msg.work.data = NULL; if (port->socket.write_ready) { nxt_port_write_handler(task, &port->socket, &msg); } else { nxt_thread_mutex_lock(&port->write_mutex); res = nxt_port_msg_push(task, port, &msg); nxt_thread_mutex_unlock(&port->write_mutex); if (res == NULL) { return NXT_ERROR; } nxt_port_use(task, port, 1); } return NXT_OK; } static void nxt_port_fd_block_write(nxt_task_t *task, nxt_port_t *port, void *data) { nxt_fd_event_block_write(task->thread->engine, &port->socket); } static void nxt_port_fd_enable_write(nxt_task_t *task, nxt_port_t *port, void *data) { nxt_fd_event_enable_write(task->thread->engine, &port->socket); } static void nxt_port_write_handler(nxt_task_t *task, void *obj, void *data) { int use_delta; size_t plain_size; ssize_t n; nxt_bool_t block_write, enable_write; nxt_port_t *port; struct iovec *iov; nxt_work_queue_t *wq; nxt_port_method_t m; nxt_port_send_msg_t *msg; nxt_sendbuf_coalesce_t sb; port = nxt_container_of(obj, nxt_port_t, socket); block_write = 0; enable_write = 0; use_delta = 0; nxt_thread_mutex_lock(&port->write_mutex); iov = port->iov; wq = &task->thread->engine->fast_work_queue; do { msg = nxt_port_msg_first(task, port, data); if (msg == NULL) { block_write = 1; goto unlock_mutex; } iov[0].iov_base = &msg->port_msg; iov[0].iov_len = sizeof(nxt_port_msg_t); sb.buf = msg->buf; sb.iobuf = &iov[1]; sb.nmax = NXT_IOBUF_MAX - 1; sb.sync = 0; sb.last = 0; sb.size = 0; sb.limit = port->max_size; sb.limit_reached = 0; sb.nmax_reached = 0; m = nxt_port_mmap_get_method(task, port, msg->buf); if (m == NXT_PORT_METHOD_MMAP) { sb.limit = (1ULL << 31) - 1; sb.nmax = nxt_min(NXT_IOBUF_MAX * 10 - 1, port->max_size / PORT_MMAP_MIN_SIZE); } if (msg->port_msg.tracking) { iov[0].iov_len += sizeof(msg->tracking_msg); } sb.limit -= iov[0].iov_len; nxt_sendbuf_mem_coalesce(task, &sb); plain_size = sb.size; /* * Send through mmap enabled only when payload * is bigger than PORT_MMAP_MIN_SIZE. */ if (m == NXT_PORT_METHOD_MMAP && plain_size > PORT_MMAP_MIN_SIZE) { nxt_port_mmap_write(task, port, msg, &sb); } else { m = NXT_PORT_METHOD_PLAIN; } msg->port_msg.last |= sb.last; msg->port_msg.mf = sb.limit_reached || sb.nmax_reached; n = nxt_socketpair_send(&port->socket, msg->fd, iov, sb.niov + 1); if (n > 0) { if (nxt_slow_path((size_t) n != sb.size + iov[0].iov_len)) { nxt_alert(task, "port %d: short write: %z instead of %uz", port->socket.fd, n, sb.size + iov[0].iov_len); goto fail; } if (msg->fd != -1 && msg->close_fd != 0) { nxt_fd_close(msg->fd); msg->fd = -1; } msg->buf = nxt_port_buf_completion(task, wq, msg->buf, plain_size, m == NXT_PORT_METHOD_MMAP); if (msg->buf != NULL) { nxt_debug(task, "port %d: frag stream #%uD", port->socket.fd, msg->port_msg.stream); /* * A file descriptor is sent only * in the first message of a stream. */ msg->fd = -1; msg->share += n; msg->port_msg.nf = 1; msg->port_msg.tracking = 0; if (msg->share >= port->max_share) { msg->share = 0; if (msg->link.next != NULL) { nxt_queue_remove(&msg->link); use_delta--; } data = NULL; if (nxt_port_msg_push(task, port, msg) != NULL) { use_delta++; } } } else { if (msg->link.next != NULL) { nxt_queue_remove(&msg->link); use_delta--; nxt_work_queue_add(wq, nxt_port_release_send_msg, task, msg, msg->work.data); } data = NULL; } } else if (nxt_slow_path(n == NXT_ERROR)) { if (msg->link.next == NULL) { if (nxt_port_msg_push(task, port, msg) != NULL) { use_delta++; } } goto fail; } /* n == NXT_AGAIN */ } while (port->socket.write_ready); if (nxt_fd_event_is_disabled(port->socket.write)) { enable_write = 1; } goto unlock_mutex; fail: use_delta++; nxt_work_queue_add(wq, nxt_port_error_handler, task, &port->socket, &port->socket); unlock_mutex: nxt_thread_mutex_unlock(&port->write_mutex); if (block_write && nxt_fd_event_is_active(port->socket.write)) { nxt_port_post(task, port, nxt_port_fd_block_write, NULL); } if (enable_write) { nxt_port_post(task, port, nxt_port_fd_enable_write, NULL); } if (use_delta != 0) { nxt_port_use(task, port, use_delta); } } static nxt_buf_t * nxt_port_buf_completion(nxt_task_t *task, nxt_work_queue_t *wq, nxt_buf_t *b, size_t sent, nxt_bool_t mmap_mode) { size_t size; while (b != NULL) { nxt_prefetch(b->next); if (!nxt_buf_is_sync(b)) { size = nxt_buf_used_size(b); if (size != 0) { if (sent == 0) { break; } if (nxt_buf_is_port_mmap(b) && mmap_mode) { /* * buffer has been sent to other side which is now * responsible for shared memory bucket release */ b->is_port_mmap_sent = 1; } if (sent < size) { if (nxt_buf_is_mem(b)) { b->mem.pos += sent; } if (nxt_buf_is_file(b)) { b->file_pos += sent; } break; } /* b->mem.free is NULL in file-only buffer. */ b->mem.pos = b->mem.free; if (nxt_buf_is_file(b)) { b->file_pos = b->file_end; } sent -= size; } } nxt_work_queue_add(wq, b->completion_handler, task, b, b->parent); b = b->next; } return b; } void nxt_port_read_enable(nxt_task_t *task, nxt_port_t *port) { port->socket.fd = port->pair[0]; port->socket.log = &nxt_main_log; port->engine = task->thread->engine; port->socket.read_work_queue = &port->engine->fast_work_queue; port->socket.read_handler = nxt_port_read_handler; port->socket.error_handler = nxt_port_error_handler; nxt_fd_event_enable_read(port->engine, &port->socket); } void nxt_port_read_close(nxt_port_t *port) { port->socket.read_ready = 0; nxt_socket_close(port->socket.task, port->pair[0]); port->pair[0] = -1; } static void nxt_port_read_handler(nxt_task_t *task, void *obj, void *data) { ssize_t n; nxt_buf_t *b; nxt_port_t *port; struct iovec iov[2]; nxt_port_recv_msg_t msg; port = msg.port = nxt_container_of(obj, nxt_port_t, socket); nxt_assert(port->engine == task->thread->engine); for ( ;; ) { b = nxt_port_buf_alloc(port); if (nxt_slow_path(b == NULL)) { /* TODO: disable event for some time */ } iov[0].iov_base = &msg.port_msg; iov[0].iov_len = sizeof(nxt_port_msg_t); iov[1].iov_base = b->mem.pos; iov[1].iov_len = port->max_size; n = nxt_socketpair_recv(&port->socket, &msg.fd, iov, 2); if (n > 0) { msg.buf = b; msg.size = n; nxt_port_read_msg_process(task, port, &msg); /* * To disable instant completion or buffer re-usage, * handler should reset 'msg.buf'. */ if (msg.buf == b) { nxt_port_buf_free(port, b); } if (port->socket.read_ready) { continue; } return; } if (n == NXT_AGAIN) { nxt_port_buf_free(port, b); nxt_fd_event_enable_read(task->thread->engine, &port->socket); return; } /* n == 0 || n == NXT_ERROR */ nxt_work_queue_add(&task->thread->engine->fast_work_queue, nxt_port_error_handler, task, &port->socket, NULL); return; } } static nxt_int_t nxt_port_lvlhsh_frag_test(nxt_lvlhsh_query_t *lhq, void *data) { nxt_port_recv_msg_t *fmsg; fmsg = data; if (lhq->key.length == sizeof(uint32_t) && *(uint32_t *) lhq->key.start == fmsg->port_msg.stream) { return NXT_OK; } return NXT_DECLINED; } static void * nxt_port_lvlhsh_frag_alloc(void *ctx, size_t size) { return nxt_mp_alloc(ctx, size); } static void nxt_port_lvlhsh_frag_free(void *ctx, void *p) { nxt_mp_free(ctx, p); } static const nxt_lvlhsh_proto_t lvlhsh_frag_proto nxt_aligned(64) = { NXT_LVLHSH_DEFAULT, nxt_port_lvlhsh_frag_test, nxt_port_lvlhsh_frag_alloc, nxt_port_lvlhsh_frag_free, }; static nxt_port_recv_msg_t * nxt_port_frag_start(nxt_task_t *task, nxt_port_t *port, nxt_port_recv_msg_t *msg) { nxt_int_t res; nxt_lvlhsh_query_t lhq; nxt_port_recv_msg_t *fmsg; nxt_debug(task, "start frag stream #%uD", msg->port_msg.stream); fmsg = nxt_mp_alloc(port->mem_pool, sizeof(nxt_port_recv_msg_t)); if (nxt_slow_path(fmsg == NULL)) { return NULL; } *fmsg = *msg; lhq.key_hash = nxt_murmur_hash2(&fmsg->port_msg.stream, sizeof(uint32_t)); lhq.key.length = sizeof(uint32_t); lhq.key.start = (u_char *) &fmsg->port_msg.stream; lhq.proto = &lvlhsh_frag_proto; lhq.replace = 0; lhq.value = fmsg; lhq.pool = port->mem_pool; res = nxt_lvlhsh_insert(&port->frags, &lhq); switch (res) { case NXT_OK: return fmsg; case NXT_DECLINED: nxt_log(task, NXT_LOG_WARN, "duplicate frag stream #%uD", fmsg->port_msg.stream); nxt_mp_free(port->mem_pool, fmsg); return NULL; default: nxt_log(task, NXT_LOG_WARN, "failed to add frag stream #%uD", fmsg->port_msg.stream); nxt_mp_free(port->mem_pool, fmsg); return NULL; } } static nxt_port_recv_msg_t * nxt_port_frag_find(nxt_task_t *task, nxt_port_t *port, uint32_t stream, nxt_bool_t last) { nxt_int_t res; nxt_lvlhsh_query_t lhq; nxt_debug(task, "%s frag stream #%uD", last ? "last" : "next", stream); lhq.key_hash = nxt_murmur_hash2(&stream, sizeof(uint32_t)); lhq.key.length = sizeof(uint32_t); lhq.key.start = (u_char *) &stream; lhq.proto = &lvlhsh_frag_proto; lhq.pool = port->mem_pool; res = last != 0 ? nxt_lvlhsh_delete(&port->frags, &lhq) : nxt_lvlhsh_find(&port->frags, &lhq); switch (res) { case NXT_OK: return lhq.value; default: nxt_log(task, NXT_LOG_INFO, "frag stream #%uD not found", stream); return NULL; } } static void nxt_port_read_msg_process(nxt_task_t *task, nxt_port_t *port, nxt_port_recv_msg_t *msg) { nxt_buf_t *b, *orig_b; nxt_port_recv_msg_t *fmsg; if (nxt_slow_path(msg->size < sizeof(nxt_port_msg_t))) { nxt_alert(task, "port %d: too small message:%uz", port->socket.fd, msg->size); if (msg->fd != -1) { nxt_fd_close(msg->fd); } return; } /* adjust size to actual buffer used size */ msg->size -= sizeof(nxt_port_msg_t); b = orig_b = msg->buf; b->mem.free += msg->size; if (msg->port_msg.tracking) { msg->cancelled = nxt_port_mmap_tracking_read(task, msg) == 0; } else { msg->cancelled = 0; } if (nxt_slow_path(msg->port_msg.nf != 0)) { fmsg = nxt_port_frag_find(task, port, msg->port_msg.stream, msg->port_msg.mf == 0); if (nxt_slow_path(fmsg == NULL)) { goto fmsg_failed; } if (nxt_fast_path(fmsg->cancelled == 0)) { if (msg->port_msg.mmap) { nxt_port_mmap_read(task, msg); } nxt_buf_chain_add(&fmsg->buf, msg->buf); fmsg->size += msg->size; msg->buf = NULL; b = NULL; if (nxt_fast_path(msg->port_msg.mf == 0)) { b = fmsg->buf; port->handler(task, fmsg); msg->buf = fmsg->buf; msg->fd = fmsg->fd; /* * To disable instant completion or buffer re-usage, * handler should reset 'msg.buf'. */ if (!msg->port_msg.mmap && msg->buf == b) { nxt_port_buf_free(port, b); } } } if (nxt_fast_path(msg->port_msg.mf == 0)) { nxt_mp_free(port->mem_pool, fmsg); } } else { if (nxt_slow_path(msg->port_msg.mf != 0)) { if (msg->port_msg.mmap && msg->cancelled == 0) { nxt_port_mmap_read(task, msg); b = msg->buf; } fmsg = nxt_port_frag_start(task, port, msg); if (nxt_slow_path(fmsg == NULL)) { goto fmsg_failed; } fmsg->port_msg.nf = 0; fmsg->port_msg.mf = 0; if (nxt_fast_path(msg->cancelled == 0)) { msg->buf = NULL; msg->fd = -1; b = NULL; } else { if (msg->fd != -1) { nxt_fd_close(msg->fd); } } } else { if (nxt_fast_path(msg->cancelled == 0)) { if (msg->port_msg.mmap) { nxt_port_mmap_read(task, msg); b = msg->buf; } port->handler(task, msg); } } } fmsg_failed: if (msg->port_msg.mmap && orig_b != b) { /* * To disable instant buffer completion, * handler should reset 'msg->buf'. */ if (msg->buf == b) { /* complete mmap buffers */ for (; b != NULL; b = b->next) { nxt_debug(task, "complete buffer %p", b); nxt_work_queue_add(port->socket.read_work_queue, b->completion_handler, task, b, b->parent); } } /* restore original buf */ msg->buf = orig_b; } } static nxt_buf_t * nxt_port_buf_alloc(nxt_port_t *port) { nxt_buf_t *b; if (port->free_bufs != NULL) { b = port->free_bufs; port->free_bufs = b->next; b->mem.pos = b->mem.start; b->mem.free = b->mem.start; b->next = NULL; } else { b = nxt_buf_mem_alloc(port->mem_pool, port->max_size, 0); if (nxt_slow_path(b == NULL)) { return NULL; } } return b; } static void nxt_port_buf_free(nxt_port_t *port, nxt_buf_t *b) { nxt_buf_chain_add(&b, port->free_bufs); port->free_bufs = b; } static void nxt_port_error_handler(nxt_task_t *task, void *obj, void *data) { int use_delta; nxt_buf_t *b; nxt_port_t *port; nxt_work_queue_t *wq; nxt_port_send_msg_t *msg; nxt_debug(task, "port error handler %p", obj); /* TODO */ port = nxt_container_of(obj, nxt_port_t, socket); use_delta = 0; if (obj == data) { use_delta--; } wq = &task->thread->engine->fast_work_queue; nxt_thread_mutex_lock(&port->write_mutex); nxt_queue_each(msg, &port->messages, nxt_port_send_msg_t, link) { for (b = msg->buf; b != NULL; b = b->next) { if (nxt_buf_is_sync(b)) { continue; } nxt_work_queue_add(wq, b->completion_handler, task, b, b->parent); } nxt_queue_remove(&msg->link); use_delta--; nxt_work_queue_add(wq, nxt_port_release_send_msg, task, msg, msg->work.data); } nxt_queue_loop; nxt_thread_mutex_unlock(&port->write_mutex); if (use_delta != 0) { nxt_port_use(task, port, use_delta); } }