883 lines
21 KiB
C
883 lines
21 KiB
C
|
|
/*
|
|
* Copyright (C) Igor Sysoev
|
|
* Copyright (C) NGINX, Inc.
|
|
*/
|
|
|
|
#include <nxt_main.h>
|
|
|
|
|
|
static void nxt_port_write_handler(nxt_task_t *task, void *obj, void *data);
|
|
static void nxt_port_read_handler(nxt_task_t *task, void *obj, void *data);
|
|
static void nxt_port_read_msg_process(nxt_task_t *task, nxt_port_t *port,
|
|
nxt_port_recv_msg_t *msg);
|
|
static nxt_buf_t *nxt_port_buf_alloc(nxt_port_t *port);
|
|
static void nxt_port_buf_free(nxt_port_t *port, nxt_buf_t *b);
|
|
static void nxt_port_error_handler(nxt_task_t *task, void *obj, void *data);
|
|
|
|
|
|
nxt_int_t
|
|
nxt_port_socket_init(nxt_task_t *task, nxt_port_t *port, size_t max_size)
|
|
{
|
|
nxt_int_t sndbuf, rcvbuf, size;
|
|
nxt_socket_t snd, rcv;
|
|
|
|
port->socket.task = task;
|
|
|
|
port->pair[0] = -1;
|
|
port->pair[1] = -1;
|
|
|
|
if (nxt_slow_path(nxt_socketpair_create(task, port->pair) != NXT_OK)) {
|
|
goto socketpair_fail;
|
|
}
|
|
|
|
snd = port->pair[1];
|
|
|
|
sndbuf = nxt_socket_getsockopt(task, snd, SOL_SOCKET, SO_SNDBUF);
|
|
if (nxt_slow_path(sndbuf < 0)) {
|
|
goto getsockopt_fail;
|
|
}
|
|
|
|
rcv = port->pair[0];
|
|
|
|
rcvbuf = nxt_socket_getsockopt(task, rcv, SOL_SOCKET, SO_RCVBUF);
|
|
if (nxt_slow_path(rcvbuf < 0)) {
|
|
goto getsockopt_fail;
|
|
}
|
|
|
|
if (max_size == 0) {
|
|
max_size = 16 * 1024;
|
|
}
|
|
|
|
if ((size_t) sndbuf < max_size) {
|
|
/*
|
|
* On Unix domain sockets
|
|
* Linux uses 224K on both send and receive directions;
|
|
* FreeBSD, MacOSX, NetBSD, and OpenBSD use 2K buffer size
|
|
* on send direction and 4K buffer size on receive direction;
|
|
* Solaris uses 16K on send direction and 5K on receive direction.
|
|
*/
|
|
(void) nxt_socket_setsockopt(task, snd, SOL_SOCKET, SO_SNDBUF,
|
|
max_size);
|
|
|
|
sndbuf = nxt_socket_getsockopt(task, snd, SOL_SOCKET, SO_SNDBUF);
|
|
if (nxt_slow_path(sndbuf < 0)) {
|
|
goto getsockopt_fail;
|
|
}
|
|
|
|
size = sndbuf * 4;
|
|
|
|
if (rcvbuf < size) {
|
|
(void) nxt_socket_setsockopt(task, rcv, SOL_SOCKET, SO_RCVBUF,
|
|
size);
|
|
|
|
rcvbuf = nxt_socket_getsockopt(task, rcv, SOL_SOCKET, SO_RCVBUF);
|
|
if (nxt_slow_path(rcvbuf < 0)) {
|
|
goto getsockopt_fail;
|
|
}
|
|
}
|
|
}
|
|
|
|
port->max_size = nxt_min(max_size, (size_t) sndbuf);
|
|
port->max_share = (64 * 1024);
|
|
|
|
return NXT_OK;
|
|
|
|
getsockopt_fail:
|
|
|
|
nxt_socket_close(task, port->pair[0]);
|
|
nxt_socket_close(task, port->pair[1]);
|
|
|
|
socketpair_fail:
|
|
|
|
return NXT_ERROR;
|
|
}
|
|
|
|
|
|
void
|
|
nxt_port_destroy(nxt_port_t *port)
|
|
{
|
|
nxt_socket_close(port->socket.task, port->socket.fd);
|
|
nxt_mp_destroy(port->mem_pool);
|
|
}
|
|
|
|
|
|
void
|
|
nxt_port_write_enable(nxt_task_t *task, nxt_port_t *port)
|
|
{
|
|
port->socket.fd = port->pair[1];
|
|
port->socket.log = &nxt_main_log;
|
|
port->socket.write_ready = 1;
|
|
|
|
port->engine = task->thread->engine;
|
|
|
|
port->socket.write_work_queue = &port->engine->fast_work_queue;
|
|
port->socket.write_handler = nxt_port_write_handler;
|
|
port->socket.error_handler = nxt_port_error_handler;
|
|
|
|
if (port->iov == NULL) {
|
|
port->iov = nxt_mp_get(port->mem_pool, sizeof(struct iovec) *
|
|
NXT_IOBUF_MAX * 10);
|
|
port->mmsg_buf = nxt_mp_get(port->mem_pool, sizeof(uint32_t) * 3 *
|
|
NXT_IOBUF_MAX * 10);
|
|
}
|
|
}
|
|
|
|
|
|
void
|
|
nxt_port_write_close(nxt_port_t *port)
|
|
{
|
|
nxt_socket_close(port->socket.task, port->pair[1]);
|
|
port->pair[1] = -1;
|
|
}
|
|
|
|
|
|
static void
|
|
nxt_port_release_send_msg(nxt_task_t *task, void *obj, void *data)
|
|
{
|
|
nxt_event_engine_t *engine;
|
|
nxt_port_send_msg_t *msg;
|
|
|
|
msg = obj;
|
|
engine = data;
|
|
|
|
#if (NXT_DEBUG)
|
|
if (nxt_slow_path(data != msg->work.data)) {
|
|
nxt_log_alert(task->log, "release msg data (%p) != msg->engine (%p)",
|
|
data, msg->work.data);
|
|
nxt_abort();
|
|
}
|
|
#endif
|
|
|
|
if (engine != task->thread->engine) {
|
|
|
|
nxt_debug(task, "current thread is %PT, expected %PT",
|
|
task->thread->tid, engine->task.thread->tid);
|
|
|
|
nxt_event_engine_post(engine, &msg->work);
|
|
|
|
return;
|
|
}
|
|
|
|
nxt_mp_release(engine->mem_pool, obj);
|
|
}
|
|
|
|
|
|
static nxt_port_send_msg_t *
|
|
nxt_port_msg_create(nxt_task_t *task, nxt_port_send_msg_t *m)
|
|
{
|
|
nxt_port_send_msg_t *msg;
|
|
|
|
msg = nxt_mp_retain(task->thread->engine->mem_pool,
|
|
sizeof(nxt_port_send_msg_t));
|
|
if (nxt_slow_path(msg == NULL)) {
|
|
return NULL;
|
|
}
|
|
|
|
msg->link.next = NULL;
|
|
msg->link.prev = NULL;
|
|
|
|
msg->buf = m->buf;
|
|
msg->fd = m->fd;
|
|
msg->close_fd = m->close_fd;
|
|
msg->port_msg = m->port_msg;
|
|
|
|
msg->work.next = NULL;
|
|
msg->work.handler = nxt_port_release_send_msg;
|
|
msg->work.task = task;
|
|
msg->work.obj = msg;
|
|
msg->work.data = task->thread->engine;
|
|
|
|
return msg;
|
|
}
|
|
|
|
|
|
static nxt_port_send_msg_t *
|
|
nxt_port_msg_push(nxt_task_t *task, nxt_port_t *port, nxt_port_send_msg_t *msg)
|
|
{
|
|
if (msg->work.data == NULL) {
|
|
msg = nxt_port_msg_create(task, msg);
|
|
}
|
|
|
|
if (msg != NULL) {
|
|
nxt_queue_insert_tail(&port->messages, &msg->link);
|
|
}
|
|
|
|
return msg;
|
|
}
|
|
|
|
|
|
static nxt_port_send_msg_t *
|
|
nxt_port_msg_first(nxt_task_t *task, nxt_port_t *port, nxt_port_send_msg_t *msg)
|
|
{
|
|
nxt_queue_link_t *lnk;
|
|
|
|
lnk = nxt_queue_first(&port->messages);
|
|
|
|
if (lnk == nxt_queue_tail(&port->messages)) {
|
|
return msg;
|
|
}
|
|
|
|
return nxt_queue_link_data(lnk, nxt_port_send_msg_t, link);
|
|
}
|
|
|
|
|
|
nxt_int_t
|
|
nxt_port_socket_twrite(nxt_task_t *task, nxt_port_t *port, nxt_uint_t type,
|
|
nxt_fd_t fd, uint32_t stream, nxt_port_id_t reply_port, nxt_buf_t *b,
|
|
void *tracking)
|
|
{
|
|
nxt_port_send_msg_t msg, *res;
|
|
|
|
msg.link.next = NULL;
|
|
msg.link.prev = NULL;
|
|
|
|
msg.buf = b;
|
|
msg.fd = fd;
|
|
msg.close_fd = (type & NXT_PORT_MSG_CLOSE_FD) != 0;
|
|
msg.share = 0;
|
|
|
|
if (tracking != NULL) {
|
|
nxt_port_mmap_tracking_write(msg.tracking_msg, tracking);
|
|
}
|
|
|
|
msg.port_msg.stream = stream;
|
|
msg.port_msg.pid = nxt_pid;
|
|
msg.port_msg.reply_port = reply_port;
|
|
msg.port_msg.type = type & NXT_PORT_MSG_MASK;
|
|
msg.port_msg.last = (type & NXT_PORT_MSG_LAST) != 0;
|
|
msg.port_msg.mmap = 0;
|
|
msg.port_msg.nf = 0;
|
|
msg.port_msg.mf = 0;
|
|
msg.port_msg.tracking = tracking != NULL;
|
|
|
|
msg.work.data = NULL;
|
|
|
|
if (port->socket.write_ready) {
|
|
nxt_port_write_handler(task, &port->socket, &msg);
|
|
} else {
|
|
nxt_thread_mutex_lock(&port->write_mutex);
|
|
|
|
res = nxt_port_msg_push(task, port, &msg);
|
|
|
|
nxt_thread_mutex_unlock(&port->write_mutex);
|
|
|
|
if (res == NULL) {
|
|
return NXT_ERROR;
|
|
}
|
|
|
|
nxt_port_use(task, port, 1);
|
|
}
|
|
|
|
return NXT_OK;
|
|
}
|
|
|
|
|
|
static void
|
|
nxt_port_fd_block_write(nxt_task_t *task, nxt_port_t *port, void *data)
|
|
{
|
|
nxt_fd_event_block_write(task->thread->engine, &port->socket);
|
|
}
|
|
|
|
|
|
static void
|
|
nxt_port_fd_enable_write(nxt_task_t *task, nxt_port_t *port, void *data)
|
|
{
|
|
nxt_fd_event_enable_write(task->thread->engine, &port->socket);
|
|
}
|
|
|
|
|
|
static void
|
|
nxt_port_write_handler(nxt_task_t *task, void *obj, void *data)
|
|
{
|
|
int use_delta;
|
|
size_t plain_size;
|
|
ssize_t n;
|
|
nxt_bool_t block_write, enable_write;
|
|
nxt_port_t *port;
|
|
struct iovec *iov;
|
|
nxt_work_queue_t *wq;
|
|
nxt_port_method_t m;
|
|
nxt_port_send_msg_t *msg;
|
|
nxt_sendbuf_coalesce_t sb;
|
|
|
|
port = nxt_container_of(obj, nxt_port_t, socket);
|
|
|
|
block_write = 0;
|
|
enable_write = 0;
|
|
use_delta = 0;
|
|
|
|
nxt_thread_mutex_lock(&port->write_mutex);
|
|
|
|
iov = port->iov;
|
|
|
|
wq = &task->thread->engine->fast_work_queue;
|
|
|
|
do {
|
|
msg = nxt_port_msg_first(task, port, data);
|
|
|
|
if (msg == NULL) {
|
|
block_write = 1;
|
|
goto unlock_mutex;
|
|
}
|
|
|
|
iov[0].iov_base = &msg->port_msg;
|
|
iov[0].iov_len = sizeof(nxt_port_msg_t);
|
|
|
|
sb.buf = msg->buf;
|
|
sb.iobuf = &iov[1];
|
|
sb.nmax = NXT_IOBUF_MAX - 1;
|
|
sb.sync = 0;
|
|
sb.last = 0;
|
|
sb.size = 0;
|
|
sb.limit = port->max_size;
|
|
|
|
sb.limit_reached = 0;
|
|
sb.nmax_reached = 0;
|
|
|
|
m = nxt_port_mmap_get_method(task, port, msg->buf);
|
|
|
|
if (m == NXT_PORT_METHOD_MMAP) {
|
|
sb.limit = (1ULL << 31) - 1;
|
|
sb.nmax = nxt_min(NXT_IOBUF_MAX * 10 - 1,
|
|
port->max_size / PORT_MMAP_MIN_SIZE);
|
|
}
|
|
|
|
if (msg->port_msg.tracking) {
|
|
iov[0].iov_len += sizeof(msg->tracking_msg);
|
|
}
|
|
|
|
nxt_sendbuf_mem_coalesce(task, &sb);
|
|
|
|
plain_size = sb.size;
|
|
|
|
/*
|
|
* Send through mmap enabled only when payload
|
|
* is bigger than PORT_MMAP_MIN_SIZE.
|
|
*/
|
|
if (m == NXT_PORT_METHOD_MMAP && plain_size > PORT_MMAP_MIN_SIZE) {
|
|
nxt_port_mmap_write(task, port, msg, &sb);
|
|
|
|
} else {
|
|
m = NXT_PORT_METHOD_PLAIN;
|
|
}
|
|
|
|
msg->port_msg.last |= sb.last;
|
|
msg->port_msg.mf = sb.limit_reached || sb.nmax_reached;
|
|
|
|
n = nxt_socketpair_send(&port->socket, msg->fd, iov, sb.niov + 1);
|
|
|
|
if (n > 0) {
|
|
if (nxt_slow_path((size_t) n != sb.size + iov[0].iov_len)) {
|
|
nxt_log(task, NXT_LOG_CRIT,
|
|
"port %d: short write: %z instead of %uz",
|
|
port->socket.fd, n, sb.size + iov[0].iov_len);
|
|
goto fail;
|
|
}
|
|
|
|
if (msg->fd != -1 && msg->close_fd != 0) {
|
|
nxt_fd_close(msg->fd);
|
|
|
|
msg->fd = -1;
|
|
}
|
|
|
|
msg->buf = nxt_sendbuf_completion(task, wq, msg->buf, plain_size,
|
|
m == NXT_PORT_METHOD_MMAP);
|
|
|
|
if (msg->buf != NULL) {
|
|
nxt_debug(task, "port %d: frag stream #%uD", port->socket.fd,
|
|
msg->port_msg.stream);
|
|
|
|
/*
|
|
* A file descriptor is sent only
|
|
* in the first message of a stream.
|
|
*/
|
|
msg->fd = -1;
|
|
msg->share += n;
|
|
msg->port_msg.nf = 1;
|
|
msg->port_msg.tracking = 0;
|
|
|
|
if (msg->share >= port->max_share) {
|
|
msg->share = 0;
|
|
|
|
if (msg->link.next != NULL) {
|
|
nxt_queue_remove(&msg->link);
|
|
use_delta--;
|
|
}
|
|
data = NULL;
|
|
|
|
if (nxt_port_msg_push(task, port, msg) != NULL) {
|
|
use_delta++;
|
|
}
|
|
}
|
|
|
|
} else {
|
|
if (msg->link.next != NULL) {
|
|
nxt_queue_remove(&msg->link);
|
|
use_delta--;
|
|
nxt_work_queue_add(wq, nxt_port_release_send_msg, task, msg,
|
|
msg->work.data);
|
|
}
|
|
data = NULL;
|
|
}
|
|
|
|
} else if (nxt_slow_path(n == NXT_ERROR)) {
|
|
if (msg->link.next == NULL) {
|
|
if (nxt_port_msg_push(task, port, msg) != NULL) {
|
|
use_delta++;
|
|
}
|
|
}
|
|
goto fail;
|
|
}
|
|
|
|
/* n == NXT_AGAIN */
|
|
|
|
} while (port->socket.write_ready);
|
|
|
|
if (nxt_fd_event_is_disabled(port->socket.write)) {
|
|
enable_write = 1;
|
|
}
|
|
|
|
goto unlock_mutex;
|
|
|
|
fail:
|
|
|
|
use_delta++;
|
|
|
|
nxt_work_queue_add(wq, nxt_port_error_handler, task, &port->socket,
|
|
&port->socket);
|
|
|
|
unlock_mutex:
|
|
nxt_thread_mutex_unlock(&port->write_mutex);
|
|
|
|
if (block_write && nxt_fd_event_is_active(port->socket.write)) {
|
|
nxt_port_post(task, port, nxt_port_fd_block_write, NULL);
|
|
}
|
|
|
|
if (enable_write) {
|
|
nxt_port_post(task, port, nxt_port_fd_enable_write, NULL);
|
|
}
|
|
|
|
if (use_delta != 0) {
|
|
nxt_port_use(task, port, use_delta);
|
|
}
|
|
}
|
|
|
|
|
|
void
|
|
nxt_port_read_enable(nxt_task_t *task, nxt_port_t *port)
|
|
{
|
|
port->socket.fd = port->pair[0];
|
|
port->socket.log = &nxt_main_log;
|
|
|
|
port->engine = task->thread->engine;
|
|
|
|
port->socket.read_work_queue = &port->engine->fast_work_queue;
|
|
port->socket.read_handler = nxt_port_read_handler;
|
|
port->socket.error_handler = nxt_port_error_handler;
|
|
|
|
nxt_fd_event_enable_read(port->engine, &port->socket);
|
|
}
|
|
|
|
|
|
void
|
|
nxt_port_read_close(nxt_port_t *port)
|
|
{
|
|
port->socket.read_ready = 0;
|
|
nxt_socket_close(port->socket.task, port->pair[0]);
|
|
port->pair[0] = -1;
|
|
}
|
|
|
|
|
|
static void
|
|
nxt_port_read_handler(nxt_task_t *task, void *obj, void *data)
|
|
{
|
|
ssize_t n;
|
|
nxt_buf_t *b;
|
|
nxt_port_t *port;
|
|
struct iovec iov[2];
|
|
nxt_port_recv_msg_t msg;
|
|
|
|
port = msg.port = nxt_container_of(obj, nxt_port_t, socket);
|
|
|
|
nxt_assert(port->engine == task->thread->engine);
|
|
|
|
for ( ;; ) {
|
|
|
|
b = nxt_port_buf_alloc(port);
|
|
|
|
if (nxt_slow_path(b == NULL)) {
|
|
/* TODO: disable event for some time */
|
|
}
|
|
|
|
iov[0].iov_base = &msg.port_msg;
|
|
iov[0].iov_len = sizeof(nxt_port_msg_t);
|
|
|
|
iov[1].iov_base = b->mem.pos;
|
|
iov[1].iov_len = port->max_size;
|
|
|
|
n = nxt_socketpair_recv(&port->socket, &msg.fd, iov, 2);
|
|
|
|
if (n > 0) {
|
|
|
|
msg.buf = b;
|
|
msg.size = n;
|
|
|
|
nxt_port_read_msg_process(task, port, &msg);
|
|
|
|
/*
|
|
* To disable instant completion or buffer re-usage,
|
|
* handler should reset 'msg.buf'.
|
|
*/
|
|
if (msg.buf == b) {
|
|
nxt_port_buf_free(port, b);
|
|
}
|
|
|
|
if (port->socket.read_ready) {
|
|
continue;
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
if (n == NXT_AGAIN) {
|
|
nxt_port_buf_free(port, b);
|
|
|
|
nxt_fd_event_enable_read(task->thread->engine, &port->socket);
|
|
return;
|
|
}
|
|
|
|
/* n == 0 || n == NXT_ERROR */
|
|
|
|
nxt_work_queue_add(&task->thread->engine->fast_work_queue,
|
|
nxt_port_error_handler, task, &port->socket, NULL);
|
|
return;
|
|
}
|
|
}
|
|
|
|
|
|
static nxt_int_t
|
|
nxt_port_lvlhsh_frag_test(nxt_lvlhsh_query_t *lhq, void *data)
|
|
{
|
|
nxt_port_recv_msg_t *fmsg;
|
|
|
|
fmsg = data;
|
|
|
|
if (lhq->key.length == sizeof(uint32_t)
|
|
&& *(uint32_t *) lhq->key.start == fmsg->port_msg.stream)
|
|
{
|
|
return NXT_OK;
|
|
}
|
|
|
|
return NXT_DECLINED;
|
|
}
|
|
|
|
|
|
static void *
|
|
nxt_port_lvlhsh_frag_alloc(void *ctx, size_t size)
|
|
{
|
|
return nxt_mp_alloc(ctx, size);
|
|
}
|
|
|
|
|
|
static void
|
|
nxt_port_lvlhsh_frag_free(void *ctx, void *p)
|
|
{
|
|
nxt_mp_free(ctx, p);
|
|
}
|
|
|
|
|
|
static const nxt_lvlhsh_proto_t lvlhsh_frag_proto nxt_aligned(64) = {
|
|
NXT_LVLHSH_DEFAULT,
|
|
nxt_port_lvlhsh_frag_test,
|
|
nxt_port_lvlhsh_frag_alloc,
|
|
nxt_port_lvlhsh_frag_free,
|
|
};
|
|
|
|
|
|
static nxt_port_recv_msg_t *
|
|
nxt_port_frag_start(nxt_task_t *task, nxt_port_t *port,
|
|
nxt_port_recv_msg_t *msg)
|
|
{
|
|
nxt_int_t res;
|
|
nxt_lvlhsh_query_t lhq;
|
|
nxt_port_recv_msg_t *fmsg;
|
|
|
|
nxt_debug(task, "start frag stream #%uD", msg->port_msg.stream);
|
|
|
|
fmsg = nxt_mp_alloc(port->mem_pool, sizeof(nxt_port_recv_msg_t));
|
|
|
|
if (nxt_slow_path(fmsg == NULL)) {
|
|
return NULL;
|
|
}
|
|
|
|
*fmsg = *msg;
|
|
|
|
lhq.key_hash = nxt_murmur_hash2(&fmsg->port_msg.stream, sizeof(uint32_t));
|
|
lhq.key.length = sizeof(uint32_t);
|
|
lhq.key.start = (u_char *) &fmsg->port_msg.stream;
|
|
lhq.proto = &lvlhsh_frag_proto;
|
|
lhq.replace = 0;
|
|
lhq.value = fmsg;
|
|
lhq.pool = port->mem_pool;
|
|
|
|
res = nxt_lvlhsh_insert(&port->frags, &lhq);
|
|
|
|
switch (res) {
|
|
|
|
case NXT_OK:
|
|
return fmsg;
|
|
|
|
case NXT_DECLINED:
|
|
nxt_log(task, NXT_LOG_WARN, "duplicate frag stream #%uD",
|
|
fmsg->port_msg.stream);
|
|
nxt_mp_free(port->mem_pool, fmsg);
|
|
|
|
return NULL;
|
|
|
|
default:
|
|
nxt_log(task, NXT_LOG_WARN, "failed to add frag stream #%uD",
|
|
fmsg->port_msg.stream);
|
|
|
|
nxt_mp_free(port->mem_pool, fmsg);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
}
|
|
|
|
|
|
static nxt_port_recv_msg_t *
|
|
nxt_port_frag_find(nxt_task_t *task, nxt_port_t *port, uint32_t stream,
|
|
nxt_bool_t last)
|
|
{
|
|
nxt_int_t res;
|
|
nxt_lvlhsh_query_t lhq;
|
|
|
|
nxt_debug(task, "%s frag stream #%uD", last ? "last" : "next", stream);
|
|
|
|
lhq.key_hash = nxt_murmur_hash2(&stream, sizeof(uint32_t));
|
|
lhq.key.length = sizeof(uint32_t);
|
|
lhq.key.start = (u_char *) &stream;
|
|
lhq.proto = &lvlhsh_frag_proto;
|
|
lhq.pool = port->mem_pool;
|
|
|
|
res = last != 0 ? nxt_lvlhsh_delete(&port->frags, &lhq) :
|
|
nxt_lvlhsh_find(&port->frags, &lhq);
|
|
|
|
switch (res) {
|
|
|
|
case NXT_OK:
|
|
return lhq.value;
|
|
|
|
default:
|
|
nxt_log(task, NXT_LOG_WARN, "frag stream #%uD not found", stream);
|
|
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
|
|
static void
|
|
nxt_port_read_msg_process(nxt_task_t *task, nxt_port_t *port,
|
|
nxt_port_recv_msg_t *msg)
|
|
{
|
|
nxt_buf_t *b, *orig_b;
|
|
nxt_port_recv_msg_t *fmsg;
|
|
|
|
if (nxt_slow_path(msg->size < sizeof(nxt_port_msg_t))) {
|
|
nxt_log(task, NXT_LOG_CRIT,
|
|
"port %d: too small message:%uz", port->socket.fd, msg->size);
|
|
|
|
if (msg->fd != -1) {
|
|
nxt_fd_close(msg->fd);
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
/* adjust size to actual buffer used size */
|
|
msg->size -= sizeof(nxt_port_msg_t);
|
|
|
|
b = orig_b = msg->buf;
|
|
b->mem.free += msg->size;
|
|
|
|
if (msg->port_msg.tracking) {
|
|
msg->cancelled = nxt_port_mmap_tracking_read(task, msg) == 0;
|
|
|
|
} else {
|
|
msg->cancelled = 0;
|
|
}
|
|
|
|
if (nxt_slow_path(msg->port_msg.nf != 0)) {
|
|
|
|
fmsg = nxt_port_frag_find(task, port, msg->port_msg.stream,
|
|
msg->port_msg.mf == 0);
|
|
|
|
nxt_assert(fmsg != NULL);
|
|
|
|
if (nxt_fast_path(fmsg->cancelled == 0)) {
|
|
|
|
if (msg->port_msg.mmap) {
|
|
nxt_port_mmap_read(task, msg);
|
|
}
|
|
|
|
nxt_buf_chain_add(&fmsg->buf, msg->buf);
|
|
|
|
fmsg->size += msg->size;
|
|
msg->buf = NULL;
|
|
b = NULL;
|
|
|
|
if (nxt_fast_path(msg->port_msg.mf == 0)) {
|
|
|
|
b = fmsg->buf;
|
|
|
|
port->handler(task, fmsg);
|
|
|
|
msg->buf = fmsg->buf;
|
|
msg->fd = fmsg->fd;
|
|
}
|
|
}
|
|
|
|
if (nxt_fast_path(msg->port_msg.mf == 0)) {
|
|
nxt_mp_free(port->mem_pool, fmsg);
|
|
}
|
|
} else {
|
|
if (nxt_slow_path(msg->port_msg.mf != 0)) {
|
|
|
|
if (msg->port_msg.mmap && msg->cancelled == 0) {
|
|
nxt_port_mmap_read(task, msg);
|
|
b = msg->buf;
|
|
}
|
|
|
|
fmsg = nxt_port_frag_start(task, port, msg);
|
|
|
|
nxt_assert(fmsg != NULL);
|
|
|
|
fmsg->port_msg.nf = 0;
|
|
fmsg->port_msg.mf = 0;
|
|
|
|
if (nxt_fast_path(msg->cancelled == 0)) {
|
|
msg->buf = NULL;
|
|
msg->fd = -1;
|
|
b = NULL;
|
|
|
|
} else {
|
|
if (msg->fd != -1) {
|
|
nxt_fd_close(msg->fd);
|
|
}
|
|
}
|
|
} else {
|
|
if (nxt_fast_path(msg->cancelled == 0)) {
|
|
|
|
if (msg->port_msg.mmap) {
|
|
nxt_port_mmap_read(task, msg);
|
|
b = msg->buf;
|
|
}
|
|
|
|
port->handler(task, msg);
|
|
}
|
|
}
|
|
}
|
|
|
|
if (msg->port_msg.mmap && orig_b != b) {
|
|
|
|
/*
|
|
* To disable instant buffer completion,
|
|
* handler should reset 'msg->buf'.
|
|
*/
|
|
if (msg->buf == b) {
|
|
/* complete mmap buffers */
|
|
for (; b != NULL; b = b->next) {
|
|
nxt_debug(task, "complete buffer %p", b);
|
|
|
|
nxt_work_queue_add(port->socket.read_work_queue,
|
|
b->completion_handler, task, b, b->parent);
|
|
}
|
|
}
|
|
|
|
/* restore original buf */
|
|
msg->buf = orig_b;
|
|
}
|
|
}
|
|
|
|
|
|
static nxt_buf_t *
|
|
nxt_port_buf_alloc(nxt_port_t *port)
|
|
{
|
|
nxt_buf_t *b;
|
|
|
|
if (port->free_bufs != NULL) {
|
|
b = port->free_bufs;
|
|
port->free_bufs = b->next;
|
|
|
|
b->mem.pos = b->mem.start;
|
|
b->mem.free = b->mem.start;
|
|
b->next = NULL;
|
|
} else {
|
|
b = nxt_buf_mem_alloc(port->mem_pool, port->max_size, 0);
|
|
if (nxt_slow_path(b == NULL)) {
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
return b;
|
|
}
|
|
|
|
|
|
static void
|
|
nxt_port_buf_free(nxt_port_t *port, nxt_buf_t *b)
|
|
{
|
|
b->next = port->free_bufs;
|
|
port->free_bufs = b;
|
|
}
|
|
|
|
|
|
static void
|
|
nxt_port_error_handler(nxt_task_t *task, void *obj, void *data)
|
|
{
|
|
int use_delta;
|
|
nxt_buf_t *b;
|
|
nxt_port_t *port;
|
|
nxt_work_queue_t *wq;
|
|
nxt_port_send_msg_t *msg;
|
|
|
|
nxt_debug(task, "port error handler %p", obj);
|
|
/* TODO */
|
|
|
|
port = nxt_container_of(obj, nxt_port_t, socket);
|
|
|
|
use_delta = 0;
|
|
|
|
if (obj == data) {
|
|
use_delta--;
|
|
}
|
|
|
|
wq = &task->thread->engine->fast_work_queue;
|
|
|
|
nxt_thread_mutex_lock(&port->write_mutex);
|
|
|
|
nxt_queue_each(msg, &port->messages, nxt_port_send_msg_t, link) {
|
|
|
|
for(b = msg->buf; b != NULL; b = b->next) {
|
|
if (nxt_buf_is_sync(b)) {
|
|
continue;
|
|
}
|
|
|
|
nxt_work_queue_add(wq, b->completion_handler, task, b, b->parent);
|
|
}
|
|
|
|
nxt_queue_remove(&msg->link);
|
|
use_delta--;
|
|
nxt_work_queue_add(wq, nxt_port_release_send_msg, task, msg,
|
|
msg->work.data);
|
|
|
|
} nxt_queue_loop;
|
|
|
|
nxt_thread_mutex_unlock(&port->write_mutex);
|
|
|
|
if (use_delta != 0) {
|
|
nxt_port_use(task, port, use_delta);
|
|
}
|
|
}
|