2019-11-05 13:55:54 -08:00
|
|
|
/*
|
|
|
|
* Copyright (C) Internet Systems Consortium, Inc. ("ISC")
|
|
|
|
*
|
|
|
|
* This Source Code Form is subject to the terms of the Mozilla Public
|
|
|
|
* License, v. 2.0. If a copy of the MPL was not distributed with this
|
|
|
|
* file, You can obtain one at http://mozilla.org/MPL/2.0/.
|
|
|
|
*
|
|
|
|
* See the COPYRIGHT file distributed with this work for additional
|
|
|
|
* information regarding copyright ownership.
|
|
|
|
*/
|
|
|
|
|
2019-11-25 09:10:29 +01:00
|
|
|
#pragma once
|
|
|
|
|
2019-11-05 13:55:54 -08:00
|
|
|
#include <unistd.h>
|
|
|
|
#include <uv.h>
|
|
|
|
|
|
|
|
#include <isc/astack.h>
|
|
|
|
#include <isc/atomic.h>
|
|
|
|
#include <isc/buffer.h>
|
|
|
|
#include <isc/condition.h>
|
|
|
|
#include <isc/magic.h>
|
|
|
|
#include <isc/mem.h>
|
|
|
|
#include <isc/netmgr.h>
|
|
|
|
#include <isc/queue.h>
|
|
|
|
#include <isc/random.h>
|
|
|
|
#include <isc/refcount.h>
|
|
|
|
#include <isc/region.h>
|
|
|
|
#include <isc/result.h>
|
|
|
|
#include <isc/sockaddr.h>
|
|
|
|
#include <isc/thread.h>
|
|
|
|
#include <isc/util.h>
|
2020-01-07 12:02:41 +01:00
|
|
|
#include "uv-compat.h"
|
2019-11-05 13:55:54 -08:00
|
|
|
|
|
|
|
#define ISC_NETMGR_TID_UNKNOWN -1
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Single network event loop worker.
|
|
|
|
*/
|
|
|
|
typedef struct isc__networker {
|
|
|
|
isc_nm_t * mgr;
|
|
|
|
int id; /* thread id */
|
|
|
|
uv_loop_t loop; /* libuv loop structure */
|
|
|
|
uv_async_t async; /* async channel to send
|
|
|
|
* data to this networker */
|
|
|
|
isc_mutex_t lock;
|
|
|
|
isc_condition_t cond;
|
|
|
|
bool paused;
|
|
|
|
bool finished;
|
|
|
|
isc_thread_t thread;
|
2019-12-02 13:54:44 +01:00
|
|
|
isc_queue_t *ievents; /* incoming async events */
|
|
|
|
isc_queue_t *ievents_prio; /* priority async events
|
|
|
|
* used for listening etc.
|
|
|
|
* can be processed while
|
|
|
|
* worker is paused */
|
2019-11-05 13:55:54 -08:00
|
|
|
isc_refcount_t references;
|
|
|
|
atomic_int_fast64_t pktcount;
|
2019-11-15 13:22:13 -08:00
|
|
|
char recvbuf[65536];
|
|
|
|
bool recvbuf_inuse;
|
2019-11-05 13:55:54 -08:00
|
|
|
} isc__networker_t;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* A general handle for a connection bound to a networker. For UDP
|
|
|
|
* connections we have peer address here, so both TCP and UDP can be
|
|
|
|
* handled with a simple send-like function
|
|
|
|
*/
|
|
|
|
#define NMHANDLE_MAGIC ISC_MAGIC('N', 'M', 'H', 'D')
|
|
|
|
#define VALID_NMHANDLE(t) ISC_MAGIC_VALID(t, \
|
|
|
|
NMHANDLE_MAGIC)
|
|
|
|
|
|
|
|
typedef void (*isc__nm_closecb)(isc_nmhandle_t *);
|
|
|
|
|
|
|
|
struct isc_nmhandle {
|
|
|
|
int magic;
|
|
|
|
isc_refcount_t references;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The socket is not 'attached' in the traditional
|
|
|
|
* reference-counting sense. Instead, we keep all handles in an
|
|
|
|
* array in the socket object. This way, we don't have circular
|
|
|
|
* dependencies and we can close all handles when we're destroying
|
|
|
|
* the socket.
|
|
|
|
*/
|
|
|
|
isc_nmsocket_t *sock;
|
|
|
|
size_t ah_pos; /* Position in the socket's
|
|
|
|
* 'active handles' array */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The handle is 'inflight' if netmgr is not currently processing
|
|
|
|
* it in any way - it might mean that e.g. a recursive resolution
|
|
|
|
* is happening. For an inflight handle we must wait for the
|
|
|
|
* calling code to finish before we can free it.
|
|
|
|
*/
|
|
|
|
atomic_bool inflight;
|
|
|
|
|
|
|
|
isc_sockaddr_t peer;
|
|
|
|
isc_sockaddr_t local;
|
2019-11-08 10:52:49 -08:00
|
|
|
isc_nm_opaquecb_t doreset; /* reset extra callback, external */
|
|
|
|
isc_nm_opaquecb_t dofree; /* free extra callback, external */
|
2019-11-05 13:55:54 -08:00
|
|
|
void * opaque;
|
|
|
|
char extra[];
|
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* An interface - an address we can listen on.
|
|
|
|
*/
|
|
|
|
struct isc_nmiface {
|
|
|
|
isc_sockaddr_t addr;
|
|
|
|
};
|
|
|
|
|
|
|
|
typedef enum isc__netievent_type {
|
|
|
|
netievent_udpsend,
|
|
|
|
netievent_udprecv,
|
|
|
|
netievent_tcpconnect,
|
|
|
|
netievent_tcpsend,
|
|
|
|
netievent_tcprecv,
|
|
|
|
netievent_tcpstartread,
|
|
|
|
netievent_tcppauseread,
|
2019-11-28 10:21:34 +01:00
|
|
|
netievent_tcpchildlisten,
|
2019-12-09 12:24:46 -08:00
|
|
|
netievent_tcpchildstop,
|
2019-11-19 11:56:00 +01:00
|
|
|
netievent_closecb,
|
2019-11-22 14:13:19 +01:00
|
|
|
netievent_shutdown,
|
2019-12-02 13:54:44 +01:00
|
|
|
netievent_stop,
|
2019-12-09 12:24:46 -08:00
|
|
|
netievent_udpstop,
|
|
|
|
netievent_tcpstop,
|
2019-12-02 13:54:44 +01:00
|
|
|
netievent_tcpclose,
|
2019-12-06 22:25:52 +01:00
|
|
|
netievent_tcpdnsclose,
|
2019-12-03 00:07:59 -08:00
|
|
|
netievent_prio = 0xff, /* event type values higher than this
|
|
|
|
* will be treated as high-priority
|
|
|
|
* events, which can be processed
|
|
|
|
* while the netmgr is paused.
|
|
|
|
*/
|
2019-12-02 13:54:44 +01:00
|
|
|
netievent_udplisten,
|
|
|
|
netievent_tcplisten,
|
2019-11-05 13:55:54 -08:00
|
|
|
} isc__netievent_type;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We have to split it because we can read and write on a socket
|
|
|
|
* simultaneously.
|
|
|
|
*/
|
|
|
|
typedef union {
|
|
|
|
isc_nm_recv_cb_t recv;
|
|
|
|
isc_nm_cb_t accept;
|
|
|
|
} isc__nm_readcb_t;
|
|
|
|
|
|
|
|
typedef union {
|
|
|
|
isc_nm_cb_t send;
|
|
|
|
isc_nm_cb_t connect;
|
|
|
|
} isc__nm_writecb_t;
|
|
|
|
|
|
|
|
typedef union {
|
|
|
|
isc_nm_recv_cb_t recv;
|
|
|
|
isc_nm_cb_t accept;
|
|
|
|
isc_nm_cb_t send;
|
|
|
|
isc_nm_cb_t connect;
|
|
|
|
} isc__nm_cb_t;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Wrapper around uv_req_t with 'our' fields in it. req->data should
|
|
|
|
* always point to its parent. Note that we always allocate more than
|
|
|
|
* sizeof(struct) because we make room for different req types;
|
|
|
|
*/
|
|
|
|
#define UVREQ_MAGIC ISC_MAGIC('N', 'M', 'U', 'R')
|
|
|
|
#define VALID_UVREQ(t) ISC_MAGIC_VALID(t, UVREQ_MAGIC)
|
|
|
|
|
|
|
|
typedef struct isc__nm_uvreq {
|
|
|
|
int magic;
|
|
|
|
isc_nmsocket_t * sock;
|
|
|
|
isc_nmhandle_t * handle;
|
|
|
|
uv_buf_t uvbuf; /* translated isc_region_t, to be
|
2019-12-07 01:48:11 +01:00
|
|
|
* sent or received */
|
2019-11-05 13:55:54 -08:00
|
|
|
isc_sockaddr_t local; /* local address */
|
|
|
|
isc_sockaddr_t peer; /* peer address */
|
|
|
|
isc__nm_cb_t cb; /* callback */
|
|
|
|
void * cbarg; /* callback argument */
|
2019-12-07 01:48:11 +01:00
|
|
|
uv_pipe_t ipc; /* used for sending socket
|
|
|
|
* uv_handles to other threads */
|
2019-11-05 13:55:54 -08:00
|
|
|
union {
|
|
|
|
uv_req_t req;
|
|
|
|
uv_getaddrinfo_t getaddrinfo;
|
|
|
|
uv_getnameinfo_t getnameinfo;
|
|
|
|
uv_shutdown_t shutdown;
|
|
|
|
uv_write_t write;
|
|
|
|
uv_connect_t connect;
|
|
|
|
uv_udp_send_t udp_send;
|
|
|
|
uv_fs_t fs;
|
|
|
|
uv_work_t work;
|
|
|
|
} uv_req;
|
|
|
|
} isc__nm_uvreq_t;
|
|
|
|
|
|
|
|
typedef struct isc__netievent__socket {
|
|
|
|
isc__netievent_type type;
|
|
|
|
isc_nmsocket_t *sock;
|
|
|
|
} isc__netievent__socket_t;
|
|
|
|
|
|
|
|
typedef isc__netievent__socket_t isc__netievent_udplisten_t;
|
2019-12-09 12:24:46 -08:00
|
|
|
typedef isc__netievent__socket_t isc__netievent_udpstop_t;
|
|
|
|
typedef isc__netievent__socket_t isc__netievent_tcpstop_t;
|
|
|
|
typedef isc__netievent__socket_t isc__netievent_tcpchildstop_t;
|
2019-11-05 13:55:54 -08:00
|
|
|
typedef isc__netievent__socket_t isc__netievent_tcpclose_t;
|
2019-12-06 22:25:52 +01:00
|
|
|
typedef isc__netievent__socket_t isc__netievent_tcpdnsclose_t;
|
2019-11-05 13:55:54 -08:00
|
|
|
typedef isc__netievent__socket_t isc__netievent_startread_t;
|
|
|
|
typedef isc__netievent__socket_t isc__netievent_pauseread_t;
|
|
|
|
|
|
|
|
typedef struct isc__netievent__socket_req {
|
|
|
|
isc__netievent_type type;
|
|
|
|
isc_nmsocket_t *sock;
|
|
|
|
isc__nm_uvreq_t *req;
|
|
|
|
} isc__netievent__socket_req_t;
|
|
|
|
|
|
|
|
typedef isc__netievent__socket_req_t isc__netievent_tcpconnect_t;
|
|
|
|
typedef isc__netievent__socket_req_t isc__netievent_tcplisten_t;
|
|
|
|
typedef isc__netievent__socket_req_t isc__netievent_tcpsend_t;
|
|
|
|
|
2020-01-07 12:02:41 +01:00
|
|
|
typedef struct isc__netievent__socket_streaminfo {
|
|
|
|
isc__netievent_type type;
|
|
|
|
isc_nmsocket_t *sock;
|
|
|
|
isc_uv_stream_info_t streaminfo;
|
|
|
|
} isc__netievent__socket_streaminfo_t;
|
|
|
|
|
|
|
|
typedef isc__netievent__socket_streaminfo_t isc__netievent_tcpchildlisten_t;
|
|
|
|
|
|
|
|
|
2019-12-08 22:44:08 +01:00
|
|
|
typedef struct isc__netievent__socket_handle {
|
|
|
|
isc__netievent_type type;
|
|
|
|
isc_nmsocket_t *sock;
|
|
|
|
isc_nmhandle_t *handle;
|
|
|
|
} isc__netievent__socket_handle_t;
|
|
|
|
|
|
|
|
typedef isc__netievent__socket_handle_t isc__netievent_closecb_t;
|
|
|
|
|
|
|
|
|
2019-11-05 13:55:54 -08:00
|
|
|
typedef struct isc__netievent_udpsend {
|
|
|
|
isc__netievent_type type;
|
|
|
|
isc_nmsocket_t *sock;
|
|
|
|
isc_sockaddr_t peer;
|
|
|
|
isc__nm_uvreq_t *req;
|
|
|
|
} isc__netievent_udpsend_t;
|
|
|
|
|
|
|
|
typedef struct isc__netievent {
|
|
|
|
isc__netievent_type type;
|
|
|
|
} isc__netievent_t;
|
|
|
|
|
2019-11-22 14:13:19 +01:00
|
|
|
typedef isc__netievent_t isc__netievent_shutdown_t;
|
|
|
|
typedef isc__netievent_t isc__netievent_stop_t;
|
|
|
|
|
2019-11-05 13:55:54 -08:00
|
|
|
typedef union {
|
|
|
|
isc__netievent_t ni;
|
2019-11-22 14:13:19 +01:00
|
|
|
isc__netievent__socket_t nis;
|
|
|
|
isc__netievent__socket_req_t nisr;
|
2019-11-05 13:55:54 -08:00
|
|
|
isc__netievent_udpsend_t nius;
|
|
|
|
} isc__netievent_storage_t;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Network manager
|
|
|
|
*/
|
|
|
|
#define NM_MAGIC ISC_MAGIC('N', 'E', 'T', 'M')
|
|
|
|
#define VALID_NM(t) ISC_MAGIC_VALID(t, NM_MAGIC)
|
|
|
|
|
|
|
|
struct isc_nm {
|
|
|
|
int magic;
|
|
|
|
isc_refcount_t references;
|
|
|
|
isc_mem_t *mctx;
|
|
|
|
uint32_t nworkers;
|
|
|
|
isc_mutex_t lock;
|
|
|
|
isc_condition_t wkstatecond;
|
|
|
|
isc__networker_t *workers;
|
2019-11-21 17:08:06 -08:00
|
|
|
|
|
|
|
isc_mempool_t *reqpool;
|
|
|
|
isc_mutex_t reqlock;
|
|
|
|
|
|
|
|
isc_mempool_t *evpool;
|
|
|
|
isc_mutex_t evlock;
|
|
|
|
|
2019-11-05 13:55:54 -08:00
|
|
|
atomic_uint_fast32_t workers_running;
|
|
|
|
atomic_uint_fast32_t workers_paused;
|
|
|
|
atomic_uint_fast32_t maxudp;
|
|
|
|
atomic_bool paused;
|
|
|
|
|
2019-11-22 15:57:42 -08:00
|
|
|
/*
|
|
|
|
* Acive connections are being closed and new connections are
|
|
|
|
* no longer allowed.
|
|
|
|
*/
|
|
|
|
atomic_bool closing;
|
|
|
|
|
2019-11-05 13:55:54 -08:00
|
|
|
/*
|
|
|
|
* A worker is actively waiting for other workers, for example to
|
|
|
|
* stop listening; that means no other thread can do the same thing
|
|
|
|
* or pause, or we'll deadlock. We have to either re-enqueue our
|
|
|
|
* event or wait for the other one to finish if we want to pause.
|
|
|
|
*/
|
|
|
|
atomic_bool interlocked;
|
2019-11-20 22:33:35 +01:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Timeout values for TCP connections, coresponding to
|
|
|
|
* tcp-intiial-timeout, tcp-idle-timeout, tcp-keepalive-timeout,
|
|
|
|
* and tcp-advertised-timeout. Note that these are stored in
|
|
|
|
* milliseconds so they can be used directly with the libuv timer,
|
|
|
|
* but they are configured in tenths of seconds.
|
|
|
|
*/
|
|
|
|
uint32_t init;
|
|
|
|
uint32_t idle;
|
|
|
|
uint32_t keepalive;
|
|
|
|
uint32_t advertised;
|
2019-11-05 13:55:54 -08:00
|
|
|
};
|
|
|
|
|
|
|
|
typedef enum isc_nmsocket_type {
|
|
|
|
isc_nm_udpsocket,
|
|
|
|
isc_nm_udplistener, /* Aggregate of nm_udpsocks */
|
|
|
|
isc_nm_tcpsocket,
|
|
|
|
isc_nm_tcplistener,
|
2019-11-28 10:21:34 +01:00
|
|
|
isc_nm_tcpchildlistener,
|
2019-11-05 13:55:54 -08:00
|
|
|
isc_nm_tcpdnslistener,
|
|
|
|
isc_nm_tcpdnssocket
|
|
|
|
} isc_nmsocket_type;
|
|
|
|
|
|
|
|
/*%
|
|
|
|
* A universal structure for either a single socket or a group of
|
|
|
|
* dup'd/SO_REUSE_PORT-using sockets listening on the same interface.
|
|
|
|
*/
|
|
|
|
#define NMSOCK_MAGIC ISC_MAGIC('N', 'M', 'S', 'K')
|
|
|
|
#define VALID_NMSOCK(t) ISC_MAGIC_VALID(t, NMSOCK_MAGIC)
|
|
|
|
|
|
|
|
struct isc_nmsocket {
|
|
|
|
/*% Unlocked, RO */
|
|
|
|
int magic;
|
|
|
|
int tid;
|
|
|
|
isc_nmsocket_type type;
|
|
|
|
isc_nm_t *mgr;
|
|
|
|
isc_nmsocket_t *parent;
|
2019-11-22 13:19:45 +01:00
|
|
|
|
2019-12-03 00:07:59 -08:00
|
|
|
/*%
|
2019-11-22 13:19:45 +01:00
|
|
|
* quota is the TCP client, attached when a TCP connection
|
|
|
|
* is established. pquota is a non-attached pointer to the
|
|
|
|
* TCP client quota, stored in listening sockets but only
|
|
|
|
* attached in connected sockets.
|
|
|
|
*/
|
2019-11-05 13:55:54 -08:00
|
|
|
isc_quota_t *quota;
|
2019-11-22 13:19:45 +01:00
|
|
|
isc_quota_t *pquota;
|
2019-11-05 13:55:54 -08:00
|
|
|
bool overquota;
|
2019-11-22 13:19:45 +01:00
|
|
|
|
2019-12-03 00:07:59 -08:00
|
|
|
/*%
|
2019-11-22 13:19:45 +01:00
|
|
|
* TCP read timeout timer.
|
|
|
|
*/
|
2019-11-19 11:56:00 +01:00
|
|
|
uv_timer_t timer;
|
|
|
|
bool timer_initialized;
|
|
|
|
uint64_t read_timeout;
|
2019-11-05 13:55:54 -08:00
|
|
|
|
|
|
|
/*% outer socket is for 'wrapped' sockets - e.g. tcpdns in tcp */
|
|
|
|
isc_nmsocket_t *outer;
|
|
|
|
|
|
|
|
/*% server socket for connections */
|
|
|
|
isc_nmsocket_t *server;
|
|
|
|
|
2019-12-03 00:07:59 -08:00
|
|
|
/*% Child sockets for multi-socket setups */
|
2019-11-05 13:55:54 -08:00
|
|
|
isc_nmsocket_t *children;
|
|
|
|
int nchildren;
|
|
|
|
isc_nmiface_t *iface;
|
|
|
|
isc_nmhandle_t *tcphandle;
|
|
|
|
|
2019-12-03 00:07:59 -08:00
|
|
|
/*% Extra data allocated at the end of each isc_nmhandle_t */
|
2019-11-05 13:55:54 -08:00
|
|
|
size_t extrahandlesize;
|
|
|
|
|
2019-11-22 13:19:45 +01:00
|
|
|
/*% TCP backlog */
|
|
|
|
int backlog;
|
|
|
|
|
2019-11-05 13:55:54 -08:00
|
|
|
/*% libuv data */
|
|
|
|
uv_os_sock_t fd;
|
|
|
|
union uv_any_handle uv_handle;
|
|
|
|
|
2019-12-03 00:07:59 -08:00
|
|
|
/*% Peer address */
|
2019-11-05 13:55:54 -08:00
|
|
|
isc_sockaddr_t peer;
|
|
|
|
|
|
|
|
/* Atomic */
|
2019-12-03 00:07:59 -08:00
|
|
|
/*% Number of running (e.g. listening) child sockets */
|
2019-11-05 13:55:54 -08:00
|
|
|
atomic_int_fast32_t rchildren;
|
|
|
|
|
|
|
|
/*%
|
2019-12-03 00:07:59 -08:00
|
|
|
* Socket is active if it's listening, working, etc. If it's
|
|
|
|
* closing, then it doesn't make a sense, for example, to
|
|
|
|
* push handles or reqs for reuse.
|
2019-11-05 13:55:54 -08:00
|
|
|
*/
|
|
|
|
atomic_bool active;
|
|
|
|
atomic_bool destroying;
|
|
|
|
|
|
|
|
/*%
|
|
|
|
* Socket is closed if it's not active and all the possible
|
|
|
|
* callbacks were fired, there are no active handles, etc.
|
2019-12-03 00:07:59 -08:00
|
|
|
* If active==false but closed==false, that means the socket
|
|
|
|
* is closing.
|
2019-11-05 13:55:54 -08:00
|
|
|
*/
|
|
|
|
atomic_bool closed;
|
|
|
|
atomic_bool listening;
|
2020-01-10 14:25:30 -08:00
|
|
|
atomic_bool listen_error;
|
2019-11-05 13:55:54 -08:00
|
|
|
isc_refcount_t references;
|
|
|
|
|
|
|
|
/*%
|
2019-11-08 10:52:49 -08:00
|
|
|
* TCPDNS socket has been set not to pipeliine.
|
2019-11-05 13:55:54 -08:00
|
|
|
*/
|
|
|
|
atomic_bool sequential;
|
2019-11-08 10:52:49 -08:00
|
|
|
|
|
|
|
/*%
|
|
|
|
* TCPDNS socket has exceeded the maximum number of
|
|
|
|
* simultaneous requests per connecton, so will be temporarily
|
|
|
|
* restricted from pipelining.
|
|
|
|
*/
|
|
|
|
atomic_bool overlimit;
|
|
|
|
|
2019-11-05 13:55:54 -08:00
|
|
|
/*%
|
|
|
|
* TCPDNS socket in sequential mode is currently processing a packet,
|
|
|
|
* we need to wait until it finishes.
|
|
|
|
*/
|
|
|
|
atomic_bool processing;
|
|
|
|
|
2019-11-08 10:52:49 -08:00
|
|
|
/*%
|
|
|
|
* A TCP socket has had isc_nm_pauseread() called.
|
|
|
|
*/
|
|
|
|
atomic_bool readpaused;
|
|
|
|
|
2019-11-20 22:33:35 +01:00
|
|
|
/*%
|
|
|
|
* A TCP or TCPDNS socket has been set to use the keepalive
|
|
|
|
* timeout instead of the default idle timeout.
|
|
|
|
*/
|
|
|
|
atomic_bool keepalive;
|
|
|
|
|
2019-11-05 13:55:54 -08:00
|
|
|
/*%
|
|
|
|
* 'spare' handles for that can be reused to avoid allocations,
|
|
|
|
* for UDP.
|
|
|
|
*/
|
|
|
|
isc_astack_t *inactivehandles;
|
|
|
|
isc_astack_t *inactivereqs;
|
|
|
|
|
2019-12-03 00:07:59 -08:00
|
|
|
/*%
|
|
|
|
* Used to wait for TCP listening events to complete, and
|
|
|
|
* for the number of running children to reach zero during
|
|
|
|
* shutdown.
|
2019-12-02 13:54:44 +01:00
|
|
|
*/
|
2019-11-05 13:55:54 -08:00
|
|
|
isc_mutex_t lock;
|
|
|
|
isc_condition_t cond;
|
|
|
|
|
2019-12-03 00:07:59 -08:00
|
|
|
/*%
|
|
|
|
* Used to pass a result back from TCP listening events.
|
|
|
|
*/
|
2019-12-02 13:54:44 +01:00
|
|
|
isc_result_t result;
|
|
|
|
|
2019-11-05 13:55:54 -08:00
|
|
|
/*%
|
|
|
|
* List of active handles.
|
2019-11-08 10:52:49 -08:00
|
|
|
* ah - current position in 'ah_frees'; this represents the
|
|
|
|
* current number of active handles;
|
|
|
|
* ah_size - size of the 'ah_frees' and 'ah_handles' arrays
|
|
|
|
* ah_handles - array pointers to active handles
|
|
|
|
*
|
2019-11-05 13:55:54 -08:00
|
|
|
* Adding a handle
|
2019-11-08 10:52:49 -08:00
|
|
|
* - if ah == ah_size, reallocate
|
|
|
|
* - x = ah_frees[ah]
|
|
|
|
* - ah_frees[ah++] = 0;
|
2019-11-05 13:55:54 -08:00
|
|
|
* - ah_handles[x] = handle
|
|
|
|
* - x must be stored with the handle!
|
|
|
|
* Removing a handle:
|
2019-11-08 10:52:49 -08:00
|
|
|
* - ah_frees[--ah] = x
|
2019-11-05 13:55:54 -08:00
|
|
|
* - ah_handles[x] = NULL;
|
|
|
|
*
|
2019-11-15 13:22:13 -08:00
|
|
|
* XXX: for now this is locked with socket->lock, but we
|
|
|
|
* might want to change it to something lockless in the
|
|
|
|
* future.
|
2019-11-05 13:55:54 -08:00
|
|
|
*/
|
2019-11-19 11:56:00 +01:00
|
|
|
atomic_int_fast32_t ah;
|
2019-11-05 13:55:54 -08:00
|
|
|
size_t ah_size;
|
|
|
|
size_t *ah_frees;
|
|
|
|
isc_nmhandle_t **ah_handles;
|
|
|
|
|
2019-12-03 00:07:59 -08:00
|
|
|
/*% Buffer for TCPDNS processing */
|
2019-11-05 13:55:54 -08:00
|
|
|
size_t buf_size;
|
|
|
|
size_t buf_len;
|
|
|
|
unsigned char *buf;
|
|
|
|
|
2019-12-03 00:07:59 -08:00
|
|
|
/*%
|
2019-11-08 10:52:49 -08:00
|
|
|
* This function will be called with handle->sock
|
|
|
|
* as the argument whenever a handle's references drop
|
|
|
|
* to zero, after its reset callback has been called.
|
|
|
|
*/
|
|
|
|
isc_nm_opaquecb_t closehandle_cb;
|
|
|
|
|
2019-11-05 13:55:54 -08:00
|
|
|
isc__nm_readcb_t rcb;
|
|
|
|
void *rcbarg;
|
2019-11-25 18:36:14 -03:00
|
|
|
|
|
|
|
isc__nm_cb_t accept_cb;
|
|
|
|
void *accept_cbarg;
|
2019-11-05 13:55:54 -08:00
|
|
|
};
|
|
|
|
|
|
|
|
bool
|
|
|
|
isc__nm_in_netthread(void);
|
|
|
|
/*%
|
|
|
|
* Returns 'true' if we're in the network thread.
|
|
|
|
*/
|
|
|
|
|
|
|
|
void *
|
|
|
|
isc__nm_get_ievent(isc_nm_t *mgr, isc__netievent_type type);
|
|
|
|
/*%<
|
|
|
|
* Allocate an ievent and set the type.
|
|
|
|
*/
|
2019-11-19 11:56:00 +01:00
|
|
|
void
|
|
|
|
isc__nm_put_ievent(isc_nm_t *mgr, void *ievent);
|
2019-11-05 13:55:54 -08:00
|
|
|
|
|
|
|
void
|
|
|
|
isc__nm_enqueue_ievent(isc__networker_t *worker, isc__netievent_t *event);
|
|
|
|
/*%<
|
|
|
|
* Enqueue an ievent onto a specific worker queue. (This the only safe
|
|
|
|
* way to use an isc__networker_t from another thread.)
|
|
|
|
*/
|
|
|
|
|
|
|
|
void
|
|
|
|
isc__nm_alloc_cb(uv_handle_t *handle, size_t size, uv_buf_t *buf);
|
|
|
|
/*%<
|
|
|
|
* Allocator for recv operations.
|
|
|
|
*
|
|
|
|
* Note that as currently implemented, this doesn't actually
|
|
|
|
* allocate anything, it just assigns the the isc__networker's UDP
|
|
|
|
* receive buffer to a socket, and marks it as "in use".
|
|
|
|
*/
|
|
|
|
|
|
|
|
void
|
|
|
|
isc__nm_free_uvbuf(isc_nmsocket_t *sock, const uv_buf_t *buf);
|
|
|
|
/*%<
|
|
|
|
* Free a buffer allocated for a receive operation.
|
|
|
|
*
|
|
|
|
* Note that as currently implemented, this doesn't actually
|
|
|
|
* free anything, marks the isc__networker's UDP receive buffer
|
|
|
|
* as "not in use".
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
|
|
isc_nmhandle_t *
|
|
|
|
isc__nmhandle_get(isc_nmsocket_t *sock, isc_sockaddr_t *peer,
|
|
|
|
isc_sockaddr_t *local);
|
|
|
|
/*%<
|
|
|
|
* Get a handle for the socket 'sock', allocating a new one
|
|
|
|
* if there isn't one availbale in 'sock->inactivehandles'.
|
|
|
|
*
|
|
|
|
* If 'peer' is not NULL, set the handle's peer address to 'peer',
|
|
|
|
* otherwise set it to 'sock->peer'.
|
|
|
|
*
|
|
|
|
* If 'local' is not NULL, set the handle's local address to 'local',
|
|
|
|
* otherwise set it to 'sock->iface->addr'.
|
|
|
|
*/
|
|
|
|
|
|
|
|
isc__nm_uvreq_t *
|
|
|
|
isc__nm_uvreq_get(isc_nm_t *mgr, isc_nmsocket_t *sock);
|
|
|
|
/*%<
|
|
|
|
* Get a UV request structure for the socket 'sock', allocating a
|
|
|
|
* new one if there isn't one availbale in 'sock->inactivereqs'.
|
|
|
|
*/
|
|
|
|
|
|
|
|
void
|
|
|
|
isc__nm_uvreq_put(isc__nm_uvreq_t **req, isc_nmsocket_t *sock);
|
|
|
|
/*%<
|
|
|
|
* Completes the use of a UV request structure, setting '*req' to NULL.
|
|
|
|
*
|
|
|
|
* The UV request is pushed onto the 'sock->inactivereqs' stack or,
|
|
|
|
* if that doesn't work, freed.
|
|
|
|
*/
|
|
|
|
|
|
|
|
void
|
|
|
|
isc__nmsocket_init(isc_nmsocket_t *sock, isc_nm_t *mgr,
|
|
|
|
isc_nmsocket_type type);
|
|
|
|
/*%<
|
|
|
|
* Initialize socket 'sock', attach it to 'mgr', and set it to type 'type'.
|
|
|
|
*/
|
|
|
|
|
|
|
|
void
|
|
|
|
isc__nmsocket_prep_destroy(isc_nmsocket_t *sock);
|
|
|
|
/*%<
|
|
|
|
* Market 'sock' as inactive, close it if necessary, and destroy it
|
|
|
|
* if there are no remaining references or active handles.
|
|
|
|
*/
|
|
|
|
|
2019-11-19 11:56:00 +01:00
|
|
|
void
|
2019-12-09 12:24:46 -08:00
|
|
|
isc__nm_async_closecb(isc__networker_t *worker, isc__netievent_t *ev0);
|
2019-11-19 11:56:00 +01:00
|
|
|
/*%<
|
|
|
|
* Issue a 'handle closed' callback on the socket.
|
|
|
|
*/
|
|
|
|
|
2019-11-22 14:13:19 +01:00
|
|
|
void
|
2019-12-09 12:24:46 -08:00
|
|
|
isc__nm_async_shutdown(isc__networker_t *worker, isc__netievent_t *ev0);
|
2019-11-22 14:13:19 +01:00
|
|
|
/*%<
|
|
|
|
* Walk through all uv handles, get the underlying sockets and issue
|
|
|
|
* close on them.
|
|
|
|
*/
|
|
|
|
|
2019-11-05 13:55:54 -08:00
|
|
|
isc_result_t
|
|
|
|
isc__nm_udp_send(isc_nmhandle_t *handle, isc_region_t *region,
|
|
|
|
isc_nm_cb_t cb, void *cbarg);
|
|
|
|
/*%<
|
|
|
|
* Back-end implemenation of isc_nm_send() for UDP handles.
|
|
|
|
*/
|
|
|
|
|
|
|
|
void
|
2019-12-09 12:24:46 -08:00
|
|
|
isc__nm_async_udplisten(isc__networker_t *worker, isc__netievent_t *ev0);
|
2019-11-05 13:55:54 -08:00
|
|
|
|
|
|
|
void
|
2019-12-09 12:24:46 -08:00
|
|
|
isc__nm_async_udpstop(isc__networker_t *worker, isc__netievent_t *ev0);
|
2019-11-05 13:55:54 -08:00
|
|
|
void
|
2019-12-09 12:24:46 -08:00
|
|
|
isc__nm_async_udpsend(isc__networker_t *worker, isc__netievent_t *ev0);
|
2019-11-05 13:55:54 -08:00
|
|
|
/*%<
|
|
|
|
* Callback handlers for asynchronous UDP events (listen, stoplisten, send).
|
|
|
|
*/
|
|
|
|
|
|
|
|
isc_result_t
|
|
|
|
isc__nm_tcp_send(isc_nmhandle_t *handle, isc_region_t *region,
|
|
|
|
isc_nm_cb_t cb, void *cbarg);
|
|
|
|
/*%<
|
|
|
|
* Back-end implemenation of isc_nm_send() for TCP handles.
|
|
|
|
*/
|
|
|
|
|
|
|
|
void
|
|
|
|
isc__nm_tcp_close(isc_nmsocket_t *sock);
|
|
|
|
/*%<
|
|
|
|
* Close a TCP socket.
|
|
|
|
*/
|
|
|
|
|
2019-11-22 14:13:19 +01:00
|
|
|
void
|
|
|
|
isc__nm_tcp_shutdown(isc_nmsocket_t *sock);
|
|
|
|
/*%<
|
|
|
|
* Called on shutdown to close and clean up a listening TCP socket.
|
|
|
|
*/
|
|
|
|
|
2019-11-05 13:55:54 -08:00
|
|
|
void
|
2019-12-09 12:24:46 -08:00
|
|
|
isc__nm_async_tcpconnect(isc__networker_t *worker, isc__netievent_t *ev0);
|
2019-11-05 13:55:54 -08:00
|
|
|
void
|
2019-12-09 12:24:46 -08:00
|
|
|
isc__nm_async_tcplisten(isc__networker_t *worker, isc__netievent_t *ev0);
|
2019-11-05 13:55:54 -08:00
|
|
|
void
|
2019-12-09 12:24:46 -08:00
|
|
|
isc__nm_async_tcpchildlisten(isc__networker_t *worker, isc__netievent_t *ev0);
|
2019-11-28 10:21:34 +01:00
|
|
|
void
|
2019-12-09 12:24:46 -08:00
|
|
|
isc__nm_async_tcpstop(isc__networker_t *worker, isc__netievent_t *ev0);
|
2019-11-05 13:55:54 -08:00
|
|
|
void
|
2019-12-09 12:24:46 -08:00
|
|
|
isc__nm_async_tcpchildstop(isc__networker_t *worker, isc__netievent_t *ev0);
|
2019-11-28 10:21:34 +01:00
|
|
|
void
|
2019-12-09 12:24:46 -08:00
|
|
|
isc__nm_async_tcpsend(isc__networker_t *worker, isc__netievent_t *ev0);
|
2019-11-05 13:55:54 -08:00
|
|
|
void
|
2019-12-09 12:24:46 -08:00
|
|
|
isc__nm_async_startread(isc__networker_t *worker, isc__netievent_t *ev0);
|
2019-11-05 13:55:54 -08:00
|
|
|
void
|
2019-12-09 12:24:46 -08:00
|
|
|
isc__nm_async_pauseread(isc__networker_t *worker, isc__netievent_t *ev0);
|
2019-11-05 13:55:54 -08:00
|
|
|
void
|
2019-12-09 12:24:46 -08:00
|
|
|
isc__nm_async_tcpclose(isc__networker_t *worker, isc__netievent_t *ev0);
|
2019-11-05 13:55:54 -08:00
|
|
|
/*%<
|
|
|
|
* Callback handlers for asynchronous TCP events (connect, listen,
|
2019-11-22 15:57:42 -08:00
|
|
|
* stoplisten, send, read, pause, close).
|
2019-11-05 13:55:54 -08:00
|
|
|
*/
|
|
|
|
|
|
|
|
isc_result_t
|
|
|
|
isc__nm_tcpdns_send(isc_nmhandle_t *handle, isc_region_t *region,
|
|
|
|
isc_nm_cb_t cb, void *cbarg);
|
|
|
|
/*%<
|
|
|
|
* Back-end implemenation of isc_nm_send() for TCPDNS handles.
|
|
|
|
*/
|
|
|
|
|
|
|
|
void
|
|
|
|
isc__nm_tcpdns_close(isc_nmsocket_t *sock);
|
|
|
|
/*%<
|
|
|
|
* Close a TCPDNS socket.
|
|
|
|
*/
|
|
|
|
|
2019-12-06 22:25:52 +01:00
|
|
|
void
|
2019-12-09 12:24:46 -08:00
|
|
|
isc__nm_async_tcpdnsclose(isc__networker_t *worker, isc__netievent_t *ev0);
|
2019-12-06 22:25:52 +01:00
|
|
|
|
2019-11-05 13:55:54 -08:00
|
|
|
#define isc__nm_uverr2result(x) \
|
|
|
|
isc___nm_uverr2result(x, true, __FILE__, __LINE__)
|
|
|
|
isc_result_t
|
|
|
|
isc___nm_uverr2result(int uverr, bool dolog,
|
|
|
|
const char *file, unsigned int line);
|
|
|
|
/*%<
|
|
|
|
* Convert a libuv error value into an isc_result_t. The
|
|
|
|
* list of supported error values is not complete; new users
|
|
|
|
* of this function should add any expected errors that are
|
|
|
|
* not already there.
|
|
|
|
*/
|
|
|
|
|
|
|
|
bool
|
|
|
|
isc__nm_acquire_interlocked(isc_nm_t *mgr);
|
|
|
|
/*%<
|
|
|
|
* Try to acquire interlocked state; return true if successful.
|
|
|
|
*/
|
|
|
|
|
|
|
|
void
|
|
|
|
isc__nm_drop_interlocked(isc_nm_t *mgr);
|
|
|
|
/*%<
|
|
|
|
* Drop interlocked state; signal waiters.
|
|
|
|
*/
|
|
|
|
|
|
|
|
void
|
|
|
|
isc__nm_acquire_interlocked_force(isc_nm_t *mgr);
|
|
|
|
/*%<
|
|
|
|
* Actively wait for interlocked state.
|
|
|
|
*/
|