/* $NetBSD: netmgr-int.h,v 1.8.2.2 2024/03/11 17:25:20 martin Exp $ */ /* * Copyright (C) Internet Systems Consortium, Inc. ("ISC") * * SPDX-License-Identifier: MPL-2.0 * * This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this * file, you can obtain one at https://mozilla.org/MPL/2.0/. * * See the COPYRIGHT file distributed with this work for additional * information regarding copyright ownership. */ #pragma once #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "uv-compat.h" #define ISC_NETMGR_TID_UNKNOWN -1 /* Must be different from ISC_NETMGR_TID_UNKNOWN */ #define ISC_NETMGR_NON_INTERLOCKED -2 /* * Receive buffers */ #if HAVE_DECL_UV_UDP_MMSG_CHUNK /* * The value 20 here is UV__MMSG_MAXWIDTH taken from the current libuv source, * libuv will not receive more that 20 datagrams in a single recvmmsg call. */ #define ISC_NETMGR_UDP_RECVBUF_SIZE (20 * UINT16_MAX) #else /* * A single DNS message size */ #define ISC_NETMGR_UDP_RECVBUF_SIZE UINT16_MAX #endif /* * The TCP receive buffer can fit one maximum sized DNS message plus its size, * the receive buffer here affects TCP, DoT and DoH. */ #define ISC_NETMGR_TCP_RECVBUF_SIZE (sizeof(uint16_t) + UINT16_MAX) /* Pick the larger buffer */ #define ISC_NETMGR_RECVBUF_SIZE \ (ISC_NETMGR_UDP_RECVBUF_SIZE >= ISC_NETMGR_TCP_RECVBUF_SIZE \ ? ISC_NETMGR_UDP_RECVBUF_SIZE \ : ISC_NETMGR_TCP_RECVBUF_SIZE) /* * Send buffer */ #define ISC_NETMGR_SENDBUF_SIZE (sizeof(uint16_t) + UINT16_MAX) /* * Make sure our RECVBUF size is large enough */ STATIC_ASSERT(ISC_NETMGR_UDP_RECVBUF_SIZE <= ISC_NETMGR_RECVBUF_SIZE, "UDP receive buffer size must be smaller or equal than worker " "receive buffer size"); STATIC_ASSERT(ISC_NETMGR_TCP_RECVBUF_SIZE <= ISC_NETMGR_RECVBUF_SIZE, "TCP receive buffer size must be smaller or equal than worker " "receive buffer size"); /*% * Regular TCP buffer size. */ #define NM_REG_BUF 4096 /*% * Larger buffer for when the regular one isn't enough; this will * hold two full DNS packets with lengths. netmgr receives 64k at * most in TCPDNS or TLSDNS connections, so there's no risk of overrun * when using a buffer this size. */ #define NM_BIG_BUF ISC_NETMGR_TCP_RECVBUF_SIZE * 2 /*% * Maximum segment size (MSS) of TCP socket on which the server responds to * queries. Value lower than common MSS on Ethernet (1220, that is 1280 (IPv6 * minimum link MTU) - 40 (IPv6 fixed header) - 20 (TCP fixed header)) will * address path MTU problem. */ #define NM_MAXSEG (1280 - 20 - 40) /* * Define NETMGR_TRACE to activate tracing of handles and sockets. * This will impair performance but enables us to quickly determine, * if netmgr resources haven't been cleaned up on shutdown, which ones * are still in use. */ #ifdef NETMGR_TRACE #define TRACE_SIZE 8 void isc__nm_dump_active(isc_nm_t *nm); #if defined(__linux__) #include #define gettid() (uint32_t) syscall(SYS_gettid) #else #define gettid() (uint32_t) pthread_self() #endif #ifdef NETMGR_TRACE_VERBOSE #define NETMGR_TRACE_LOG(format, ...) \ fprintf(stderr, "%" PRIu32 ":%d:%s:%u:%s:" format, gettid(), \ isc_nm_tid(), file, line, func, __VA_ARGS__) #else #define NETMGR_TRACE_LOG(format, ...) \ (void)file; \ (void)line; \ (void)func; #endif #define FLARG_PASS , file, line, func #define FLARG \ , const char *file __attribute__((unused)), \ unsigned int line __attribute__((unused)), \ const char *func __attribute__((unused)) #define FLARG_IEVENT(ievent) \ const char *file = ievent->file; \ unsigned int line = ievent->line; \ const char *func = ievent->func; #define FLARG_IEVENT_PASS(ievent) \ ievent->file = file; \ ievent->line = line; \ ievent->func = func; #define isc__nm_uvreq_get(req, sock) \ isc___nm_uvreq_get(req, sock, __FILE__, __LINE__, __func__) #define isc__nm_uvreq_put(req, sock) \ isc___nm_uvreq_put(req, sock, __FILE__, __LINE__, __func__) #define isc__nmsocket_init(sock, mgr, type, iface) \ isc___nmsocket_init(sock, mgr, type, iface, __FILE__, __LINE__, \ __func__) #define isc__nmsocket_put(sockp) \ isc___nmsocket_put(sockp, __FILE__, __LINE__, __func__) #define isc__nmsocket_attach(sock, target) \ isc___nmsocket_attach(sock, target, __FILE__, __LINE__, __func__) #define isc__nmsocket_detach(socketp) \ isc___nmsocket_detach(socketp, __FILE__, __LINE__, __func__) #define isc__nmsocket_close(socketp) \ isc___nmsocket_close(socketp, __FILE__, __LINE__, __func__) #define isc__nmhandle_get(sock, peer, local) \ isc___nmhandle_get(sock, peer, local, __FILE__, __LINE__, __func__) #define isc__nmsocket_prep_destroy(sock) \ isc___nmsocket_prep_destroy(sock, __FILE__, __LINE__, __func__) #else #define NETMGR_TRACE_LOG(format, ...) #define FLARG_PASS #define FLARG #define FLARG_IEVENT(ievent) #define FLARG_IEVENT_PASS(ievent) #define isc__nm_uvreq_get(req, sock) isc___nm_uvreq_get(req, sock) #define isc__nm_uvreq_put(req, sock) isc___nm_uvreq_put(req, sock) #define isc__nmsocket_init(sock, mgr, type, iface) \ isc___nmsocket_init(sock, mgr, type, iface) #define isc__nmsocket_put(sockp) isc___nmsocket_put(sockp) #define isc__nmsocket_attach(sock, target) isc___nmsocket_attach(sock, target) #define isc__nmsocket_detach(socketp) isc___nmsocket_detach(socketp) #define isc__nmsocket_close(socketp) isc___nmsocket_close(socketp) #define isc__nmhandle_get(sock, peer, local) \ isc___nmhandle_get(sock, peer, local) #define isc__nmsocket_prep_destroy(sock) isc___nmsocket_prep_destroy(sock) #endif /* * Queue types in the order of processing priority. */ typedef enum { NETIEVENT_PRIORITY = 0, NETIEVENT_PRIVILEGED = 1, NETIEVENT_TASK = 2, NETIEVENT_NORMAL = 3, NETIEVENT_MAX = 4, } netievent_type_t; typedef struct isc__nm_uvreq isc__nm_uvreq_t; typedef struct isc__netievent isc__netievent_t; typedef ISC_LIST(isc__netievent_t) isc__netievent_list_t; typedef struct ievent { isc_mutex_t lock; isc_condition_t cond; isc__netievent_list_t list; } ievent_t; /* * Single network event loop worker. */ typedef struct isc__networker { isc_nm_t *mgr; int id; /* thread id */ uv_loop_t loop; /* libuv loop structure */ uv_async_t async; /* async channel to send * data to this networker */ bool paused; bool finished; isc_thread_t thread; ievent_t ievents[NETIEVENT_MAX]; isc_refcount_t references; atomic_int_fast64_t pktcount; char *recvbuf; char *sendbuf; bool recvbuf_inuse; } isc__networker_t; /* * A general handle for a connection bound to a networker. For UDP * connections we have peer address here, so both TCP and UDP can be * handled with a simple send-like function */ #define NMHANDLE_MAGIC ISC_MAGIC('N', 'M', 'H', 'D') #define VALID_NMHANDLE(t) \ (ISC_MAGIC_VALID(t, NMHANDLE_MAGIC) && \ atomic_load(&(t)->references) > 0) typedef void (*isc__nm_closecb)(isc_nmhandle_t *); typedef struct isc_nm_http_session isc_nm_http_session_t; struct isc_nmhandle { int magic; isc_refcount_t references; /* * The socket is not 'attached' in the traditional * reference-counting sense. Instead, we keep all handles in an * array in the socket object. This way, we don't have circular * dependencies and we can close all handles when we're destroying * the socket. */ isc_nmsocket_t *sock; isc_nm_http_session_t *httpsession; isc_sockaddr_t peer; isc_sockaddr_t local; isc_nm_opaquecb_t doreset; /* reset extra callback, external */ isc_nm_opaquecb_t dofree; /* free extra callback, external */ #ifdef NETMGR_TRACE void *backtrace[TRACE_SIZE]; int backtrace_size; LINK(isc_nmhandle_t) active_link; #endif void *opaque; max_align_t extra[]; }; typedef enum isc__netievent_type { netievent_udpconnect, netievent_udpclose, netievent_udpsend, netievent_udpread, netievent_udpcancel, netievent_routeconnect, netievent_tcpconnect, netievent_tcpclose, netievent_tcpsend, netievent_tcpstartread, netievent_tcppauseread, netievent_tcpaccept, netievent_tcpcancel, netievent_tcpdnsaccept, netievent_tcpdnsconnect, netievent_tcpdnsclose, netievent_tcpdnssend, netievent_tcpdnsread, netievent_tcpdnscancel, netievent_tlsclose, netievent_tlssend, netievent_tlsstartread, netievent_tlsconnect, netievent_tlsdobio, netievent_tlscancel, netievent_tlsdnsaccept, netievent_tlsdnsconnect, netievent_tlsdnsclose, netievent_tlsdnssend, netievent_tlsdnsread, netievent_tlsdnscancel, netievent_tlsdnscycle, netievent_tlsdnsshutdown, netievent_httpclose, netievent_httpsend, netievent_httpendpoints, netievent_shutdown, netievent_stop, netievent_pause, netievent_connectcb, netievent_readcb, netievent_sendcb, netievent_detach, netievent_close, netievent_task, netievent_privilegedtask, netievent_settlsctx, /* * event type values higher than this will be treated * as high-priority events, which can be processed * while the netmgr is pausing or paused. */ netievent_prio = 0xff, netievent_udplisten, netievent_udpstop, netievent_tcplisten, netievent_tcpstop, netievent_tcpdnslisten, netievent_tcpdnsstop, netievent_tlsdnslisten, netievent_tlsdnsstop, netievent_sockstop, /* for multilayer sockets */ netievent_resume, } isc__netievent_type; typedef union { isc_nm_recv_cb_t recv; isc_nm_cb_t send; isc_nm_cb_t connect; isc_nm_accept_cb_t accept; } isc__nm_cb_t; /* * Wrapper around uv_req_t with 'our' fields in it. req->data should * always point to its parent. Note that we always allocate more than * sizeof(struct) because we make room for different req types; */ #define UVREQ_MAGIC ISC_MAGIC('N', 'M', 'U', 'R') #define VALID_UVREQ(t) ISC_MAGIC_VALID(t, UVREQ_MAGIC) struct isc__nm_uvreq { int magic; isc_nmsocket_t *sock; isc_nmhandle_t *handle; char tcplen[2]; /* The TCP DNS message length */ uv_buf_t uvbuf; /* translated isc_region_t, to be * sent or received */ isc_sockaddr_t local; /* local address */ isc_sockaddr_t peer; /* peer address */ isc__nm_cb_t cb; /* callback */ void *cbarg; /* callback argument */ isc_nm_timer_t *timer; /* TCP write timer */ int connect_tries; /* connect retries */ union { uv_handle_t handle; uv_req_t req; uv_getaddrinfo_t getaddrinfo; uv_getnameinfo_t getnameinfo; uv_shutdown_t shutdown; uv_write_t write; uv_connect_t connect; uv_udp_send_t udp_send; uv_fs_t fs; uv_work_t work; } uv_req; ISC_LINK(isc__nm_uvreq_t) link; }; void * isc__nm_get_netievent(isc_nm_t *mgr, isc__netievent_type type); /*%< * Allocate an ievent and set the type. */ void isc__nm_put_netievent(isc_nm_t *mgr, void *ievent); /* * The macros here are used to simulate the "inheritance" in C, there's the base * netievent structure that contains just its own type and socket, and there are * extended netievent types that also have handles or requests or other data. * * The macros here ensure that: * * 1. every netievent type has matching definition, declaration and * implementation * * 2. we handle all the netievent types of same subclass the same, e.g. if the * extended netievent contains handle, we always attach to the handle in * the ctor and detach from the handle in dtor. * * There are three macros here for each netievent subclass: * * 1. NETIEVENT_*_TYPE(type) creates the typedef for each type; used below in * this header * * 2. NETIEVENT_*_DECL(type) generates the declaration of the get and put * functions (isc__nm_get_netievent_* and isc__nm_put_netievent_*); used * below in this header * * 3. NETIEVENT_*_DEF(type) generates the definition of the functions; used * either in netmgr.c or matching protocol file (e.g. udp.c, tcp.c, etc.) */ #define NETIEVENT__SOCKET \ isc__netievent_type type; \ ISC_LINK(isc__netievent_t) link; \ isc_nmsocket_t *sock; \ const char *file; \ unsigned int line; \ const char *func; typedef struct isc__netievent__socket { NETIEVENT__SOCKET; } isc__netievent__socket_t; #define NETIEVENT_SOCKET_TYPE(type) \ typedef isc__netievent__socket_t isc__netievent_##type##_t #define NETIEVENT_SOCKET_DECL(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_nmsocket_t *sock); \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent) #define NETIEVENT_SOCKET_DEF(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_nmsocket_t *sock) { \ isc__netievent_##type##_t *ievent = \ isc__nm_get_netievent(nm, netievent_##type); \ isc__nmsocket_attach(sock, &ievent->sock); \ \ return (ievent); \ } \ \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent) { \ isc__nmsocket_detach(&ievent->sock); \ isc__nm_put_netievent(nm, ievent); \ } typedef struct isc__netievent__socket_req { NETIEVENT__SOCKET; isc__nm_uvreq_t *req; } isc__netievent__socket_req_t; #define NETIEVENT_SOCKET_REQ_TYPE(type) \ typedef isc__netievent__socket_req_t isc__netievent_##type##_t #define NETIEVENT_SOCKET_REQ_DECL(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_nmsocket_t *sock, isc__nm_uvreq_t *req); \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent) #define NETIEVENT_SOCKET_REQ_DEF(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_nmsocket_t *sock, isc__nm_uvreq_t *req) { \ isc__netievent_##type##_t *ievent = \ isc__nm_get_netievent(nm, netievent_##type); \ isc__nmsocket_attach(sock, &ievent->sock); \ ievent->req = req; \ \ return (ievent); \ } \ \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent) { \ isc__nmsocket_detach(&ievent->sock); \ isc__nm_put_netievent(nm, ievent); \ } typedef struct isc__netievent__socket_req_result { NETIEVENT__SOCKET; isc__nm_uvreq_t *req; isc_result_t result; } isc__netievent__socket_req_result_t; #define NETIEVENT_SOCKET_REQ_RESULT_TYPE(type) \ typedef isc__netievent__socket_req_result_t isc__netievent_##type##_t #define NETIEVENT_SOCKET_REQ_RESULT_DECL(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_nmsocket_t *sock, isc__nm_uvreq_t *req, \ isc_result_t result); \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent) #define NETIEVENT_SOCKET_REQ_RESULT_DEF(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_nmsocket_t *sock, isc__nm_uvreq_t *req, \ isc_result_t result) { \ isc__netievent_##type##_t *ievent = \ isc__nm_get_netievent(nm, netievent_##type); \ isc__nmsocket_attach(sock, &ievent->sock); \ ievent->req = req; \ ievent->result = result; \ \ return (ievent); \ } \ \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent) { \ isc__nmsocket_detach(&ievent->sock); \ isc__nm_put_netievent(nm, ievent); \ } typedef struct isc__netievent__socket_handle { NETIEVENT__SOCKET; isc_nmhandle_t *handle; } isc__netievent__socket_handle_t; #define NETIEVENT_SOCKET_HANDLE_TYPE(type) \ typedef isc__netievent__socket_handle_t isc__netievent_##type##_t #define NETIEVENT_SOCKET_HANDLE_DECL(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_nmsocket_t *sock, isc_nmhandle_t *handle); \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent) #define NETIEVENT_SOCKET_HANDLE_DEF(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_nmsocket_t *sock, isc_nmhandle_t *handle) { \ isc__netievent_##type##_t *ievent = \ isc__nm_get_netievent(nm, netievent_##type); \ isc__nmsocket_attach(sock, &ievent->sock); \ isc_nmhandle_attach(handle, &ievent->handle); \ \ return (ievent); \ } \ \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent) { \ isc__nmsocket_detach(&ievent->sock); \ isc_nmhandle_detach(&ievent->handle); \ isc__nm_put_netievent(nm, ievent); \ } typedef struct isc__netievent__socket_quota { NETIEVENT__SOCKET; isc_quota_t *quota; } isc__netievent__socket_quota_t; #define NETIEVENT_SOCKET_QUOTA_TYPE(type) \ typedef isc__netievent__socket_quota_t isc__netievent_##type##_t #define NETIEVENT_SOCKET_QUOTA_DECL(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_nmsocket_t *sock, isc_quota_t *quota); \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent) #define NETIEVENT_SOCKET_QUOTA_DEF(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_nmsocket_t *sock, isc_quota_t *quota) { \ isc__netievent_##type##_t *ievent = \ isc__nm_get_netievent(nm, netievent_##type); \ isc__nmsocket_attach(sock, &ievent->sock); \ ievent->quota = quota; \ \ return (ievent); \ } \ \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent) { \ isc__nmsocket_detach(&ievent->sock); \ isc__nm_put_netievent(nm, ievent); \ } typedef struct isc__netievent__task { isc__netievent_type type; ISC_LINK(isc__netievent_t) link; isc_task_t *task; } isc__netievent__task_t; #define NETIEVENT_TASK_TYPE(type) \ typedef isc__netievent__task_t isc__netievent_##type##_t; #define NETIEVENT_TASK_DECL(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_task_t *task); \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent); #define NETIEVENT_TASK_DEF(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_task_t *task) { \ isc__netievent_##type##_t *ievent = \ isc__nm_get_netievent(nm, netievent_##type); \ ievent->task = task; \ \ return (ievent); \ } \ \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent) { \ ievent->task = NULL; \ isc__nm_put_netievent(nm, ievent); \ } typedef struct isc__netievent_udpsend { NETIEVENT__SOCKET; isc_sockaddr_t peer; isc__nm_uvreq_t *req; } isc__netievent_udpsend_t; typedef struct isc__netievent_tlsconnect { NETIEVENT__SOCKET; SSL_CTX *ctx; isc_sockaddr_t local; /* local address */ isc_sockaddr_t peer; /* peer address */ } isc__netievent_tlsconnect_t; typedef struct isc__netievent { isc__netievent_type type; ISC_LINK(isc__netievent_t) link; } isc__netievent_t; #define NETIEVENT_TYPE(type) typedef isc__netievent_t isc__netievent_##type##_t #define NETIEVENT_DECL(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type(isc_nm_t *nm); \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent) #define NETIEVENT_DEF(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm) { \ isc__netievent_##type##_t *ievent = \ isc__nm_get_netievent(nm, netievent_##type); \ \ return (ievent); \ } \ \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent) { \ isc__nm_put_netievent(nm, ievent); \ } typedef struct isc__netievent__tlsctx { NETIEVENT__SOCKET; isc_tlsctx_t *tlsctx; } isc__netievent__tlsctx_t; #define NETIEVENT_SOCKET_TLSCTX_TYPE(type) \ typedef isc__netievent__tlsctx_t isc__netievent_##type##_t; #define NETIEVENT_SOCKET_TLSCTX_DECL(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_nmsocket_t *sock, isc_tlsctx_t *tlsctx); \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent); #define NETIEVENT_SOCKET_TLSCTX_DEF(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_nmsocket_t *sock, isc_tlsctx_t *tlsctx) { \ isc__netievent_##type##_t *ievent = \ isc__nm_get_netievent(nm, netievent_##type); \ isc__nmsocket_attach(sock, &ievent->sock); \ isc_tlsctx_attach(tlsctx, &ievent->tlsctx); \ \ return (ievent); \ } \ \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent) { \ isc_tlsctx_free(&ievent->tlsctx); \ isc__nmsocket_detach(&ievent->sock); \ isc__nm_put_netievent(nm, ievent); \ } #ifdef HAVE_LIBNGHTTP2 typedef struct isc__netievent__http_eps { NETIEVENT__SOCKET; isc_nm_http_endpoints_t *endpoints; } isc__netievent__http_eps_t; #define NETIEVENT_SOCKET_HTTP_EPS_TYPE(type) \ typedef isc__netievent__http_eps_t isc__netievent_##type##_t; #define NETIEVENT_SOCKET_HTTP_EPS_DECL(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_nmsocket_t *sock, \ isc_nm_http_endpoints_t *endpoints); \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent); #define NETIEVENT_SOCKET_HTTP_EPS_DEF(type) \ isc__netievent_##type##_t *isc__nm_get_netievent_##type( \ isc_nm_t *nm, isc_nmsocket_t *sock, \ isc_nm_http_endpoints_t *endpoints) { \ isc__netievent_##type##_t *ievent = \ isc__nm_get_netievent(nm, netievent_##type); \ isc__nmsocket_attach(sock, &ievent->sock); \ isc_nm_http_endpoints_attach(endpoints, &ievent->endpoints); \ \ return (ievent); \ } \ \ void isc__nm_put_netievent_##type(isc_nm_t *nm, \ isc__netievent_##type##_t *ievent) { \ isc_nm_http_endpoints_detach(&ievent->endpoints); \ isc__nmsocket_detach(&ievent->sock); \ isc__nm_put_netievent(nm, ievent); \ } #endif /* HAVE_LIBNGHTTP2 */ typedef union { isc__netievent_t ni; isc__netievent__socket_t nis; isc__netievent__socket_req_t nisr; isc__netievent_udpsend_t nius; isc__netievent__socket_quota_t nisq; isc__netievent_tlsconnect_t nitc; isc__netievent__tlsctx_t nitls; #ifdef HAVE_LIBNGHTTP2 isc__netievent__http_eps_t nihttpeps; #endif /* HAVE_LIBNGHTTP2 */ } isc__netievent_storage_t; /* * Work item for a uv_work threadpool. */ typedef struct isc__nm_work { isc_nm_t *netmgr; uv_work_t req; isc_nm_workcb_t cb; isc_nm_after_workcb_t after_cb; void *data; } isc__nm_work_t; /* * Network manager */ #define NM_MAGIC ISC_MAGIC('N', 'E', 'T', 'M') #define VALID_NM(t) ISC_MAGIC_VALID(t, NM_MAGIC) struct isc_nm { int magic; isc_refcount_t references; isc_mem_t *mctx; int nworkers; int nlisteners; isc_mutex_t lock; isc_condition_t wkstatecond; isc_condition_t wkpausecond; isc__networker_t *workers; isc_stats_t *stats; uint_fast32_t workers_running; atomic_uint_fast32_t workers_paused; atomic_uint_fast32_t maxudp; bool load_balance_sockets; atomic_bool paused; /* * Active connections are being closed and new connections are * no longer allowed. */ atomic_bool closing; /* * A worker is actively waiting for other workers, for example to * stop listening; that means no other thread can do the same thing * or pause, or we'll deadlock. We have to either re-enqueue our * event or wait for the other one to finish if we want to pause. */ atomic_int interlocked; /* * Timeout values for TCP connections, corresponding to * tcp-intiial-timeout, tcp-idle-timeout, tcp-keepalive-timeout, * and tcp-advertised-timeout. Note that these are stored in * milliseconds so they can be used directly with the libuv timer, * but they are configured in tenths of seconds. */ atomic_uint_fast32_t init; atomic_uint_fast32_t idle; atomic_uint_fast32_t keepalive; atomic_uint_fast32_t advertised; isc_barrier_t pausing; isc_barrier_t resuming; /* * Socket SO_RCVBUF and SO_SNDBUF values */ atomic_int_fast32_t recv_udp_buffer_size; atomic_int_fast32_t send_udp_buffer_size; atomic_int_fast32_t recv_tcp_buffer_size; atomic_int_fast32_t send_tcp_buffer_size; #ifdef NETMGR_TRACE ISC_LIST(isc_nmsocket_t) active_sockets; #endif }; /*% * A universal structure for either a single socket or a group of * dup'd/SO_REUSE_PORT-using sockets listening on the same interface. */ #define NMSOCK_MAGIC ISC_MAGIC('N', 'M', 'S', 'K') #define VALID_NMSOCK(t) ISC_MAGIC_VALID(t, NMSOCK_MAGIC) /*% * Index into socket stat counter arrays. */ typedef enum { STATID_OPEN = 0, STATID_OPENFAIL = 1, STATID_CLOSE = 2, STATID_BINDFAIL = 3, STATID_CONNECTFAIL = 4, STATID_CONNECT = 5, STATID_ACCEPTFAIL = 6, STATID_ACCEPT = 7, STATID_SENDFAIL = 8, STATID_RECVFAIL = 9, STATID_ACTIVE = 10, STATID_MAX = 11, } isc__nm_statid_t; #if HAVE_LIBNGHTTP2 typedef struct isc_nmsocket_tls_send_req { isc_nmsocket_t *tlssock; isc_region_t data; isc_nm_cb_t cb; void *cbarg; isc_nmhandle_t *handle; bool finish; uint8_t smallbuf[512]; } isc_nmsocket_tls_send_req_t; typedef enum isc_http_request_type { ISC_HTTP_REQ_GET, ISC_HTTP_REQ_POST, ISC_HTTP_REQ_UNSUPPORTED } isc_http_request_type_t; typedef enum isc_http_scheme_type { ISC_HTTP_SCHEME_HTTP, ISC_HTTP_SCHEME_HTTP_SECURE, ISC_HTTP_SCHEME_UNSUPPORTED } isc_http_scheme_type_t; typedef struct isc_nm_httpcbarg { isc_nm_recv_cb_t cb; void *cbarg; LINK(struct isc_nm_httpcbarg) link; } isc_nm_httpcbarg_t; typedef struct isc_nm_httphandler { char *path; isc_nm_recv_cb_t cb; void *cbarg; size_t extrahandlesize; LINK(struct isc_nm_httphandler) link; } isc_nm_httphandler_t; struct isc_nm_http_endpoints { uint32_t magic; isc_mem_t *mctx; ISC_LIST(isc_nm_httphandler_t) handlers; ISC_LIST(isc_nm_httpcbarg_t) handler_cbargs; isc_refcount_t references; atomic_bool in_use; }; typedef struct isc_nmsocket_h2 { isc_nmsocket_t *psock; /* owner of the structure */ char *request_path; char *query_data; size_t query_data_len; bool query_too_large; isc_nm_httphandler_t *handler; isc_buffer_t rbuf; isc_buffer_t wbuf; int32_t stream_id; isc_nm_http_session_t *session; isc_nmsocket_t *httpserver; /* maximum concurrent streams (server-side) */ atomic_uint_fast32_t max_concurrent_streams; uint32_t min_ttl; /* used to set "max-age" in responses */ isc_http_request_type_t request_type; isc_http_scheme_type_t request_scheme; size_t content_length; char clenbuf[128]; char cache_control_buf[128]; int headers_error_code; size_t headers_data_processed; isc_nm_recv_cb_t cb; void *cbarg; LINK(struct isc_nmsocket_h2) link; isc_nm_http_endpoints_t **listener_endpoints; size_t n_listener_endpoints; bool response_submitted; struct { char *uri; bool post; isc_tlsctx_t *tlsctx; isc_sockaddr_t local_interface; void *cstream; const char *tls_peer_verify_string; } connect; } isc_nmsocket_h2_t; #endif /* HAVE_LIBNGHTTP2 */ typedef void (*isc_nm_closehandlecb_t)(void *arg); /*%< * Opaque callback function, used for isc_nmhandle 'reset' and 'free' * callbacks. */ struct isc_nmsocket { /*% Unlocked, RO */ int magic; int tid; isc_nmsocket_type type; isc_nm_t *mgr; /*% Parent socket for multithreaded listeners */ isc_nmsocket_t *parent; /*% Listener socket this connection was accepted on */ isc_nmsocket_t *listener; /*% Self socket */ isc_nmsocket_t *self; isc_barrier_t startlistening; isc_barrier_t stoplistening; /*% TLS stuff */ struct tls { isc_tls_t *tls; isc_tlsctx_t *ctx; isc_tlsctx_client_session_cache_t *client_sess_cache; bool client_session_saved; BIO *app_rbio; BIO *app_wbio; BIO *ssl_rbio; BIO *ssl_wbio; enum { TLS_STATE_NONE, TLS_STATE_HANDSHAKE, TLS_STATE_IO, TLS_STATE_ERROR, TLS_STATE_CLOSING } state; isc_region_t senddata; ISC_LIST(isc__nm_uvreq_t) sendreqs; bool cycle; isc_result_t pending_error; /* List of active send requests. */ isc__nm_uvreq_t *pending_req; bool alpn_negotiated; const char *tls_verify_errmsg; } tls; #if HAVE_LIBNGHTTP2 /*% TLS stuff */ struct tlsstream { bool server; BIO *bio_in; BIO *bio_out; isc_tls_t *tls; isc_tlsctx_t *ctx; isc_tlsctx_t **listener_tls_ctx; /*%< A context reference per worker */ size_t n_listener_tls_ctx; isc_tlsctx_client_session_cache_t *client_sess_cache; bool client_session_saved; isc_nmsocket_t *tlslistener; isc_nmsocket_t *tlssocket; atomic_bool result_updated; enum { TLS_INIT, TLS_HANDSHAKE, TLS_IO, TLS_CLOSED } state; /*%< The order of these is significant */ size_t nsending; bool reading; } tlsstream; isc_nmsocket_h2_t h2; #endif /* HAVE_LIBNGHTTP2 */ /*% * quota is the TCP client, attached when a TCP connection * is established. pquota is a non-attached pointer to the * TCP client quota, stored in listening sockets but only * attached in connected sockets. */ isc_quota_t *quota; isc_quota_t *pquota; isc_quota_cb_t quotacb; /*% * Socket statistics */ const isc_statscounter_t *statsindex; /*% * TCP read/connect timeout timers. */ uv_timer_t read_timer; uint64_t read_timeout; uint64_t connect_timeout; /*% * TCP write timeout timer. */ uint64_t write_timeout; /*% outer socket is for 'wrapped' sockets - e.g. tcpdns in tcp */ isc_nmsocket_t *outer; /*% server socket for connections */ isc_nmsocket_t *server; /*% Child sockets for multi-socket setups */ isc_nmsocket_t *children; uint_fast32_t nchildren; isc_sockaddr_t iface; isc_nmhandle_t *statichandle; isc_nmhandle_t *outerhandle; /*% Extra data allocated at the end of each isc_nmhandle_t */ size_t extrahandlesize; /*% TCP backlog */ int backlog; /*% libuv data */ uv_os_sock_t fd; union uv_any_handle uv_handle; /*% Peer address */ isc_sockaddr_t peer; /* Atomic */ /*% Number of running (e.g. listening) child sockets */ atomic_uint_fast32_t rchildren; /*% * Socket is active if it's listening, working, etc. If it's * closing, then it doesn't make a sense, for example, to * push handles or reqs for reuse. */ atomic_bool active; atomic_bool destroying; bool route_sock; /*% * Socket is closed if it's not active and all the possible * callbacks were fired, there are no active handles, etc. * If active==false but closed==false, that means the socket * is closing. */ atomic_bool closing; atomic_bool closed; atomic_bool listening; atomic_bool connecting; atomic_bool connected; atomic_bool accepting; atomic_bool reading; atomic_bool timedout; isc_refcount_t references; /*% * Established an outgoing connection, as client not server. */ atomic_bool client; /*% * TCPDNS socket has been set not to pipeline. */ atomic_bool sequential; /*% * The socket is processing read callback, this is guard to not read * data before the readcb is back. */ bool processing; /*% * A TCP socket has had isc_nm_pauseread() called. */ atomic_bool readpaused; /*% * A TCP or TCPDNS socket has been set to use the keepalive * timeout instead of the default idle timeout. */ atomic_bool keepalive; /*% * 'spare' handles for that can be reused to avoid allocations, * for UDP. */ isc_astack_t *inactivehandles; isc_astack_t *inactivereqs; /*% * Used to wait for TCP listening events to complete, and * for the number of running children to reach zero during * shutdown. * * We use two condition variables to prevent the race where the netmgr * threads would be able to finish and destroy the socket before it's * unlocked by the isc_nm_listen() function. So, the flow is as * follows: * * 1. parent thread creates all children sockets and passes then to * netthreads, looks at the signaling variable and WAIT(cond) until * the childrens are done initializing * * 2. the events get picked by netthreads, calls the libuv API (and * either succeeds or fails) and WAIT(scond) until all other * children sockets in netthreads are initialized and the listening * socket lock is unlocked * * 3. the control is given back to the parent thread which now either * returns success or shutdowns the listener if an error has * occured in the children netthread * * NOTE: The other approach would be doing an extra attach to the parent * listening socket, and then detach it in the parent thread, but that * breaks the promise that once the libuv socket is initialized on the * nmsocket, the nmsocket needs to be handled only by matching * netthread, so in fact that would add a complexity in a way that * isc__nmsocket_detach would have to be converted to use an * asynchrounous netievent. */ isc_mutex_t lock; isc_condition_t cond; isc_condition_t scond; /*% * Used to pass a result back from listen or connect events. */ isc_result_t result; /*% * Current number of active handles. */ atomic_int_fast32_t ah; /*% Buffer for TCPDNS processing */ size_t buf_size; size_t buf_len; unsigned char *buf; /*% * This function will be called with handle->sock * as the argument whenever a handle's references drop * to zero, after its reset callback has been called. */ isc_nm_closehandlecb_t closehandle_cb; isc_nmhandle_t *recv_handle; isc_nm_recv_cb_t recv_cb; void *recv_cbarg; bool recv_read; isc_nm_cb_t connect_cb; void *connect_cbarg; isc_nm_accept_cb_t accept_cb; void *accept_cbarg; atomic_int_fast32_t active_child_connections; isc_barrier_t barrier; bool barrier_initialised; #ifdef NETMGR_TRACE void *backtrace[TRACE_SIZE]; int backtrace_size; LINK(isc_nmsocket_t) active_link; ISC_LIST(isc_nmhandle_t) active_handles; #endif }; bool isc__nm_in_netthread(void); /*%< * Returns 'true' if we're in the network thread. */ void isc__nm_maybe_enqueue_ievent(isc__networker_t *worker, isc__netievent_t *event); /*%< * If the caller is already in the matching nmthread, process the netievent * directly, if not enqueue using isc__nm_enqueue_ievent(). */ void isc__nm_enqueue_ievent(isc__networker_t *worker, isc__netievent_t *event); /*%< * Enqueue an ievent onto a specific worker queue. (This the only safe * way to use an isc__networker_t from another thread.) */ void isc__nm_free_uvbuf(isc_nmsocket_t *sock, const uv_buf_t *buf); /*%< * Free a buffer allocated for a receive operation. * * Note that as currently implemented, this doesn't actually * free anything, marks the isc__networker's UDP receive buffer * as "not in use". */ isc_nmhandle_t * isc___nmhandle_get(isc_nmsocket_t *sock, isc_sockaddr_t *peer, isc_sockaddr_t *local FLARG); /*%< * Get a handle for the socket 'sock', allocating a new one * if there isn't one available in 'sock->inactivehandles'. * * If 'peer' is not NULL, set the handle's peer address to 'peer', * otherwise set it to 'sock->peer'. * * If 'local' is not NULL, set the handle's local address to 'local', * otherwise set it to 'sock->iface->addr'. * * 'sock' will be attached to 'handle->sock'. The caller may need * to detach the socket afterward. */ isc__nm_uvreq_t * isc___nm_uvreq_get(isc_nm_t *mgr, isc_nmsocket_t *sock FLARG); /*%< * Get a UV request structure for the socket 'sock', allocating a * new one if there isn't one available in 'sock->inactivereqs'. */ void isc___nm_uvreq_put(isc__nm_uvreq_t **req, isc_nmsocket_t *sock FLARG); /*%< * Completes the use of a UV request structure, setting '*req' to NULL. * * The UV request is pushed onto the 'sock->inactivereqs' stack or, * if that doesn't work, freed. */ void isc___nmsocket_init(isc_nmsocket_t *sock, isc_nm_t *mgr, isc_nmsocket_type type, isc_sockaddr_t *iface FLARG); /*%< * Initialize socket 'sock', attach it to 'mgr', and set it to type 'type' * and its interface to 'iface'. */ void isc___nmsocket_attach(isc_nmsocket_t *sock, isc_nmsocket_t **target FLARG); /*%< * Attach to a socket, increasing refcount */ void isc___nmsocket_detach(isc_nmsocket_t **socketp FLARG); /*%< * Detach from socket, decreasing refcount and possibly destroying the * socket if it's no longer referenced. */ void isc___nmsocket_prep_destroy(isc_nmsocket_t *sock FLARG); /*%< * Market 'sock' as inactive, close it if necessary, and destroy it * if there are no remaining references or active handles. */ void isc__nmsocket_shutdown(isc_nmsocket_t *sock); /*%< * Initiate the socket shutdown which actively calls the active * callbacks. */ void isc__nmsocket_reset(isc_nmsocket_t *sock); /*%< * Reset and close the socket. */ bool isc__nmsocket_active(isc_nmsocket_t *sock); /*%< * Determine whether 'sock' is active by checking 'sock->active' * or, for child sockets, 'sock->parent->active'. */ bool isc__nmsocket_deactivate(isc_nmsocket_t *sock); /*%< * @brief Deactivate active socket * * Atomically deactive the socket by setting @p sock->active or, for child * sockets, @p sock->parent->active to @c false * * @param[in] sock - valid nmsocket * @return @c false if the socket was already inactive, @c true otherwise */ void isc__nmsocket_clearcb(isc_nmsocket_t *sock); /*%< * Clear the recv and accept callbacks in 'sock'. */ void isc__nmsocket_timer_stop(isc_nmsocket_t *sock); void isc__nmsocket_timer_start(isc_nmsocket_t *sock); void isc__nmsocket_timer_restart(isc_nmsocket_t *sock); bool isc__nmsocket_timer_running(isc_nmsocket_t *sock); /*%< * Start/stop/restart/check the timeout on the socket */ void isc__nm_connectcb(isc_nmsocket_t *sock, isc__nm_uvreq_t *uvreq, isc_result_t eresult, bool async); void isc__nm_async_connectcb(isc__networker_t *worker, isc__netievent_t *ev0); /*%< * Issue a connect callback on the socket, used to call the callback */ void isc__nm_readcb(isc_nmsocket_t *sock, isc__nm_uvreq_t *uvreq, isc_result_t eresult); void isc__nm_async_readcb(isc__networker_t *worker, isc__netievent_t *ev0); /*%< * Issue a read callback on the socket, used to call the callback * on failed conditions when the event can't be scheduled on the uv loop. * */ void isc__nm_sendcb(isc_nmsocket_t *sock, isc__nm_uvreq_t *uvreq, isc_result_t eresult, bool async); void isc__nm_async_sendcb(isc__networker_t *worker, isc__netievent_t *ev0); /*%< * Issue a write callback on the socket, used to call the callback * on failed conditions when the event can't be scheduled on the uv loop. */ void isc__nm_async_shutdown(isc__networker_t *worker, isc__netievent_t *ev0); /*%< * Walk through all uv handles, get the underlying sockets and issue * close on them. */ void isc__nm_udp_send(isc_nmhandle_t *handle, const isc_region_t *region, isc_nm_cb_t cb, void *cbarg); /*%< * Back-end implementation of isc_nm_send() for UDP handles. */ void isc__nm_udp_read(isc_nmhandle_t *handle, isc_nm_recv_cb_t cb, void *cbarg); /* * Back-end implementation of isc_nm_read() for UDP handles. */ void isc__nm_udp_close(isc_nmsocket_t *sock); /*%< * Close a UDP socket. */ void isc__nm_udp_cancelread(isc_nmhandle_t *handle); /*%< * Stop reading on a connected UDP handle. */ void isc__nm_udp_shutdown(isc_nmsocket_t *sock); /*%< * Called during the shutdown process to close and clean up connected * sockets. */ void isc__nm_udp_stoplistening(isc_nmsocket_t *sock); /*%< * Stop listening on 'sock'. */ void isc__nm_udp_settimeout(isc_nmhandle_t *handle, uint32_t timeout); /*%< * Set or clear the recv timeout for the UDP socket associated with 'handle'. */ void isc__nm_async_udplisten(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_udpconnect(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_udpstop(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_udpsend(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_udpread(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_udpcancel(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_udpclose(isc__networker_t *worker, isc__netievent_t *ev0); /*%< * Callback handlers for asynchronous UDP events (listen, stoplisten, send). */ void isc__nm_async_routeconnect(isc__networker_t *worker, isc__netievent_t *ev0); /*%< * Callback handler for route socket events. */ void isc__nm_tcp_send(isc_nmhandle_t *handle, const isc_region_t *region, isc_nm_cb_t cb, void *cbarg); /*%< * Back-end implementation of isc_nm_send() for TCP handles. */ void isc__nm_tcp_read(isc_nmhandle_t *handle, isc_nm_recv_cb_t cb, void *cbarg); /* * Back-end implementation of isc_nm_read() for TCP handles. */ void isc__nm_tcp_close(isc_nmsocket_t *sock); /*%< * Close a TCP socket. */ void isc__nm_tcp_pauseread(isc_nmhandle_t *handle); /*%< * Pause reading on this handle, while still remembering the callback. */ void isc__nm_tcp_resumeread(isc_nmhandle_t *handle); /*%< * Resume reading from socket. * */ void isc__nm_tcp_shutdown(isc_nmsocket_t *sock); /*%< * Called during the shutdown process to close and clean up connected * sockets. */ void isc__nm_tcp_cancelread(isc_nmhandle_t *handle); /*%< * Stop reading on a connected TCP handle. */ void isc__nm_tcp_stoplistening(isc_nmsocket_t *sock); /*%< * Stop listening on 'sock'. */ int_fast32_t isc__nm_tcp_listener_nactive(isc_nmsocket_t *sock); /*%< * Returns the number of active connections for the TCP listener socket. */ void isc__nm_tcp_settimeout(isc_nmhandle_t *handle, uint32_t timeout); /*%< * Set the read timeout for the TCP socket associated with 'handle'. */ void isc__nm_async_tcpconnect(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tcplisten(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tcpaccept(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tcpstop(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tcpsend(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_startread(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_pauseread(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tcpstartread(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tcppauseread(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tcpcancel(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tcpclose(isc__networker_t *worker, isc__netievent_t *ev0); /*%< * Callback handlers for asynchronous TCP events (connect, listen, * stoplisten, send, read, pause, close). */ void isc__nm_async_tlsclose(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tlssend(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tlsstartread(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tlsdobio(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tlscancel(isc__networker_t *worker, isc__netievent_t *ev0); /*%< * Callback handlers for asynchronous TLS events. */ void isc__nm_tcpdns_send(isc_nmhandle_t *handle, isc_region_t *region, isc_nm_cb_t cb, void *cbarg); /*%< * Back-end implementation of isc_nm_send() for TCPDNS handles. */ void isc__nm_tcpdns_shutdown(isc_nmsocket_t *sock); void isc__nm_tcpdns_close(isc_nmsocket_t *sock); /*%< * Close a TCPDNS socket. */ void isc__nm_tcpdns_stoplistening(isc_nmsocket_t *sock); /*%< * Stop listening on 'sock'. */ void isc__nm_tcpdns_settimeout(isc_nmhandle_t *handle, uint32_t timeout); /*%< * Set the read timeout and reset the timer for the TCPDNS socket * associated with 'handle', and the TCP socket it wraps around. */ void isc__nm_async_tcpdnsaccept(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tcpdnsconnect(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tcpdnslisten(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tcpdnscancel(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tcpdnsclose(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tcpdnssend(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tcpdnsstop(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tcpdnsread(isc__networker_t *worker, isc__netievent_t *ev0); /*%< * Callback handlers for asynchronous TCPDNS events. */ void isc__nm_tcpdns_read(isc_nmhandle_t *handle, isc_nm_recv_cb_t cb, void *cbarg); /* * Back-end implementation of isc_nm_read() for TCPDNS handles. */ void isc__nm_tcpdns_cancelread(isc_nmhandle_t *handle); /*%< * Stop reading on a connected TCPDNS handle. */ void isc__nm_tlsdns_send(isc_nmhandle_t *handle, isc_region_t *region, isc_nm_cb_t cb, void *cbarg); void isc__nm_tlsdns_shutdown(isc_nmsocket_t *sock); void isc__nm_tlsdns_close(isc_nmsocket_t *sock); /*%< * Close a TLSDNS socket. */ void isc__nm_tlsdns_stoplistening(isc_nmsocket_t *sock); /*%< * Stop listening on 'sock'. */ void isc__nm_tlsdns_settimeout(isc_nmhandle_t *handle, uint32_t timeout); /*%< * Set the read timeout and reset the timer for the TLSDNS socket * associated with 'handle', and the TCP socket it wraps around. */ void isc__nm_tlsdns_read(isc_nmhandle_t *handle, isc_nm_recv_cb_t cb, void *cbarg); /* * Back-end implementation of isc_nm_read() for TLSDNS handles. */ void isc__nm_tlsdns_cancelread(isc_nmhandle_t *handle); /*%< * Stop reading on a connected TLSDNS handle. */ const char * isc__nm_tlsdns_verify_tls_peer_result_string(const isc_nmhandle_t *handle); void isc__nm_async_tlsdnscycle(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tlsdnsaccept(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tlsdnsconnect(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tlsdnslisten(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tlsdnscancel(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tlsdnsclose(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tlsdnssend(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tlsdnsstop(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tlsdnsshutdown(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tlsdnsread(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_tlsdns_set_tlsctx(isc_nmsocket_t *listener, isc_tlsctx_t *tlsctx, const int tid); /*%< * Callback handlers for asynchronous TLSDNS events. */ isc_result_t isc__nm_tlsdns_xfr_checkperm(isc_nmsocket_t *sock); /*%< * Check if it is permitted to do a zone transfer over the given TLSDNS * socket. * * Returns: * \li #ISC_R_SUCCESS Success, permission check passed successfully * \li #ISC_R_DOTALPNERROR No permission because of ALPN tag mismatch * \li any other result indicates failure (i.e. no permission) * * Requires: * \li 'sock' is a valid TLSDNS socket. */ void isc__nm_tlsdns_cleanup_data(isc_nmsocket_t *sock); #if HAVE_LIBNGHTTP2 void isc__nm_tls_send(isc_nmhandle_t *handle, const isc_region_t *region, isc_nm_cb_t cb, void *cbarg); void isc__nm_tls_cancelread(isc_nmhandle_t *handle); /*%< * Back-end implementation of isc_nm_send() for TLSDNS handles. */ void isc__nm_tls_read(isc_nmhandle_t *handle, isc_nm_recv_cb_t cb, void *cbarg); void isc__nm_tls_close(isc_nmsocket_t *sock); /*%< * Close a TLS socket. */ void isc__nm_tls_pauseread(isc_nmhandle_t *handle); /*%< * Pause reading on this handle, while still remembering the callback. */ void isc__nm_tls_resumeread(isc_nmhandle_t *handle); /*%< * Resume reading from the handle. * */ void isc__nm_tls_cleanup_data(isc_nmsocket_t *sock); void isc__nm_tls_stoplistening(isc_nmsocket_t *sock); void isc__nm_tls_settimeout(isc_nmhandle_t *handle, uint32_t timeout); void isc__nm_tls_cleartimeout(isc_nmhandle_t *handle); /*%< * Set the read timeout and reset the timer for the socket * associated with 'handle', and the TCP socket it wraps * around. */ const char * isc__nm_tls_verify_tls_peer_result_string(const isc_nmhandle_t *handle); void isc__nmhandle_tls_keepalive(isc_nmhandle_t *handle, bool value); /*%< * Set the keepalive value on the underlying TCP handle. */ void isc__nm_async_tls_set_tlsctx(isc_nmsocket_t *listener, isc_tlsctx_t *tlsctx, const int tid); void isc__nmhandle_tls_setwritetimeout(isc_nmhandle_t *handle, uint64_t write_timeout); void isc__nm_http_stoplistening(isc_nmsocket_t *sock); void isc__nm_http_settimeout(isc_nmhandle_t *handle, uint32_t timeout); void isc__nm_http_cleartimeout(isc_nmhandle_t *handle); /*%< * Set the read timeout and reset the timer for the socket * associated with 'handle', and the TLS/TCP socket it wraps * around. */ void isc__nmhandle_http_keepalive(isc_nmhandle_t *handle, bool value); /*%< * Set the keepalive value on the underlying session handle */ void isc__nm_http_initsocket(isc_nmsocket_t *sock); void isc__nm_http_cleanup_data(isc_nmsocket_t *sock); isc_result_t isc__nm_http_request(isc_nmhandle_t *handle, isc_region_t *region, isc_nm_recv_cb_t reply_cb, void *cbarg); void isc__nm_http_send(isc_nmhandle_t *handle, const isc_region_t *region, isc_nm_cb_t cb, void *cbarg); void isc__nm_http_read(isc_nmhandle_t *handle, isc_nm_recv_cb_t cb, void *cbarg); void isc__nm_http_close(isc_nmsocket_t *sock); void isc__nm_http_bad_request(isc_nmhandle_t *handle); /*%< * Respond to the request with 400 "Bad Request" status. * * Requires: * \li 'handle' is a valid HTTP netmgr handle object, referencing a server-side * socket */ bool isc__nm_http_has_encryption(const isc_nmhandle_t *handle); void isc__nm_http_set_maxage(isc_nmhandle_t *handle, const uint32_t ttl); const char * isc__nm_http_verify_tls_peer_result_string(const isc_nmhandle_t *handle); void isc__nm_async_httpsend(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_httpclose(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_async_httpendpoints(isc__networker_t *worker, isc__netievent_t *ev0); bool isc__nm_parse_httpquery(const char *query_string, const char **start, size_t *len); char * isc__nm_base64url_to_base64(isc_mem_t *mem, const char *base64url, const size_t base64url_len, size_t *res_len); char * isc__nm_base64_to_base64url(isc_mem_t *mem, const char *base64, const size_t base64_len, size_t *res_len); void isc__nm_httpsession_attach(isc_nm_http_session_t *source, isc_nm_http_session_t **targetp); void isc__nm_httpsession_detach(isc_nm_http_session_t **sessionp); void isc__nm_http_set_tlsctx(isc_nmsocket_t *sock, isc_tlsctx_t *tlsctx); void isc__nm_http_set_max_streams(isc_nmsocket_t *listener, const uint32_t max_concurrent_streams); #endif void isc__nm_async_settlsctx(isc__networker_t *worker, isc__netievent_t *ev0); #define isc__nm_uverr2result(x) \ isc___nm_uverr2result(x, true, __FILE__, __LINE__, __func__) isc_result_t isc___nm_uverr2result(int uverr, bool dolog, const char *file, unsigned int line, const char *func); /*%< * Convert a libuv error value into an isc_result_t. The * list of supported error values is not complete; new users * of this function should add any expected errors that are * not already there. */ bool isc__nm_acquire_interlocked(isc_nm_t *mgr); /*%< * Try to acquire interlocked state; return true if successful. */ void isc__nm_drop_interlocked(isc_nm_t *mgr); /*%< * Drop interlocked state; signal waiters. */ void isc__nm_acquire_interlocked_force(isc_nm_t *mgr); /*%< * Actively wait for interlocked state. */ void isc__nm_async_sockstop(isc__networker_t *worker, isc__netievent_t *ev0); void isc__nm_incstats(isc_nmsocket_t *sock, isc__nm_statid_t id); /*%< * Increment socket-related statistics counters. */ void isc__nm_decstats(isc_nmsocket_t *sock, isc__nm_statid_t id); /*%< * Decrement socket-related statistics counters. */ isc_result_t isc__nm_socket(int domain, int type, int protocol, uv_os_sock_t *sockp); /*%< * Platform independent socket() version */ void isc__nm_closesocket(uv_os_sock_t sock); /*%< * Platform independent closesocket() version */ isc_result_t isc__nm_socket_freebind(uv_os_sock_t fd, sa_family_t sa_family); /*%< * Set the IP_FREEBIND (or equivalent) socket option on the uv_handle */ isc_result_t isc__nm_socket_reuse(uv_os_sock_t fd); /*%< * Set the SO_REUSEADDR or SO_REUSEPORT (or equivalent) socket option on the fd */ isc_result_t isc__nm_socket_reuse_lb(uv_os_sock_t fd); /*%< * Set the SO_REUSEPORT_LB (or equivalent) socket option on the fd */ isc_result_t isc__nm_socket_incoming_cpu(uv_os_sock_t fd); /*%< * Set the SO_INCOMING_CPU socket option on the fd if available */ isc_result_t isc__nm_socket_disable_pmtud(uv_os_sock_t fd, sa_family_t sa_family); /*%< * Disable the Path MTU Discovery, either by disabling IP(V6)_DONTFRAG socket * option, or setting the IP(V6)_MTU_DISCOVER socket option to IP_PMTUDISC_OMIT */ isc_result_t isc__nm_socket_v6only(uv_os_sock_t fd, sa_family_t sa_family); /*%< * Restrict the socket to sending and receiving IPv6 packets only */ isc_result_t isc__nm_socket_connectiontimeout(uv_os_sock_t fd, int timeout_ms); /*%< * Set the connection timeout in milliseconds, on non-Linux platforms, * the minimum value must be at least 1000 (1 second). */ isc_result_t isc__nm_socket_tcp_nodelay(uv_os_sock_t fd); /*%< * Disables Nagle's algorithm on a TCP socket (sets TCP_NODELAY). */ isc_result_t isc__nm_socket_tcp_maxseg(uv_os_sock_t fd, int size); /*%< * Set the TCP maximum segment size */ isc_result_t isc__nm_socket_min_mtu(uv_os_sock_t fd, sa_family_t sa_family); /*%< * Use minimum MTU on IPv6 sockets */ void isc__nm_set_network_buffers(isc_nm_t *nm, uv_handle_t *handle); /*%> * Sets the pre-configured network buffers size on the handle. */ void isc__nmsocket_barrier_init(isc_nmsocket_t *listener); /*%> * Initialise the socket synchronisation barrier according to the * number of children. */ void isc__nmsocket_stop(isc_nmsocket_t *listener); /*%> * Broadcast "stop" event for a listener socket across all workers and * wait its processing completion - then, stop and close the underlying * transport listener socket. * * The primitive is used in multi-layer transport listener sockets to * implement shutdown properly: after the broadcasted events has been * processed it is safe to destroy the shared data within the listener * socket (including shutting down the underlying transport listener * socket). */ /* * typedef all the netievent types */ NETIEVENT_SOCKET_TYPE(close); NETIEVENT_SOCKET_TYPE(tcpclose); NETIEVENT_SOCKET_TYPE(tcplisten); NETIEVENT_SOCKET_TYPE(tcppauseread); NETIEVENT_SOCKET_TYPE(tcpstop); NETIEVENT_SOCKET_TYPE(tlsclose); /* NETIEVENT_SOCKET_TYPE(tlsconnect); */ /* unique type, defined independently */ NETIEVENT_SOCKET_TYPE(tlsdobio); NETIEVENT_SOCKET_TYPE(tlsstartread); NETIEVENT_SOCKET_HANDLE_TYPE(tlscancel); NETIEVENT_SOCKET_TYPE(udpclose); NETIEVENT_SOCKET_TYPE(udplisten); NETIEVENT_SOCKET_TYPE(udpread); /* NETIEVENT_SOCKET_TYPE(udpsend); */ /* unique type, defined independently */ NETIEVENT_SOCKET_TYPE(udpstop); NETIEVENT_SOCKET_TYPE(tcpdnsclose); NETIEVENT_SOCKET_TYPE(tcpdnsread); NETIEVENT_SOCKET_TYPE(tcpdnsstop); NETIEVENT_SOCKET_TYPE(tcpdnslisten); NETIEVENT_SOCKET_REQ_TYPE(tcpdnsconnect); NETIEVENT_SOCKET_REQ_TYPE(tcpdnssend); NETIEVENT_SOCKET_HANDLE_TYPE(tcpdnscancel); NETIEVENT_SOCKET_QUOTA_TYPE(tcpdnsaccept); NETIEVENT_SOCKET_TYPE(tlsdnsclose); NETIEVENT_SOCKET_TYPE(tlsdnsread); NETIEVENT_SOCKET_TYPE(tlsdnsstop); NETIEVENT_SOCKET_TYPE(tlsdnsshutdown); NETIEVENT_SOCKET_TYPE(tlsdnslisten); NETIEVENT_SOCKET_REQ_TYPE(tlsdnsconnect); NETIEVENT_SOCKET_REQ_TYPE(tlsdnssend); NETIEVENT_SOCKET_HANDLE_TYPE(tlsdnscancel); NETIEVENT_SOCKET_QUOTA_TYPE(tlsdnsaccept); NETIEVENT_SOCKET_TYPE(tlsdnscycle); #ifdef HAVE_LIBNGHTTP2 NETIEVENT_SOCKET_REQ_TYPE(httpsend); NETIEVENT_SOCKET_TYPE(httpclose); NETIEVENT_SOCKET_HTTP_EPS_TYPE(httpendpoints); #endif /* HAVE_LIBNGHTTP2 */ NETIEVENT_SOCKET_REQ_TYPE(tcpconnect); NETIEVENT_SOCKET_REQ_TYPE(tcpsend); NETIEVENT_SOCKET_TYPE(tcpstartread); NETIEVENT_SOCKET_REQ_TYPE(tlssend); NETIEVENT_SOCKET_REQ_TYPE(udpconnect); NETIEVENT_SOCKET_REQ_TYPE(routeconnect); NETIEVENT_SOCKET_REQ_RESULT_TYPE(connectcb); NETIEVENT_SOCKET_REQ_RESULT_TYPE(readcb); NETIEVENT_SOCKET_REQ_RESULT_TYPE(sendcb); NETIEVENT_SOCKET_HANDLE_TYPE(detach); NETIEVENT_SOCKET_HANDLE_TYPE(tcpcancel); NETIEVENT_SOCKET_HANDLE_TYPE(udpcancel); NETIEVENT_SOCKET_QUOTA_TYPE(tcpaccept); NETIEVENT_TYPE(pause); NETIEVENT_TYPE(resume); NETIEVENT_TYPE(shutdown); NETIEVENT_TYPE(stop); NETIEVENT_TASK_TYPE(task); NETIEVENT_TASK_TYPE(privilegedtask); NETIEVENT_SOCKET_TLSCTX_TYPE(settlsctx); NETIEVENT_SOCKET_TYPE(sockstop); /* Now declared the helper functions */ NETIEVENT_SOCKET_DECL(close); NETIEVENT_SOCKET_DECL(tcpclose); NETIEVENT_SOCKET_DECL(tcplisten); NETIEVENT_SOCKET_DECL(tcppauseread); NETIEVENT_SOCKET_DECL(tcpstartread); NETIEVENT_SOCKET_DECL(tcpstop); NETIEVENT_SOCKET_DECL(tlsclose); NETIEVENT_SOCKET_DECL(tlsconnect); NETIEVENT_SOCKET_DECL(tlsdobio); NETIEVENT_SOCKET_DECL(tlsstartread); NETIEVENT_SOCKET_HANDLE_DECL(tlscancel); NETIEVENT_SOCKET_DECL(udpclose); NETIEVENT_SOCKET_DECL(udplisten); NETIEVENT_SOCKET_DECL(udpread); NETIEVENT_SOCKET_DECL(udpsend); NETIEVENT_SOCKET_DECL(udpstop); NETIEVENT_SOCKET_DECL(tcpdnsclose); NETIEVENT_SOCKET_DECL(tcpdnsread); NETIEVENT_SOCKET_DECL(tcpdnsstop); NETIEVENT_SOCKET_DECL(tcpdnslisten); NETIEVENT_SOCKET_REQ_DECL(tcpdnsconnect); NETIEVENT_SOCKET_REQ_DECL(tcpdnssend); NETIEVENT_SOCKET_HANDLE_DECL(tcpdnscancel); NETIEVENT_SOCKET_QUOTA_DECL(tcpdnsaccept); NETIEVENT_SOCKET_DECL(tlsdnsclose); NETIEVENT_SOCKET_DECL(tlsdnsread); NETIEVENT_SOCKET_DECL(tlsdnsstop); NETIEVENT_SOCKET_DECL(tlsdnsshutdown); NETIEVENT_SOCKET_DECL(tlsdnslisten); NETIEVENT_SOCKET_REQ_DECL(tlsdnsconnect); NETIEVENT_SOCKET_REQ_DECL(tlsdnssend); NETIEVENT_SOCKET_HANDLE_DECL(tlsdnscancel); NETIEVENT_SOCKET_QUOTA_DECL(tlsdnsaccept); NETIEVENT_SOCKET_DECL(tlsdnscycle); #ifdef HAVE_LIBNGHTTP2 NETIEVENT_SOCKET_REQ_DECL(httpsend); NETIEVENT_SOCKET_DECL(httpclose); NETIEVENT_SOCKET_HTTP_EPS_DECL(httpendpoints); #endif /* HAVE_LIBNGHTTP2 */ NETIEVENT_SOCKET_REQ_DECL(tcpconnect); NETIEVENT_SOCKET_REQ_DECL(tcpsend); NETIEVENT_SOCKET_REQ_DECL(tlssend); NETIEVENT_SOCKET_REQ_DECL(udpconnect); NETIEVENT_SOCKET_REQ_DECL(routeconnect); NETIEVENT_SOCKET_REQ_RESULT_DECL(connectcb); NETIEVENT_SOCKET_REQ_RESULT_DECL(readcb); NETIEVENT_SOCKET_REQ_RESULT_DECL(sendcb); NETIEVENT_SOCKET_HANDLE_DECL(udpcancel); NETIEVENT_SOCKET_HANDLE_DECL(tcpcancel); NETIEVENT_SOCKET_DECL(detach); NETIEVENT_SOCKET_QUOTA_DECL(tcpaccept); NETIEVENT_DECL(pause); NETIEVENT_DECL(resume); NETIEVENT_DECL(shutdown); NETIEVENT_DECL(stop); NETIEVENT_TASK_DECL(task); NETIEVENT_TASK_DECL(privilegedtask); NETIEVENT_SOCKET_TLSCTX_DECL(settlsctx); NETIEVENT_SOCKET_DECL(sockstop); void isc__nm_udp_failed_read_cb(isc_nmsocket_t *sock, isc_result_t result); void isc__nm_tcp_failed_read_cb(isc_nmsocket_t *sock, isc_result_t result); void isc__nm_tcpdns_failed_read_cb(isc_nmsocket_t *sock, isc_result_t result); void isc__nm_tlsdns_failed_read_cb(isc_nmsocket_t *sock, isc_result_t result, bool async); isc_result_t isc__nm_tcpdns_processbuffer(isc_nmsocket_t *sock); isc_result_t isc__nm_tlsdns_processbuffer(isc_nmsocket_t *sock); isc__nm_uvreq_t * isc__nm_get_read_req(isc_nmsocket_t *sock, isc_sockaddr_t *sockaddr); void isc__nm_alloc_cb(uv_handle_t *handle, size_t size, uv_buf_t *buf); void isc__nm_udp_read_cb(uv_udp_t *handle, ssize_t nrecv, const uv_buf_t *buf, const struct sockaddr *addr, unsigned flags); void isc__nm_tcp_read_cb(uv_stream_t *stream, ssize_t nread, const uv_buf_t *buf); void isc__nm_tcpdns_read_cb(uv_stream_t *stream, ssize_t nread, const uv_buf_t *buf); void isc__nm_tlsdns_read_cb(uv_stream_t *stream, ssize_t nread, const uv_buf_t *buf); isc_result_t isc__nm_start_reading(isc_nmsocket_t *sock); void isc__nm_stop_reading(isc_nmsocket_t *sock); isc_result_t isc__nm_process_sock_buffer(isc_nmsocket_t *sock); void isc__nm_resume_processing(void *arg); bool isc__nmsocket_closing(isc_nmsocket_t *sock); bool isc__nm_closing(isc_nmsocket_t *sock); void isc__nm_alloc_dnsbuf(isc_nmsocket_t *sock, size_t len); void isc__nm_failed_send_cb(isc_nmsocket_t *sock, isc__nm_uvreq_t *req, isc_result_t eresult); void isc__nm_failed_accept_cb(isc_nmsocket_t *sock, isc_result_t eresult); void isc__nm_failed_connect_cb(isc_nmsocket_t *sock, isc__nm_uvreq_t *req, isc_result_t eresult, bool async); void isc__nm_failed_read_cb(isc_nmsocket_t *sock, isc_result_t result, bool async); void isc__nm_accept_connection_log(isc_result_t result, bool can_log_quota); /* * Timeout callbacks */ void isc__nmsocket_connecttimeout_cb(uv_timer_t *timer); void isc__nmsocket_readtimeout_cb(uv_timer_t *timer); void isc__nmsocket_writetimeout_cb(void *data, isc_result_t eresult); #define UV_RUNTIME_CHECK(func, ret) \ if (ret != 0) { \ FATAL_ERROR("%s failed: %s\n", #func, uv_strerror(ret)); \ } void isc__nmsocket_log_tls_session_reuse(isc_nmsocket_t *sock, isc_tls_t *tls);