2
0
mirror of https://gitlab.isc.org/isc-projects/bind9 synced 2025-08-24 19:18:50 +00:00
bind/lib/isc/app.c

441 lines
10 KiB
C
Raw Normal View History

1999-05-12 09:44:35 +00:00
/*
* Copyright (C) Internet Systems Consortium, Inc. ("ISC")
*
* SPDX-License-Identifier: MPL-2.0
*
* This Source Code Form is subject to the terms of the Mozilla Public
* License, v. 2.0. If a copy of the MPL was not distributed with this
* file, you can obtain one at https://mozilla.org/MPL/2.0/.
*
* See the COPYRIGHT file distributed with this work for additional
* information regarding copyright ownership.
1999-05-12 09:44:35 +00:00
*/
/*! \file */
2000-06-22 22:00:42 +00:00
#include <errno.h>
#include <inttypes.h>
#include <pthread.h>
#include <signal.h>
#include <stdbool.h>
1999-05-12 09:44:35 +00:00
#include <stddef.h>
#include <stdlib.h>
#include <sys/time.h>
#include <sys/types.h>
#include <unistd.h>
1999-05-12 09:44:35 +00:00
#include <isc/app.h>
#include <isc/atomic.h>
#include <isc/condition.h>
#include <isc/event.h>
#include <isc/mem.h>
#include <isc/mutex.h>
#include <isc/strerr.h>
#include <isc/string.h>
#include <isc/task.h>
#include <isc/thread.h>
#include <isc/time.h>
1999-12-16 22:24:22 +00:00
#include <isc/util.h>
1999-05-12 09:44:35 +00:00
/*%
* For BIND9 applications built with threads, we use a single app
* context and let multiple taskmgr and netmgr threads do actual jobs.
*/
static isc_thread_t blockedthread;
static atomic_bool is_running = 0;
/*
* The application context of this module.
*/
2020-02-13 14:44:37 -08:00
#define APPCTX_MAGIC ISC_MAGIC('A', 'p', 'c', 'x')
#define VALID_APPCTX(c) ISC_MAGIC_VALID(c, APPCTX_MAGIC)
struct isc_appctx {
2020-02-13 14:44:37 -08:00
unsigned int magic;
isc_mem_t *mctx;
isc_mutex_t lock;
isc_eventlist_t on_run;
2020-02-13 14:44:37 -08:00
atomic_bool shutdown_requested;
atomic_bool running;
atomic_bool want_shutdown;
atomic_bool want_reload;
atomic_bool blocked;
isc_mutex_t readylock;
isc_condition_t ready;
};
static isc_appctx_t isc_g_appctx;
2000-01-22 01:39:35 +00:00
static void
2020-02-13 14:44:37 -08:00
handle_signal(int sig, void (*handler)(int)) {
1999-05-12 22:54:46 +00:00
struct sigaction sa;
2001-11-27 01:56:32 +00:00
memset(&sa, 0, sizeof(sa));
1999-05-12 22:54:46 +00:00
sa.sa_handler = handler;
if (sigfillset(&sa.sa_mask) != 0 || sigaction(sig, &sa, NULL) < 0) {
char strbuf[ISC_STRERRORSIZE];
strerror_r(errno, strbuf, sizeof(strbuf));
isc_error_fatal(__FILE__, __LINE__,
"handle_signal() %d setup: %s", sig, strbuf);
1999-05-12 22:54:46 +00:00
}
}
isc_result_t
2020-02-13 14:44:37 -08:00
isc_app_ctxstart(isc_appctx_t *ctx) {
REQUIRE(VALID_APPCTX(ctx));
1999-05-12 09:44:35 +00:00
/*
* Start an ISC library application.
*/
isc_mutex_init(&ctx->lock);
isc_mutex_init(&ctx->readylock);
2018-11-15 17:20:36 +01:00
isc_condition_init(&ctx->ready);
ISC_LIST_INIT(ctx->on_run);
atomic_init(&ctx->shutdown_requested, false);
atomic_init(&ctx->running, false);
atomic_init(&ctx->want_shutdown, false);
atomic_init(&ctx->want_reload, false);
atomic_init(&ctx->blocked, false);
1999-05-12 09:44:35 +00:00
2020-02-13 14:44:37 -08:00
int presult;
sigset_t sset;
char strbuf[ISC_STRERRORSIZE];
1999-05-12 22:54:46 +00:00
/*
* Always ignore SIGPIPE.
*/
handle_signal(SIGPIPE, SIG_IGN);
1999-05-12 22:54:46 +00:00
handle_signal(SIGHUP, SIG_DFL);
handle_signal(SIGTERM, SIG_DFL);
handle_signal(SIGINT, SIG_DFL);
1999-05-12 09:44:35 +00:00
/*
2000-01-22 01:39:35 +00:00
* Block SIGHUP, SIGINT, SIGTERM.
1999-05-12 09:44:35 +00:00
*
* If isc_app_start() is called from the main thread before any other
* threads have been created, then the pthread_sigmask() call below
2000-03-18 01:29:48 +00:00
* will result in all threads having SIGHUP, SIGINT and SIGTERM
* blocked by default, ensuring that only the thread that calls
* sigwait() for them will get those signals.
1999-05-12 09:44:35 +00:00
*/
if (sigemptyset(&sset) != 0 || sigaddset(&sset, SIGHUP) != 0 ||
2020-02-13 14:44:37 -08:00
sigaddset(&sset, SIGINT) != 0 || sigaddset(&sset, SIGTERM) != 0)
{
strerror_r(errno, strbuf, sizeof(strbuf));
isc_error_fatal(__FILE__, __LINE__,
"isc_app_start() sigsetops: %s", strbuf);
1999-05-12 09:44:35 +00:00
}
presult = pthread_sigmask(SIG_BLOCK, &sset, NULL);
if (presult != 0) {
strerror_r(presult, strbuf, sizeof(strbuf));
isc_error_fatal(__FILE__, __LINE__,
"isc_app_start() pthread_sigmask: %s", strbuf);
1999-05-12 09:44:35 +00:00
}
return (ISC_R_SUCCESS);
}
isc_result_t
2020-02-13 14:44:37 -08:00
isc_app_start(void) {
isc_g_appctx.magic = APPCTX_MAGIC;
isc_g_appctx.mctx = NULL;
/* The remaining members will be initialized in ctxstart() */
return (isc_app_ctxstart(&isc_g_appctx));
1999-05-12 09:44:35 +00:00
}
isc_result_t
isc_app_onrun(isc_mem_t *mctx, isc_task_t *task, isc_taskaction_t action,
2020-02-13 14:44:37 -08:00
void *arg) {
return (isc_app_ctxonrun(&isc_g_appctx, mctx, task, action, arg));
}
isc_result_t
isc_app_ctxonrun(isc_appctx_t *ctx, isc_mem_t *mctx, isc_task_t *task,
2020-02-13 14:44:37 -08:00
isc_taskaction_t action, void *arg) {
1999-07-14 02:03:44 +00:00
isc_event_t *event;
2020-02-13 14:44:37 -08:00
isc_task_t *cloned_task = NULL;
1999-07-14 02:03:44 +00:00
if (atomic_load_acquire(&ctx->running)) {
return (ISC_R_ALREADYRUNNING);
1999-07-14 02:03:44 +00:00
}
/*
* Note that we store the task to which we're going to send the event
* in the event's "sender" field.
*/
isc_task_attach(task, &cloned_task);
event = isc_event_allocate(mctx, cloned_task, ISC_APPEVENT_SHUTDOWN,
2001-11-27 01:56:32 +00:00
action, arg, sizeof(*event));
LOCK(&ctx->lock);
ISC_LINK_INIT(event, ev_link);
ISC_LIST_APPEND(ctx->on_run, event, ev_link);
UNLOCK(&ctx->lock);
1999-07-14 02:03:44 +00:00
return (ISC_R_SUCCESS);
1999-07-14 02:03:44 +00:00
}
isc_result_t
2020-02-13 14:44:37 -08:00
isc_app_ctxrun(isc_appctx_t *ctx) {
1999-05-12 09:44:35 +00:00
isc_event_t *event, *next_event;
2020-02-13 14:44:37 -08:00
isc_task_t *task;
REQUIRE(VALID_APPCTX(ctx));
1999-05-12 09:44:35 +00:00
if (atomic_compare_exchange_strong_acq_rel(&ctx->running,
&(bool){ false }, true))
2020-02-13 14:44:37 -08:00
{
2000-01-22 01:39:35 +00:00
/*
* Post any on-run events (in FIFO order).
*/
LOCK(&ctx->lock);
for (event = ISC_LIST_HEAD(ctx->on_run); event != NULL;
2000-01-22 01:39:35 +00:00
event = next_event) {
next_event = ISC_LIST_NEXT(event, ev_link);
ISC_LIST_UNLINK(ctx->on_run, event, ev_link);
task = event->ev_sender;
event->ev_sender = NULL;
2000-01-22 01:39:35 +00:00
isc_task_sendanddetach(&task, &event);
}
UNLOCK(&ctx->lock);
1999-05-12 09:44:35 +00:00
}
1999-07-14 02:03:44 +00:00
/*
* There is no danger if isc_app_shutdown() is called before we
* wait for signals. Signals are blocked, so any such signal will
* simply be made pending and we will get it when we call
* sigwait().
1999-07-14 02:03:44 +00:00
*/
while (!atomic_load_acquire(&ctx->want_shutdown)) {
if (ctx == &isc_g_appctx) {
sigset_t sset;
2020-02-13 14:44:37 -08:00
int sig;
/*
* Wait for SIGHUP, SIGINT, or SIGTERM.
*/
if (sigemptyset(&sset) != 0 ||
sigaddset(&sset, SIGHUP) != 0 ||
sigaddset(&sset, SIGINT) != 0 ||
2020-02-13 14:44:37 -08:00
sigaddset(&sset, SIGTERM) != 0)
{
char strbuf[ISC_STRERRORSIZE];
strerror_r(errno, strbuf, sizeof(strbuf));
isc_error_fatal(__FILE__, __LINE__,
"isc_app_run() sigsetops: %s",
strbuf);
}
2000-05-18 22:39:24 +00:00
if (sigwait(&sset, &sig) == 0) {
switch (sig) {
case SIGINT:
case SIGTERM:
atomic_store_release(
&ctx->want_shutdown, true);
break;
case SIGHUP:
atomic_store_release(&ctx->want_reload,
true);
break;
default:
UNREACHABLE();
}
}
} else {
/*
* Tools using multiple contexts don't
* rely on a signal, just wait until woken
* up.
*/
if (atomic_load_acquire(&ctx->want_shutdown)) {
break;
}
if (!atomic_load_acquire(&ctx->want_reload)) {
LOCK(&ctx->readylock);
WAIT(&ctx->ready, &ctx->readylock);
UNLOCK(&ctx->readylock);
}
}
if (atomic_compare_exchange_strong_acq_rel(
&ctx->want_reload, &(bool){ true }, false))
{
return (ISC_R_RELOAD);
}
if (atomic_load_acquire(&ctx->want_shutdown) &&
2020-02-13 14:44:37 -08:00
atomic_load_acquire(&ctx->blocked))
{
2001-03-14 06:31:17 +00:00
exit(1);
}
2000-01-22 01:39:35 +00:00
}
1999-05-12 09:44:35 +00:00
return (ISC_R_SUCCESS);
}
isc_result_t
2020-02-13 14:44:37 -08:00
isc_app_run(void) {
isc_result_t result;
REQUIRE(atomic_compare_exchange_strong_acq_rel(&is_running,
&(bool){ false }, true));
result = isc_app_ctxrun(&isc_g_appctx);
atomic_store_release(&is_running, false);
return (result);
}
bool
Complete rewrite the BIND 9 build system The rewrite of BIND 9 build system is a large work and cannot be reasonable split into separate merge requests. Addition of the automake has a positive effect on the readability and maintainability of the build system as it is more declarative, it allows conditional and we are able to drop all of the custom make code that BIND 9 developed over the years to overcome the deficiencies of autoconf + custom Makefile.in files. This squashed commit contains following changes: - conversion (or rather fresh rewrite) of all Makefile.in files to Makefile.am by using automake - the libtool is now properly integrated with automake (the way we used it was rather hackish as the only official way how to use libtool is via automake - the dynamic module loading was rewritten from a custom patchwork to libtool's libltdl (which includes the patchwork to support module loading on different systems internally) - conversion of the unit test executor from kyua to automake parallel driver - conversion of the system test executor from custom make/shell to automake parallel driver - The GSSAPI has been refactored, the custom SPNEGO on the basis that all major KRB5/GSSAPI (mit-krb5, heimdal and Windows) implementations support SPNEGO mechanism. - The various defunct tests from bin/tests have been removed: bin/tests/optional and bin/tests/pkcs11 - The text files generated from the MD files have been removed, the MarkDown has been designed to be readable by both humans and computers - The xsl header is now generated by a simple sed command instead of perl helper - The <irs/platform.h> header has been removed - cleanups of configure.ac script to make it more simpler, addition of multiple macros (there's still work to be done though) - the tarball can now be prepared with `make dist` - the system tests are partially able to run in oot build Here's a list of unfinished work that needs to be completed in subsequent merge requests: - `make distcheck` doesn't yet work (because of system tests oot run is not yet finished) - documentation is not yet built, there's a different merge request with docbook to sphinx-build rst conversion that needs to be rebased and adapted on top of the automake - msvc build is non functional yet and we need to decide whether we will just cross-compile bind9 using mingw-w64 or fix the msvc build - contributed dlz modules are not included neither in the autoconf nor automake
2018-08-07 16:46:53 +02:00
isc_app_isrunning(void) {
return (atomic_load_acquire(&is_running));
}
void
2020-02-13 14:44:37 -08:00
isc_app_ctxshutdown(isc_appctx_t *ctx) {
REQUIRE(VALID_APPCTX(ctx));
REQUIRE(atomic_load_acquire(&ctx->running));
1999-05-12 09:44:35 +00:00
/* If ctx->shutdown_requested == true, we are already shutting
* down and we want to just bail out.
*/
if (atomic_compare_exchange_strong_acq_rel(&ctx->shutdown_requested,
&(bool){ false }, true))
2020-02-13 14:44:37 -08:00
{
if (ctx != &isc_g_appctx) {
/* Tool using multiple contexts */
atomic_store_release(&ctx->want_shutdown, true);
SIGNAL(&ctx->ready);
} else {
/* Normal single BIND9 context */
if (kill(getpid(), SIGTERM) < 0) {
char strbuf[ISC_STRERRORSIZE];
strerror_r(errno, strbuf, sizeof(strbuf));
isc_error_fatal(__FILE__, __LINE__,
"isc_app_shutdown() "
"kill: %s",
strbuf);
}
1999-05-12 22:35:40 +00:00
}
1999-05-12 09:44:35 +00:00
}
}
void
2020-02-13 14:44:37 -08:00
isc_app_shutdown(void) {
isc_app_ctxshutdown(&isc_g_appctx);
}
void
2020-02-13 14:44:37 -08:00
isc_app_ctxsuspend(isc_appctx_t *ctx) {
REQUIRE(VALID_APPCTX(ctx));
REQUIRE(atomic_load(&ctx->running));
2000-01-22 01:39:35 +00:00
/*
* Don't send the reload signal if we're shutting down.
*/
if (!atomic_load_acquire(&ctx->shutdown_requested)) {
if (ctx != &isc_g_appctx) {
/* Tool using multiple contexts */
atomic_store_release(&ctx->want_reload, true);
SIGNAL(&ctx->ready);
} else {
/* Normal single BIND9 context */
if (kill(getpid(), SIGHUP) < 0) {
char strbuf[ISC_STRERRORSIZE];
strerror_r(errno, strbuf, sizeof(strbuf));
isc_error_fatal(__FILE__, __LINE__,
"isc_app_reload() "
"kill: %s",
strbuf);
}
}
}
2000-01-22 01:39:35 +00:00
}
void
2020-02-13 14:44:37 -08:00
isc_app_reload(void) {
isc_app_ctxsuspend(&isc_g_appctx);
1999-05-12 09:44:35 +00:00
}
void
2020-02-13 14:44:37 -08:00
isc_app_ctxfinish(isc_appctx_t *ctx) {
REQUIRE(VALID_APPCTX(ctx));
isc_mutex_destroy(&ctx->lock);
isc_mutex_destroy(&ctx->readylock);
isc_condition_destroy(&ctx->ready);
}
void
2020-02-13 14:44:37 -08:00
isc_app_finish(void) {
isc_app_ctxfinish(&isc_g_appctx);
}
void
2020-02-13 14:44:37 -08:00
isc_app_block(void) {
REQUIRE(atomic_load_acquire(&isc_g_appctx.running));
REQUIRE(atomic_compare_exchange_strong_acq_rel(&isc_g_appctx.blocked,
&(bool){ false }, true));
sigset_t sset;
blockedthread = pthread_self();
RUNTIME_CHECK(sigemptyset(&sset) == 0 &&
sigaddset(&sset, SIGINT) == 0 &&
sigaddset(&sset, SIGTERM) == 0);
RUNTIME_CHECK(pthread_sigmask(SIG_UNBLOCK, &sset, NULL) == 0);
}
void
2020-02-13 14:44:37 -08:00
isc_app_unblock(void) {
REQUIRE(atomic_load_acquire(&isc_g_appctx.running));
REQUIRE(atomic_compare_exchange_strong_acq_rel(&isc_g_appctx.blocked,
&(bool){ true }, false));
REQUIRE(blockedthread == pthread_self());
sigset_t sset;
RUNTIME_CHECK(sigemptyset(&sset) == 0 &&
2008-01-18 23:46:58 +00:00
sigaddset(&sset, SIGINT) == 0 &&
sigaddset(&sset, SIGTERM) == 0);
RUNTIME_CHECK(pthread_sigmask(SIG_BLOCK, &sset, NULL) == 0);
}
isc_result_t
2020-02-13 14:44:37 -08:00
isc_appctx_create(isc_mem_t *mctx, isc_appctx_t **ctxp) {
isc_appctx_t *ctx;
REQUIRE(mctx != NULL);
REQUIRE(ctxp != NULL && *ctxp == NULL);
ctx = isc_mem_get(mctx, sizeof(*ctx));
Refactor taskmgr to run on top of netmgr This commit changes the taskmgr to run the individual tasks on the netmgr internal workers. While an effort has been put into keeping the taskmgr interface intact, couple of changes have been made: * The taskmgr has no concept of universal privileged mode - rather the tasks are either privileged or unprivileged (normal). The privileged tasks are run as a first thing when the netmgr is unpaused. There are now four different queues in in the netmgr: 1. priority queue - netievent on the priority queue are run even when the taskmgr enter exclusive mode and netmgr is paused. This is needed to properly start listening on the interfaces, free resources and resume. 2. privileged task queue - only privileged tasks are queued here and this is the first queue that gets processed when network manager is unpaused using isc_nm_resume(). All netmgr workers need to clean the privileged task queue before they all proceed normal operation. Both task queues are processed when the workers are finished. 3. task queue - only (traditional) task are scheduled here and this queue along with privileged task queues are process when the netmgr workers are finishing. This is needed to process the task shutdown events. 4. normal queue - this is the queue with netmgr events, e.g. reading, sending, callbacks and pretty much everything is processed here. * The isc_taskmgr_create() now requires initialized netmgr (isc_nm_t) object. * The isc_nm_destroy() function now waits for indefinite time, but it will print out the active objects when in tracing mode (-DNETMGR_TRACE=1 and -DNETMGR_TRACE_VERBOSE=1), the netmgr has been made a little bit more asynchronous and it might take longer time to shutdown all the active networking connections. * Previously, the isc_nm_stoplistening() was a synchronous operation. This has been changed and the isc_nm_stoplistening() just schedules the child sockets to stop listening and exits. This was needed to prevent a deadlock as the the (traditional) tasks are now executed on the netmgr threads. * The socket selection logic in isc__nm_udp_send() was flawed, but fortunatelly, it was broken, so we never hit the problem where we created uvreq_t on a socket from nmhandle_t, but then a different socket could be picked up and then we were trying to run the send callback on a socket that had different threadid than currently running.
2021-04-09 11:31:19 +02:00
*ctx = (isc_appctx_t){ .magic = 0 };
isc_mem_attach(mctx, &ctx->mctx);
Refactor taskmgr to run on top of netmgr This commit changes the taskmgr to run the individual tasks on the netmgr internal workers. While an effort has been put into keeping the taskmgr interface intact, couple of changes have been made: * The taskmgr has no concept of universal privileged mode - rather the tasks are either privileged or unprivileged (normal). The privileged tasks are run as a first thing when the netmgr is unpaused. There are now four different queues in in the netmgr: 1. priority queue - netievent on the priority queue are run even when the taskmgr enter exclusive mode and netmgr is paused. This is needed to properly start listening on the interfaces, free resources and resume. 2. privileged task queue - only privileged tasks are queued here and this is the first queue that gets processed when network manager is unpaused using isc_nm_resume(). All netmgr workers need to clean the privileged task queue before they all proceed normal operation. Both task queues are processed when the workers are finished. 3. task queue - only (traditional) task are scheduled here and this queue along with privileged task queues are process when the netmgr workers are finishing. This is needed to process the task shutdown events. 4. normal queue - this is the queue with netmgr events, e.g. reading, sending, callbacks and pretty much everything is processed here. * The isc_taskmgr_create() now requires initialized netmgr (isc_nm_t) object. * The isc_nm_destroy() function now waits for indefinite time, but it will print out the active objects when in tracing mode (-DNETMGR_TRACE=1 and -DNETMGR_TRACE_VERBOSE=1), the netmgr has been made a little bit more asynchronous and it might take longer time to shutdown all the active networking connections. * Previously, the isc_nm_stoplistening() was a synchronous operation. This has been changed and the isc_nm_stoplistening() just schedules the child sockets to stop listening and exits. This was needed to prevent a deadlock as the the (traditional) tasks are now executed on the netmgr threads. * The socket selection logic in isc__nm_udp_send() was flawed, but fortunatelly, it was broken, so we never hit the problem where we created uvreq_t on a socket from nmhandle_t, but then a different socket could be picked up and then we were trying to run the send callback on a socket that had different threadid than currently running.
2021-04-09 11:31:19 +02:00
ctx->magic = APPCTX_MAGIC;
*ctxp = ctx;
return (ISC_R_SUCCESS);
}
void
2020-02-13 14:44:37 -08:00
isc_appctx_destroy(isc_appctx_t **ctxp) {
isc_appctx_t *ctx;
REQUIRE(ctxp != NULL);
ctx = *ctxp;
*ctxp = NULL;
REQUIRE(VALID_APPCTX(ctx));
ctx->magic = 0;
isc_mem_putanddetach(&ctx->mctx, ctx, sizeof(*ctx));
}