simple-ipc: split async server initialization and running
To start an async ipc server, you call ipc_server_run_async(). That initializes the ipc_server_data object, and starts all of the threads running, which may immediately start serving clients. This can create some awkward timing problems, though. In the fsmonitor daemon (the sole user of the simple-ipc system), we want to create the ipc server early in the process, which means we may start serving clients before the rest of the daemon is fully initialized. To solve this, let's break run_async() into two parts: an initialization which allocates all data and spawns the threads (without letting them run), and a start function which actually lets them begin work. Since we have two simple-ipc implementations, we have to handle this twice: - in ipc-unix-socket.c, we have a central listener thread which hands connections off to worker threads using a work_available mutex. We can hold that mutex after init, and release it when we're ready to start. We do need an extra "started" flag so that we know whether the main thread is holding the mutex or not (e.g., if we prematurely stop the server, we want to make sure all of the worker threads are released to hear about the shutdown). - in ipc-win32.c, we don't have a central mutex. So we'll introduce a new startup_barrier mutex, which we'll similarly hold until we're ready to let the threads proceed. We again need a "started" flag here to make sure that we release the barrier mutex when shutting down, so that the sub-threads can proceed to the finish. I've renamed the run_async() function to init_async() to make sure we catch all callers, since they'll now need to call the matching start_async(). We could leave run_async() as a wrapper that does both, but there's not much point. There are only two callers, one of which is fsmonitor, which will want to actually do work between the two calls. And the other is just a test-tool wrapper. For now I've added the start_async() calls in fsmonitor where they would otherwise have happened, so there should be no behavior change with this patch. Signed-off-by: Jeff King <peff@peff.net> Acked-by: Koji Nakamaru <koji.nakamaru@gree.net> Signed-off-by: Junio C Hamano <gitster@pobox.com>
This commit is contained in:

committed by
Junio C Hamano

parent
4f71522dfb
commit
766fce69e9
@ -371,6 +371,9 @@ struct ipc_server_data {
|
||||
HANDLE hEventStopRequested;
|
||||
struct ipc_server_thread_data *thread_list;
|
||||
int is_stopped;
|
||||
|
||||
pthread_mutex_t startup_barrier;
|
||||
int started;
|
||||
};
|
||||
|
||||
enum connect_result {
|
||||
@ -526,6 +529,16 @@ static int use_connection(struct ipc_server_thread_data *server_thread_data)
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void wait_for_startup_barrier(struct ipc_server_data *server_data)
|
||||
{
|
||||
/*
|
||||
* Temporarily hold the startup_barrier mutex before starting,
|
||||
* which lets us know that it's OK to start serving requests.
|
||||
*/
|
||||
pthread_mutex_lock(&server_data->startup_barrier);
|
||||
pthread_mutex_unlock(&server_data->startup_barrier);
|
||||
}
|
||||
|
||||
/*
|
||||
* Thread proc for an IPC server worker thread. It handles a series of
|
||||
* connections from clients. It cleans and reuses the hPipe between each
|
||||
@ -550,6 +563,8 @@ static void *server_thread_proc(void *_server_thread_data)
|
||||
memset(&oConnect, 0, sizeof(oConnect));
|
||||
oConnect.hEvent = hEventConnected;
|
||||
|
||||
wait_for_startup_barrier(server_thread_data->server_data);
|
||||
|
||||
for (;;) {
|
||||
cr = wait_for_connection(server_thread_data, &oConnect);
|
||||
|
||||
@ -752,10 +767,10 @@ static HANDLE create_new_pipe(wchar_t *wpath, int is_first)
|
||||
return hPipe;
|
||||
}
|
||||
|
||||
int ipc_server_run_async(struct ipc_server_data **returned_server_data,
|
||||
const char *path, const struct ipc_server_opts *opts,
|
||||
ipc_server_application_cb *application_cb,
|
||||
void *application_data)
|
||||
int ipc_server_init_async(struct ipc_server_data **returned_server_data,
|
||||
const char *path, const struct ipc_server_opts *opts,
|
||||
ipc_server_application_cb *application_cb,
|
||||
void *application_data)
|
||||
{
|
||||
struct ipc_server_data *server_data;
|
||||
wchar_t wpath[MAX_PATH];
|
||||
@ -787,6 +802,13 @@ int ipc_server_run_async(struct ipc_server_data **returned_server_data,
|
||||
strbuf_addstr(&server_data->buf_path, path);
|
||||
wcscpy(server_data->wpath, wpath);
|
||||
|
||||
/*
|
||||
* Hold the startup_barrier lock so that no threads will progress
|
||||
* until ipc_server_start_async() is called.
|
||||
*/
|
||||
pthread_mutex_init(&server_data->startup_barrier, NULL);
|
||||
pthread_mutex_lock(&server_data->startup_barrier);
|
||||
|
||||
if (nr_threads < 1)
|
||||
nr_threads = 1;
|
||||
|
||||
@ -837,6 +859,15 @@ int ipc_server_run_async(struct ipc_server_data **returned_server_data,
|
||||
return 0;
|
||||
}
|
||||
|
||||
void ipc_server_start_async(struct ipc_server_data *server_data)
|
||||
{
|
||||
if (!server_data || server_data->started)
|
||||
return;
|
||||
|
||||
server_data->started = 1;
|
||||
pthread_mutex_unlock(&server_data->startup_barrier);
|
||||
}
|
||||
|
||||
int ipc_server_stop_async(struct ipc_server_data *server_data)
|
||||
{
|
||||
if (!server_data)
|
||||
@ -850,6 +881,13 @@ int ipc_server_stop_async(struct ipc_server_data *server_data)
|
||||
* We DO NOT attempt to force them to drop an active connection.
|
||||
*/
|
||||
SetEvent(server_data->hEventStopRequested);
|
||||
|
||||
/*
|
||||
* If we haven't yet told the threads they are allowed to run,
|
||||
* do so now, so they can receive the shutdown event.
|
||||
*/
|
||||
ipc_server_start_async(server_data);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
@ -900,5 +938,7 @@ void ipc_server_free(struct ipc_server_data *server_data)
|
||||
free(std);
|
||||
}
|
||||
|
||||
pthread_mutex_destroy(&server_data->startup_barrier);
|
||||
|
||||
free(server_data);
|
||||
}
|
||||
|
Reference in New Issue
Block a user