mirror of
https://github.com/haproxy/haproxy.git
synced 2026-03-01 12:51:09 -05:00
Many changes have been made to do so. First, the fd_updt array, where all pending FDs for polling are stored, is now a thread-local array. Then 3 locks have been added to protect, respectively, the fdtab array, the fd_cache array and poll information. In addition, a lock for each entry in the fdtab array has been added to protect all accesses to a specific FD or its information. For pollers, according to the poller, the way to manage the concurrency is different. There is a poller loop on each thread. So the set of monitored FDs may need to be protected. epoll and kqueue are thread-safe per-se, so there few things to do to protect these pollers. This is not possible with select and poll, so there is no sharing between the threads. The poller on each thread is independant from others. Finally, per-thread init/deinit functions are used for each pollers and for FD part for manage thread-local ressources. Now, you must be carefull when a FD is created during the HAProxy startup. All update on the FD state must be made in the threads context and never before their creation. This is mandatory because fd_updt array is thread-local and initialized only for threads. Because there is no pollers for the main one, this array remains uninitialized in this context. For this reason, listeners are now enabled in run_thread_poll_loop function, just like the worker pipe.
271 lines
5.6 KiB
C
271 lines
5.6 KiB
C
/*
|
|
* FD polling functions for FreeBSD kqueue()
|
|
*
|
|
* Copyright 2000-2014 Willy Tarreau <w@1wt.eu>
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation; either version
|
|
* 2 of the License, or (at your option) any later version.
|
|
*
|
|
*/
|
|
|
|
#include <unistd.h>
|
|
#include <sys/time.h>
|
|
#include <sys/types.h>
|
|
|
|
#include <sys/event.h>
|
|
#include <sys/time.h>
|
|
|
|
#include <common/compat.h>
|
|
#include <common/config.h>
|
|
#include <common/ticks.h>
|
|
#include <common/time.h>
|
|
#include <common/tools.h>
|
|
|
|
#include <types/global.h>
|
|
|
|
#include <proto/fd.h>
|
|
|
|
|
|
/* private data */
|
|
static int kqueue_fd;
|
|
static THREAD_LOCAL struct kevent *kev = NULL;
|
|
|
|
/*
|
|
* kqueue() poller
|
|
*/
|
|
REGPRM2 static void _do_poll(struct poller *p, int exp)
|
|
{
|
|
int status;
|
|
int count, fd, delta_ms;
|
|
struct timespec timeout;
|
|
int updt_idx, en, eo;
|
|
int changes = 0;
|
|
|
|
/* first, scan the update list to find changes */
|
|
for (updt_idx = 0; updt_idx < fd_nbupdt; updt_idx++) {
|
|
fd = fd_updt[updt_idx];
|
|
|
|
if (!fdtab[fd].owner)
|
|
continue;
|
|
|
|
SPIN_LOCK(FD_LOCK, &fdtab[fd].lock);
|
|
fdtab[fd].updated = 0;
|
|
fdtab[fd].new = 0;
|
|
|
|
eo = fdtab[fd].state;
|
|
en = fd_compute_new_polled_status(eo);
|
|
fdtab[fd].state = en;
|
|
SPIN_UNLOCK(FD_LOCK, &fdtab[fd].lock);
|
|
|
|
if ((eo ^ en) & FD_EV_POLLED_RW) {
|
|
/* poll status changed */
|
|
if ((eo ^ en) & FD_EV_POLLED_R) {
|
|
/* read poll status changed */
|
|
if (en & FD_EV_POLLED_R) {
|
|
EV_SET(&kev[changes], fd, EVFILT_READ, EV_ADD, 0, 0, NULL);
|
|
changes++;
|
|
}
|
|
else {
|
|
EV_SET(&kev[changes], fd, EVFILT_READ, EV_DELETE, 0, 0, NULL);
|
|
changes++;
|
|
}
|
|
}
|
|
|
|
if ((eo ^ en) & FD_EV_POLLED_W) {
|
|
/* write poll status changed */
|
|
if (en & FD_EV_POLLED_W) {
|
|
EV_SET(&kev[changes], fd, EVFILT_WRITE, EV_ADD, 0, 0, NULL);
|
|
changes++;
|
|
}
|
|
else {
|
|
EV_SET(&kev[changes], fd, EVFILT_WRITE, EV_DELETE, 0, 0, NULL);
|
|
changes++;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
if (changes)
|
|
kevent(kqueue_fd, kev, changes, NULL, 0, NULL);
|
|
fd_nbupdt = 0;
|
|
|
|
delta_ms = 0;
|
|
timeout.tv_sec = 0;
|
|
timeout.tv_nsec = 0;
|
|
|
|
if (!exp) {
|
|
delta_ms = MAX_DELAY_MS;
|
|
timeout.tv_sec = (MAX_DELAY_MS / 1000);
|
|
timeout.tv_nsec = (MAX_DELAY_MS % 1000) * 1000000;
|
|
}
|
|
else if (!tick_is_expired(exp, now_ms)) {
|
|
delta_ms = TICKS_TO_MS(tick_remain(now_ms, exp)) + 1;
|
|
if (delta_ms > MAX_DELAY_MS)
|
|
delta_ms = MAX_DELAY_MS;
|
|
timeout.tv_sec = (delta_ms / 1000);
|
|
timeout.tv_nsec = (delta_ms % 1000) * 1000000;
|
|
}
|
|
|
|
fd = MIN(maxfd, global.tune.maxpollevents);
|
|
gettimeofday(&before_poll, NULL);
|
|
status = kevent(kqueue_fd, // int kq
|
|
NULL, // const struct kevent *changelist
|
|
0, // int nchanges
|
|
kev, // struct kevent *eventlist
|
|
fd, // int nevents
|
|
&timeout); // const struct timespec *timeout
|
|
tv_update_date(delta_ms, status);
|
|
measure_idle();
|
|
|
|
for (count = 0; count < status; count++) {
|
|
unsigned int n = 0;
|
|
fd = kev[count].ident;
|
|
|
|
if (!fdtab[fd].owner)
|
|
continue;
|
|
|
|
if (kev[count].filter == EVFILT_READ) {
|
|
if (kev[count].data)
|
|
n |= FD_POLL_IN;
|
|
if (kev[count].flags & EV_EOF)
|
|
n |= FD_POLL_HUP;
|
|
}
|
|
else if (kev[count].filter == EVFILT_WRITE) {
|
|
n |= FD_POLL_OUT;
|
|
if (kev[count].flags & EV_EOF)
|
|
n |= FD_POLL_ERR;
|
|
}
|
|
|
|
fd_update_events(fd, n);
|
|
}
|
|
}
|
|
|
|
|
|
static int init_kqueue_per_thread()
|
|
{
|
|
/* we can have up to two events per fd (*/
|
|
kev = calloc(1, sizeof(struct kevent) * 2 * global.maxsock);
|
|
if (kev == NULL)
|
|
return 0;
|
|
return 1;
|
|
}
|
|
|
|
static void deinit_kqueue_per_thread()
|
|
{
|
|
free(kev);
|
|
}
|
|
|
|
/*
|
|
* Initialization of the kqueue() poller.
|
|
* Returns 0 in case of failure, non-zero in case of success. If it fails, it
|
|
* disables the poller by setting its pref to 0.
|
|
*/
|
|
REGPRM1 static int _do_init(struct poller *p)
|
|
{
|
|
p->private = NULL;
|
|
|
|
kqueue_fd = kqueue();
|
|
if (kqueue_fd < 0)
|
|
goto fail_fd;
|
|
|
|
if (global.nbthread > 1) {
|
|
hap_register_per_thread_init(init_kqueue_per_thread);
|
|
hap_register_per_thread_deinit(deinit_kqueue_per_thread);
|
|
}
|
|
else if (!init_kqueue_per_thread())
|
|
goto fail_kev;
|
|
|
|
return 1;
|
|
|
|
fail_kev:
|
|
close(kqueue_fd);
|
|
kqueue_fd = -1;
|
|
fail_fd:
|
|
p->pref = 0;
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Termination of the kqueue() poller.
|
|
* Memory is released and the poller is marked as unselectable.
|
|
*/
|
|
REGPRM1 static void _do_term(struct poller *p)
|
|
{
|
|
free(kev);
|
|
|
|
if (kqueue_fd >= 0) {
|
|
close(kqueue_fd);
|
|
kqueue_fd = -1;
|
|
}
|
|
|
|
p->private = NULL;
|
|
p->pref = 0;
|
|
}
|
|
|
|
/*
|
|
* Check that the poller works.
|
|
* Returns 1 if OK, otherwise 0.
|
|
*/
|
|
REGPRM1 static int _do_test(struct poller *p)
|
|
{
|
|
int fd;
|
|
|
|
fd = kqueue();
|
|
if (fd < 0)
|
|
return 0;
|
|
close(fd);
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Recreate the kqueue file descriptor after a fork(). Returns 1 if OK,
|
|
* otherwise 0. Note that some pollers need to be reopened after a fork()
|
|
* (such as kqueue), and some others may fail to do so in a chroot.
|
|
*/
|
|
REGPRM1 static int _do_fork(struct poller *p)
|
|
{
|
|
if (kqueue_fd >= 0)
|
|
close(kqueue_fd);
|
|
kqueue_fd = kqueue();
|
|
if (kqueue_fd < 0)
|
|
return 0;
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* It is a constructor, which means that it will automatically be called before
|
|
* main(). This is GCC-specific but it works at least since 2.95.
|
|
* Special care must be taken so that it does not need any uninitialized data.
|
|
*/
|
|
__attribute__((constructor))
|
|
static void _do_register(void)
|
|
{
|
|
struct poller *p;
|
|
|
|
if (nbpollers >= MAX_POLLERS)
|
|
return;
|
|
|
|
kqueue_fd = -1;
|
|
p = &pollers[nbpollers++];
|
|
|
|
p->name = "kqueue";
|
|
p->pref = 300;
|
|
p->flags = HAP_POLL_F_RDHUP;
|
|
p->private = NULL;
|
|
|
|
p->clo = NULL;
|
|
p->test = _do_test;
|
|
p->init = _do_init;
|
|
p->term = _do_term;
|
|
p->poll = _do_poll;
|
|
p->fork = _do_fork;
|
|
}
|
|
|
|
|
|
/*
|
|
* Local variables:
|
|
* c-indent-level: 8
|
|
* c-basic-offset: 8
|
|
* End:
|
|
*/
|