Implement backlog for balancer

personal/stbuehler/wip
Stefan Bühler 13 years ago
parent 7bb671de4c
commit 3b202faf01

@ -9,17 +9,14 @@
* Options:
* none
* Actions:
* balance.rr <actions> - balance between actions (list or single action)
*
* balance.rr <actions> - balance between actions (list or single action) with RoundRobin
* balance.sqf <actions> - balance between actions (list or single action) with SQF
*
* Example config:
* balance.rr { fastcgi "127.0.0.1:9090"; };
*
* Todo:
* - add some select function (current: always take first)
* - support backlogs
*
* Author:
* Copyright (c) 2009 Stefan Bühler
* Copyright (c) 2009-2010 Stefan Bühler
*/
@ -46,34 +43,69 @@ typedef enum {
BM_ROUNDROBIN
} balancer_method;
typedef struct backend backend;
typedef struct balancer balancer;
typedef struct bcontext bcontext;
struct backend {
liAction *act;
guint load;
backend_state state;
ev_tstamp wake;
};
typedef struct backend backend;
struct balancer {
GMutex *lock;
liWorker *wrk;
GMutex *lock; /* balancer functions with "_" prefix need to be called with the lock being locked */
GArray *backends;
balancer_state state;
balancer_method method;
gint next_ndx;
ev_tstamp wake;
ev_async async;
gboolean delete_later; /* marked as "delete later in srv event loop" */
GQueue backlog;
gint backlog_limit;
ev_timer backlog_timer;
gint backlog_reactivate_now;
liPlugin *p;
};
typedef struct balancer balancer;
static balancer* balancer_new(liPlugin *p, balancer_method method) {
balancer *b = g_slice_new(balancer);
struct bcontext { /* context for a balancer in a vrequest */
gint selected; /* selected backend */
GList backlog_link;
liVRequestRef *ref;
gboolean scheduled;
};
static void balancer_timer_cb(struct ev_loop *loop, ev_timer *w, int revents);
static void balancer_async_cb(struct ev_loop *loop, ev_async *w, int revents);
static balancer* balancer_new(liWorker *wrk, liPlugin *p, balancer_method method) {
balancer *b = g_slice_new0(balancer);
b->wrk = wrk;
b->lock = g_mutex_new();
b->backends = g_array_new(FALSE, TRUE, sizeof(backend));
b->method = method;
b->state = BAL_ALIVE;
b->next_ndx = 0;
b->p = p;
b->backlog_limit = -1;
ev_init(&b->backlog_timer, balancer_timer_cb);
b->backlog_timer.data = b;
ev_init(&b->async, balancer_async_cb);
b->async.data = b;
ev_async_start(wrk->loop, &b->async);
ev_unref(wrk->loop);
return b;
}
@ -81,6 +113,10 @@ static void balancer_free(liServer *srv, balancer *b) {
guint i;
if (!b) return;
g_mutex_free(b->lock);
ev_timer_stop(b->wrk->loop, &b->backlog_timer);
li_ev_safe_ref_and_stop(ev_async_stop, b->wrk->loop, &b->async);
for (i = 0; i < b->backends->len; i++) {
backend *be = &g_array_index(b->backends, backend, i);
li_action_release(srv, be->act);
@ -122,8 +158,176 @@ static gboolean balancer_fill_backends(balancer *b, liServer *srv, liValue *val)
}
}
static void _balancer_context_backlog_unlink(balancer *b, bcontext *bc) {
if (NULL != bc->backlog_link.data) {
g_queue_unlink(&b->backlog, &bc->backlog_link);
li_vrequest_ref_release(bc->ref);
bc->backlog_link.data = NULL;
bc->backlog_link.next = bc->backlog_link.prev = NULL;
}
}
static void _balancer_context_backlog_push(balancer *b, gpointer *context, liVRequest *vr) {
bcontext *bc = *context;
if (NULL == bc) {
*context = bc = g_slice_new0(bcontext);
bc->selected = -1;
}
if (NULL == bc->backlog_link.data) {
bc->ref = li_vrequest_get_ref(vr);
bc->backlog_link.data = bc;
if (bc->scheduled) {
/* higher priority: has already been waiting and was scheduled */
g_queue_push_head_link(&b->backlog, &bc->backlog_link);
} else {
g_queue_push_tail_link(&b->backlog, &bc->backlog_link);
}
bc->scheduled = 0; /* reset scheduled flag */
}
}
/* returns FALSE if b was destroyed (only possible in event loop) */
static gboolean _balancer_backlog_update_watcher(liWorker *wrk, balancer *b) {
if (wrk != b->wrk) {
ev_async_send(b->wrk->loop, &b->async);
return TRUE;
}
if (b->delete_later) {
g_mutex_unlock(b->lock);
balancer_free(wrk->srv, b);
return FALSE;
}
if (b->state == BAL_ALIVE) {
ev_timer_stop(wrk->loop, &b->backlog_timer);
} else {
ev_timer_stop(wrk->loop, &b->backlog_timer);
ev_timer_set(&b->backlog_timer, b->wake - ev_now(wrk->loop), 0);
ev_timer_start(wrk->loop, &b->backlog_timer);
}
return TRUE;
}
/* returns FALSE if b was destroyed (only possible in event loop) */
static gboolean _balancer_backlog_schedule(liWorker *wrk, balancer *b) {
GList *it;
while (b->backlog_reactivate_now > 0) {
bcontext *bc;
it = g_queue_peek_head_link(&b->backlog);
if (NULL == it) {
/* backlog done */
b->state = BAL_ALIVE;
b->backlog_reactivate_now = 0;
b->wake = 0;
return _balancer_backlog_update_watcher(wrk, b);
}
bc = it->data;
bc->scheduled = 1;
li_vrequest_joblist_append_async(bc->ref);
g_queue_unlink(&b->backlog, it);
li_vrequest_ref_release(bc->ref);
it->data = NULL;
it->next = it->prev = NULL;
}
return _balancer_backlog_update_watcher(wrk, b);
}
static void balancer_timer_cb(struct ev_loop *loop, ev_timer *w, int revents) {
balancer *b = w->data;
int n;
UNUSED(loop);
UNUSED(revents);
g_mutex_lock(b->lock);
n = b->backends->len / 2;
if (n == 0) n = 1;
b->backlog_reactivate_now += n;
if (!_balancer_backlog_schedule(b->wrk, b)) return;
g_mutex_unlock(b->lock);
}
static void balancer_async_cb(struct ev_loop *loop, ev_async *w, int revents) {
balancer *b = w->data;
UNUSED(loop);
UNUSED(revents);
g_mutex_lock(b->lock);
if (!_balancer_backlog_update_watcher(b->wrk, b)) return;
g_mutex_unlock(b->lock);
}
static void balancer_context_free(liVRequest *vr, balancer *b, gpointer *context, gboolean success) {
bcontext *bc = *context;
if (!bc) return;
*context = NULL;
g_mutex_lock(b->lock);
_balancer_context_backlog_unlink(b, bc);
if (bc->selected >= 0) {
backend *be = &g_array_index(b->backends, backend, bc->selected);
be->load--;
bc->selected = -1;
if (success) {
/* reactivate it (if not alive), as it obviously isn't completely down */
be->state = BE_ALIVE;
b->backlog_reactivate_now++;
_balancer_backlog_schedule(vr->wrk, b);
}
}
g_mutex_unlock(b->lock);
g_slice_free(bcontext, bc);
}
static void _balancer_context_select_backend(balancer *b, gpointer *context, gint ndx) {
bcontext *bc = *context;
if (NULL == bc) {
*context = bc = g_slice_new0(bcontext);
bc->selected = -1;
}
_balancer_context_backlog_unlink(b, bc);
if (bc->selected >= 0) {
backend *be = &g_array_index(b->backends, backend, bc->selected);
be->load--;
}
bc->selected = ndx;
if (bc->selected >= 0) {
backend *be = &g_array_index(b->backends, backend, bc->selected);
be->load++;
b->next_ndx = ndx + 1;
}
}
static liHandlerResult balancer_act_select(liVRequest *vr, gboolean backlog_provided, gpointer param, gpointer *context) {
balancer *b = (balancer*) param;
bcontext *bc = *context;
gint be_ndx, load;
guint i, j;
backend *be;
@ -131,12 +335,42 @@ static liHandlerResult balancer_act_select(liVRequest *vr, gboolean backlog_prov
gboolean all_dead = TRUE;
gboolean debug = _OPTION(vr, b->p, 0).boolean;
UNUSED(backlog_provided);
be_ndx = -1;
g_mutex_lock(b->lock);
if (b->state != BAL_ALIVE && backlog_provided) {
/* don't use own backlog if someone else before us does provide it */
if (b->state == BAL_DOWN) {
li_vrequest_backend_dead(vr);
} else {
li_vrequest_backend_overloaded(vr);
}
}
/* don't backlog scheduled requests */
if ((NULL == bc || !bc->scheduled) && b->backlog.length > 0) {
if (-1 == b->backlog_limit || (gint)b->backlog.length < b->backlog_limit) {
/* backlog not full yet */
_balancer_context_backlog_push(b, context, vr);
g_mutex_unlock(b->lock);
return LI_HANDLER_WAIT_FOR_EVENT;
}
/* backlog full / no backlog */
if (b->state == BAL_DOWN) {
li_vrequest_backend_dead(vr);
} else {
li_vrequest_backend_overloaded(vr);
}
g_mutex_unlock(b->lock);
return LI_HANDLER_GO_ON;
}
switch (b->method) {
case BM_SQF:
load = -1;
@ -172,7 +406,31 @@ static liHandlerResult balancer_act_select(liVRequest *vr, gboolean backlog_prov
}
if (-1 == be_ndx) {
/* Couldn't find a active backend */
/* Couldn't find active backend */
if (b->state == BAL_ALIVE) {
b->state = all_dead ? BAL_DOWN : BAL_OVERLOADED;
b->wake = ev_now(vr->wrk->loop) + 10;
for (i = 0; i < b->backends->len; i++) {
be = &g_array_index(b->backends, backend, i);
if (b->wake > be->wake) b->wake = be->wake;
}
/* start backlog now */
b->backlog_reactivate_now = 0;
_balancer_backlog_update_watcher(vr->wrk, b);
}
if (-1 == b->backlog_limit || (gint)b->backlog.length < b->backlog_limit) {
_balancer_context_backlog_push(b, context, vr);
g_mutex_unlock(b->lock);
return LI_HANDLER_WAIT_FOR_EVENT;
}
if (all_dead) {
li_vrequest_backend_dead(vr);
} else {
@ -184,9 +442,7 @@ static liHandlerResult balancer_act_select(liVRequest *vr, gboolean backlog_prov
return LI_HANDLER_GO_ON;
}
be = &g_array_index(b->backends, backend, be_ndx);
be->load++;
b->next_ndx = be_ndx + 1;
_balancer_context_select_backend(b, context, be_ndx);
g_mutex_unlock(b->lock);
@ -195,29 +451,26 @@ static liHandlerResult balancer_act_select(liVRequest *vr, gboolean backlog_prov
}
li_action_enter(vr, be->act);
*context = GINT_TO_POINTER(be_ndx);
return LI_HANDLER_GO_ON;
}
static liHandlerResult balancer_act_fallback(liVRequest *vr, gboolean backlog_provided, gpointer param, gpointer *context, liBackendError error) {
balancer *b = (balancer*) param;
gint be_ndx = GPOINTER_TO_INT(*context);
bcontext *bc = *context;
backend *be;
gboolean debug = _OPTION(vr, b->p, 0).boolean;
UNUSED(backlog_provided);
if (be_ndx < 0) return LI_HANDLER_GO_ON;
be = &g_array_index(b->backends, backend, be_ndx);
if (!bc || bc->selected < 0) return LI_HANDLER_GO_ON;
be = &g_array_index(b->backends, backend, bc->selected);
if (debug || CORE_OPTION(LI_CORE_OPTION_DEBUG_REQUEST_HANDLING).boolean){
VR_DEBUG(vr, "balancer fallback: %i (error: %i)", be_ndx, error);
VR_DEBUG(vr, "balancer fallback: %i (error: %i)", bc->selected, error);
}
g_mutex_lock(b->lock);
be->load--;
_balancer_context_select_backend(b, context, -1);
if (error == BACKEND_OVERLOAD || be->load > 0) {
/* long timeout for overload - we will enable the backend anyway if another request finishs */
@ -231,49 +484,44 @@ static liHandlerResult balancer_act_fallback(liVRequest *vr, gboolean backlog_pr
be->state = BE_DOWN;
}
g_mutex_unlock(b->lock);
*context = GINT_TO_POINTER(-1);
if (b->wake > be->wake) b->wake = be->wake;
g_mutex_unlock(b->lock);
return balancer_act_select(vr, backlog_provided, param, context);
}
static liHandlerResult balancer_act_finished(liVRequest *vr, gpointer param, gpointer context) {
balancer *b = (balancer*) param;
gint be_ndx = GPOINTER_TO_INT(context);
backend *be;
bcontext *bc = context;
gboolean debug = _OPTION(vr, b->p, 0).boolean;
UNUSED(vr);
if (be_ndx < 0) return LI_HANDLER_GO_ON;
be = &g_array_index(b->backends, backend, be_ndx);
if (!bc) return LI_HANDLER_GO_ON;
if (debug){
VR_DEBUG(vr, "balancer finished: %i", be_ndx);
if (debug && bc->selected >= 0) {
VR_DEBUG(vr, "balancer finished: %i", bc->selected);
}
g_mutex_lock(b->lock);
/* TODO implement backlog */
be->load--;
/* reactivate it (if not alive), as it obviously isn't completely down */
be->state = BE_ALIVE;
g_mutex_unlock(b->lock);
balancer_context_free(vr, b, &context, TRUE);
return LI_HANDLER_GO_ON;
}
static void balancer_act_free(liServer *srv, gpointer param) {
balancer_free(srv, (balancer*) param);
balancer *b = param;
UNUSED(srv);
g_mutex_lock(b->lock);
b->delete_later = TRUE;
_balancer_backlog_update_watcher(NULL, b);
g_mutex_unlock(b->lock);
}
static liAction* balancer_create(liServer *srv, liWorker *wrk, liPlugin* p, liValue *val, gpointer userdata) {
balancer *b;
UNUSED(wrk); UNUSED(p); UNUSED(userdata);
if (!val) {
ERROR(srv, "%s", "need parameter");
@ -281,7 +529,7 @@ static liAction* balancer_create(liServer *srv, liWorker *wrk, liPlugin* p, liVa
}
/* userdata contains the method */
b = balancer_new(p, GPOINTER_TO_INT(userdata));
b = balancer_new(wrk, p, GPOINTER_TO_INT(userdata));
if (!balancer_fill_backends(b, srv, val)) {
balancer_free(srv, b);
return NULL;

Loading…
Cancel
Save