mirror of
https://git.haproxy.org/git/haproxy.git/
synced 2025-11-15 16:01:02 +01:00
MAJOR: task: make use of the scope-aware ebtree functions
Currently the task scheduler suffers from an O(n) lookup when
skipping tasks that are not for the current thread. The reason
is that eb32_lookup_ge() has no information about the current
thread so it always revisits many tasks for other threads before
finding its own tasks.
This is particularly visible with HTTP/2 since the number of
concurrent streams created at once causes long series of tasks
for the same stream in the scheduler. With only 10 connections
and 100 streams each, by running on two threads, the performance
drops from 640kreq/s to 11.2kreq/s! Lookup metrics show that for
only 200000 task lookups, 430 million skips had to be performed,
which means that on average, each lookup leads to 2150 nodes to
be visited.
This commit backports the principle of scope lookups for ebtrees
from the ebtree_v7 development tree. The idea is that each node
contains a mask indicating the union of the scopes for the nodes
below it, which is fed during insertion, and used during lookups.
Then during lookups, branches that do not contain any leaf matching
the requested scope are simply ignored. This perfectly matches a
thread mask, allowing a thread to only extract the tasks it cares
about from the run queue, and to always find them in O(log(n))
instead of O(n). Thus the scheduler uses tid_bit and
task->thread_mask as the ebtree scope here.
Doing this has recovered most of the performance, as can be seen on
the test below with two threads, 10 connections, 100 streams each,
and 1 million requests total :
Before After Gain
test duration : 89.6s 4.73s x19
HTTP requests/s (DEBUG) : 11200 211300 x19
HTTP requests/s (PROD) : 15900 447000 x28
spin_lock time : 85.2s 0.46s /185
time per lookup : 13us 40ns /325
Even when going to 6 threads (on 3 hyperthreaded CPU cores), the
performance stays around 284000 req/s, showing that the contention
is much lower.
A test showed that there's no benefit in using this for the wait queue
though.
This commit is contained in:
parent
8878b6c4cb
commit
8d38805d3d
@ -32,6 +32,7 @@
|
||||
#include <common/ticks.h>
|
||||
#include <common/hathreads.h>
|
||||
|
||||
#include <eb32sctree.h>
|
||||
#include <eb32tree.h>
|
||||
|
||||
#include <types/global.h>
|
||||
@ -163,7 +164,7 @@ static inline struct task *task_unlink_wq(struct task *t)
|
||||
*/
|
||||
static inline struct task *__task_unlink_rq(struct task *t)
|
||||
{
|
||||
eb32_delete(&t->rq);
|
||||
eb32sc_delete(&t->rq);
|
||||
tasks_run_queue--;
|
||||
if (likely(t->nice))
|
||||
niced_tasks--;
|
||||
|
||||
@ -26,6 +26,7 @@
|
||||
|
||||
#include <common/config.h>
|
||||
#include <common/mini-clist.h>
|
||||
#include <eb32sctree.h>
|
||||
#include <eb32tree.h>
|
||||
|
||||
/* values for task->state */
|
||||
@ -63,7 +64,7 @@ struct notification {
|
||||
|
||||
/* The base for all tasks */
|
||||
struct task {
|
||||
struct eb32_node rq; /* ebtree node used to hold the task in the run queue */
|
||||
struct eb32sc_node rq; /* ebtree node used to hold the task in the run queue */
|
||||
unsigned short state; /* task state : bit field of TASK_* */
|
||||
unsigned short pending_state; /* pending states for running talk */
|
||||
short nice; /* the task's current nice value from -1024 to +1024 */
|
||||
|
||||
15
src/task.c
15
src/task.c
@ -17,6 +17,7 @@
|
||||
#include <common/mini-clist.h>
|
||||
#include <common/standard.h>
|
||||
#include <common/time.h>
|
||||
#include <eb32sctree.h>
|
||||
#include <eb32tree.h>
|
||||
|
||||
#include <proto/proxy.h>
|
||||
@ -74,7 +75,7 @@ struct task *__task_wakeup(struct task *t)
|
||||
* if task is running
|
||||
*/
|
||||
t->state = t->pending_state;
|
||||
eb32_insert(&rqueue, &t->rq);
|
||||
eb32sc_insert(&rqueue, &t->rq, t->thread_mask);
|
||||
return t;
|
||||
}
|
||||
|
||||
@ -188,7 +189,7 @@ void process_runnable_tasks()
|
||||
struct task *t;
|
||||
int i;
|
||||
int max_processed;
|
||||
struct eb32_node *rq_next;
|
||||
struct eb32sc_node *rq_next;
|
||||
int rewind;
|
||||
struct task *local_tasks[16];
|
||||
int local_tasks_count;
|
||||
@ -212,13 +213,13 @@ void process_runnable_tasks()
|
||||
*/
|
||||
|
||||
rewind = 0;
|
||||
rq_next = eb32_lookup_ge(&rqueue, rqueue_ticks - TIMER_LOOK_BACK);
|
||||
rq_next = eb32sc_lookup_ge(&rqueue, rqueue_ticks - TIMER_LOOK_BACK, tid_bit);
|
||||
if (!rq_next) {
|
||||
/* we might have reached the end of the tree, typically because
|
||||
* <rqueue_ticks> is in the first half and we're first scanning
|
||||
* the last half. Let's loop back to the beginning of the tree now.
|
||||
*/
|
||||
rq_next = eb32_first(&rqueue);
|
||||
rq_next = eb32sc_first(&rqueue, tid_bit);
|
||||
if (!rq_next) {
|
||||
break;
|
||||
}
|
||||
@ -227,8 +228,8 @@ void process_runnable_tasks()
|
||||
|
||||
local_tasks_count = 0;
|
||||
while (local_tasks_count < 16) {
|
||||
t = eb32_entry(rq_next, struct task, rq);
|
||||
rq_next = eb32_next(rq_next);
|
||||
t = eb32sc_entry(rq_next, struct task, rq);
|
||||
rq_next = eb32sc_next(rq_next, tid_bit);
|
||||
if (t->thread_mask & tid_bit) {
|
||||
/* detach the task from the queue */
|
||||
__task_unlink_rq(t);
|
||||
@ -238,7 +239,7 @@ void process_runnable_tasks()
|
||||
local_tasks[local_tasks_count++] = t;
|
||||
}
|
||||
if (!rq_next) {
|
||||
if (rewind || !(rq_next = eb32_first(&rqueue))) {
|
||||
if (rewind || !(rq_next = eb32sc_first(&rqueue, tid_bit))) {
|
||||
break;
|
||||
}
|
||||
rewind = 1;
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user