mirror of
https://git.haproxy.org/git/haproxy.git/
synced 2025-11-27 13:51:00 +01:00
CLEANUP: activity: make the number of sched activity entries more configurable
This removes all the hard-coded 8-bit and 256 entries to use a pair of macros instead so that we can more easily experiment with larger table sizes if needed.
This commit is contained in:
parent
a9a2384612
commit
a3423873fe
@ -115,6 +115,9 @@ struct activity {
|
|||||||
char __end[0] __attribute__((aligned(64))); // align size to 64.
|
char __end[0] __attribute__((aligned(64))); // align size to 64.
|
||||||
};
|
};
|
||||||
|
|
||||||
|
/* 256 entries for callers * callees should be highly sufficient (~45 seen usually) */
|
||||||
|
#define SCHED_ACT_HASH_BITS 8
|
||||||
|
#define SCHED_ACT_HASH_BUCKETS (1U << SCHED_ACT_HASH_BITS)
|
||||||
|
|
||||||
/* global profiling stats from the scheduler: each entry corresponds to a
|
/* global profiling stats from the scheduler: each entry corresponds to a
|
||||||
* task or tasklet ->process function pointer, with a number of calls and
|
* task or tasklet ->process function pointer, with a number of calls and
|
||||||
|
|||||||
@ -27,7 +27,7 @@
|
|||||||
|
|
||||||
extern unsigned int profiling;
|
extern unsigned int profiling;
|
||||||
extern struct activity activity[MAX_THREADS];
|
extern struct activity activity[MAX_THREADS];
|
||||||
extern struct sched_activity sched_activity[256];
|
extern struct sched_activity sched_activity[SCHED_ACT_HASH_BUCKETS];
|
||||||
|
|
||||||
void report_stolen_time(uint64_t stolen);
|
void report_stolen_time(uint64_t stolen);
|
||||||
void activity_count_runtime(uint32_t run_time);
|
void activity_count_runtime(uint32_t run_time);
|
||||||
|
|||||||
@ -45,8 +45,8 @@ unsigned int profiling __read_mostly = HA_PROF_TASKS_AOFF;
|
|||||||
/* One struct per thread containing all collected measurements */
|
/* One struct per thread containing all collected measurements */
|
||||||
struct activity activity[MAX_THREADS] __attribute__((aligned(64))) = { };
|
struct activity activity[MAX_THREADS] __attribute__((aligned(64))) = { };
|
||||||
|
|
||||||
/* One struct per function pointer hash entry (256 values, 0=collision) */
|
/* One struct per function pointer hash entry (SCHED_ACT_HASH_BUCKETS values, 0=collision) */
|
||||||
struct sched_activity sched_activity[256] __attribute__((aligned(64))) = { };
|
struct sched_activity sched_activity[SCHED_ACT_HASH_BUCKETS] __attribute__((aligned(64))) = { };
|
||||||
|
|
||||||
|
|
||||||
#ifdef USE_MEMORY_PROFILING
|
#ifdef USE_MEMORY_PROFILING
|
||||||
@ -440,7 +440,7 @@ static int cli_parse_set_profiling(char **args, char *payload, struct appctx *ap
|
|||||||
while (!_HA_ATOMIC_CAS(&profiling, &old, (old & ~HA_PROF_TASKS_MASK) | HA_PROF_TASKS_ON))
|
while (!_HA_ATOMIC_CAS(&profiling, &old, (old & ~HA_PROF_TASKS_MASK) | HA_PROF_TASKS_ON))
|
||||||
;
|
;
|
||||||
/* also flush current profiling stats */
|
/* also flush current profiling stats */
|
||||||
for (i = 0; i < 256; i++) {
|
for (i = 0; i < SCHED_ACT_HASH_BUCKETS; i++) {
|
||||||
HA_ATOMIC_STORE(&sched_activity[i].calls, 0);
|
HA_ATOMIC_STORE(&sched_activity[i].calls, 0);
|
||||||
HA_ATOMIC_STORE(&sched_activity[i].cpu_time, 0);
|
HA_ATOMIC_STORE(&sched_activity[i].cpu_time, 0);
|
||||||
HA_ATOMIC_STORE(&sched_activity[i].lat_time, 0);
|
HA_ATOMIC_STORE(&sched_activity[i].lat_time, 0);
|
||||||
@ -531,7 +531,7 @@ static int cmp_memprof_addr(const void *a, const void *b)
|
|||||||
*/
|
*/
|
||||||
struct sched_activity *sched_activity_entry(struct sched_activity *array, const void *func)
|
struct sched_activity *sched_activity_entry(struct sched_activity *array, const void *func)
|
||||||
{
|
{
|
||||||
uint32_t hash = ptr_hash(func, 8);
|
uint32_t hash = ptr_hash(func, SCHED_ACT_HASH_BITS);
|
||||||
struct sched_activity *ret;
|
struct sched_activity *ret;
|
||||||
const void *old = NULL;
|
const void *old = NULL;
|
||||||
|
|
||||||
@ -564,7 +564,7 @@ struct sched_activity *sched_activity_entry(struct sched_activity *array, const
|
|||||||
static int cli_io_handler_show_profiling(struct appctx *appctx)
|
static int cli_io_handler_show_profiling(struct appctx *appctx)
|
||||||
{
|
{
|
||||||
struct show_prof_ctx *ctx = appctx->svcctx;
|
struct show_prof_ctx *ctx = appctx->svcctx;
|
||||||
struct sched_activity tmp_activity[256] __attribute__((aligned(64)));
|
struct sched_activity tmp_activity[SCHED_ACT_HASH_BUCKETS] __attribute__((aligned(64)));
|
||||||
#ifdef USE_MEMORY_PROFILING
|
#ifdef USE_MEMORY_PROFILING
|
||||||
struct memprof_stats tmp_memstats[MEMPROF_HASH_BUCKETS + 1];
|
struct memprof_stats tmp_memstats[MEMPROF_HASH_BUCKETS + 1];
|
||||||
unsigned long long tot_alloc_calls, tot_free_calls;
|
unsigned long long tot_alloc_calls, tot_free_calls;
|
||||||
@ -611,9 +611,9 @@ static int cli_io_handler_show_profiling(struct appctx *appctx)
|
|||||||
|
|
||||||
memcpy(tmp_activity, sched_activity, sizeof(tmp_activity));
|
memcpy(tmp_activity, sched_activity, sizeof(tmp_activity));
|
||||||
if (ctx->by_addr)
|
if (ctx->by_addr)
|
||||||
qsort(tmp_activity, 256, sizeof(tmp_activity[0]), cmp_sched_activity_addr);
|
qsort(tmp_activity, SCHED_ACT_HASH_BUCKETS, sizeof(tmp_activity[0]), cmp_sched_activity_addr);
|
||||||
else
|
else
|
||||||
qsort(tmp_activity, 256, sizeof(tmp_activity[0]), cmp_sched_activity_calls);
|
qsort(tmp_activity, SCHED_ACT_HASH_BUCKETS, sizeof(tmp_activity[0]), cmp_sched_activity_calls);
|
||||||
|
|
||||||
if (!ctx->linenum)
|
if (!ctx->linenum)
|
||||||
chunk_appendf(&trash, "Tasks activity:\n"
|
chunk_appendf(&trash, "Tasks activity:\n"
|
||||||
@ -621,7 +621,7 @@ static int cli_io_handler_show_profiling(struct appctx *appctx)
|
|||||||
|
|
||||||
max_lines = ctx->maxcnt;
|
max_lines = ctx->maxcnt;
|
||||||
if (!max_lines)
|
if (!max_lines)
|
||||||
max_lines = 256;
|
max_lines = SCHED_ACT_HASH_BUCKETS;
|
||||||
|
|
||||||
for (i = ctx->linenum; i < max_lines && tmp_activity[i].calls; i++) {
|
for (i = ctx->linenum; i < max_lines && tmp_activity[i].calls; i++) {
|
||||||
ctx->linenum = i;
|
ctx->linenum = i;
|
||||||
@ -794,7 +794,7 @@ static int cli_parse_show_profiling(char **args, char *payload, struct appctx *a
|
|||||||
*/
|
*/
|
||||||
static int cli_io_handler_show_tasks(struct appctx *appctx)
|
static int cli_io_handler_show_tasks(struct appctx *appctx)
|
||||||
{
|
{
|
||||||
struct sched_activity tmp_activity[256] __attribute__((aligned(64)));
|
struct sched_activity tmp_activity[SCHED_ACT_HASH_BUCKETS] __attribute__((aligned(64)));
|
||||||
struct stconn *sc = appctx_sc(appctx);
|
struct stconn *sc = appctx_sc(appctx);
|
||||||
struct buffer *name_buffer = get_trash_chunk();
|
struct buffer *name_buffer = get_trash_chunk();
|
||||||
struct sched_activity *entry;
|
struct sched_activity *entry;
|
||||||
@ -894,16 +894,16 @@ static int cli_io_handler_show_tasks(struct appctx *appctx)
|
|||||||
chunk_reset(&trash);
|
chunk_reset(&trash);
|
||||||
|
|
||||||
tot_calls = 0;
|
tot_calls = 0;
|
||||||
for (i = 0; i < 256; i++)
|
for (i = 0; i < SCHED_ACT_HASH_BUCKETS; i++)
|
||||||
tot_calls += tmp_activity[i].calls;
|
tot_calls += tmp_activity[i].calls;
|
||||||
|
|
||||||
qsort(tmp_activity, 256, sizeof(tmp_activity[0]), cmp_sched_activity_calls);
|
qsort(tmp_activity, SCHED_ACT_HASH_BUCKETS, sizeof(tmp_activity[0]), cmp_sched_activity_calls);
|
||||||
|
|
||||||
chunk_appendf(&trash, "Running tasks: %d (%d threads)\n"
|
chunk_appendf(&trash, "Running tasks: %d (%d threads)\n"
|
||||||
" function places %% lat_tot lat_avg\n",
|
" function places %% lat_tot lat_avg\n",
|
||||||
(int)tot_calls, global.nbthread);
|
(int)tot_calls, global.nbthread);
|
||||||
|
|
||||||
for (i = 0; i < 256 && tmp_activity[i].calls; i++) {
|
for (i = 0; i < SCHED_ACT_HASH_BUCKETS && tmp_activity[i].calls; i++) {
|
||||||
chunk_reset(name_buffer);
|
chunk_reset(name_buffer);
|
||||||
|
|
||||||
if (!tmp_activity[i].func)
|
if (!tmp_activity[i].func)
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user