mirror of
https://git.haproxy.org/git/haproxy.git/
synced 2025-09-20 21:31:28 +02:00
Insert some missing includes statement in QUIC source files. This was detected after the next commit which adjust the include list used in quic_conn-t.h file.
137 lines
4.5 KiB
C
137 lines
4.5 KiB
C
/*
|
|
* Congestion controller handling.
|
|
*
|
|
* This file contains definitions for QUIC congestion control.
|
|
*
|
|
* Copyright 2019 HAProxy Technologies, Frederic Lecaille <flecaille@haproxy.com>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation, version 2.1
|
|
* exclusively.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include <haproxy/global.h>
|
|
#include <haproxy/proto_quic.h>
|
|
#include <haproxy/quic_cc.h>
|
|
#include <haproxy/quic_pacing.h>
|
|
#include <haproxy/thread.h>
|
|
|
|
struct quic_cc_algo *default_quic_cc_algo = &quic_cc_algo_cubic;
|
|
|
|
/*
|
|
* Initialize <cc> congestion control with <algo> as algorithm depending on <ipv4>
|
|
* a boolean which is true for an IPv4 path.
|
|
*/
|
|
void quic_cc_init(struct quic_cc *cc,
|
|
struct quic_cc_algo *algo, struct quic_conn *qc)
|
|
{
|
|
cc->qc = qc;
|
|
cc->algo = algo;
|
|
if (cc->algo->init)
|
|
(cc->algo->init(cc));
|
|
}
|
|
|
|
/* Send <ev> event to <cc> congestion controller. */
|
|
void quic_cc_event(struct quic_cc *cc, struct quic_cc_event *ev)
|
|
{
|
|
if (cc->algo->event)
|
|
cc->algo->event(cc, ev);
|
|
}
|
|
|
|
void quic_cc_state_trace(struct buffer *buf, const struct quic_cc *cc)
|
|
{
|
|
cc->algo->state_trace(buf, cc);
|
|
}
|
|
|
|
/* Return interval in nanoseconds between each datagram emission for a smooth pacing. */
|
|
uint quic_cc_default_pacing_inter(const struct quic_cc *cc)
|
|
{
|
|
struct quic_cc_path *path = container_of(cc, struct quic_cc_path, cc);
|
|
return path->loss.srtt * 1000000 / (path->cwnd / path->mtu + 1) + 1;
|
|
}
|
|
|
|
/* Returns true if congestion window on path ought to be increased. */
|
|
static int quic_cwnd_may_increase(const struct quic_cc_path *path)
|
|
{
|
|
/* RFC 9002 7.8. Underutilizing the Congestion Window
|
|
*
|
|
* When bytes in flight is smaller than the congestion window and
|
|
* sending is not pacing limited, the congestion window is
|
|
* underutilized. This can happen due to insufficient application data
|
|
* or flow control limits. When this occurs, the congestion window
|
|
* SHOULD NOT be increased in either slow start or congestion avoidance.
|
|
*/
|
|
|
|
/* Consider that congestion window can be increased if it is at least
|
|
* half full or window size is less than 16k. These conditions should
|
|
* not be restricted too much to prevent slow window growing.
|
|
*/
|
|
return 2 * path->in_flight >= path->cwnd || path->cwnd < 16384;
|
|
}
|
|
|
|
/* Calculate ratio of free memory relative to the maximum configured limit. */
|
|
static int quic_cc_max_win_ratio(void)
|
|
{
|
|
uint64_t tot, free = 0;
|
|
int ratio = 100;
|
|
|
|
if (global.tune.quic_frontend_max_tx_mem) {
|
|
tot = cshared_read(&quic_mem_diff);
|
|
if (global.tune.quic_frontend_max_tx_mem > tot)
|
|
free = global.tune.quic_frontend_max_tx_mem - tot;
|
|
ratio = free * 100 / global.tune.quic_frontend_max_tx_mem;
|
|
}
|
|
|
|
return ratio;
|
|
}
|
|
|
|
/* Restore congestion window for <path> to its minimal value. */
|
|
void quic_cc_path_reset(struct quic_cc_path *path)
|
|
{
|
|
const uint64_t old = path->cwnd;
|
|
path->cwnd = path->limit_min;
|
|
cshared_add(&quic_mem_diff, path->cwnd - old);
|
|
}
|
|
|
|
/* Set congestion window for <path> to <val>. Min and max limits are enforced. */
|
|
void quic_cc_path_set(struct quic_cc_path *path, uint64_t val)
|
|
{
|
|
const uint64_t old = path->cwnd;
|
|
const uint64_t limit_max = path->limit_max * quic_cc_max_win_ratio() / 100;
|
|
|
|
path->cwnd = QUIC_MIN(val, limit_max);
|
|
path->cwnd = QUIC_MAX(path->cwnd, path->limit_min);
|
|
cshared_add(&quic_mem_diff, path->cwnd - old);
|
|
|
|
path->cwnd_last_max = QUIC_MAX(path->cwnd, path->cwnd_last_max);
|
|
}
|
|
|
|
/* Increment congestion window for <path> with <val>. Min and max limits are
|
|
* enforced. Contrary to quic_cc_path_set(), increase is performed only if a
|
|
* certain minimal level of the window was already filled.
|
|
*/
|
|
void quic_cc_path_inc(struct quic_cc_path *path, uint64_t val)
|
|
{
|
|
const uint64_t old = path->cwnd;
|
|
uint64_t limit_max;
|
|
|
|
if (quic_cwnd_may_increase(path)) {
|
|
limit_max = path->limit_max * quic_cc_max_win_ratio() / 100;
|
|
path->cwnd = QUIC_MIN(path->cwnd + val, limit_max);
|
|
path->cwnd = QUIC_MAX(path->cwnd, path->limit_min);
|
|
cshared_add(&quic_mem_diff, path->cwnd - old);
|
|
|
|
path->cwnd_last_max = QUIC_MAX(path->cwnd, path->cwnd_last_max);
|
|
}
|
|
}
|