mirror of
https://git.haproxy.org/git/haproxy.git/
synced 2025-08-07 07:37:02 +02:00
Add a condition to release asap the quic_conn memory when the connection is in "connection close" state from ->close() QUIC xprt callback.
176 lines
4.7 KiB
C
176 lines
4.7 KiB
C
/*
|
|
* QUIC xprt layer. Act as an abstraction between quic_conn and MUX layers.
|
|
*
|
|
* Copyright 2020 HAProxy Technologies, Frederic Lecaille <flecaille@haproxy.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation; either version
|
|
* 2 of the License, or (at your option) any later version.
|
|
*
|
|
*/
|
|
|
|
#include <haproxy/api.h>
|
|
#include <haproxy/connection.h>
|
|
#include <haproxy/quic_conn.h>
|
|
#include <haproxy/ssl_sock.h>
|
|
#include <haproxy/trace.h>
|
|
|
|
#define TRACE_SOURCE &trace_quic
|
|
|
|
static void quic_close(struct connection *conn, void *xprt_ctx)
|
|
{
|
|
struct ssl_sock_ctx *conn_ctx = xprt_ctx;
|
|
struct quic_conn *qc = conn_ctx->qc;
|
|
|
|
TRACE_ENTER(QUIC_EV_CONN_CLOSE, qc);
|
|
|
|
/* Next application data can be dropped. */
|
|
qc->mux_state = QC_MUX_RELEASED;
|
|
|
|
/* If the quic-conn timer has already expired or if already in "connection close"
|
|
* state, free the quic-conn.
|
|
*/
|
|
if (qc->flags & (QUIC_FL_CONN_EXP_TIMER|QUIC_FL_CONN_CLOSING)) {
|
|
quic_conn_release(qc);
|
|
qc = NULL;
|
|
goto leave;
|
|
}
|
|
|
|
/* Schedule a CONNECTION_CLOSE emission. If process stopping is in
|
|
* progress, quic-conn idle-timer will be scheduled immediately after
|
|
* its emission to ensure an immediate connection closing.
|
|
*/
|
|
qc_check_close_on_released_mux(qc);
|
|
leave:
|
|
TRACE_LEAVE(QUIC_EV_CONN_CLOSE, qc);
|
|
}
|
|
|
|
/* Called from the upper layer, to subscribe <es> to events <event_type>. The
|
|
* event subscriber <es> is not allowed to change from a previous call as long
|
|
* as at least one event is still subscribed. The <event_type> must only be a
|
|
* combination of SUB_RETRY_RECV and SUB_RETRY_SEND. It always returns 0.
|
|
*/
|
|
static int quic_conn_subscribe(struct connection *conn, void *xprt_ctx, int event_type, struct wait_event *es)
|
|
{
|
|
struct quic_conn *qc = conn->handle.qc;
|
|
|
|
TRACE_ENTER(QUIC_EV_CONN_SUB, qc);
|
|
|
|
BUG_ON(event_type & ~(SUB_RETRY_SEND|SUB_RETRY_RECV));
|
|
BUG_ON(qc->subs && qc->subs != es);
|
|
|
|
es->events |= event_type;
|
|
qc->subs = es;
|
|
|
|
/* TODO implement a check_events to detect if subscriber should be
|
|
* woken up immediately ?
|
|
*/
|
|
|
|
if (event_type & SUB_RETRY_RECV)
|
|
TRACE_DEVEL("subscribe(recv)", QUIC_EV_CONN_XPRTRECV, qc);
|
|
|
|
if (event_type & SUB_RETRY_SEND)
|
|
TRACE_DEVEL("subscribe(send)", QUIC_EV_CONN_XPRTSEND, qc);
|
|
|
|
TRACE_LEAVE(QUIC_EV_CONN_SUB, qc);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Called from the upper layer, to unsubscribe <es> from events <event_type>.
|
|
* The <es> pointer is not allowed to differ from the one passed to the
|
|
* subscribe() call. It always returns zero.
|
|
*/
|
|
static int quic_conn_unsubscribe(struct connection *conn, void *xprt_ctx, int event_type, struct wait_event *es)
|
|
{
|
|
struct quic_conn *qc = conn->handle.qc;
|
|
|
|
TRACE_ENTER(QUIC_EV_CONN_SUB, qc);
|
|
|
|
if (event_type & SUB_RETRY_RECV)
|
|
TRACE_DEVEL("unsubscribe(recv)", QUIC_EV_CONN_XPRTRECV, qc);
|
|
if (event_type & SUB_RETRY_SEND)
|
|
TRACE_DEVEL("unsubscribe(send)", QUIC_EV_CONN_XPRTSEND, qc);
|
|
|
|
es->events &= ~event_type;
|
|
if (!es->events)
|
|
qc->subs = NULL;
|
|
|
|
/* TODO implement ignore_events similar to conn_unsubscribe() ? */
|
|
|
|
TRACE_LEAVE(QUIC_EV_CONN_SUB, qc);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Store in <xprt_ctx> the context attached to <conn>.
|
|
* Returns always 0.
|
|
*/
|
|
static int qc_conn_init(struct connection *conn, void **xprt_ctx)
|
|
{
|
|
struct quic_conn *qc = NULL;
|
|
|
|
TRACE_ENTER(QUIC_EV_CONN_NEW, qc);
|
|
|
|
/* do not store the context if already set */
|
|
if (*xprt_ctx)
|
|
goto out;
|
|
|
|
*xprt_ctx = conn->handle.qc->xprt_ctx;
|
|
|
|
out:
|
|
TRACE_LEAVE(QUIC_EV_CONN_NEW, qc);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Start the QUIC transport layer */
|
|
static int qc_xprt_start(struct connection *conn, void *ctx)
|
|
{
|
|
int ret = 0;
|
|
struct quic_conn *qc;
|
|
|
|
qc = conn->handle.qc;
|
|
TRACE_ENTER(QUIC_EV_CONN_NEW, qc);
|
|
|
|
if (qc->flags & QUIC_FL_CONN_AFFINITY_CHANGED)
|
|
qc_finalize_affinity_rebind(qc);
|
|
|
|
/* mux-quic can now be considered ready. */
|
|
qc->mux_state = QC_MUX_READY;
|
|
|
|
ret = 1;
|
|
out:
|
|
TRACE_LEAVE(QUIC_EV_CONN_NEW, qc);
|
|
return ret;
|
|
}
|
|
|
|
static struct ssl_sock_ctx *qc_get_ssl_sock_ctx(struct connection *conn)
|
|
{
|
|
if (!conn || conn->xprt != xprt_get(XPRT_QUIC) || !conn->handle.qc || !conn->xprt_ctx)
|
|
return NULL;
|
|
|
|
return conn->handle.qc->xprt_ctx;
|
|
}
|
|
|
|
/* transport-layer operations for QUIC connections. */
|
|
static struct xprt_ops ssl_quic = {
|
|
.close = quic_close,
|
|
.subscribe = quic_conn_subscribe,
|
|
.unsubscribe = quic_conn_unsubscribe,
|
|
.init = qc_conn_init,
|
|
.start = qc_xprt_start,
|
|
.prepare_bind_conf = ssl_sock_prepare_bind_conf,
|
|
.destroy_bind_conf = ssl_sock_destroy_bind_conf,
|
|
.get_alpn = ssl_sock_get_alpn,
|
|
.get_ssl_sock_ctx = qc_get_ssl_sock_ctx,
|
|
.name = "QUIC",
|
|
};
|
|
|
|
static void __quic_conn_init(void)
|
|
{
|
|
xprt_register(XPRT_QUIC, &ssl_quic);
|
|
}
|
|
INITCALL0(STG_REGISTER, __quic_conn_init);
|