blob: 9b132482df3e4c715ba3b20ed80a0b31d553519c [file] [log] [blame]
/*
* NewReno congestion control algorithm.
*
* This file contains definitions for QUIC congestion control.
*
* Copyright 2019 HAProxy Technologies, Frederic Lecaille <flecaille@haproxy.com>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation, version 2.1
* exclusively.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with this library; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include <haproxy/api-t.h>
#include <haproxy/buf.h>
#include <haproxy/chunk.h>
#include <haproxy/quic_cc.h>
#include <haproxy/quic_conn-t.h>
#include <haproxy/trace.h>
#define TRACE_SOURCE &trace_quic
/* Newreno state */
struct nr {
uint32_t state;
uint32_t ssthresh;
uint32_t recovery_start_time;
uint32_t remain_acked;
};
static int quic_cc_nr_init(struct quic_cc *cc)
{
struct nr *nr = quic_cc_priv(cc);
nr->state = QUIC_CC_ST_SS;
nr->ssthresh = QUIC_CC_INFINITE_SSTHESH;
nr->recovery_start_time = 0;
nr->remain_acked = 0;
return 1;
}
/* Re-enter slow start state. */
static void quic_cc_nr_slow_start(struct quic_cc *cc)
{
struct quic_path *path;
struct nr *nr = quic_cc_priv(cc);
path = container_of(cc, struct quic_path, cc);
path->cwnd = path->min_cwnd;
/* Re-entering slow start state. */
nr->state = QUIC_CC_ST_SS;
/* Recovery start time reset */
nr->recovery_start_time = 0;
}
/* Enter a recovery period. */
static void quic_cc_nr_enter_recovery(struct quic_cc *cc)
{
struct quic_path *path;
struct nr *nr = quic_cc_priv(cc);
path = container_of(cc, struct quic_path, cc);
nr->recovery_start_time = now_ms;
nr->ssthresh = path->cwnd >> 1;
path->cwnd = QUIC_MAX(nr->ssthresh, (uint32_t)path->min_cwnd);
nr->state = QUIC_CC_ST_RP;
}
/* Slow start callback. */
static void quic_cc_nr_ss_cb(struct quic_cc *cc, struct quic_cc_event *ev)
{
struct quic_path *path;
struct nr *nr = quic_cc_priv(cc);
TRACE_ENTER(QUIC_EV_CONN_CC, cc->qc);
TRACE_PROTO("CC reno", QUIC_EV_CONN_CC, cc->qc, ev);
path = container_of(cc, struct quic_path, cc);
switch (ev->type) {
case QUIC_CC_EVT_ACK:
path->cwnd += ev->ack.acked;
path->mcwnd = QUIC_MAX(path->cwnd, path->mcwnd);
/* Exit to congestion avoidance if slow start threshold is reached. */
if (path->cwnd > nr->ssthresh)
nr->state = QUIC_CC_ST_CA;
break;
case QUIC_CC_EVT_LOSS:
quic_cc_nr_enter_recovery(cc);
break;
case QUIC_CC_EVT_ECN_CE:
/* XXX TO DO XXX */
break;
}
TRACE_PROTO("CC reno", QUIC_EV_CONN_CC, cc->qc, NULL, cc);
TRACE_LEAVE(QUIC_EV_CONN_CC, cc->qc);
}
/* Congestion avoidance callback. */
static void quic_cc_nr_ca_cb(struct quic_cc *cc, struct quic_cc_event *ev)
{
struct quic_path *path;
struct nr *nr = quic_cc_priv(cc);
TRACE_ENTER(QUIC_EV_CONN_CC, cc->qc);
TRACE_PROTO("CC reno", QUIC_EV_CONN_CC, cc->qc, ev);
path = container_of(cc, struct quic_path, cc);
switch (ev->type) {
case QUIC_CC_EVT_ACK:
{
uint64_t acked;
/* Increasing the congestion window by (acked / cwnd)
*/
acked = ev->ack.acked * path->mtu + nr->remain_acked;
nr->remain_acked = acked % path->cwnd;
path->cwnd += acked / path->cwnd;
path->mcwnd = QUIC_MAX(path->cwnd, path->mcwnd);
break;
}
case QUIC_CC_EVT_LOSS:
quic_cc_nr_enter_recovery(cc);
break;
case QUIC_CC_EVT_ECN_CE:
/* XXX TO DO XXX */
break;
}
out:
TRACE_PROTO("CC reno", QUIC_EV_CONN_CC, cc->qc, NULL, cc);
TRACE_LEAVE(QUIC_EV_CONN_CC, cc->qc);
}
/* Recovery period callback. */
static void quic_cc_nr_rp_cb(struct quic_cc *cc, struct quic_cc_event *ev)
{
struct quic_path *path;
struct nr *nr = quic_cc_priv(cc);
TRACE_ENTER(QUIC_EV_CONN_CC, cc->qc);
TRACE_PROTO("CC reno", QUIC_EV_CONN_CC, cc->qc, ev);
path = container_of(cc, struct quic_path, cc);
switch (ev->type) {
case QUIC_CC_EVT_ACK:
/* RFC 9022 7.3.2. Recovery
* A recovery period ends and the sender enters congestion avoidance when a
* packet sent during the recovery period is acknowledged.
*/
if (tick_is_le(ev->ack.time_sent, nr->recovery_start_time)) {
TRACE_PROTO("CC reno (still in recovery period)", QUIC_EV_CONN_CC, cc->qc, ev);
goto leave;
}
nr->state = QUIC_CC_ST_CA;
nr->recovery_start_time = TICK_ETERNITY;
path->cwnd = nr->ssthresh;
break;
case QUIC_CC_EVT_LOSS:
/* Do nothing */
break;
case QUIC_CC_EVT_ECN_CE:
/* XXX TO DO XXX */
break;
}
leave:
TRACE_PROTO("CC reno", QUIC_EV_CONN_CC, cc->qc, ev);
TRACE_ENTER(QUIC_EV_CONN_CC, cc->qc, ev);
}
static void quic_cc_nr_state_trace(struct buffer *buf, const struct quic_cc *cc)
{
struct quic_path *path;
struct nr *nr = quic_cc_priv(cc);
path = container_of(cc, struct quic_path, cc);
chunk_appendf(buf, " state=%s cwnd=%llu mcwnd=%llu ssthresh=%ld rpst=%dms pktloss=%llu",
quic_cc_state_str(nr->state),
(unsigned long long)path->cwnd,
(unsigned long long)path->mcwnd,
(long)nr->ssthresh,
!tick_isset(nr->recovery_start_time) ? -1 :
TICKS_TO_MS(tick_remain(nr->recovery_start_time, now_ms)),
(unsigned long long)path->loss.nb_lost_pkt);
}
static void (*quic_cc_nr_state_cbs[])(struct quic_cc *cc,
struct quic_cc_event *ev) = {
[QUIC_CC_ST_SS] = quic_cc_nr_ss_cb,
[QUIC_CC_ST_CA] = quic_cc_nr_ca_cb,
[QUIC_CC_ST_RP] = quic_cc_nr_rp_cb,
};
static void quic_cc_nr_event(struct quic_cc *cc, struct quic_cc_event *ev)
{
struct nr *nr = quic_cc_priv(cc);
return quic_cc_nr_state_cbs[nr->state](cc, ev);
}
struct quic_cc_algo quic_cc_algo_nr = {
.type = QUIC_CC_ALGO_TP_NEWRENO,
.init = quic_cc_nr_init,
.event = quic_cc_nr_event,
.slow_start = quic_cc_nr_slow_start,
.state_trace = quic_cc_nr_state_trace,
};