| /* |
| * NewReno congestion control algorithm. |
| * |
| * This file contains definitions for QUIC congestion control. |
| * |
| * Copyright 2019 HAProxy Technologies, Frederic Lecaille <flecaille@haproxy.com> |
| * |
| * This library is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU Lesser General Public |
| * License as published by the Free Software Foundation, version 2.1 |
| * exclusively. |
| * |
| * This library is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * Lesser General Public License for more details. |
| * |
| * You should have received a copy of the GNU Lesser General Public |
| * License along with this library; if not, write to the Free Software |
| * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| */ |
| |
| #include <haproxy/api-t.h> |
| #include <haproxy/buf.h> |
| #include <haproxy/chunk.h> |
| #include <haproxy/quic_cc.h> |
| #include <haproxy/quic_conn-t.h> |
| #include <haproxy/trace.h> |
| |
| #define TRACE_SOURCE &trace_quic |
| |
| /* Newreno state */ |
| struct nr { |
| uint32_t state; |
| uint32_t ssthresh; |
| uint32_t recovery_start_time; |
| uint32_t remain_acked; |
| }; |
| |
| static int quic_cc_nr_init(struct quic_cc *cc) |
| { |
| struct nr *nr = quic_cc_priv(cc); |
| |
| nr->state = QUIC_CC_ST_SS; |
| nr->ssthresh = QUIC_CC_INFINITE_SSTHESH; |
| nr->recovery_start_time = 0; |
| nr->remain_acked = 0; |
| |
| return 1; |
| } |
| |
| /* Re-enter slow start state. */ |
| static void quic_cc_nr_slow_start(struct quic_cc *cc) |
| { |
| struct quic_path *path; |
| struct nr *nr = quic_cc_priv(cc); |
| |
| path = container_of(cc, struct quic_path, cc); |
| path->cwnd = path->min_cwnd; |
| /* Re-entering slow start state. */ |
| nr->state = QUIC_CC_ST_SS; |
| /* Recovery start time reset */ |
| nr->recovery_start_time = 0; |
| } |
| |
| /* Enter a recovery period. */ |
| static void quic_cc_nr_enter_recovery(struct quic_cc *cc) |
| { |
| struct quic_path *path; |
| struct nr *nr = quic_cc_priv(cc); |
| |
| path = container_of(cc, struct quic_path, cc); |
| nr->recovery_start_time = now_ms; |
| nr->ssthresh = path->cwnd >> 1; |
| path->cwnd = QUIC_MAX(nr->ssthresh, (uint32_t)path->min_cwnd); |
| nr->state = QUIC_CC_ST_RP; |
| } |
| |
| /* Slow start callback. */ |
| static void quic_cc_nr_ss_cb(struct quic_cc *cc, struct quic_cc_event *ev) |
| { |
| struct quic_path *path; |
| struct nr *nr = quic_cc_priv(cc); |
| |
| TRACE_ENTER(QUIC_EV_CONN_CC, cc->qc); |
| TRACE_PROTO("CC reno", QUIC_EV_CONN_CC, cc->qc, ev); |
| path = container_of(cc, struct quic_path, cc); |
| switch (ev->type) { |
| case QUIC_CC_EVT_ACK: |
| path->cwnd += ev->ack.acked; |
| path->mcwnd = QUIC_MAX(path->cwnd, path->mcwnd); |
| /* Exit to congestion avoidance if slow start threshold is reached. */ |
| if (path->cwnd > nr->ssthresh) |
| nr->state = QUIC_CC_ST_CA; |
| break; |
| |
| case QUIC_CC_EVT_LOSS: |
| quic_cc_nr_enter_recovery(cc); |
| break; |
| |
| case QUIC_CC_EVT_ECN_CE: |
| /* XXX TO DO XXX */ |
| break; |
| } |
| TRACE_PROTO("CC reno", QUIC_EV_CONN_CC, cc->qc, NULL, cc); |
| TRACE_LEAVE(QUIC_EV_CONN_CC, cc->qc); |
| } |
| |
| /* Congestion avoidance callback. */ |
| static void quic_cc_nr_ca_cb(struct quic_cc *cc, struct quic_cc_event *ev) |
| { |
| struct quic_path *path; |
| struct nr *nr = quic_cc_priv(cc); |
| |
| TRACE_ENTER(QUIC_EV_CONN_CC, cc->qc); |
| TRACE_PROTO("CC reno", QUIC_EV_CONN_CC, cc->qc, ev); |
| path = container_of(cc, struct quic_path, cc); |
| switch (ev->type) { |
| case QUIC_CC_EVT_ACK: |
| { |
| uint64_t acked; |
| |
| /* Increasing the congestion window by (acked / cwnd) |
| */ |
| acked = ev->ack.acked * path->mtu + nr->remain_acked; |
| nr->remain_acked = acked % path->cwnd; |
| path->cwnd += acked / path->cwnd; |
| path->mcwnd = QUIC_MAX(path->cwnd, path->mcwnd); |
| break; |
| } |
| |
| case QUIC_CC_EVT_LOSS: |
| quic_cc_nr_enter_recovery(cc); |
| break; |
| |
| case QUIC_CC_EVT_ECN_CE: |
| /* XXX TO DO XXX */ |
| break; |
| } |
| |
| out: |
| TRACE_PROTO("CC reno", QUIC_EV_CONN_CC, cc->qc, NULL, cc); |
| TRACE_LEAVE(QUIC_EV_CONN_CC, cc->qc); |
| } |
| |
| /* Recovery period callback. */ |
| static void quic_cc_nr_rp_cb(struct quic_cc *cc, struct quic_cc_event *ev) |
| { |
| struct quic_path *path; |
| struct nr *nr = quic_cc_priv(cc); |
| |
| TRACE_ENTER(QUIC_EV_CONN_CC, cc->qc); |
| TRACE_PROTO("CC reno", QUIC_EV_CONN_CC, cc->qc, ev); |
| path = container_of(cc, struct quic_path, cc); |
| switch (ev->type) { |
| case QUIC_CC_EVT_ACK: |
| /* RFC 9022 7.3.2. Recovery |
| * A recovery period ends and the sender enters congestion avoidance when a |
| * packet sent during the recovery period is acknowledged. |
| */ |
| if (tick_is_le(ev->ack.time_sent, nr->recovery_start_time)) { |
| TRACE_PROTO("CC reno (still in recovery period)", QUIC_EV_CONN_CC, cc->qc, ev); |
| goto leave; |
| } |
| |
| nr->state = QUIC_CC_ST_CA; |
| nr->recovery_start_time = TICK_ETERNITY; |
| path->cwnd = nr->ssthresh; |
| break; |
| case QUIC_CC_EVT_LOSS: |
| /* Do nothing */ |
| break; |
| case QUIC_CC_EVT_ECN_CE: |
| /* XXX TO DO XXX */ |
| break; |
| } |
| |
| leave: |
| TRACE_PROTO("CC reno", QUIC_EV_CONN_CC, cc->qc, ev); |
| TRACE_ENTER(QUIC_EV_CONN_CC, cc->qc, ev); |
| } |
| static void quic_cc_nr_state_trace(struct buffer *buf, const struct quic_cc *cc) |
| { |
| struct quic_path *path; |
| struct nr *nr = quic_cc_priv(cc); |
| |
| path = container_of(cc, struct quic_path, cc); |
| chunk_appendf(buf, " state=%s cwnd=%llu mcwnd=%llu ssthresh=%ld rpst=%dms pktloss=%llu", |
| quic_cc_state_str(nr->state), |
| (unsigned long long)path->cwnd, |
| (unsigned long long)path->mcwnd, |
| (long)nr->ssthresh, |
| !tick_isset(nr->recovery_start_time) ? -1 : |
| TICKS_TO_MS(tick_remain(nr->recovery_start_time, now_ms)), |
| (unsigned long long)path->loss.nb_lost_pkt); |
| } |
| |
| static void (*quic_cc_nr_state_cbs[])(struct quic_cc *cc, |
| struct quic_cc_event *ev) = { |
| [QUIC_CC_ST_SS] = quic_cc_nr_ss_cb, |
| [QUIC_CC_ST_CA] = quic_cc_nr_ca_cb, |
| [QUIC_CC_ST_RP] = quic_cc_nr_rp_cb, |
| }; |
| |
| static void quic_cc_nr_event(struct quic_cc *cc, struct quic_cc_event *ev) |
| { |
| struct nr *nr = quic_cc_priv(cc); |
| |
| return quic_cc_nr_state_cbs[nr->state](cc, ev); |
| } |
| |
| struct quic_cc_algo quic_cc_algo_nr = { |
| .type = QUIC_CC_ALGO_TP_NEWRENO, |
| .init = quic_cc_nr_init, |
| .event = quic_cc_nr_event, |
| .slow_start = quic_cc_nr_slow_start, |
| .state_trace = quic_cc_nr_state_trace, |
| }; |
| |