blob: ffb2b790403d84f21e2cbf11013b5c950e5c1f46 [file] [log] [blame]
Christopher Faulet1329f2a2021-12-16 17:32:56 +01001/*
Willy Tarreau4596fe22022-05-17 19:07:51 +02002 * stream connector management functions
Christopher Faulet1329f2a2021-12-16 17:32:56 +01003 *
4 * Copyright 2021 Christopher Faulet <cfaulet@haproxy.com>
5 *
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
10 *
11 */
12
13#include <haproxy/api.h>
Christopher Faulet37046632022-04-01 11:36:58 +020014#include <haproxy/applet.h>
Christopher Faulet1329f2a2021-12-16 17:32:56 +010015#include <haproxy/connection.h>
16#include <haproxy/conn_stream.h>
Christopher Faulet19bd7282022-04-01 13:58:09 +020017#include <haproxy/cs_utils.h>
Christopher Faulet5e29b762022-04-04 08:58:34 +020018#include <haproxy/check.h>
19#include <haproxy/http_ana.h>
20#include <haproxy/pipe.h>
Christopher Faulet1329f2a2021-12-16 17:32:56 +010021#include <haproxy/pool.h>
Christopher Faulet1329f2a2021-12-16 17:32:56 +010022
Willy Tarreau4596fe22022-05-17 19:07:51 +020023DECLARE_POOL(pool_head_connstream, "stconn", sizeof(struct stconn));
Willy Tarreauea59b022022-05-17 17:53:22 +020024DECLARE_POOL(pool_head_sedesc, "sedesc", sizeof(struct sedesc));
Christopher Faulet1329f2a2021-12-16 17:32:56 +010025
Willy Tarreau3a3f4802022-05-17 18:28:19 +020026/* functions used by default on a detached stream connector */
Willy Tarreau4596fe22022-05-17 19:07:51 +020027static void sc_app_shutr(struct stconn *cs);
28static void sc_app_shutw(struct stconn *cs);
29static void sc_app_chk_rcv(struct stconn *cs);
30static void sc_app_chk_snd(struct stconn *cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +020031
Willy Tarreau3a3f4802022-05-17 18:28:19 +020032/* functions used on a mux-based stream connector */
Willy Tarreau4596fe22022-05-17 19:07:51 +020033static void sc_app_shutr_conn(struct stconn *cs);
34static void sc_app_shutw_conn(struct stconn *cs);
35static void sc_app_chk_rcv_conn(struct stconn *cs);
36static void sc_app_chk_snd_conn(struct stconn *cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +020037
Willy Tarreau3a3f4802022-05-17 18:28:19 +020038/* functions used on an applet-based stream connector */
Willy Tarreau4596fe22022-05-17 19:07:51 +020039static void sc_app_shutr_applet(struct stconn *cs);
40static void sc_app_shutw_applet(struct stconn *cs);
41static void sc_app_chk_rcv_applet(struct stconn *cs);
42static void sc_app_chk_snd_applet(struct stconn *cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +020043
Willy Tarreau2f2318d2022-05-18 10:17:16 +020044static int cs_conn_process(struct stconn *cs);
45static int cs_conn_recv(struct stconn *cs);
46static int cs_conn_send(struct stconn *cs);
47static int cs_applet_process(struct stconn *cs);
48
Willy Tarreau3a3f4802022-05-17 18:28:19 +020049/* stream connector operations for connections */
50struct sc_app_ops sc_app_conn_ops = {
51 .chk_rcv = sc_app_chk_rcv_conn,
52 .chk_snd = sc_app_chk_snd_conn,
53 .shutr = sc_app_shutr_conn,
54 .shutw = sc_app_shutw_conn,
Willy Tarreau2f2318d2022-05-18 10:17:16 +020055 .wake = cs_conn_process,
56 .name = "STRM",
Christopher Faulet9ffddd52022-04-01 14:04:29 +020057};
58
Willy Tarreau3a3f4802022-05-17 18:28:19 +020059/* stream connector operations for embedded tasks */
60struct sc_app_ops sc_app_embedded_ops = {
61 .chk_rcv = sc_app_chk_rcv,
62 .chk_snd = sc_app_chk_snd,
63 .shutr = sc_app_shutr,
64 .shutw = sc_app_shutw,
Willy Tarreau2f2318d2022-05-18 10:17:16 +020065 .wake = NULL, /* may never be used */
66 .name = "NONE", /* may never be used */
Christopher Faulet9ffddd52022-04-01 14:04:29 +020067};
68
Willy Tarreau2f2318d2022-05-18 10:17:16 +020069/* stream connector operations for applets */
Willy Tarreau3a3f4802022-05-17 18:28:19 +020070struct sc_app_ops sc_app_applet_ops = {
71 .chk_rcv = sc_app_chk_rcv_applet,
72 .chk_snd = sc_app_chk_snd_applet,
73 .shutr = sc_app_shutr_applet,
74 .shutw = sc_app_shutw_applet,
Christopher Faulet5e29b762022-04-04 08:58:34 +020075 .wake = cs_applet_process,
76 .name = "STRM",
77};
78
Willy Tarreau2f2318d2022-05-18 10:17:16 +020079/* stream connector for health checks on connections */
80struct sc_app_ops sc_app_check_ops = {
81 .chk_rcv = NULL,
82 .chk_snd = NULL,
83 .shutr = NULL,
84 .shutw = NULL,
85 .wake = wake_srv_chk,
86 .name = "CHCK",
87};
Christopher Faulet5e29b762022-04-04 08:58:34 +020088
Christopher Faulet9ed77422022-04-12 08:51:15 +020089/* Initializes an endpoint */
Willy Tarreauea59b022022-05-17 17:53:22 +020090void sedesc_init(struct sedesc *sedesc)
Christopher Fauletdb90f2a2022-03-22 16:06:25 +010091{
Willy Tarreauea59b022022-05-17 17:53:22 +020092 sedesc->se = NULL;
93 sedesc->conn = NULL;
Willy Tarreauc1054922022-05-18 07:43:52 +020094 sedesc->sc = NULL;
Willy Tarreauea59b022022-05-17 17:53:22 +020095 se_fl_setall(sedesc, SE_FL_NONE);
Christopher Fauletdb90f2a2022-03-22 16:06:25 +010096}
97
Christopher Faulet9ed77422022-04-12 08:51:15 +020098/* Tries to alloc an endpoint and initialize it. Returns NULL on failure. */
Willy Tarreauea59b022022-05-17 17:53:22 +020099struct sedesc *sedesc_new()
Christopher Fauletdb90f2a2022-03-22 16:06:25 +0100100{
Willy Tarreauea59b022022-05-17 17:53:22 +0200101 struct sedesc *sedesc;
Christopher Fauletdb90f2a2022-03-22 16:06:25 +0100102
Willy Tarreauea59b022022-05-17 17:53:22 +0200103 sedesc = pool_alloc(pool_head_sedesc);
104 if (unlikely(!sedesc))
Christopher Fauletdb90f2a2022-03-22 16:06:25 +0100105 return NULL;
106
Willy Tarreauea59b022022-05-17 17:53:22 +0200107 sedesc_init(sedesc);
108 return sedesc;
Christopher Fauletdb90f2a2022-03-22 16:06:25 +0100109}
110
Christopher Faulet9ed77422022-04-12 08:51:15 +0200111/* Releases an endpoint. It is the caller responsibility to be sure it is safe
112 * and it is not shared with another entity
113 */
Willy Tarreauea59b022022-05-17 17:53:22 +0200114void sedesc_free(struct sedesc *sedesc)
Christopher Fauletdb90f2a2022-03-22 16:06:25 +0100115{
Willy Tarreauea59b022022-05-17 17:53:22 +0200116 pool_free(pool_head_sedesc, sedesc);
Christopher Fauletdb90f2a2022-03-22 16:06:25 +0100117}
Christopher Faulet1329f2a2021-12-16 17:32:56 +0100118
Willy Tarreau4596fe22022-05-17 19:07:51 +0200119/* Tries to allocate a new stconn and initialize its main fields. On
Christopher Faulet9ed77422022-04-12 08:51:15 +0200120 * failure, nothing is allocated and NULL is returned. It is an internal
Willy Tarreaub605c422022-05-17 17:04:55 +0200121 * function. The caller must, at least, set the SE_FL_ORPHAN or SE_FL_DETACHED
Christopher Faulet9ed77422022-04-12 08:51:15 +0200122 * flag.
Christopher Faulet1329f2a2021-12-16 17:32:56 +0100123 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200124static struct stconn *cs_new(struct sedesc *sedesc)
Christopher Faulet1329f2a2021-12-16 17:32:56 +0100125{
Willy Tarreau4596fe22022-05-17 19:07:51 +0200126 struct stconn *cs;
Christopher Faulet1329f2a2021-12-16 17:32:56 +0100127
128 cs = pool_alloc(pool_head_connstream);
Christopher Fauletdb90f2a2022-03-22 16:06:25 +0100129
Christopher Faulet1329f2a2021-12-16 17:32:56 +0100130 if (unlikely(!cs))
Christopher Fauletdb90f2a2022-03-22 16:06:25 +0100131 goto alloc_error;
Christopher Fauletbb772d02022-03-22 15:28:36 +0100132
133 cs->obj_type = OBJ_TYPE_CS;
Willy Tarreaucb041662022-05-17 19:44:42 +0200134 cs->flags = SC_FL_NONE;
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200135 cs->state = SC_ST_INI;
Christopher Faulet1d987772022-03-29 18:03:35 +0200136 cs->hcto = TICK_ETERNITY;
Christopher Fauletbb772d02022-03-22 15:28:36 +0100137 cs->app = NULL;
Willy Tarreau2f2318d2022-05-18 10:17:16 +0200138 cs->app_ops = NULL;
Christopher Faulet8da67aa2022-03-29 17:53:09 +0200139 cs->src = NULL;
140 cs->dst = NULL;
Christopher Faulet2f35e7b2022-03-31 11:09:28 +0200141 cs->wait_event.tasklet = NULL;
142 cs->wait_event.events = 0;
143
Christopher Faulet9ed77422022-04-12 08:51:15 +0200144 /* If there is no endpoint, allocate a new one now */
Willy Tarreauea59b022022-05-17 17:53:22 +0200145 if (!sedesc) {
146 sedesc = sedesc_new();
147 if (unlikely(!sedesc))
Christopher Fauletb669d682022-03-22 18:37:19 +0100148 goto alloc_error;
149 }
Willy Tarreau798465b2022-05-17 18:20:02 +0200150 cs->sedesc = sedesc;
Willy Tarreauc1054922022-05-18 07:43:52 +0200151 sedesc->sc = cs;
Christopher Fauletdb90f2a2022-03-22 16:06:25 +0100152
Christopher Faulet1329f2a2021-12-16 17:32:56 +0100153 return cs;
Christopher Fauletdb90f2a2022-03-22 16:06:25 +0100154
155 alloc_error:
156 pool_free(pool_head_connstream, cs);
157 return NULL;
Christopher Faulet1329f2a2021-12-16 17:32:56 +0100158}
159
Willy Tarreau4596fe22022-05-17 19:07:51 +0200160/* Creates a new stream connector and its associated stream from a mux. <endp> must be
161 * defined. It returns NULL on error. On success, the new stream connector is
Willy Tarreaub605c422022-05-17 17:04:55 +0200162 * returned. In this case, SE_FL_ORPHAN flag is removed.
Christopher Faulet9ed77422022-04-12 08:51:15 +0200163 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200164struct stconn *cs_new_from_endp(struct sedesc *sedesc, struct session *sess, struct buffer *input)
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100165{
Willy Tarreau4596fe22022-05-17 19:07:51 +0200166 struct stconn *cs;
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100167
Willy Tarreauea59b022022-05-17 17:53:22 +0200168 cs = cs_new(sedesc);
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100169 if (unlikely(!cs))
170 return NULL;
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100171 if (unlikely(!stream_new(sess, cs, input))) {
172 pool_free(pool_head_connstream, cs);
173 cs = NULL;
174 }
Willy Tarreauea59b022022-05-17 17:53:22 +0200175 se_fl_clr(sedesc, SE_FL_ORPHAN);
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100176 return cs;
177}
178
Willy Tarreau4596fe22022-05-17 19:07:51 +0200179/* Creates a new stream connector from an stream. There is no endpoint here, thus it
Willy Tarreaub605c422022-05-17 17:04:55 +0200180 * will be created by cs_new(). So the SE_FL_DETACHED flag is set. It returns
Willy Tarreau4596fe22022-05-17 19:07:51 +0200181 * NULL on error. On success, the new stream connector is returned.
Christopher Faulet9ed77422022-04-12 08:51:15 +0200182 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200183struct stconn *cs_new_from_strm(struct stream *strm, unsigned int flags)
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100184{
Willy Tarreau4596fe22022-05-17 19:07:51 +0200185 struct stconn *cs;
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100186
187 cs = cs_new(NULL);
188 if (unlikely(!cs))
189 return NULL;
190 cs->flags |= flags;
Willy Tarreaub605c422022-05-17 17:04:55 +0200191 sc_ep_set(cs, SE_FL_DETACHED);
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100192 cs->app = &strm->obj_type;
Willy Tarreau2f2318d2022-05-18 10:17:16 +0200193 cs->app_ops = &sc_app_embedded_ops;
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100194 return cs;
195}
196
Willy Tarreau4596fe22022-05-17 19:07:51 +0200197/* Creates a new stream connector from an health-check. There is no endpoint here,
Willy Tarreaub605c422022-05-17 17:04:55 +0200198 * thus it will be created by cs_new(). So the SE_FL_DETACHED flag is set. It
Willy Tarreau4596fe22022-05-17 19:07:51 +0200199 * returns NULL on error. On success, the new stream connector is returned.
Christopher Faulet9ed77422022-04-12 08:51:15 +0200200 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200201struct stconn *cs_new_from_check(struct check *check, unsigned int flags)
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100202{
Willy Tarreau4596fe22022-05-17 19:07:51 +0200203 struct stconn *cs;
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100204
205 cs = cs_new(NULL);
206 if (unlikely(!cs))
207 return NULL;
208 cs->flags |= flags;
Willy Tarreaub605c422022-05-17 17:04:55 +0200209 sc_ep_set(cs, SE_FL_DETACHED);
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100210 cs->app = &check->obj_type;
Willy Tarreau2f2318d2022-05-18 10:17:16 +0200211 cs->app_ops = &sc_app_check_ops;
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100212 return cs;
213}
214
Willy Tarreau4596fe22022-05-17 19:07:51 +0200215/* Releases a stconn previously allocated by cs_new(), as well as its
Christopher Faulet9ed77422022-04-12 08:51:15 +0200216 * endpoint, if it exists. This function is called internally or on error path.
Christopher Faulet1329f2a2021-12-16 17:32:56 +0100217 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200218void cs_free(struct stconn *cs)
Christopher Faulet1329f2a2021-12-16 17:32:56 +0100219{
Christopher Faulet8da67aa2022-03-29 17:53:09 +0200220 sockaddr_free(&cs->src);
221 sockaddr_free(&cs->dst);
Willy Tarreau798465b2022-05-17 18:20:02 +0200222 if (cs->sedesc) {
Willy Tarreaub605c422022-05-17 17:04:55 +0200223 BUG_ON(!sc_ep_test(cs, SE_FL_DETACHED));
Willy Tarreau798465b2022-05-17 18:20:02 +0200224 sedesc_free(cs->sedesc);
Christopher Fauletdb90f2a2022-03-22 16:06:25 +0100225 }
Christopher Faulet2f35e7b2022-03-31 11:09:28 +0200226 if (cs->wait_event.tasklet)
227 tasklet_free(cs->wait_event.tasklet);
Christopher Faulet1329f2a2021-12-16 17:32:56 +0100228 pool_free(pool_head_connstream, cs);
229}
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100230
Willy Tarreau4596fe22022-05-17 19:07:51 +0200231/* Conditionally removes a stream connector if it is detached and if there is no app
Christopher Fauleteb50c012022-04-21 14:22:53 +0200232 * layer defined. Except on error path, this one must be used. if release, the
233 * pointer on the CS is set to NULL.
Christopher Fauletaa69d8f2022-04-12 18:09:48 +0200234 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200235static void cs_free_cond(struct stconn **csp)
Christopher Fauletaa69d8f2022-04-12 18:09:48 +0200236{
Willy Tarreau4596fe22022-05-17 19:07:51 +0200237 struct stconn *cs = *csp;
Christopher Fauleteb50c012022-04-21 14:22:53 +0200238
Willy Tarreau798465b2022-05-17 18:20:02 +0200239 if (!cs->app && (!cs->sedesc || sc_ep_test(cs, SE_FL_DETACHED))) {
Christopher Fauletaa69d8f2022-04-12 18:09:48 +0200240 cs_free(cs);
Christopher Fauleteb50c012022-04-21 14:22:53 +0200241 *csp = NULL;
242 }
Christopher Fauletaa69d8f2022-04-12 18:09:48 +0200243}
244
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100245
Willy Tarreau4596fe22022-05-17 19:07:51 +0200246/* Attaches a stconn to a mux endpoint and sets the endpoint ctx. Returns
Willy Tarreaub605c422022-05-17 17:04:55 +0200247 * -1 on error and 0 on sucess. SE_FL_DETACHED flag is removed. This function is
Christopher Faulet9ed77422022-04-12 08:51:15 +0200248 * called from a mux when it is attached to a stream or a health-check.
249 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200250int cs_attach_mux(struct stconn *cs, void *endp, void *ctx)
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100251{
Christopher Faulet93882042022-01-19 14:56:50 +0100252 struct connection *conn = ctx;
Willy Tarreau798465b2022-05-17 18:20:02 +0200253 struct sedesc *sedesc = cs->sedesc;
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100254
Willy Tarreau798465b2022-05-17 18:20:02 +0200255 sedesc->se = endp;
256 sedesc->conn = ctx;
257 se_fl_set(sedesc, SE_FL_T_MUX);
258 se_fl_clr(sedesc, SE_FL_DETACHED);
Christopher Faulet93882042022-01-19 14:56:50 +0100259 if (!conn->ctx)
260 conn->ctx = cs;
261 if (cs_strm(cs)) {
Christopher Faulet2f35e7b2022-03-31 11:09:28 +0200262 if (!cs->wait_event.tasklet) {
263 cs->wait_event.tasklet = tasklet_new();
264 if (!cs->wait_event.tasklet)
265 return -1;
Christopher Faulet4a7764a2022-04-01 16:58:52 +0200266 cs->wait_event.tasklet->process = cs_conn_io_cb;
267 cs->wait_event.tasklet->context = cs;
Christopher Faulet2f35e7b2022-03-31 11:09:28 +0200268 cs->wait_event.events = 0;
269 }
270
Willy Tarreau2f2318d2022-05-18 10:17:16 +0200271 cs->app_ops = &sc_app_conn_ops;
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100272 }
Christopher Fauletc95eaef2022-05-18 15:57:15 +0200273 else if (cs_check(cs)) {
274 if (!cs->wait_event.tasklet) {
275 cs->wait_event.tasklet = tasklet_new();
276 if (!cs->wait_event.tasklet)
277 return -1;
278 cs->wait_event.tasklet->process = srv_chk_io_cb;
279 cs->wait_event.tasklet->context = cs;
280 cs->wait_event.events = 0;
281 }
282
Willy Tarreau2f2318d2022-05-18 10:17:16 +0200283 cs->app_ops = &sc_app_check_ops;
Christopher Fauletc95eaef2022-05-18 15:57:15 +0200284 }
Christopher Faulet070b91b2022-03-31 19:27:18 +0200285 return 0;
Christopher Faulet93882042022-01-19 14:56:50 +0100286}
287
Willy Tarreau4596fe22022-05-17 19:07:51 +0200288/* Attaches a stconn to an applet endpoint and sets the endpoint
Willy Tarreaub605c422022-05-17 17:04:55 +0200289 * ctx. Returns -1 on error and 0 on sucess. SE_FL_DETACHED flag is
Christopher Faulet9ed77422022-04-12 08:51:15 +0200290 * removed. This function is called by a stream when a backend applet is
291 * registered.
292 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200293static void cs_attach_applet(struct stconn *cs, void *endp)
Christopher Faulet93882042022-01-19 14:56:50 +0100294{
Willy Tarreau798465b2022-05-17 18:20:02 +0200295 cs->sedesc->se = endp;
Willy Tarreaub605c422022-05-17 17:04:55 +0200296 sc_ep_set(cs, SE_FL_T_APPLET);
297 sc_ep_clr(cs, SE_FL_DETACHED);
Willy Tarreau2f2318d2022-05-18 10:17:16 +0200298 if (cs_strm(cs))
299 cs->app_ops = &sc_app_applet_ops;
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100300}
301
Willy Tarreau4596fe22022-05-17 19:07:51 +0200302/* Attaches a stconn to a app layer and sets the relevant
Willy Tarreaub605c422022-05-17 17:04:55 +0200303 * callbacks. Returns -1 on error and 0 on success. SE_FL_ORPHAN flag is
Christopher Faulet9ed77422022-04-12 08:51:15 +0200304 * removed. This function is called by a stream when it is created to attach it
Willy Tarreau4596fe22022-05-17 19:07:51 +0200305 * on the stream connector on the client side.
Christopher Faulet9ed77422022-04-12 08:51:15 +0200306 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200307int cs_attach_strm(struct stconn *cs, struct stream *strm)
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100308{
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100309 cs->app = &strm->obj_type;
Willy Tarreaub605c422022-05-17 17:04:55 +0200310 sc_ep_clr(cs, SE_FL_ORPHAN);
311 if (sc_ep_test(cs, SE_FL_T_MUX)) {
Christopher Faulet2f35e7b2022-03-31 11:09:28 +0200312 cs->wait_event.tasklet = tasklet_new();
Christopher Faulet582a2262022-04-04 11:25:59 +0200313 if (!cs->wait_event.tasklet)
Christopher Faulet2f35e7b2022-03-31 11:09:28 +0200314 return -1;
Christopher Faulet4a7764a2022-04-01 16:58:52 +0200315 cs->wait_event.tasklet->process = cs_conn_io_cb;
316 cs->wait_event.tasklet->context = cs;
Christopher Faulet2f35e7b2022-03-31 11:09:28 +0200317 cs->wait_event.events = 0;
318
Willy Tarreau2f2318d2022-05-18 10:17:16 +0200319 cs->app_ops = &sc_app_conn_ops;
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100320 }
Willy Tarreaub605c422022-05-17 17:04:55 +0200321 else if (sc_ep_test(cs, SE_FL_T_APPLET)) {
Willy Tarreau2f2318d2022-05-18 10:17:16 +0200322 cs->app_ops = &sc_app_applet_ops;
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100323 }
324 else {
Willy Tarreau2f2318d2022-05-18 10:17:16 +0200325 cs->app_ops = &sc_app_embedded_ops;
Christopher Fauleta9e8b392022-03-23 11:01:09 +0100326 }
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100327 return 0;
328}
329
Willy Tarreau4596fe22022-05-17 19:07:51 +0200330/* Detaches the stconn from the endpoint, if any. For a connecrion, if a
Christopher Faulet9ed77422022-04-12 08:51:15 +0200331 * mux owns the connection ->detach() callback is called. Otherwise, it means
Willy Tarreau4596fe22022-05-17 19:07:51 +0200332 * the stream connector owns the connection. In this case the connection is closed
Christopher Faulet9ed77422022-04-12 08:51:15 +0200333 * and released. For an applet, the appctx is released. If still allocated, the
334 * endpoint is reset and flag as detached. If the app layer is also detached,
Willy Tarreau4596fe22022-05-17 19:07:51 +0200335 * the stream connector is released.
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100336 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200337static void cs_detach_endp(struct stconn **csp)
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100338{
Willy Tarreau4596fe22022-05-17 19:07:51 +0200339 struct stconn *cs = *csp;
Christopher Fauleteb50c012022-04-21 14:22:53 +0200340
341 if (!cs)
342 return;
343
Willy Tarreau798465b2022-05-17 18:20:02 +0200344 if (!cs->sedesc)
Christopher Fauletb041b232022-03-24 10:27:02 +0100345 goto reset_cs;
346
Willy Tarreaub605c422022-05-17 17:04:55 +0200347 if (sc_ep_test(cs, SE_FL_T_MUX)) {
Christopher Fauletcea05432022-04-14 11:40:12 +0200348 struct connection *conn = __cs_conn(cs);
Willy Tarreau798465b2022-05-17 18:20:02 +0200349 struct sedesc *sedesc = cs->sedesc;
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100350
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100351 if (conn->mux) {
Christopher Faulet2f35e7b2022-03-31 11:09:28 +0200352 if (cs->wait_event.events != 0)
353 conn->mux->unsubscribe(cs, cs->wait_event.events, &cs->wait_event);
Willy Tarreau798465b2022-05-17 18:20:02 +0200354 se_fl_set(sedesc, SE_FL_ORPHAN);
Willy Tarreauc1054922022-05-18 07:43:52 +0200355 sedesc->sc = NULL;
Willy Tarreau798465b2022-05-17 18:20:02 +0200356 cs->sedesc = NULL;
357 conn->mux->detach(sedesc);
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100358 }
359 else {
360 /* It's too early to have a mux, let's just destroy
361 * the connection
362 */
363 conn_stop_tracking(conn);
364 conn_full_close(conn);
365 if (conn->destroy_cb)
366 conn->destroy_cb(conn);
367 conn_free(conn);
368 }
369 }
Willy Tarreaub605c422022-05-17 17:04:55 +0200370 else if (sc_ep_test(cs, SE_FL_T_APPLET)) {
Christopher Fauletcea05432022-04-14 11:40:12 +0200371 struct appctx *appctx = __cs_appctx(cs);
Christopher Fauletdb90f2a2022-03-22 16:06:25 +0100372
Willy Tarreaub605c422022-05-17 17:04:55 +0200373 sc_ep_set(cs, SE_FL_ORPHAN);
Willy Tarreauc1054922022-05-18 07:43:52 +0200374 cs->sedesc->sc = NULL;
Willy Tarreau798465b2022-05-17 18:20:02 +0200375 cs->sedesc = NULL;
Willy Tarreau1c3ead42022-05-10 19:42:22 +0200376 appctx_shut(appctx);
377 appctx_free(appctx);
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100378 }
379
Willy Tarreau798465b2022-05-17 18:20:02 +0200380 if (cs->sedesc) {
Christopher Faulet9ec2f4d2022-03-23 15:15:29 +0100381 /* the cs is the only one one the endpoint */
Willy Tarreau798465b2022-05-17 18:20:02 +0200382 cs->sedesc->se = NULL;
383 cs->sedesc->conn = NULL;
Willy Tarreaub605c422022-05-17 17:04:55 +0200384 sc_ep_clr(cs, ~SE_FL_APP_MASK);
385 sc_ep_set(cs, SE_FL_DETACHED);
Christopher Fauletdb90f2a2022-03-22 16:06:25 +0100386 }
387
Christopher Fauletb041b232022-03-24 10:27:02 +0100388 reset_cs:
Christopher Fauletc36de9d2022-01-06 08:44:58 +0100389 /* FIXME: Rest CS for now but must be reviewed. CS flags are only
390 * connection related for now but this will evolved
391 */
Willy Tarreaucb041662022-05-17 19:44:42 +0200392 cs->flags &= SC_FL_ISBACK;
Christopher Faulet582a2262022-04-04 11:25:59 +0200393 if (cs_strm(cs))
Willy Tarreau2f2318d2022-05-18 10:17:16 +0200394 cs->app_ops = &sc_app_embedded_ops;
395 else
396 cs->app_ops = NULL;
Christopher Fauleteb50c012022-04-21 14:22:53 +0200397 cs_free_cond(csp);
Christopher Fauletc36de9d2022-01-06 08:44:58 +0100398}
399
Willy Tarreau4596fe22022-05-17 19:07:51 +0200400/* Detaches the stconn from the app layer. If there is no endpoint attached
401 * to the stconn
Christopher Faulet9ed77422022-04-12 08:51:15 +0200402 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200403static void cs_detach_app(struct stconn **csp)
Christopher Fauletc36de9d2022-01-06 08:44:58 +0100404{
Willy Tarreau4596fe22022-05-17 19:07:51 +0200405 struct stconn *cs = *csp;
Christopher Fauleteb50c012022-04-21 14:22:53 +0200406
407 if (!cs)
408 return;
409
Christopher Fauletc36de9d2022-01-06 08:44:58 +0100410 cs->app = NULL;
Willy Tarreau2f2318d2022-05-18 10:17:16 +0200411 cs->app_ops = NULL;
Christopher Faulet8da67aa2022-03-29 17:53:09 +0200412 sockaddr_free(&cs->src);
413 sockaddr_free(&cs->dst);
Christopher Faulet2f35e7b2022-03-31 11:09:28 +0200414
415 if (cs->wait_event.tasklet)
416 tasklet_free(cs->wait_event.tasklet);
417 cs->wait_event.tasklet = NULL;
418 cs->wait_event.events = 0;
Christopher Fauleteb50c012022-04-21 14:22:53 +0200419 cs_free_cond(csp);
420}
421
Willy Tarreau4596fe22022-05-17 19:07:51 +0200422/* Destroy the stconn. It is detached from its endpoint and its
423 * application. After this call, the stconn must be considered as released.
Christopher Fauleteb50c012022-04-21 14:22:53 +0200424 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200425void cs_destroy(struct stconn *cs)
Christopher Fauleteb50c012022-04-21 14:22:53 +0200426{
427 cs_detach_endp(&cs);
428 cs_detach_app(&cs);
429 BUG_ON_HOT(cs);
Christopher Fauletcda94ac2021-12-23 17:28:17 +0100430}
Christopher Faulet9ec2f4d2022-03-23 15:15:29 +0100431
Willy Tarreau4596fe22022-05-17 19:07:51 +0200432/* Resets the stream connector endpoint. It happens when the app layer want to renew
Christopher Faulet9ed77422022-04-12 08:51:15 +0200433 * its endpoint. For a connection retry for instance. If a mux or an applet is
434 * attached, a new endpoint is created. Returns -1 on error and 0 on sucess.
Christopher Fauleta6c4a482022-04-28 18:25:24 +0200435 *
Willy Tarreaub605c422022-05-17 17:04:55 +0200436 * Only SE_FL_ERROR flag is removed on the endpoint. Orther flags are preserved.
Christopher Fauleta6c4a482022-04-28 18:25:24 +0200437 * It is the caller responsibility to remove other flags if needed.
Christopher Faulet9ed77422022-04-12 08:51:15 +0200438 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200439int cs_reset_endp(struct stconn *cs)
Christopher Faulet9ec2f4d2022-03-23 15:15:29 +0100440{
Willy Tarreauea59b022022-05-17 17:53:22 +0200441 struct sedesc *new_endp;
Christopher Fauletb041b232022-03-24 10:27:02 +0100442
Christopher Faulet9ec2f4d2022-03-23 15:15:29 +0100443 BUG_ON(!cs->app);
Christopher Fauleta6c4a482022-04-28 18:25:24 +0200444
Willy Tarreaub605c422022-05-17 17:04:55 +0200445 sc_ep_clr(cs, SE_FL_ERROR);
Christopher Fauletb041b232022-03-24 10:27:02 +0100446 if (!__cs_endp_target(cs)) {
447 /* endpoint not attached or attached to a mux with no
448 * target. Thus the endpoint will not be release but just
Christopher Fauleteb50c012022-04-21 14:22:53 +0200449 * reset. The app is still attached, the cs will not be
450 * released.
Christopher Fauletb041b232022-03-24 10:27:02 +0100451 */
Christopher Fauleteb50c012022-04-21 14:22:53 +0200452 cs_detach_endp(&cs);
Christopher Fauletb041b232022-03-24 10:27:02 +0100453 return 0;
Christopher Faulet9ec2f4d2022-03-23 15:15:29 +0100454 }
Christopher Fauletb041b232022-03-24 10:27:02 +0100455
456 /* allocate the new endpoint first to be able to set error if it
457 * fails */
Willy Tarreauea59b022022-05-17 17:53:22 +0200458 new_endp = sedesc_new();
Christopher Fauletb041b232022-03-24 10:27:02 +0100459 if (!unlikely(new_endp)) {
Willy Tarreaub605c422022-05-17 17:04:55 +0200460 sc_ep_set(cs, SE_FL_ERROR);
Christopher Fauletb041b232022-03-24 10:27:02 +0100461 return -1;
462 }
Willy Tarreaub605c422022-05-17 17:04:55 +0200463 se_fl_setall(new_endp, sc_ep_get(cs) & SE_FL_APP_MASK);
Christopher Fauletb041b232022-03-24 10:27:02 +0100464
Christopher Fauleteb50c012022-04-21 14:22:53 +0200465 /* The app is still attached, the cs will not be released */
466 cs_detach_endp(&cs);
Willy Tarreau798465b2022-05-17 18:20:02 +0200467 BUG_ON(cs->sedesc);
468 cs->sedesc = new_endp;
Willy Tarreauc1054922022-05-18 07:43:52 +0200469 cs->sedesc->sc = cs;
Willy Tarreaub605c422022-05-17 17:04:55 +0200470 sc_ep_set(cs, SE_FL_DETACHED);
Christopher Faulet9ec2f4d2022-03-23 15:15:29 +0100471 return 0;
472}
Christopher Faulet37046632022-04-01 11:36:58 +0200473
474
Willy Tarreau4596fe22022-05-17 19:07:51 +0200475/* Create an applet to handle a stream connector as a new appctx. The CS will
Christopher Faulet37046632022-04-01 11:36:58 +0200476 * wake it up every time it is solicited. The appctx must be deleted by the task
477 * handler using cs_detach_endp(), possibly from within the function itself.
478 * It also pre-initializes the applet's context and returns it (or NULL in case
479 * it could not be allocated).
480 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200481struct appctx *cs_applet_create(struct stconn *cs, struct applet *app)
Christopher Faulet37046632022-04-01 11:36:58 +0200482{
483 struct appctx *appctx;
484
485 DPRINTF(stderr, "registering handler %p for cs %p (was %p)\n", app, cs, cs_strm_task(cs));
486
Willy Tarreau798465b2022-05-17 18:20:02 +0200487 appctx = appctx_new_here(app, cs->sedesc);
Christopher Faulet37046632022-04-01 11:36:58 +0200488 if (!appctx)
489 return NULL;
Christopher Faulet2d9cc852022-05-16 17:29:37 +0200490 cs_attach_applet(cs, appctx);
Christopher Faulet37046632022-04-01 11:36:58 +0200491 appctx->t->nice = __cs_strm(cs)->task->nice;
Christopher Fauleta0bdec32022-04-04 07:51:21 +0200492 cs_cant_get(cs);
Christopher Faulet37046632022-04-01 11:36:58 +0200493 appctx_wakeup(appctx);
Christopher Fauleta33ff7a2022-04-21 11:52:07 +0200494
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200495 cs->state = SC_ST_RDY;
Christopher Faulet37046632022-04-01 11:36:58 +0200496 return appctx;
497}
498
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200499/*
Willy Tarreau4596fe22022-05-17 19:07:51 +0200500 * This function performs a shutdown-read on a detached stream connector in a
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200501 * connected or init state (it does nothing for other states). It either shuts
502 * the read side or marks itself as closed. The buffer flags are updated to
Willy Tarreaucb041662022-05-17 19:44:42 +0200503 * reflect the new state. If the stream connector has SC_FL_NOHALF, we also
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200504 * forward the close to the write side. The owner task is woken up if it exists.
505 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200506static void sc_app_shutr(struct stconn *cs)
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200507{
508 struct channel *ic = cs_ic(cs);
509
Christopher Fauleta0bdec32022-04-04 07:51:21 +0200510 cs_rx_shut_blk(cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200511 if (ic->flags & CF_SHUTR)
512 return;
513 ic->flags |= CF_SHUTR;
514 ic->rex = TICK_ETERNITY;
515
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200516 if (!cs_state_in(cs->state, SC_SB_CON|SC_SB_RDY|SC_SB_EST))
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200517 return;
518
519 if (cs_oc(cs)->flags & CF_SHUTW) {
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200520 cs->state = SC_ST_DIS;
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200521 __cs_strm(cs)->conn_exp = TICK_ETERNITY;
522 }
Willy Tarreaucb041662022-05-17 19:44:42 +0200523 else if (cs->flags & SC_FL_NOHALF) {
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200524 /* we want to immediately forward this close to the write side */
Willy Tarreau3a3f4802022-05-17 18:28:19 +0200525 return sc_app_shutw(cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200526 }
527
528 /* note that if the task exists, it must unregister itself once it runs */
Willy Tarreaucb041662022-05-17 19:44:42 +0200529 if (!(cs->flags & SC_FL_DONT_WAKE))
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200530 task_wakeup(cs_strm_task(cs), TASK_WOKEN_IO);
531}
532
533/*
Willy Tarreau4596fe22022-05-17 19:07:51 +0200534 * This function performs a shutdown-write on a detached stream connector in a
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200535 * connected or init state (it does nothing for other states). It either shuts
536 * the write side or marks itself as closed. The buffer flags are updated to
Christopher Faulet6b0a0fb2022-04-04 11:29:28 +0200537 * reflect the new state. It does also close everything if the CS was marked as
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200538 * being in error state. The owner task is woken up if it exists.
539 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200540static void sc_app_shutw(struct stconn *cs)
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200541{
542 struct channel *ic = cs_ic(cs);
543 struct channel *oc = cs_oc(cs);
544
545 oc->flags &= ~CF_SHUTW_NOW;
546 if (oc->flags & CF_SHUTW)
547 return;
548 oc->flags |= CF_SHUTW;
549 oc->wex = TICK_ETERNITY;
Christopher Fauleta0bdec32022-04-04 07:51:21 +0200550 cs_done_get(cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200551
552 if (tick_isset(cs->hcto)) {
553 ic->rto = cs->hcto;
554 ic->rex = tick_add(now_ms, ic->rto);
555 }
556
557 switch (cs->state) {
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200558 case SC_ST_RDY:
559 case SC_ST_EST:
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200560 /* we have to shut before closing, otherwise some short messages
561 * may never leave the system, especially when there are remaining
562 * unread data in the socket input buffer, or when nolinger is set.
Willy Tarreaucb041662022-05-17 19:44:42 +0200563 * However, if SC_FL_NOLINGER is explicitly set, we know there is
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200564 * no risk so we close both sides immediately.
565 */
Willy Tarreaucb041662022-05-17 19:44:42 +0200566 if (!sc_ep_test(cs, SE_FL_ERROR) && !(cs->flags & SC_FL_NOLINGER) &&
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200567 !(ic->flags & (CF_SHUTR|CF_DONT_READ)))
568 return;
569
570 /* fall through */
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200571 case SC_ST_CON:
572 case SC_ST_CER:
573 case SC_ST_QUE:
574 case SC_ST_TAR:
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200575 /* Note that none of these states may happen with applets */
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200576 cs->state = SC_ST_DIS;
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200577 /* fall through */
578 default:
Willy Tarreaucb041662022-05-17 19:44:42 +0200579 cs->flags &= ~SC_FL_NOLINGER;
Christopher Fauleta0bdec32022-04-04 07:51:21 +0200580 cs_rx_shut_blk(cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200581 ic->flags |= CF_SHUTR;
582 ic->rex = TICK_ETERNITY;
583 __cs_strm(cs)->conn_exp = TICK_ETERNITY;
584 }
585
586 /* note that if the task exists, it must unregister itself once it runs */
Willy Tarreaucb041662022-05-17 19:44:42 +0200587 if (!(cs->flags & SC_FL_DONT_WAKE))
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200588 task_wakeup(cs_strm_task(cs), TASK_WOKEN_IO);
589}
590
591/* default chk_rcv function for scheduled tasks */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200592static void sc_app_chk_rcv(struct stconn *cs)
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200593{
594 struct channel *ic = cs_ic(cs);
595
596 DPRINTF(stderr, "%s: cs=%p, cs->state=%d ic->flags=%08x oc->flags=%08x\n",
597 __FUNCTION__,
598 cs, cs->state, ic->flags, cs_oc(cs)->flags);
599
600 if (ic->pipe) {
601 /* stop reading */
Christopher Fauleta0bdec32022-04-04 07:51:21 +0200602 cs_rx_room_blk(cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200603 }
604 else {
605 /* (re)start reading */
Willy Tarreaucb041662022-05-17 19:44:42 +0200606 if (!(cs->flags & SC_FL_DONT_WAKE))
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200607 task_wakeup(cs_strm_task(cs), TASK_WOKEN_IO);
608 }
609}
610
611/* default chk_snd function for scheduled tasks */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200612static void sc_app_chk_snd(struct stconn *cs)
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200613{
614 struct channel *oc = cs_oc(cs);
615
616 DPRINTF(stderr, "%s: cs=%p, cs->state=%d ic->flags=%08x oc->flags=%08x\n",
617 __FUNCTION__,
618 cs, cs->state, cs_ic(cs)->flags, oc->flags);
619
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200620 if (unlikely(cs->state != SC_ST_EST || (oc->flags & CF_SHUTW)))
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200621 return;
622
Willy Tarreaub605c422022-05-17 17:04:55 +0200623 if (!sc_ep_test(cs, SE_FL_WAIT_DATA) || /* not waiting for data */
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200624 channel_is_empty(oc)) /* called with nothing to send ! */
625 return;
626
627 /* Otherwise there are remaining data to be sent in the buffer,
628 * so we tell the handler.
629 */
Willy Tarreaub605c422022-05-17 17:04:55 +0200630 sc_ep_clr(cs, SE_FL_WAIT_DATA);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200631 if (!tick_isset(oc->wex))
632 oc->wex = tick_add_ifset(now_ms, oc->wto);
633
Willy Tarreaucb041662022-05-17 19:44:42 +0200634 if (!(cs->flags & SC_FL_DONT_WAKE))
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200635 task_wakeup(cs_strm_task(cs), TASK_WOKEN_IO);
636}
637
638/*
Willy Tarreau3a3f4802022-05-17 18:28:19 +0200639 * This function performs a shutdown-read on a stream connector attached to
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200640 * a connection in a connected or init state (it does nothing for other
641 * states). It either shuts the read side or marks itself as closed. The buffer
Willy Tarreau3a3f4802022-05-17 18:28:19 +0200642 * flags are updated to reflect the new state. If the stream connector has
Willy Tarreaucb041662022-05-17 19:44:42 +0200643 * SC_FL_NOHALF, we also forward the close to the write side. If a control
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200644 * layer is defined, then it is supposed to be a socket layer and file
645 * descriptors are then shutdown or closed accordingly. The function
646 * automatically disables polling if needed.
647 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200648static void sc_app_shutr_conn(struct stconn *cs)
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200649{
650 struct channel *ic = cs_ic(cs);
651
652 BUG_ON(!cs_conn(cs));
653
Christopher Fauleta0bdec32022-04-04 07:51:21 +0200654 cs_rx_shut_blk(cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200655 if (ic->flags & CF_SHUTR)
656 return;
657 ic->flags |= CF_SHUTR;
658 ic->rex = TICK_ETERNITY;
659
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200660 if (!cs_state_in(cs->state, SC_SB_CON|SC_SB_RDY|SC_SB_EST))
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200661 return;
662
663 if (cs_oc(cs)->flags & CF_SHUTW) {
Christopher Fauletff022a22022-04-21 08:38:54 +0200664 cs_conn_shut(cs);
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200665 cs->state = SC_ST_DIS;
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200666 __cs_strm(cs)->conn_exp = TICK_ETERNITY;
667 }
Willy Tarreaucb041662022-05-17 19:44:42 +0200668 else if (cs->flags & SC_FL_NOHALF) {
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200669 /* we want to immediately forward this close to the write side */
Willy Tarreau3a3f4802022-05-17 18:28:19 +0200670 return sc_app_shutw_conn(cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200671 }
672}
673
674/*
Willy Tarreau3a3f4802022-05-17 18:28:19 +0200675 * This function performs a shutdown-write on a stream connector attached to
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200676 * a connection in a connected or init state (it does nothing for other
677 * states). It either shuts the write side or marks itself as closed. The
678 * buffer flags are updated to reflect the new state. It does also close
Christopher Faulet6b0a0fb2022-04-04 11:29:28 +0200679 * everything if the CS was marked as being in error state. If there is a
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200680 * data-layer shutdown, it is called.
681 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200682static void sc_app_shutw_conn(struct stconn *cs)
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200683{
684 struct channel *ic = cs_ic(cs);
685 struct channel *oc = cs_oc(cs);
686
687 BUG_ON(!cs_conn(cs));
688
689 oc->flags &= ~CF_SHUTW_NOW;
690 if (oc->flags & CF_SHUTW)
691 return;
692 oc->flags |= CF_SHUTW;
693 oc->wex = TICK_ETERNITY;
Christopher Fauleta0bdec32022-04-04 07:51:21 +0200694 cs_done_get(cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200695
696 if (tick_isset(cs->hcto)) {
697 ic->rto = cs->hcto;
698 ic->rex = tick_add(now_ms, ic->rto);
699 }
700
701 switch (cs->state) {
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200702 case SC_ST_RDY:
703 case SC_ST_EST:
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200704 /* we have to shut before closing, otherwise some short messages
705 * may never leave the system, especially when there are remaining
706 * unread data in the socket input buffer, or when nolinger is set.
Willy Tarreaucb041662022-05-17 19:44:42 +0200707 * However, if SC_FL_NOLINGER is explicitly set, we know there is
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200708 * no risk so we close both sides immediately.
709 */
710
Willy Tarreaub605c422022-05-17 17:04:55 +0200711 if (sc_ep_test(cs, SE_FL_ERROR)) {
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200712 /* quick close, the socket is already shut anyway */
713 }
Willy Tarreaucb041662022-05-17 19:44:42 +0200714 else if (cs->flags & SC_FL_NOLINGER) {
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200715 /* unclean data-layer shutdown, typically an aborted request
716 * or a forwarded shutdown from a client to a server due to
717 * option abortonclose. No need for the TLS layer to try to
718 * emit a shutdown message.
719 */
720 cs_conn_shutw(cs, CO_SHW_SILENT);
721 }
722 else {
723 /* clean data-layer shutdown. This only happens on the
724 * frontend side, or on the backend side when forwarding
725 * a client close in TCP mode or in HTTP TUNNEL mode
726 * while option abortonclose is set. We want the TLS
727 * layer to try to signal it to the peer before we close.
728 */
729 cs_conn_shutw(cs, CO_SHW_NORMAL);
730
731 if (!(ic->flags & (CF_SHUTR|CF_DONT_READ)))
732 return;
733 }
734
735 /* fall through */
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200736 case SC_ST_CON:
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200737 /* we may have to close a pending connection, and mark the
738 * response buffer as shutr
739 */
Christopher Fauletff022a22022-04-21 08:38:54 +0200740 cs_conn_shut(cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200741 /* fall through */
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200742 case SC_ST_CER:
743 case SC_ST_QUE:
744 case SC_ST_TAR:
745 cs->state = SC_ST_DIS;
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200746 /* fall through */
747 default:
Willy Tarreaucb041662022-05-17 19:44:42 +0200748 cs->flags &= ~SC_FL_NOLINGER;
Christopher Fauleta0bdec32022-04-04 07:51:21 +0200749 cs_rx_shut_blk(cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200750 ic->flags |= CF_SHUTR;
751 ic->rex = TICK_ETERNITY;
752 __cs_strm(cs)->conn_exp = TICK_ETERNITY;
753 }
754}
755
Willy Tarreau3a3f4802022-05-17 18:28:19 +0200756/* This function is used for inter-stream connector calls. It is called by the
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200757 * consumer to inform the producer side that it may be interested in checking
758 * for free space in the buffer. Note that it intentionally does not update
759 * timeouts, so that we can still check them later at wake-up. This function is
Willy Tarreau3a3f4802022-05-17 18:28:19 +0200760 * dedicated to connection-based stream connectors.
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200761 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200762static void sc_app_chk_rcv_conn(struct stconn *cs)
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200763{
764 BUG_ON(!cs_conn(cs));
765
766 /* (re)start reading */
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200767 if (cs_state_in(cs->state, SC_SB_CON|SC_SB_RDY|SC_SB_EST))
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200768 tasklet_wakeup(cs->wait_event.tasklet);
769}
770
771
Willy Tarreau3a3f4802022-05-17 18:28:19 +0200772/* This function is used for inter-stream connector calls. It is called by the
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200773 * producer to inform the consumer side that it may be interested in checking
774 * for data in the buffer. Note that it intentionally does not update timeouts,
775 * so that we can still check them later at wake-up.
776 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200777static void sc_app_chk_snd_conn(struct stconn *cs)
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200778{
779 struct channel *oc = cs_oc(cs);
780
781 BUG_ON(!cs_conn(cs));
782
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200783 if (unlikely(!cs_state_in(cs->state, SC_SB_RDY|SC_SB_EST) ||
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200784 (oc->flags & CF_SHUTW)))
785 return;
786
787 if (unlikely(channel_is_empty(oc))) /* called with nothing to send ! */
788 return;
789
790 if (!oc->pipe && /* spliced data wants to be forwarded ASAP */
Willy Tarreaub605c422022-05-17 17:04:55 +0200791 !sc_ep_test(cs, SE_FL_WAIT_DATA)) /* not waiting for data */
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200792 return;
793
794 if (!(cs->wait_event.events & SUB_RETRY_SEND) && !channel_is_empty(cs_oc(cs)))
Christopher Faulet000ba3e2022-04-01 17:06:32 +0200795 cs_conn_send(cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200796
Willy Tarreaub605c422022-05-17 17:04:55 +0200797 if (sc_ep_test(cs, SE_FL_ERROR | SE_FL_ERR_PENDING) || cs_is_conn_error(cs)) {
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200798 /* Write error on the file descriptor */
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200799 if (cs->state >= SC_ST_CON)
Willy Tarreaub605c422022-05-17 17:04:55 +0200800 sc_ep_set(cs, SE_FL_ERROR);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200801 goto out_wakeup;
802 }
803
804 /* OK, so now we know that some data might have been sent, and that we may
805 * have to poll first. We have to do that too if the buffer is not empty.
806 */
807 if (channel_is_empty(oc)) {
808 /* the connection is established but we can't write. Either the
809 * buffer is empty, or we just refrain from sending because the
810 * ->o limit was reached. Maybe we just wrote the last
811 * chunk and need to close.
812 */
813 if (((oc->flags & (CF_SHUTW|CF_AUTO_CLOSE|CF_SHUTW_NOW)) ==
814 (CF_AUTO_CLOSE|CF_SHUTW_NOW)) &&
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200815 cs_state_in(cs->state, SC_SB_RDY|SC_SB_EST)) {
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200816 cs_shutw(cs);
817 goto out_wakeup;
818 }
819
820 if ((oc->flags & (CF_SHUTW|CF_SHUTW_NOW)) == 0)
Willy Tarreaub605c422022-05-17 17:04:55 +0200821 sc_ep_set(cs, SE_FL_WAIT_DATA);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200822 oc->wex = TICK_ETERNITY;
823 }
824 else {
825 /* Otherwise there are remaining data to be sent in the buffer,
826 * which means we have to poll before doing so.
827 */
Willy Tarreaub605c422022-05-17 17:04:55 +0200828 sc_ep_clr(cs, SE_FL_WAIT_DATA);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200829 if (!tick_isset(oc->wex))
830 oc->wex = tick_add_ifset(now_ms, oc->wto);
831 }
832
833 if (likely(oc->flags & CF_WRITE_ACTIVITY)) {
834 struct channel *ic = cs_ic(cs);
835
836 /* update timeout if we have written something */
837 if ((oc->flags & (CF_SHUTW|CF_WRITE_PARTIAL)) == CF_WRITE_PARTIAL &&
838 !channel_is_empty(oc))
839 oc->wex = tick_add_ifset(now_ms, oc->wto);
840
Willy Tarreaucb041662022-05-17 19:44:42 +0200841 if (tick_isset(ic->rex) && !(cs->flags & SC_FL_INDEP_STR)) {
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200842 /* Note: to prevent the client from expiring read timeouts
843 * during writes, we refresh it. We only do this if the
844 * interface is not configured for "independent streams",
845 * because for some applications it's better not to do this,
846 * for instance when continuously exchanging small amounts
847 * of data which can full the socket buffers long before a
848 * write timeout is detected.
849 */
850 ic->rex = tick_add_ifset(now_ms, ic->rto);
851 }
852 }
853
854 /* in case of special condition (error, shutdown, end of write...), we
855 * have to notify the task.
856 */
857 if (likely((oc->flags & (CF_WRITE_NULL|CF_WRITE_ERROR|CF_SHUTW)) ||
858 ((oc->flags & CF_WAKE_WRITE) &&
859 ((channel_is_empty(oc) && !oc->to_forward) ||
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200860 !cs_state_in(cs->state, SC_SB_EST))))) {
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200861 out_wakeup:
Willy Tarreaucb041662022-05-17 19:44:42 +0200862 if (!(cs->flags & SC_FL_DONT_WAKE))
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200863 task_wakeup(cs_strm_task(cs), TASK_WOKEN_IO);
864 }
865}
866
867/*
Willy Tarreau3a3f4802022-05-17 18:28:19 +0200868 * This function performs a shutdown-read on a stream connector attached to an
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200869 * applet in a connected or init state (it does nothing for other states). It
870 * either shuts the read side or marks itself as closed. The buffer flags are
Willy Tarreaucb041662022-05-17 19:44:42 +0200871 * updated to reflect the new state. If the stream connector has SC_FL_NOHALF,
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200872 * we also forward the close to the write side. The owner task is woken up if
873 * it exists.
874 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200875static void sc_app_shutr_applet(struct stconn *cs)
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200876{
877 struct channel *ic = cs_ic(cs);
878
879 BUG_ON(!cs_appctx(cs));
880
Christopher Fauleta0bdec32022-04-04 07:51:21 +0200881 cs_rx_shut_blk(cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200882 if (ic->flags & CF_SHUTR)
883 return;
884 ic->flags |= CF_SHUTR;
885 ic->rex = TICK_ETERNITY;
886
887 /* Note: on shutr, we don't call the applet */
888
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200889 if (!cs_state_in(cs->state, SC_SB_CON|SC_SB_RDY|SC_SB_EST))
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200890 return;
891
892 if (cs_oc(cs)->flags & CF_SHUTW) {
Willy Tarreau1c3ead42022-05-10 19:42:22 +0200893 appctx_shut(__cs_appctx(cs));
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200894 cs->state = SC_ST_DIS;
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200895 __cs_strm(cs)->conn_exp = TICK_ETERNITY;
896 }
Willy Tarreaucb041662022-05-17 19:44:42 +0200897 else if (cs->flags & SC_FL_NOHALF) {
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200898 /* we want to immediately forward this close to the write side */
Willy Tarreau3a3f4802022-05-17 18:28:19 +0200899 return sc_app_shutw_applet(cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200900 }
901}
902
903/*
Willy Tarreau3a3f4802022-05-17 18:28:19 +0200904 * This function performs a shutdown-write on a stream connector attached to an
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200905 * applet in a connected or init state (it does nothing for other states). It
906 * either shuts the write side or marks itself as closed. The buffer flags are
907 * updated to reflect the new state. It does also close everything if the SI
908 * was marked as being in error state. The owner task is woken up if it exists.
909 */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200910static void sc_app_shutw_applet(struct stconn *cs)
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200911{
912 struct channel *ic = cs_ic(cs);
913 struct channel *oc = cs_oc(cs);
914
915 BUG_ON(!cs_appctx(cs));
916
917 oc->flags &= ~CF_SHUTW_NOW;
918 if (oc->flags & CF_SHUTW)
919 return;
920 oc->flags |= CF_SHUTW;
921 oc->wex = TICK_ETERNITY;
Christopher Fauleta0bdec32022-04-04 07:51:21 +0200922 cs_done_get(cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200923
924 if (tick_isset(cs->hcto)) {
925 ic->rto = cs->hcto;
926 ic->rex = tick_add(now_ms, ic->rto);
927 }
928
929 /* on shutw we always wake the applet up */
930 appctx_wakeup(__cs_appctx(cs));
931
932 switch (cs->state) {
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200933 case SC_ST_RDY:
934 case SC_ST_EST:
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200935 /* we have to shut before closing, otherwise some short messages
936 * may never leave the system, especially when there are remaining
937 * unread data in the socket input buffer, or when nolinger is set.
Willy Tarreaucb041662022-05-17 19:44:42 +0200938 * However, if SC_FL_NOLINGER is explicitly set, we know there is
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200939 * no risk so we close both sides immediately.
940 */
Willy Tarreaucb041662022-05-17 19:44:42 +0200941 if (!sc_ep_test(cs, SE_FL_ERROR) && !(cs->flags & SC_FL_NOLINGER) &&
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200942 !(ic->flags & (CF_SHUTR|CF_DONT_READ)))
943 return;
944
945 /* fall through */
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200946 case SC_ST_CON:
947 case SC_ST_CER:
948 case SC_ST_QUE:
949 case SC_ST_TAR:
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200950 /* Note that none of these states may happen with applets */
Willy Tarreau1c3ead42022-05-10 19:42:22 +0200951 appctx_shut(__cs_appctx(cs));
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200952 cs->state = SC_ST_DIS;
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200953 /* fall through */
954 default:
Willy Tarreaucb041662022-05-17 19:44:42 +0200955 cs->flags &= ~SC_FL_NOLINGER;
Christopher Fauleta0bdec32022-04-04 07:51:21 +0200956 cs_rx_shut_blk(cs);
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200957 ic->flags |= CF_SHUTR;
958 ic->rex = TICK_ETERNITY;
959 __cs_strm(cs)->conn_exp = TICK_ETERNITY;
960 }
961}
962
963/* chk_rcv function for applets */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200964static void sc_app_chk_rcv_applet(struct stconn *cs)
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200965{
966 struct channel *ic = cs_ic(cs);
967
968 BUG_ON(!cs_appctx(cs));
969
970 DPRINTF(stderr, "%s: cs=%p, cs->state=%d ic->flags=%08x oc->flags=%08x\n",
971 __FUNCTION__,
972 cs, cs->state, ic->flags, cs_oc(cs)->flags);
973
974 if (!ic->pipe) {
975 /* (re)start reading */
976 appctx_wakeup(__cs_appctx(cs));
977 }
978}
979
980/* chk_snd function for applets */
Willy Tarreau4596fe22022-05-17 19:07:51 +0200981static void sc_app_chk_snd_applet(struct stconn *cs)
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200982{
983 struct channel *oc = cs_oc(cs);
984
985 BUG_ON(!cs_appctx(cs));
986
987 DPRINTF(stderr, "%s: cs=%p, cs->state=%d ic->flags=%08x oc->flags=%08x\n",
988 __FUNCTION__,
989 cs, cs->state, cs_ic(cs)->flags, oc->flags);
990
Willy Tarreau026e8fb2022-05-17 19:47:17 +0200991 if (unlikely(cs->state != SC_ST_EST || (oc->flags & CF_SHUTW)))
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200992 return;
993
994 /* we only wake the applet up if it was waiting for some data */
995
Willy Tarreaub605c422022-05-17 17:04:55 +0200996 if (!sc_ep_test(cs, SE_FL_WAIT_DATA))
Christopher Faulet9ffddd52022-04-01 14:04:29 +0200997 return;
998
999 if (!tick_isset(oc->wex))
1000 oc->wex = tick_add_ifset(now_ms, oc->wto);
1001
1002 if (!channel_is_empty(oc)) {
1003 /* (re)start sending */
1004 appctx_wakeup(__cs_appctx(cs));
1005 }
1006}
Christopher Faulet13045f02022-04-01 14:23:38 +02001007
1008
1009/* This function is designed to be called from within the stream handler to
Willy Tarreau4596fe22022-05-17 19:07:51 +02001010 * update the input channel's expiration timer and the stream connector's
Christopher Faulet13045f02022-04-01 14:23:38 +02001011 * Rx flags based on the channel's flags. It needs to be called only once
1012 * after the channel's flags have settled down, and before they are cleared,
1013 * though it doesn't harm to call it as often as desired (it just slightly
1014 * hurts performance). It must not be called from outside of the stream
1015 * handler, as what it does will be used to compute the stream task's
1016 * expiration.
1017 */
Willy Tarreau4596fe22022-05-17 19:07:51 +02001018void cs_update_rx(struct stconn *cs)
Christopher Faulet13045f02022-04-01 14:23:38 +02001019{
1020 struct channel *ic = cs_ic(cs);
1021
1022 if (ic->flags & CF_SHUTR) {
Christopher Fauleta0bdec32022-04-04 07:51:21 +02001023 cs_rx_shut_blk(cs);
Christopher Faulet13045f02022-04-01 14:23:38 +02001024 return;
1025 }
1026
1027 /* Read not closed, update FD status and timeout for reads */
1028 if (ic->flags & CF_DONT_READ)
Christopher Fauleta0bdec32022-04-04 07:51:21 +02001029 cs_rx_chan_blk(cs);
Christopher Faulet13045f02022-04-01 14:23:38 +02001030 else
Christopher Fauleta0bdec32022-04-04 07:51:21 +02001031 cs_rx_chan_rdy(cs);
Christopher Faulet13045f02022-04-01 14:23:38 +02001032
1033 if (!channel_is_empty(ic) || !channel_may_recv(ic)) {
1034 /* stop reading, imposed by channel's policy or contents */
Christopher Fauleta0bdec32022-04-04 07:51:21 +02001035 cs_rx_room_blk(cs);
Christopher Faulet13045f02022-04-01 14:23:38 +02001036 }
1037 else {
1038 /* (re)start reading and update timeout. Note: we don't recompute the timeout
1039 * every time we get here, otherwise it would risk never to expire. We only
1040 * update it if is was not yet set. The stream socket handler will already
1041 * have updated it if there has been a completed I/O.
1042 */
Christopher Fauleta0bdec32022-04-04 07:51:21 +02001043 cs_rx_room_rdy(cs);
Christopher Faulet13045f02022-04-01 14:23:38 +02001044 }
Willy Tarreaub605c422022-05-17 17:04:55 +02001045 if (sc_ep_test(cs, SE_FL_RXBLK_ANY))
Christopher Faulet13045f02022-04-01 14:23:38 +02001046 ic->rex = TICK_ETERNITY;
1047 else if (!(ic->flags & CF_READ_NOEXP) && !tick_isset(ic->rex))
1048 ic->rex = tick_add_ifset(now_ms, ic->rto);
1049
1050 cs_chk_rcv(cs);
1051}
1052
1053/* This function is designed to be called from within the stream handler to
Willy Tarreau4596fe22022-05-17 19:07:51 +02001054 * update the output channel's expiration timer and the stream connector's
Christopher Faulet13045f02022-04-01 14:23:38 +02001055 * Tx flags based on the channel's flags. It needs to be called only once
1056 * after the channel's flags have settled down, and before they are cleared,
1057 * though it doesn't harm to call it as often as desired (it just slightly
1058 * hurts performance). It must not be called from outside of the stream
1059 * handler, as what it does will be used to compute the stream task's
1060 * expiration.
1061 */
Willy Tarreau4596fe22022-05-17 19:07:51 +02001062void cs_update_tx(struct stconn *cs)
Christopher Faulet13045f02022-04-01 14:23:38 +02001063{
1064 struct channel *oc = cs_oc(cs);
1065 struct channel *ic = cs_ic(cs);
1066
1067 if (oc->flags & CF_SHUTW)
1068 return;
1069
1070 /* Write not closed, update FD status and timeout for writes */
1071 if (channel_is_empty(oc)) {
1072 /* stop writing */
Willy Tarreaub605c422022-05-17 17:04:55 +02001073 if (!sc_ep_test(cs, SE_FL_WAIT_DATA)) {
Christopher Faulet13045f02022-04-01 14:23:38 +02001074 if ((oc->flags & CF_SHUTW_NOW) == 0)
Willy Tarreaub605c422022-05-17 17:04:55 +02001075 sc_ep_set(cs, SE_FL_WAIT_DATA);
Christopher Faulet13045f02022-04-01 14:23:38 +02001076 oc->wex = TICK_ETERNITY;
1077 }
1078 return;
1079 }
1080
1081 /* (re)start writing and update timeout. Note: we don't recompute the timeout
1082 * every time we get here, otherwise it would risk never to expire. We only
1083 * update it if is was not yet set. The stream socket handler will already
1084 * have updated it if there has been a completed I/O.
1085 */
Willy Tarreaub605c422022-05-17 17:04:55 +02001086 sc_ep_clr(cs, SE_FL_WAIT_DATA);
Christopher Faulet13045f02022-04-01 14:23:38 +02001087 if (!tick_isset(oc->wex)) {
1088 oc->wex = tick_add_ifset(now_ms, oc->wto);
Willy Tarreaucb041662022-05-17 19:44:42 +02001089 if (tick_isset(ic->rex) && !(cs->flags & SC_FL_INDEP_STR)) {
Christopher Faulet13045f02022-04-01 14:23:38 +02001090 /* Note: depending on the protocol, we don't know if we're waiting
1091 * for incoming data or not. So in order to prevent the socket from
1092 * expiring read timeouts during writes, we refresh the read timeout,
1093 * except if it was already infinite or if we have explicitly setup
1094 * independent streams.
1095 */
1096 ic->rex = tick_add_ifset(now_ms, ic->rto);
Christopher Faulet5e29b762022-04-04 08:58:34 +02001097 }
1098 }
1099}
1100
1101/* This function is the equivalent to cs_update() except that it's
1102 * designed to be called from outside the stream handlers, typically the lower
1103 * layers (applets, connections) after I/O completion. After updating the stream
1104 * interface and timeouts, it will try to forward what can be forwarded, then to
1105 * wake the associated task up if an important event requires special handling.
Willy Tarreaub605c422022-05-17 17:04:55 +02001106 * It may update SE_FL_WAIT_DATA and/or SE_FL_RXBLK_ROOM, that the callers are
Christopher Faulet5e29b762022-04-04 08:58:34 +02001107 * encouraged to watch to take appropriate action.
1108 * It should not be called from within the stream itself, cs_update()
1109 * is designed for this.
1110 */
Willy Tarreau4596fe22022-05-17 19:07:51 +02001111static void cs_notify(struct stconn *cs)
Christopher Faulet5e29b762022-04-04 08:58:34 +02001112{
1113 struct channel *ic = cs_ic(cs);
1114 struct channel *oc = cs_oc(cs);
Willy Tarreau4596fe22022-05-17 19:07:51 +02001115 struct stconn *cso = cs_opposite(cs);
Christopher Faulet5e29b762022-04-04 08:58:34 +02001116 struct task *task = cs_strm_task(cs);
1117
1118 /* process consumer side */
1119 if (channel_is_empty(oc)) {
1120 struct connection *conn = cs_conn(cs);
1121
1122 if (((oc->flags & (CF_SHUTW|CF_SHUTW_NOW)) == CF_SHUTW_NOW) &&
Willy Tarreau026e8fb2022-05-17 19:47:17 +02001123 (cs->state == SC_ST_EST) && (!conn || !(conn->flags & (CO_FL_WAIT_XPRT | CO_FL_EARLY_SSL_HS))))
Christopher Faulet5e29b762022-04-04 08:58:34 +02001124 cs_shutw(cs);
1125 oc->wex = TICK_ETERNITY;
1126 }
1127
1128 /* indicate that we may be waiting for data from the output channel or
1129 * we're about to close and can't expect more data if SHUTW_NOW is there.
1130 */
1131 if (!(oc->flags & (CF_SHUTW|CF_SHUTW_NOW)))
Willy Tarreaub605c422022-05-17 17:04:55 +02001132 sc_ep_set(cs, SE_FL_WAIT_DATA);
Christopher Faulet5e29b762022-04-04 08:58:34 +02001133 else if ((oc->flags & (CF_SHUTW|CF_SHUTW_NOW)) == CF_SHUTW_NOW)
Willy Tarreaub605c422022-05-17 17:04:55 +02001134 sc_ep_clr(cs, SE_FL_WAIT_DATA);
Christopher Faulet5e29b762022-04-04 08:58:34 +02001135
1136 /* update OC timeouts and wake the other side up if it's waiting for room */
1137 if (oc->flags & CF_WRITE_ACTIVITY) {
1138 if ((oc->flags & (CF_SHUTW|CF_WRITE_PARTIAL)) == CF_WRITE_PARTIAL &&
1139 !channel_is_empty(oc))
1140 if (tick_isset(oc->wex))
1141 oc->wex = tick_add_ifset(now_ms, oc->wto);
1142
Willy Tarreaucb041662022-05-17 19:44:42 +02001143 if (!(cs->flags & SC_FL_INDEP_STR))
Christopher Faulet5e29b762022-04-04 08:58:34 +02001144 if (tick_isset(ic->rex))
1145 ic->rex = tick_add_ifset(now_ms, ic->rto);
1146 }
1147
1148 if (oc->flags & CF_DONT_READ)
1149 cs_rx_chan_blk(cso);
1150 else
1151 cs_rx_chan_rdy(cso);
1152
1153 /* Notify the other side when we've injected data into the IC that
1154 * needs to be forwarded. We can do fast-forwarding as soon as there
1155 * are output data, but we avoid doing this if some of the data are
1156 * not yet scheduled for being forwarded, because it is very likely
1157 * that it will be done again immediately afterwards once the following
Willy Tarreaub605c422022-05-17 17:04:55 +02001158 * data are parsed (eg: HTTP chunking). We only SE_FL_RXBLK_ROOM once
Christopher Faulet5e29b762022-04-04 08:58:34 +02001159 * we've emptied *some* of the output buffer, and not just when there
1160 * is available room, because applets are often forced to stop before
1161 * the buffer is full. We must not stop based on input data alone because
1162 * an HTTP parser might need more data to complete the parsing.
1163 */
1164 if (!channel_is_empty(ic) &&
Willy Tarreaub605c422022-05-17 17:04:55 +02001165 sc_ep_test(cso, SE_FL_WAIT_DATA) &&
Christopher Faulet5e29b762022-04-04 08:58:34 +02001166 (!(ic->flags & CF_EXPECT_MORE) || c_full(ic) || ci_data(ic) == 0 || ic->pipe)) {
1167 int new_len, last_len;
1168
1169 last_len = co_data(ic);
1170 if (ic->pipe)
1171 last_len += ic->pipe->data;
1172
1173 cs_chk_snd(cso);
1174
1175 new_len = co_data(ic);
1176 if (ic->pipe)
1177 new_len += ic->pipe->data;
1178
1179 /* check if the consumer has freed some space either in the
1180 * buffer or in the pipe.
1181 */
1182 if (new_len < last_len)
1183 cs_rx_room_rdy(cs);
1184 }
1185
1186 if (!(ic->flags & CF_DONT_READ))
1187 cs_rx_chan_rdy(cs);
1188
1189 cs_chk_rcv(cs);
1190 cs_chk_rcv(cso);
1191
1192 if (cs_rx_blocked(cs)) {
1193 ic->rex = TICK_ETERNITY;
1194 }
1195 else if ((ic->flags & (CF_SHUTR|CF_READ_PARTIAL)) == CF_READ_PARTIAL) {
1196 /* we must re-enable reading if cs_chk_snd() has freed some space */
1197 if (!(ic->flags & CF_READ_NOEXP) && tick_isset(ic->rex))
1198 ic->rex = tick_add_ifset(now_ms, ic->rto);
1199 }
1200
1201 /* wake the task up only when needed */
1202 if (/* changes on the production side */
1203 (ic->flags & (CF_READ_NULL|CF_READ_ERROR)) ||
Willy Tarreau026e8fb2022-05-17 19:47:17 +02001204 !cs_state_in(cs->state, SC_SB_CON|SC_SB_RDY|SC_SB_EST) ||
Willy Tarreaub605c422022-05-17 17:04:55 +02001205 sc_ep_test(cs, SE_FL_ERROR) ||
Christopher Faulet5e29b762022-04-04 08:58:34 +02001206 ((ic->flags & CF_READ_PARTIAL) &&
Willy Tarreau026e8fb2022-05-17 19:47:17 +02001207 ((ic->flags & CF_EOI) || !ic->to_forward || cso->state != SC_ST_EST)) ||
Christopher Faulet5e29b762022-04-04 08:58:34 +02001208
1209 /* changes on the consumption side */
1210 (oc->flags & (CF_WRITE_NULL|CF_WRITE_ERROR)) ||
1211 ((oc->flags & CF_WRITE_ACTIVITY) &&
1212 ((oc->flags & CF_SHUTW) ||
1213 (((oc->flags & CF_WAKE_WRITE) ||
1214 !(oc->flags & (CF_AUTO_CLOSE|CF_SHUTW_NOW|CF_SHUTW))) &&
Willy Tarreau026e8fb2022-05-17 19:47:17 +02001215 (cso->state != SC_ST_EST ||
Christopher Faulet5e29b762022-04-04 08:58:34 +02001216 (channel_is_empty(oc) && !oc->to_forward)))))) {
1217 task_wakeup(task, TASK_WOKEN_IO);
1218 }
1219 else {
1220 /* Update expiration date for the task and requeue it */
1221 task->expire = tick_first((tick_is_expired(task->expire, now_ms) ? 0 : task->expire),
1222 tick_first(tick_first(ic->rex, ic->wex),
1223 tick_first(oc->rex, oc->wex)));
1224
1225 task->expire = tick_first(task->expire, ic->analyse_exp);
1226 task->expire = tick_first(task->expire, oc->analyse_exp);
1227 task->expire = tick_first(task->expire, __cs_strm(cs)->conn_exp);
1228
1229 task_queue(task);
1230 }
1231 if (ic->flags & CF_READ_ACTIVITY)
1232 ic->flags &= ~CF_READ_DONTWAIT;
1233}
1234
1235/*
1236 * This function propagates a null read received on a socket-based connection.
Willy Tarreaucb041662022-05-17 19:44:42 +02001237 * It updates the stream connector. If the stream connector has SC_FL_NOHALF,
Christopher Faulet5e29b762022-04-04 08:58:34 +02001238 * the close is also forwarded to the write side as an abort.
1239 */
Willy Tarreau4596fe22022-05-17 19:07:51 +02001240static void cs_conn_read0(struct stconn *cs)
Christopher Faulet5e29b762022-04-04 08:58:34 +02001241{
1242 struct channel *ic = cs_ic(cs);
1243 struct channel *oc = cs_oc(cs);
1244
1245 BUG_ON(!cs_conn(cs));
1246
1247 cs_rx_shut_blk(cs);
1248 if (ic->flags & CF_SHUTR)
1249 return;
1250 ic->flags |= CF_SHUTR;
1251 ic->rex = TICK_ETERNITY;
1252
Willy Tarreau026e8fb2022-05-17 19:47:17 +02001253 if (!cs_state_in(cs->state, SC_SB_CON|SC_SB_RDY|SC_SB_EST))
Christopher Faulet5e29b762022-04-04 08:58:34 +02001254 return;
1255
1256 if (oc->flags & CF_SHUTW)
1257 goto do_close;
1258
Willy Tarreaucb041662022-05-17 19:44:42 +02001259 if (cs->flags & SC_FL_NOHALF) {
Christopher Faulet5e29b762022-04-04 08:58:34 +02001260 /* we want to immediately forward this close to the write side */
1261 /* force flag on ssl to keep stream in cache */
1262 cs_conn_shutw(cs, CO_SHW_SILENT);
1263 goto do_close;
1264 }
1265
1266 /* otherwise that's just a normal read shutdown */
1267 return;
1268
1269 do_close:
1270 /* OK we completely close the socket here just as if we went through cs_shut[rw]() */
Christopher Fauletff022a22022-04-21 08:38:54 +02001271 cs_conn_shut(cs);
Christopher Faulet5e29b762022-04-04 08:58:34 +02001272
1273 oc->flags &= ~CF_SHUTW_NOW;
1274 oc->flags |= CF_SHUTW;
1275 oc->wex = TICK_ETERNITY;
1276
1277 cs_done_get(cs);
1278
Willy Tarreau026e8fb2022-05-17 19:47:17 +02001279 cs->state = SC_ST_DIS;
Christopher Faulet5e29b762022-04-04 08:58:34 +02001280 __cs_strm(cs)->conn_exp = TICK_ETERNITY;
1281 return;
1282}
1283
1284/*
1285 * This is the callback which is called by the connection layer to receive data
1286 * into the buffer from the connection. It iterates over the mux layer's
1287 * rcv_buf function.
1288 */
Willy Tarreau4596fe22022-05-17 19:07:51 +02001289static int cs_conn_recv(struct stconn *cs)
Christopher Faulet5e29b762022-04-04 08:58:34 +02001290{
1291 struct connection *conn = __cs_conn(cs);
1292 struct channel *ic = cs_ic(cs);
1293 int ret, max, cur_read = 0;
1294 int read_poll = MAX_READ_POLL_LOOPS;
1295 int flags = 0;
1296
1297 /* If not established yet, do nothing. */
Willy Tarreau026e8fb2022-05-17 19:47:17 +02001298 if (cs->state != SC_ST_EST)
Christopher Faulet5e29b762022-04-04 08:58:34 +02001299 return 0;
1300
1301 /* If another call to cs_conn_recv() failed, and we subscribed to
1302 * recv events already, give up now.
1303 */
1304 if (cs->wait_event.events & SUB_RETRY_RECV)
1305 return 0;
1306
1307 /* maybe we were called immediately after an asynchronous shutr */
1308 if (ic->flags & CF_SHUTR)
1309 return 1;
1310
1311 /* we must wait because the mux is not installed yet */
1312 if (!conn->mux)
1313 return 0;
1314
1315 /* stop here if we reached the end of data */
Willy Tarreaub605c422022-05-17 17:04:55 +02001316 if (sc_ep_test(cs, SE_FL_EOS))
Christopher Faulet5e29b762022-04-04 08:58:34 +02001317 goto end_recv;
1318
1319 /* stop immediately on errors. Note that we DON'T want to stop on
1320 * POLL_ERR, as the poller might report a write error while there
1321 * are still data available in the recv buffer. This typically
1322 * happens when we send too large a request to a backend server
1323 * which rejects it before reading it all.
1324 */
Willy Tarreaub605c422022-05-17 17:04:55 +02001325 if (!sc_ep_test(cs, SE_FL_RCV_MORE)) {
Christopher Faulet5e29b762022-04-04 08:58:34 +02001326 if (!conn_xprt_ready(conn))
1327 return 0;
Willy Tarreaub605c422022-05-17 17:04:55 +02001328 if (sc_ep_test(cs, SE_FL_ERROR))
Christopher Faulet5e29b762022-04-04 08:58:34 +02001329 goto end_recv;
1330 }
1331
1332 /* prepare to detect if the mux needs more room */
Willy Tarreaub605c422022-05-17 17:04:55 +02001333 sc_ep_clr(cs, SE_FL_WANT_ROOM);
Christopher Faulet5e29b762022-04-04 08:58:34 +02001334
1335 if ((ic->flags & (CF_STREAMER | CF_STREAMER_FAST)) && !co_data(ic) &&
1336 global.tune.idle_timer &&
1337 (unsigned short)(now_ms - ic->last_read) >= global.tune.idle_timer) {
1338 /* The buffer was empty and nothing was transferred for more
1339 * than one second. This was caused by a pause and not by
1340 * congestion. Reset any streaming mode to reduce latency.
1341 */
1342 ic->xfer_small = 0;
1343 ic->xfer_large = 0;
1344 ic->flags &= ~(CF_STREAMER | CF_STREAMER_FAST);
1345 }
1346
1347 /* First, let's see if we may splice data across the channel without
1348 * using a buffer.
1349 */
Willy Tarreaub605c422022-05-17 17:04:55 +02001350 if (sc_ep_test(cs, SE_FL_MAY_SPLICE) &&
Christopher Faulet5e29b762022-04-04 08:58:34 +02001351 (ic->pipe || ic->to_forward >= MIN_SPLICE_FORWARD) &&
1352 ic->flags & CF_KERN_SPLICING) {
1353 if (c_data(ic)) {
1354 /* We're embarrassed, there are already data pending in
1355 * the buffer and we don't want to have them at two
1356 * locations at a time. Let's indicate we need some
1357 * place and ask the consumer to hurry.
1358 */
1359 flags |= CO_RFL_BUF_FLUSH;
1360 goto abort_splice;
1361 }
1362
1363 if (unlikely(ic->pipe == NULL)) {
1364 if (pipes_used >= global.maxpipes || !(ic->pipe = get_pipe())) {
1365 ic->flags &= ~CF_KERN_SPLICING;
1366 goto abort_splice;
1367 }
1368 }
1369
1370 ret = conn->mux->rcv_pipe(cs, ic->pipe, ic->to_forward);
1371 if (ret < 0) {
1372 /* splice not supported on this end, let's disable it */
1373 ic->flags &= ~CF_KERN_SPLICING;
1374 goto abort_splice;
1375 }
1376
1377 if (ret > 0) {
1378 if (ic->to_forward != CHN_INFINITE_FORWARD)
1379 ic->to_forward -= ret;
1380 ic->total += ret;
1381 cur_read += ret;
1382 ic->flags |= CF_READ_PARTIAL;
1383 }
1384
Willy Tarreaub605c422022-05-17 17:04:55 +02001385 if (sc_ep_test(cs, SE_FL_EOS | SE_FL_ERROR))
Christopher Faulet5e29b762022-04-04 08:58:34 +02001386 goto end_recv;
1387
1388 if (conn->flags & CO_FL_WAIT_ROOM) {
1389 /* the pipe is full or we have read enough data that it
1390 * could soon be full. Let's stop before needing to poll.
1391 */
1392 cs_rx_room_blk(cs);
1393 goto done_recv;
1394 }
1395
1396 /* splice not possible (anymore), let's go on on standard copy */
1397 }
1398
1399 abort_splice:
1400 if (ic->pipe && unlikely(!ic->pipe->data)) {
1401 put_pipe(ic->pipe);
1402 ic->pipe = NULL;
1403 }
1404
Willy Tarreaub605c422022-05-17 17:04:55 +02001405 if (ic->pipe && ic->to_forward && !(flags & CO_RFL_BUF_FLUSH) && sc_ep_test(cs, SE_FL_MAY_SPLICE)) {
Christopher Faulet5e29b762022-04-04 08:58:34 +02001406 /* don't break splicing by reading, but still call rcv_buf()
1407 * to pass the flag.
1408 */
1409 goto done_recv;
1410 }
1411
1412 /* now we'll need a input buffer for the stream */
1413 if (!cs_alloc_ibuf(cs, &(__cs_strm(cs)->buffer_wait)))
1414 goto end_recv;
1415
1416 /* For an HTX stream, if the buffer is stuck (no output data with some
1417 * input data) and if the HTX message is fragmented or if its free space
1418 * wraps, we force an HTX deframentation. It is a way to have a
1419 * contiguous free space nad to let the mux to copy as much data as
1420 * possible.
1421 *
1422 * NOTE: A possible optim may be to let the mux decides if defrag is
1423 * required or not, depending on amount of data to be xferred.
1424 */
1425 if (IS_HTX_STRM(__cs_strm(cs)) && !co_data(ic)) {
1426 struct htx *htx = htxbuf(&ic->buf);
1427
1428 if (htx_is_not_empty(htx) && ((htx->flags & HTX_FL_FRAGMENTED) || htx_space_wraps(htx)))
1429 htx_defrag(htx, NULL, 0);
1430 }
1431
1432 /* Instruct the mux it must subscribed for read events */
1433 flags |= ((!conn_is_back(conn) && (__cs_strm(cs)->be->options & PR_O_ABRT_CLOSE)) ? CO_RFL_KEEP_RECV : 0);
1434
1435 /* Important note : if we're called with POLL_IN|POLL_HUP, it means the read polling
1436 * was enabled, which implies that the recv buffer was not full. So we have a guarantee
1437 * that if such an event is not handled above in splice, it will be handled here by
1438 * recv().
1439 */
Willy Tarreaub605c422022-05-17 17:04:55 +02001440 while (sc_ep_test(cs, SE_FL_RCV_MORE) ||
Christopher Faulet5e29b762022-04-04 08:58:34 +02001441 (!(conn->flags & CO_FL_HANDSHAKE) &&
Willy Tarreaub605c422022-05-17 17:04:55 +02001442 (!sc_ep_test(cs, SE_FL_ERROR | SE_FL_EOS)) && !(ic->flags & CF_SHUTR))) {
Christopher Faulet5e29b762022-04-04 08:58:34 +02001443 int cur_flags = flags;
1444
1445 /* Compute transient CO_RFL_* flags */
1446 if (co_data(ic)) {
1447 cur_flags |= (CO_RFL_BUF_WET | CO_RFL_BUF_NOT_STUCK);
1448 }
1449
1450 /* <max> may be null. This is the mux responsibility to set
Willy Tarreaub605c422022-05-17 17:04:55 +02001451 * SE_FL_RCV_MORE on the CS if more space is needed.
Christopher Faulet5e29b762022-04-04 08:58:34 +02001452 */
1453 max = channel_recv_max(ic);
1454 ret = conn->mux->rcv_buf(cs, &ic->buf, max, cur_flags);
1455
Willy Tarreaub605c422022-05-17 17:04:55 +02001456 if (sc_ep_test(cs, SE_FL_WANT_ROOM)) {
1457 /* SE_FL_WANT_ROOM must not be reported if the channel's
Christopher Faulet5e29b762022-04-04 08:58:34 +02001458 * buffer is empty.
1459 */
1460 BUG_ON(c_empty(ic));
1461
1462 cs_rx_room_blk(cs);
1463 /* Add READ_PARTIAL because some data are pending but
1464 * cannot be xferred to the channel
1465 */
1466 ic->flags |= CF_READ_PARTIAL;
1467 }
1468
1469 if (ret <= 0) {
1470 /* if we refrained from reading because we asked for a
1471 * flush to satisfy rcv_pipe(), we must not subscribe
1472 * and instead report that there's not enough room
1473 * here to proceed.
1474 */
1475 if (flags & CO_RFL_BUF_FLUSH)
1476 cs_rx_room_blk(cs);
1477 break;
1478 }
1479
1480 cur_read += ret;
1481
1482 /* if we're allowed to directly forward data, we must update ->o */
1483 if (ic->to_forward && !(ic->flags & (CF_SHUTW|CF_SHUTW_NOW))) {
1484 unsigned long fwd = ret;
1485 if (ic->to_forward != CHN_INFINITE_FORWARD) {
1486 if (fwd > ic->to_forward)
1487 fwd = ic->to_forward;
1488 ic->to_forward -= fwd;
1489 }
1490 c_adv(ic, fwd);
1491 }
1492
1493 ic->flags |= CF_READ_PARTIAL;
1494 ic->total += ret;
1495
1496 /* End-of-input reached, we can leave. In this case, it is
Christopher Faulet6b0a0fb2022-04-04 11:29:28 +02001497 * important to break the loop to not block the CS because of
Christopher Faulet5e29b762022-04-04 08:58:34 +02001498 * the channel's policies.This way, we are still able to receive
1499 * shutdowns.
1500 */
Willy Tarreaub605c422022-05-17 17:04:55 +02001501 if (sc_ep_test(cs, SE_FL_EOI))
Christopher Faulet5e29b762022-04-04 08:58:34 +02001502 break;
1503
1504 if ((ic->flags & CF_READ_DONTWAIT) || --read_poll <= 0) {
1505 /* we're stopped by the channel's policy */
1506 cs_rx_chan_blk(cs);
1507 break;
1508 }
1509
1510 /* if too many bytes were missing from last read, it means that
1511 * it's pointless trying to read again because the system does
1512 * not have them in buffers.
1513 */
1514 if (ret < max) {
1515 /* if a streamer has read few data, it may be because we
1516 * have exhausted system buffers. It's not worth trying
1517 * again.
1518 */
1519 if (ic->flags & CF_STREAMER) {
1520 /* we're stopped by the channel's policy */
1521 cs_rx_chan_blk(cs);
1522 break;
1523 }
1524
1525 /* if we read a large block smaller than what we requested,
1526 * it's almost certain we'll never get anything more.
1527 */
1528 if (ret >= global.tune.recv_enough) {
1529 /* we're stopped by the channel's policy */
1530 cs_rx_chan_blk(cs);
1531 break;
1532 }
1533 }
1534
1535 /* if we are waiting for more space, don't try to read more data
1536 * right now.
1537 */
1538 if (cs_rx_blocked(cs))
1539 break;
1540 } /* while !flags */
1541
1542 done_recv:
1543 if (cur_read) {
1544 if ((ic->flags & (CF_STREAMER | CF_STREAMER_FAST)) &&
1545 (cur_read <= ic->buf.size / 2)) {
1546 ic->xfer_large = 0;
1547 ic->xfer_small++;
1548 if (ic->xfer_small >= 3) {
1549 /* we have read less than half of the buffer in
1550 * one pass, and this happened at least 3 times.
1551 * This is definitely not a streamer.
1552 */
1553 ic->flags &= ~(CF_STREAMER | CF_STREAMER_FAST);
1554 }
1555 else if (ic->xfer_small >= 2) {
1556 /* if the buffer has been at least half full twice,
1557 * we receive faster than we send, so at least it
1558 * is not a "fast streamer".
1559 */
1560 ic->flags &= ~CF_STREAMER_FAST;
1561 }
1562 }
1563 else if (!(ic->flags & CF_STREAMER_FAST) &&
1564 (cur_read >= ic->buf.size - global.tune.maxrewrite)) {
1565 /* we read a full buffer at once */
1566 ic->xfer_small = 0;
1567 ic->xfer_large++;
1568 if (ic->xfer_large >= 3) {
1569 /* we call this buffer a fast streamer if it manages
1570 * to be filled in one call 3 consecutive times.
1571 */
1572 ic->flags |= (CF_STREAMER | CF_STREAMER_FAST);
1573 }
1574 }
1575 else {
1576 ic->xfer_small = 0;
1577 ic->xfer_large = 0;
1578 }
1579 ic->last_read = now_ms;
1580 }
1581
1582 end_recv:
1583 ret = (cur_read != 0);
1584
1585 /* Report EOI on the channel if it was reached from the mux point of
1586 * view. */
Willy Tarreaub605c422022-05-17 17:04:55 +02001587 if (sc_ep_test(cs, SE_FL_EOI) && !(ic->flags & CF_EOI)) {
Christopher Faulet5e29b762022-04-04 08:58:34 +02001588 ic->flags |= (CF_EOI|CF_READ_PARTIAL);
1589 ret = 1;
1590 }
1591
Willy Tarreaub605c422022-05-17 17:04:55 +02001592 if (sc_ep_test(cs, SE_FL_ERROR))
Christopher Faulet5e29b762022-04-04 08:58:34 +02001593 ret = 1;
Willy Tarreaub605c422022-05-17 17:04:55 +02001594 else if (sc_ep_test(cs, SE_FL_EOS)) {
Christopher Faulet5e29b762022-04-04 08:58:34 +02001595 /* we received a shutdown */
1596 ic->flags |= CF_READ_NULL;
1597 if (ic->flags & CF_AUTO_CLOSE)
1598 channel_shutw_now(ic);
1599 cs_conn_read0(cs);
1600 ret = 1;
1601 }
1602 else if (!cs_rx_blocked(cs)) {
1603 /* Subscribe to receive events if we're blocking on I/O */
1604 conn->mux->subscribe(cs, SUB_RETRY_RECV, &cs->wait_event);
1605 cs_rx_endp_done(cs);
1606 } else {
1607 cs_rx_endp_more(cs);
1608 ret = 1;
1609 }
1610 return ret;
1611}
1612
Willy Tarreau4596fe22022-05-17 19:07:51 +02001613/* This tries to perform a synchronous receive on the stream connector to
Christopher Faulet5e29b762022-04-04 08:58:34 +02001614 * try to collect last arrived data. In practice it's only implemented on
Willy Tarreau4596fe22022-05-17 19:07:51 +02001615 * stconns. Returns 0 if nothing was done, non-zero if new data or a
Christopher Faulet5e29b762022-04-04 08:58:34 +02001616 * shutdown were collected. This may result on some delayed receive calls
1617 * to be programmed and performed later, though it doesn't provide any
1618 * such guarantee.
1619 */
Willy Tarreau4596fe22022-05-17 19:07:51 +02001620int cs_conn_sync_recv(struct stconn *cs)
Christopher Faulet5e29b762022-04-04 08:58:34 +02001621{
Willy Tarreau026e8fb2022-05-17 19:47:17 +02001622 if (!cs_state_in(cs->state, SC_SB_RDY|SC_SB_EST))
Christopher Faulet5e29b762022-04-04 08:58:34 +02001623 return 0;
1624
1625 if (!cs_conn_mux(cs))
Willy Tarreau4596fe22022-05-17 19:07:51 +02001626 return 0; // only stconns are supported
Christopher Faulet5e29b762022-04-04 08:58:34 +02001627
1628 if (cs->wait_event.events & SUB_RETRY_RECV)
1629 return 0; // already subscribed
1630
1631 if (!cs_rx_endp_ready(cs) || cs_rx_blocked(cs))
1632 return 0; // already failed
1633
1634 return cs_conn_recv(cs);
1635}
1636
1637/*
1638 * This function is called to send buffer data to a stream socket.
1639 * It calls the mux layer's snd_buf function. It relies on the
1640 * caller to commit polling changes. The caller should check conn->flags
1641 * for errors.
1642 */
Willy Tarreau4596fe22022-05-17 19:07:51 +02001643static int cs_conn_send(struct stconn *cs)
Christopher Faulet5e29b762022-04-04 08:58:34 +02001644{
1645 struct connection *conn = __cs_conn(cs);
1646 struct stream *s = __cs_strm(cs);
1647 struct channel *oc = cs_oc(cs);
1648 int ret;
1649 int did_send = 0;
1650
Willy Tarreaub605c422022-05-17 17:04:55 +02001651 if (sc_ep_test(cs, SE_FL_ERROR | SE_FL_ERR_PENDING) || cs_is_conn_error(cs)) {
Christopher Faulet5e29b762022-04-04 08:58:34 +02001652 /* We're probably there because the tasklet was woken up,
1653 * but process_stream() ran before, detected there were an
Willy Tarreau026e8fb2022-05-17 19:47:17 +02001654 * error and put the CS back to SC_ST_TAR. There's still
Christopher Faulet5e29b762022-04-04 08:58:34 +02001655 * CO_FL_ERROR on the connection but we don't want to add
Willy Tarreaub605c422022-05-17 17:04:55 +02001656 * SE_FL_ERROR back, so give up
Christopher Faulet5e29b762022-04-04 08:58:34 +02001657 */
Willy Tarreau026e8fb2022-05-17 19:47:17 +02001658 if (cs->state < SC_ST_CON)
Christopher Faulet5e29b762022-04-04 08:58:34 +02001659 return 0;
Willy Tarreaub605c422022-05-17 17:04:55 +02001660 sc_ep_set(cs, SE_FL_ERROR);
Christopher Faulet5e29b762022-04-04 08:58:34 +02001661 return 1;
1662 }
1663
1664 /* We're already waiting to be able to send, give up */
1665 if (cs->wait_event.events & SUB_RETRY_SEND)
1666 return 0;
1667
1668 /* we might have been called just after an asynchronous shutw */
1669 if (oc->flags & CF_SHUTW)
1670 return 1;
1671
1672 /* we must wait because the mux is not installed yet */
1673 if (!conn->mux)
1674 return 0;
1675
1676 if (oc->pipe && conn->xprt->snd_pipe && conn->mux->snd_pipe) {
1677 ret = conn->mux->snd_pipe(cs, oc->pipe);
1678 if (ret > 0)
1679 did_send = 1;
1680
1681 if (!oc->pipe->data) {
1682 put_pipe(oc->pipe);
1683 oc->pipe = NULL;
1684 }
1685
1686 if (oc->pipe)
1687 goto end;
1688 }
1689
1690 /* At this point, the pipe is empty, but we may still have data pending
1691 * in the normal buffer.
1692 */
1693 if (co_data(oc)) {
1694 /* when we're here, we already know that there is no spliced
1695 * data left, and that there are sendable buffered data.
1696 */
1697
1698 /* check if we want to inform the kernel that we're interested in
1699 * sending more data after this call. We want this if :
1700 * - we're about to close after this last send and want to merge
1701 * the ongoing FIN with the last segment.
1702 * - we know we can't send everything at once and must get back
1703 * here because of unaligned data
1704 * - there is still a finite amount of data to forward
1705 * The test is arranged so that the most common case does only 2
1706 * tests.
1707 */
1708 unsigned int send_flag = 0;
1709
1710 if ((!(oc->flags & (CF_NEVER_WAIT|CF_SEND_DONTWAIT)) &&
1711 ((oc->to_forward && oc->to_forward != CHN_INFINITE_FORWARD) ||
1712 (oc->flags & CF_EXPECT_MORE) ||
1713 (IS_HTX_STRM(s) &&
1714 (!(oc->flags & (CF_EOI|CF_SHUTR)) && htx_expect_more(htxbuf(&oc->buf)))))) ||
1715 ((oc->flags & CF_ISRESP) &&
1716 ((oc->flags & (CF_AUTO_CLOSE|CF_SHUTW_NOW)) == (CF_AUTO_CLOSE|CF_SHUTW_NOW))))
1717 send_flag |= CO_SFL_MSG_MORE;
1718
1719 if (oc->flags & CF_STREAMER)
1720 send_flag |= CO_SFL_STREAMER;
1721
1722 if (s->txn && s->txn->flags & TX_L7_RETRY && !b_data(&s->txn->l7_buffer)) {
1723 /* If we want to be able to do L7 retries, copy
1724 * the data we're about to send, so that we are able
1725 * to resend them if needed
1726 */
1727 /* Try to allocate a buffer if we had none.
1728 * If it fails, the next test will just
1729 * disable the l7 retries by setting
1730 * l7_conn_retries to 0.
1731 */
1732 if (s->txn->req.msg_state != HTTP_MSG_DONE)
1733 s->txn->flags &= ~TX_L7_RETRY;
1734 else {
1735 if (b_alloc(&s->txn->l7_buffer) == NULL)
1736 s->txn->flags &= ~TX_L7_RETRY;
1737 else {
1738 memcpy(b_orig(&s->txn->l7_buffer),
1739 b_orig(&oc->buf),
1740 b_size(&oc->buf));
1741 s->txn->l7_buffer.head = co_data(oc);
1742 b_add(&s->txn->l7_buffer, co_data(oc));
1743 }
1744
1745 }
1746 }
1747
1748 ret = conn->mux->snd_buf(cs, &oc->buf, co_data(oc), send_flag);
1749 if (ret > 0) {
1750 did_send = 1;
1751 c_rew(oc, ret);
1752 c_realign_if_empty(oc);
1753
1754 if (!co_data(oc)) {
1755 /* Always clear both flags once everything has been sent, they're one-shot */
1756 oc->flags &= ~(CF_EXPECT_MORE | CF_SEND_DONTWAIT);
1757 }
1758 /* if some data remain in the buffer, it's only because the
1759 * system buffers are full, we will try next time.
1760 */
Christopher Faulet13045f02022-04-01 14:23:38 +02001761 }
1762 }
Christopher Faulet5e29b762022-04-04 08:58:34 +02001763
1764 end:
1765 if (did_send) {
1766 oc->flags |= CF_WRITE_PARTIAL | CF_WROTE_DATA;
Willy Tarreau026e8fb2022-05-17 19:47:17 +02001767 if (cs->state == SC_ST_CON)
1768 cs->state = SC_ST_RDY;
Christopher Faulet5e29b762022-04-04 08:58:34 +02001769
1770 cs_rx_room_rdy(cs_opposite(cs));
1771 }
1772
Willy Tarreaub605c422022-05-17 17:04:55 +02001773 if (sc_ep_test(cs, SE_FL_ERROR | SE_FL_ERR_PENDING)) {
1774 sc_ep_set(cs, SE_FL_ERROR);
Christopher Faulet5e29b762022-04-04 08:58:34 +02001775 return 1;
1776 }
1777
1778 /* We couldn't send all of our data, let the mux know we'd like to send more */
1779 if (!channel_is_empty(oc))
1780 conn->mux->subscribe(cs, SUB_RETRY_SEND, &cs->wait_event);
1781 return did_send;
1782}
1783
Willy Tarreau4596fe22022-05-17 19:07:51 +02001784/* perform a synchronous send() for the stream connector. The CF_WRITE_NULL and
Christopher Faulet5e29b762022-04-04 08:58:34 +02001785 * CF_WRITE_PARTIAL flags are cleared prior to the attempt, and will possibly
1786 * be updated in case of success.
1787 */
Willy Tarreau4596fe22022-05-17 19:07:51 +02001788void cs_conn_sync_send(struct stconn *cs)
Christopher Faulet5e29b762022-04-04 08:58:34 +02001789{
1790 struct channel *oc = cs_oc(cs);
1791
1792 oc->flags &= ~(CF_WRITE_NULL|CF_WRITE_PARTIAL);
1793
1794 if (oc->flags & CF_SHUTW)
1795 return;
1796
1797 if (channel_is_empty(oc))
1798 return;
1799
Willy Tarreau026e8fb2022-05-17 19:47:17 +02001800 if (!cs_state_in(cs->state, SC_SB_CON|SC_SB_RDY|SC_SB_EST))
Christopher Faulet5e29b762022-04-04 08:58:34 +02001801 return;
1802
1803 if (!cs_conn_mux(cs))
1804 return;
1805
1806 cs_conn_send(cs);
1807}
1808
1809/* Called by I/O handlers after completion.. It propagates
Willy Tarreau4596fe22022-05-17 19:07:51 +02001810 * connection flags to the stream connector, updates the stream (which may or
Christopher Faulet5e29b762022-04-04 08:58:34 +02001811 * may not take this opportunity to try to forward data), then update the
Willy Tarreau4596fe22022-05-17 19:07:51 +02001812 * connection's polling based on the channels and stream connector's final
Christopher Faulet5e29b762022-04-04 08:58:34 +02001813 * states. The function always returns 0.
1814 */
Willy Tarreau4596fe22022-05-17 19:07:51 +02001815static int cs_conn_process(struct stconn *cs)
Christopher Faulet5e29b762022-04-04 08:58:34 +02001816{
1817 struct connection *conn = __cs_conn(cs);
1818 struct channel *ic = cs_ic(cs);
1819 struct channel *oc = cs_oc(cs);
1820
1821 BUG_ON(!conn);
1822
1823 /* If we have data to send, try it now */
1824 if (!channel_is_empty(oc) && !(cs->wait_event.events & SUB_RETRY_SEND))
1825 cs_conn_send(cs);
1826
Willy Tarreau4596fe22022-05-17 19:07:51 +02001827 /* First step, report to the stream connector what was detected at the
Christopher Faulet5e29b762022-04-04 08:58:34 +02001828 * connection layer : errors and connection establishment.
Willy Tarreaub605c422022-05-17 17:04:55 +02001829 * Only add SE_FL_ERROR if we're connected, or we're attempting to
Christopher Faulet5e29b762022-04-04 08:58:34 +02001830 * connect, we may get there because we got woken up, but only run
1831 * after process_stream() noticed there were an error, and decided
1832 * to retry to connect, the connection may still have CO_FL_ERROR,
Willy Tarreaub605c422022-05-17 17:04:55 +02001833 * and we don't want to add SE_FL_ERROR back
Christopher Faulet5e29b762022-04-04 08:58:34 +02001834 *
1835 * Note: This test is only required because cs_conn_process is also the SI
1836 * wake callback. Otherwise cs_conn_recv()/cs_conn_send() already take
1837 * care of it.
1838 */
1839
Willy Tarreau026e8fb2022-05-17 19:47:17 +02001840 if (cs->state >= SC_ST_CON) {
Christopher Faulet5e29b762022-04-04 08:58:34 +02001841 if (cs_is_conn_error(cs))
Willy Tarreaub605c422022-05-17 17:04:55 +02001842 sc_ep_set(cs, SE_FL_ERROR);
Christopher Faulet5e29b762022-04-04 08:58:34 +02001843 }
1844
1845 /* If we had early data, and the handshake ended, then
1846 * we can remove the flag, and attempt to wake the task up,
1847 * in the event there's an analyser waiting for the end of
1848 * the handshake.
1849 */
1850 if (!(conn->flags & (CO_FL_WAIT_XPRT | CO_FL_EARLY_SSL_HS)) &&
Willy Tarreaub605c422022-05-17 17:04:55 +02001851 sc_ep_test(cs, SE_FL_WAIT_FOR_HS)) {
1852 sc_ep_clr(cs, SE_FL_WAIT_FOR_HS);
Christopher Faulet5e29b762022-04-04 08:58:34 +02001853 task_wakeup(cs_strm_task(cs), TASK_WOKEN_MSG);
1854 }
1855
Willy Tarreau026e8fb2022-05-17 19:47:17 +02001856 if (!cs_state_in(cs->state, SC_SB_EST|SC_SB_DIS|SC_SB_CLO) &&
Christopher Faulet5e29b762022-04-04 08:58:34 +02001857 (conn->flags & CO_FL_WAIT_XPRT) == 0) {
1858 __cs_strm(cs)->conn_exp = TICK_ETERNITY;
1859 oc->flags |= CF_WRITE_NULL;
Willy Tarreau026e8fb2022-05-17 19:47:17 +02001860 if (cs->state == SC_ST_CON)
1861 cs->state = SC_ST_RDY;
Christopher Faulet5e29b762022-04-04 08:58:34 +02001862 }
1863
1864 /* Report EOS on the channel if it was reached from the mux point of
1865 * view.
1866 *
1867 * Note: This test is only required because cs_conn_process is also the SI
1868 * wake callback. Otherwise cs_conn_recv()/cs_conn_send() already take
1869 * care of it.
1870 */
Willy Tarreaub605c422022-05-17 17:04:55 +02001871 if (sc_ep_test(cs, SE_FL_EOS) && !(ic->flags & CF_SHUTR)) {
Christopher Faulet5e29b762022-04-04 08:58:34 +02001872 /* we received a shutdown */
1873 ic->flags |= CF_READ_NULL;
1874 if (ic->flags & CF_AUTO_CLOSE)
1875 channel_shutw_now(ic);
1876 cs_conn_read0(cs);
1877 }
1878
1879 /* Report EOI on the channel if it was reached from the mux point of
1880 * view.
1881 *
1882 * Note: This test is only required because cs_conn_process is also the SI
1883 * wake callback. Otherwise cs_conn_recv()/cs_conn_send() already take
1884 * care of it.
1885 */
Willy Tarreaub605c422022-05-17 17:04:55 +02001886 if (sc_ep_test(cs, SE_FL_EOI) && !(ic->flags & CF_EOI))
Christopher Faulet5e29b762022-04-04 08:58:34 +02001887 ic->flags |= (CF_EOI|CF_READ_PARTIAL);
1888
Willy Tarreau4596fe22022-05-17 19:07:51 +02001889 /* Second step : update the stream connector and channels, try to forward any
Christopher Faulet5e29b762022-04-04 08:58:34 +02001890 * pending data, then possibly wake the stream up based on the new
Willy Tarreau4596fe22022-05-17 19:07:51 +02001891 * stream connector status.
Christopher Faulet5e29b762022-04-04 08:58:34 +02001892 */
1893 cs_notify(cs);
1894 stream_release_buffers(__cs_strm(cs));
1895 return 0;
1896}
1897
Willy Tarreau4596fe22022-05-17 19:07:51 +02001898/* This is the ->process() function for any stream connector's wait_event task.
1899 * It's assigned during the stream connector's initialization, for any type of
1900 * stream connector. Thus it is always safe to perform a tasklet_wakeup() on a
1901 * stream connector, as the presence of the CS is checked there.
Christopher Faulet5e29b762022-04-04 08:58:34 +02001902 */
1903struct task *cs_conn_io_cb(struct task *t, void *ctx, unsigned int state)
1904{
Willy Tarreau4596fe22022-05-17 19:07:51 +02001905 struct stconn *cs = ctx;
Christopher Faulet5e29b762022-04-04 08:58:34 +02001906 int ret = 0;
1907
1908 if (!cs_conn(cs))
1909 return t;
1910
1911 if (!(cs->wait_event.events & SUB_RETRY_SEND) && !channel_is_empty(cs_oc(cs)))
1912 ret = cs_conn_send(cs);
1913 if (!(cs->wait_event.events & SUB_RETRY_RECV))
1914 ret |= cs_conn_recv(cs);
1915 if (ret != 0)
1916 cs_conn_process(cs);
1917
1918 stream_release_buffers(__cs_strm(cs));
1919 return t;
1920}
1921
1922/* Callback to be used by applet handlers upon completion. It updates the stream
1923 * (which may or may not take this opportunity to try to forward data), then
Willy Tarreau4596fe22022-05-17 19:07:51 +02001924 * may re-enable the applet's based on the channels and stream connector's final
Christopher Faulet5e29b762022-04-04 08:58:34 +02001925 * states.
1926 */
Willy Tarreau4596fe22022-05-17 19:07:51 +02001927static int cs_applet_process(struct stconn *cs)
Christopher Faulet5e29b762022-04-04 08:58:34 +02001928{
1929 struct channel *ic = cs_ic(cs);
1930
1931 BUG_ON(!cs_appctx(cs));
1932
1933 /* If the applet wants to write and the channel is closed, it's a
1934 * broken pipe and it must be reported.
1935 */
Willy Tarreaub605c422022-05-17 17:04:55 +02001936 if (!sc_ep_test(cs, SE_FL_RX_WAIT_EP) && (ic->flags & CF_SHUTR))
1937 sc_ep_set(cs, SE_FL_ERROR);
Christopher Faulet5e29b762022-04-04 08:58:34 +02001938
1939 /* automatically mark the applet having data available if it reported
1940 * begin blocked by the channel.
1941 */
1942 if (cs_rx_blocked(cs))
1943 cs_rx_endp_more(cs);
1944
Willy Tarreau4596fe22022-05-17 19:07:51 +02001945 /* update the stream connector, channels, and possibly wake the stream up */
Christopher Faulet5e29b762022-04-04 08:58:34 +02001946 cs_notify(cs);
1947 stream_release_buffers(__cs_strm(cs));
1948
1949 /* cs_notify may have passed through chk_snd and released some
1950 * RXBLK flags. Process_stream will consider those flags to wake up the
1951 * appctx but in the case the task is not in runqueue we may have to
1952 * wakeup the appctx immediately.
1953 */
1954 if ((cs_rx_endp_ready(cs) && !cs_rx_blocked(cs)) ||
1955 (cs_tx_endp_ready(cs) && !cs_tx_blocked(cs)))
1956 appctx_wakeup(__cs_appctx(cs));
1957 return 0;
Christopher Faulet13045f02022-04-01 14:23:38 +02001958}