REORG: server: uninline the idle conns management functions

The following functions are quite heavy and have no reason to be kept
inlined:

   srv_release_conn, srv_lookup_conn, srv_lookup_conn_next,
   srv_add_to_idle_list

They were moved to server.c. It's worth noting that they're a bit
at the edge between server and connection and that maybe we could
create an idle-conn file for these in the near future.
diff --git a/include/haproxy/server.h b/include/haproxy/server.h
index 0598f61..40e7863 100644
--- a/include/haproxy/server.h
+++ b/include/haproxy/server.h
@@ -74,6 +74,11 @@
 int snr_resolution_error_cb(struct resolv_requester *requester, int error_code);
 struct server *snr_check_ip_callback(struct server *srv, void *ip, unsigned char *ip_family);
 struct task *srv_cleanup_idle_conns(struct task *task, void *ctx, unsigned int state);
+void srv_release_conn(struct server *srv, struct connection *conn);
+struct connection *srv_lookup_conn(struct eb_root *tree, uint64_t hash);
+struct connection *srv_lookup_conn_next(struct connection *conn);
+
+int srv_add_to_idle_list(struct server *srv, struct connection *conn, int is_safe);
 struct task *srv_cleanup_toremove_conns(struct task *task, void *context, unsigned int state);
 
 int srv_apply_track(struct server *srv, struct proxy *curproxy);
@@ -268,134 +273,6 @@
 		srv->est_need_conns = curr;
 }
 
-
-/* removes an idle conn after updating the server idle conns counters */
-static inline void srv_release_conn(struct server *srv, struct connection *conn)
-{
-	if (conn->flags & CO_FL_LIST_MASK) {
-		/* The connection is currently in the server's idle list, so tell it
-		 * there's one less connection available in that list.
-		 */
-		_HA_ATOMIC_DEC(&srv->curr_idle_conns);
-		_HA_ATOMIC_DEC(conn->flags & CO_FL_SAFE_LIST ? &srv->curr_safe_nb : &srv->curr_idle_nb);
-		_HA_ATOMIC_DEC(&srv->curr_idle_thr[tid]);
-	}
-	else {
-		/* The connection is not private and not in any server's idle
-		 * list, so decrement the current number of used connections
-		 */
-		_HA_ATOMIC_DEC(&srv->curr_used_conns);
-	}
-
-	/* Remove the connection from any tree (safe, idle or available) */
-	HA_SPIN_LOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
-	conn_delete_from_tree(&conn->hash_node->node);
-	HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
-}
-
-/* This adds an idle connection to the server's list if the connection is
- * reusable, not held by any owner anymore, but still has available streams.
- */
-static inline int srv_add_to_idle_list(struct server *srv, struct connection *conn, int is_safe)
-{
-	/* we try to keep the connection in the server's idle list
-	 * if we don't have too many FD in use, and if the number of
-	 * idle+current conns is lower than what was observed before
-	 * last purge, or if we already don't have idle conns for the
-	 * current thread and we don't exceed last count by global.nbthread.
-	 */
-	if (!(conn->flags & CO_FL_PRIVATE) &&
-	    srv && srv->pool_purge_delay > 0 &&
-	    ((srv->proxy->options & PR_O_REUSE_MASK) != PR_O_REUSE_NEVR) &&
-	    ha_used_fds < global.tune.pool_high_count &&
-	    (srv->max_idle_conns == -1 || srv->max_idle_conns > srv->curr_idle_conns) &&
-	    ((eb_is_empty(&srv->per_thr[tid].safe_conns) &&
-	      (is_safe || eb_is_empty(&srv->per_thr[tid].idle_conns))) ||
-	     (ha_used_fds < global.tune.pool_low_count &&
-	      (srv->curr_used_conns + srv->curr_idle_conns <=
-	       MAX(srv->curr_used_conns, srv->est_need_conns) + srv->low_idle_conns))) &&
-	    !conn->mux->used_streams(conn) && conn->mux->avail_streams(conn)) {
-		int retadd;
-
-		retadd = _HA_ATOMIC_ADD_FETCH(&srv->curr_idle_conns, 1);
-		if (retadd > srv->max_idle_conns) {
-			_HA_ATOMIC_DEC(&srv->curr_idle_conns);
-			return 0;
-		}
-		_HA_ATOMIC_DEC(&srv->curr_used_conns);
-
-		HA_SPIN_LOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
-		conn_delete_from_tree(&conn->hash_node->node);
-
-		if (is_safe) {
-			conn->flags = (conn->flags & ~CO_FL_LIST_MASK) | CO_FL_SAFE_LIST;
-			ebmb_insert(&srv->per_thr[tid].safe_conns, &conn->hash_node->node, sizeof(conn->hash_node->hash));
-			_HA_ATOMIC_INC(&srv->curr_safe_nb);
-		} else {
-			conn->flags = (conn->flags & ~CO_FL_LIST_MASK) | CO_FL_IDLE_LIST;
-			ebmb_insert(&srv->per_thr[tid].idle_conns, &conn->hash_node->node, sizeof(conn->hash_node->hash));
-			_HA_ATOMIC_INC(&srv->curr_idle_nb);
-		}
-		HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
-		_HA_ATOMIC_INC(&srv->curr_idle_thr[tid]);
-
-		__ha_barrier_full();
-		if ((volatile void *)srv->idle_node.node.leaf_p == NULL) {
-			HA_SPIN_LOCK(OTHER_LOCK, &idle_conn_srv_lock);
-			if ((volatile void *)srv->idle_node.node.leaf_p == NULL) {
-				srv->idle_node.key = tick_add(srv->pool_purge_delay,
-				                              now_ms);
-				eb32_insert(&idle_conn_srv, &srv->idle_node);
-				if (!task_in_wq(idle_conn_task) && !
-				    task_in_rq(idle_conn_task)) {
-					task_schedule(idle_conn_task,
-					              srv->idle_node.key);
-				}
-
-			}
-			HA_SPIN_UNLOCK(OTHER_LOCK, &idle_conn_srv_lock);
-		}
-		return 1;
-	}
-	return 0;
-}
-
-/* retrieve a connection from its <hash> in <tree>
- * returns NULL if no connection found
- */
-static inline struct connection *srv_lookup_conn(struct eb_root *tree, uint64_t hash)
-{
-	struct ebmb_node *node = NULL;
-	struct connection *conn = NULL;
-	struct conn_hash_node *hash_node = NULL;
-
-	node = ebmb_lookup(tree, &hash, sizeof(hash_node->hash));
-	if (node) {
-		hash_node = ebmb_entry(node, struct conn_hash_node, node);
-		conn = hash_node->conn;
-	}
-
-	return conn;
-}
-
-/* retrieve the next connection sharing the same hash as <conn>
- * returns NULL if no connection found
- */
-static inline struct connection *srv_lookup_conn_next(struct connection *conn)
-{
-	struct ebmb_node *node = NULL;
-	struct connection *next_conn = NULL;
-	struct conn_hash_node *hash_node = NULL;
-
-	node = ebmb_next_dup(&conn->hash_node->node);
-	if (node) {
-		hash_node = ebmb_entry(node, struct conn_hash_node, node);
-		next_conn = hash_node->conn;
-	}
-
-	return next_conn;
-}
-
 #endif /* _HAPROXY_SERVER_H */
 
 /*
diff --git a/src/server.c b/src/server.c
index 73454bc..49cd7e2 100644
--- a/src/server.c
+++ b/src/server.c
@@ -16,6 +16,8 @@
 #include <ctype.h>
 #include <errno.h>
 
+#include <import/ebmbtree.h>
+
 #include <haproxy/api.h>
 #include <haproxy/applet-t.h>
 #include <haproxy/backend.h>
@@ -5618,6 +5620,133 @@
 		if ((i = ((i + 1 == global.nbthread) ? 0 : i + 1)) == tid)
 			break;
 	}
+}
+
+/* removes an idle conn after updating the server idle conns counters */
+void srv_release_conn(struct server *srv, struct connection *conn)
+{
+	if (conn->flags & CO_FL_LIST_MASK) {
+		/* The connection is currently in the server's idle list, so tell it
+		 * there's one less connection available in that list.
+		 */
+		_HA_ATOMIC_DEC(&srv->curr_idle_conns);
+		_HA_ATOMIC_DEC(conn->flags & CO_FL_SAFE_LIST ? &srv->curr_safe_nb : &srv->curr_idle_nb);
+		_HA_ATOMIC_DEC(&srv->curr_idle_thr[tid]);
+	}
+	else {
+		/* The connection is not private and not in any server's idle
+		 * list, so decrement the current number of used connections
+		 */
+		_HA_ATOMIC_DEC(&srv->curr_used_conns);
+	}
+
+	/* Remove the connection from any tree (safe, idle or available) */
+	HA_SPIN_LOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
+	conn_delete_from_tree(&conn->hash_node->node);
+	HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
+}
+
+/* retrieve a connection from its <hash> in <tree>
+ * returns NULL if no connection found
+ */
+struct connection *srv_lookup_conn(struct eb_root *tree, uint64_t hash)
+{
+	struct ebmb_node *node = NULL;
+	struct connection *conn = NULL;
+	struct conn_hash_node *hash_node = NULL;
+
+	node = ebmb_lookup(tree, &hash, sizeof(hash_node->hash));
+	if (node) {
+		hash_node = ebmb_entry(node, struct conn_hash_node, node);
+		conn = hash_node->conn;
+	}
+
+	return conn;
+}
+
+/* retrieve the next connection sharing the same hash as <conn>
+ * returns NULL if no connection found
+ */
+struct connection *srv_lookup_conn_next(struct connection *conn)
+{
+	struct ebmb_node *node = NULL;
+	struct connection *next_conn = NULL;
+	struct conn_hash_node *hash_node = NULL;
+
+	node = ebmb_next_dup(&conn->hash_node->node);
+	if (node) {
+		hash_node = ebmb_entry(node, struct conn_hash_node, node);
+		next_conn = hash_node->conn;
+	}
+
+	return next_conn;
+}
+
+/* This adds an idle connection to the server's list if the connection is
+ * reusable, not held by any owner anymore, but still has available streams.
+ */
+int srv_add_to_idle_list(struct server *srv, struct connection *conn, int is_safe)
+{
+	/* we try to keep the connection in the server's idle list
+	 * if we don't have too many FD in use, and if the number of
+	 * idle+current conns is lower than what was observed before
+	 * last purge, or if we already don't have idle conns for the
+	 * current thread and we don't exceed last count by global.nbthread.
+	 */
+	if (!(conn->flags & CO_FL_PRIVATE) &&
+	    srv && srv->pool_purge_delay > 0 &&
+	    ((srv->proxy->options & PR_O_REUSE_MASK) != PR_O_REUSE_NEVR) &&
+	    ha_used_fds < global.tune.pool_high_count &&
+	    (srv->max_idle_conns == -1 || srv->max_idle_conns > srv->curr_idle_conns) &&
+	    ((eb_is_empty(&srv->per_thr[tid].safe_conns) &&
+	      (is_safe || eb_is_empty(&srv->per_thr[tid].idle_conns))) ||
+	     (ha_used_fds < global.tune.pool_low_count &&
+	      (srv->curr_used_conns + srv->curr_idle_conns <=
+	       MAX(srv->curr_used_conns, srv->est_need_conns) + srv->low_idle_conns))) &&
+	    !conn->mux->used_streams(conn) && conn->mux->avail_streams(conn)) {
+		int retadd;
+
+		retadd = _HA_ATOMIC_ADD_FETCH(&srv->curr_idle_conns, 1);
+		if (retadd > srv->max_idle_conns) {
+			_HA_ATOMIC_DEC(&srv->curr_idle_conns);
+			return 0;
+		}
+		_HA_ATOMIC_DEC(&srv->curr_used_conns);
+
+		HA_SPIN_LOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
+		conn_delete_from_tree(&conn->hash_node->node);
+
+		if (is_safe) {
+			conn->flags = (conn->flags & ~CO_FL_LIST_MASK) | CO_FL_SAFE_LIST;
+			ebmb_insert(&srv->per_thr[tid].safe_conns, &conn->hash_node->node, sizeof(conn->hash_node->hash));
+			_HA_ATOMIC_INC(&srv->curr_safe_nb);
+		} else {
+			conn->flags = (conn->flags & ~CO_FL_LIST_MASK) | CO_FL_IDLE_LIST;
+			ebmb_insert(&srv->per_thr[tid].idle_conns, &conn->hash_node->node, sizeof(conn->hash_node->hash));
+			_HA_ATOMIC_INC(&srv->curr_idle_nb);
+		}
+		HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
+		_HA_ATOMIC_INC(&srv->curr_idle_thr[tid]);
+
+		__ha_barrier_full();
+		if ((volatile void *)srv->idle_node.node.leaf_p == NULL) {
+			HA_SPIN_LOCK(OTHER_LOCK, &idle_conn_srv_lock);
+			if ((volatile void *)srv->idle_node.node.leaf_p == NULL) {
+				srv->idle_node.key = tick_add(srv->pool_purge_delay,
+				                              now_ms);
+				eb32_insert(&idle_conn_srv, &srv->idle_node);
+				if (!task_in_wq(idle_conn_task) && !
+				    task_in_rq(idle_conn_task)) {
+					task_schedule(idle_conn_task,
+					              srv->idle_node.key);
+				}
+
+			}
+			HA_SPIN_UNLOCK(OTHER_LOCK, &idle_conn_srv_lock);
+		}
+		return 1;
+	}
+	return 0;
 }
 
 struct task *srv_cleanup_idle_conns(struct task *task, void *context, unsigned int state)