REORG: fd: uninline fd_update_events()

This function has become a monster (80 lines and 2/3 of a kB), it doesn't
benefit from being static nor inline anymore, let's move it to fd.c.
diff --git a/include/haproxy/fd.h b/include/haproxy/fd.h
index e1a80d7..7d9d0e6 100644
--- a/include/haproxy/fd.h
+++ b/include/haproxy/fd.h
@@ -117,6 +117,7 @@
 void fd_add_to_fd_list(volatile struct fdlist *list, int fd, int off);
 void fd_rm_from_fd_list(volatile struct fdlist *list, int fd, int off);
 void updt_fd_polling(const int fd);
+void fd_update_events(int fd, uint evts);
 
 /* Called from the poller to acknowledge we read an entry from the global
  * update list, to remove our bit from the update_mask, and remove it from
@@ -349,97 +350,6 @@
 	return _HA_ATOMIC_AND_FETCH(&fdtab[fd].running_mask, ~tid_bit);
 }
 
-/* Update events seen for FD <fd> and its state if needed. This should be
- * called by the poller, passing FD_EV_*_{R,W,RW} in <evts>. FD_EV_ERR_*
- * doesn't need to also pass FD_EV_SHUT_*, it's implied. ERR and SHUT are
- * allowed to be reported regardless of R/W readiness.
- */
-static inline void fd_update_events(int fd, uint evts)
-{
-	unsigned long locked;
-	uint old, new;
-	uint new_flags, must_stop;
-
-	ti->flags &= ~TI_FL_STUCK; // this thread is still running
-
-	/* do nothing if the FD was taken over under us */
-	if (fd_set_running(fd) == -1)
-		return;
-
-	locked = (fdtab[fd].thread_mask != tid_bit);
-
-	/* OK now we are guaranteed that our thread_mask was present and
-	 * that we're allowed to update the FD.
-	 */
-
-	new_flags =
-	      ((evts & FD_EV_READY_R) ? FD_POLL_IN  : 0) |
-	      ((evts & FD_EV_READY_W) ? FD_POLL_OUT : 0) |
-	      ((evts & FD_EV_SHUT_R)  ? FD_POLL_HUP : 0) |
-	      ((evts & FD_EV_ERR_RW)  ? FD_POLL_ERR : 0);
-
-	/* SHUTW reported while FD was active for writes is an error */
-	if ((fdtab[fd].state & FD_EV_ACTIVE_W) && (evts & FD_EV_SHUT_W))
-		new_flags |= FD_POLL_ERR;
-
-	/* compute the inactive events reported late that must be stopped */
-	must_stop = 0;
-	if (unlikely(!fd_active(fd))) {
-		/* both sides stopped */
-		must_stop = FD_POLL_IN | FD_POLL_OUT;
-	}
-	else if (unlikely(!fd_recv_active(fd) && (evts & (FD_EV_READY_R | FD_EV_SHUT_R | FD_EV_ERR_RW)))) {
-		/* only send remains */
-		must_stop = FD_POLL_IN;
-	}
-	else if (unlikely(!fd_send_active(fd) && (evts & (FD_EV_READY_W | FD_EV_SHUT_W | FD_EV_ERR_RW)))) {
-		/* only recv remains */
-		must_stop = FD_POLL_OUT;
-	}
-
-	if (new_flags & (FD_POLL_IN | FD_POLL_HUP | FD_POLL_ERR))
-		new_flags |= FD_EV_READY_R;
-
-	if (new_flags & (FD_POLL_OUT | FD_POLL_ERR))
-		new_flags |= FD_EV_READY_W;
-
-	old = fdtab[fd].state;
-	new = (old & ~FD_POLL_UPDT_MASK) | new_flags;
-
-	if (unlikely(locked)) {
-		/* Locked FDs (those with more than 2 threads) are atomically updated */
-		while (unlikely(new != old && !_HA_ATOMIC_CAS(&fdtab[fd].state, &old, new)))
-			new = (old & ~FD_POLL_UPDT_MASK) | new_flags;
-	} else {
-		if (new != old)
-			fdtab[fd].state = new;
-	}
-
-	if (fdtab[fd].iocb && fd_active(fd)) {
-		fdtab[fd].iocb(fd);
-	}
-
-	/* another thread might have attempted to close this FD in the mean
-	 * time (e.g. timeout task) striking on a previous thread and closing.
-	 * This is detected by both thread_mask and running_mask being 0 after
-	 * we remove ourselves last.
-	 */
-	if ((fdtab[fd].running_mask & tid_bit) &&
-	    fd_clr_running(fd) == 0 && !fdtab[fd].thread_mask) {
-		_fd_delete_orphan(fd);
-	}
-
-	/* we had to stop this FD and it still must be stopped after the I/O
-	 * cb's changes, so let's program an update for this.
-	 */
-	if (must_stop && !(fdtab[fd].update_mask & tid_bit)) {
-		if (((must_stop & FD_POLL_IN)  && !fd_recv_active(fd)) ||
-		    ((must_stop & FD_POLL_OUT) && !fd_send_active(fd)))
-			if (!HA_ATOMIC_BTS(&fdtab[fd].update_mask, tid))
-				fd_updt[fd_nbupdt++] = fd;
-	}
-}
-
 /* Prepares <fd> for being polled */
 static inline void fd_insert(int fd, void *owner, void (*iocb)(int fd), unsigned long thread_mask)
 {
diff --git a/src/fd.c b/src/fd.c
index 733658c..febe451 100644
--- a/src/fd.c
+++ b/src/fd.c
@@ -445,6 +445,97 @@
 	}
 }
 
+/* Update events seen for FD <fd> and its state if needed. This should be
+ * called by the poller, passing FD_EV_*_{R,W,RW} in <evts>. FD_EV_ERR_*
+ * doesn't need to also pass FD_EV_SHUT_*, it's implied. ERR and SHUT are
+ * allowed to be reported regardless of R/W readiness.
+ */
+void fd_update_events(int fd, uint evts)
+{
+	unsigned long locked;
+	uint old, new;
+	uint new_flags, must_stop;
+
+	ti->flags &= ~TI_FL_STUCK; // this thread is still running
+
+	/* do nothing if the FD was taken over under us */
+	if (fd_set_running(fd) == -1)
+		return;
+
+	locked = (fdtab[fd].thread_mask != tid_bit);
+
+	/* OK now we are guaranteed that our thread_mask was present and
+	 * that we're allowed to update the FD.
+	 */
+
+	new_flags =
+	      ((evts & FD_EV_READY_R) ? FD_POLL_IN  : 0) |
+	      ((evts & FD_EV_READY_W) ? FD_POLL_OUT : 0) |
+	      ((evts & FD_EV_SHUT_R)  ? FD_POLL_HUP : 0) |
+	      ((evts & FD_EV_ERR_RW)  ? FD_POLL_ERR : 0);
+
+	/* SHUTW reported while FD was active for writes is an error */
+	if ((fdtab[fd].state & FD_EV_ACTIVE_W) && (evts & FD_EV_SHUT_W))
+		new_flags |= FD_POLL_ERR;
+
+	/* compute the inactive events reported late that must be stopped */
+	must_stop = 0;
+	if (unlikely(!fd_active(fd))) {
+		/* both sides stopped */
+		must_stop = FD_POLL_IN | FD_POLL_OUT;
+	}
+	else if (unlikely(!fd_recv_active(fd) && (evts & (FD_EV_READY_R | FD_EV_SHUT_R | FD_EV_ERR_RW)))) {
+		/* only send remains */
+		must_stop = FD_POLL_IN;
+	}
+	else if (unlikely(!fd_send_active(fd) && (evts & (FD_EV_READY_W | FD_EV_SHUT_W | FD_EV_ERR_RW)))) {
+		/* only recv remains */
+		must_stop = FD_POLL_OUT;
+	}
+
+	if (new_flags & (FD_POLL_IN | FD_POLL_HUP | FD_POLL_ERR))
+		new_flags |= FD_EV_READY_R;
+
+	if (new_flags & (FD_POLL_OUT | FD_POLL_ERR))
+		new_flags |= FD_EV_READY_W;
+
+	old = fdtab[fd].state;
+	new = (old & ~FD_POLL_UPDT_MASK) | new_flags;
+
+	if (unlikely(locked)) {
+		/* Locked FDs (those with more than 2 threads) are atomically updated */
+		while (unlikely(new != old && !_HA_ATOMIC_CAS(&fdtab[fd].state, &old, new)))
+			new = (old & ~FD_POLL_UPDT_MASK) | new_flags;
+	} else {
+		if (new != old)
+			fdtab[fd].state = new;
+	}
+
+	if (fdtab[fd].iocb && fd_active(fd)) {
+		fdtab[fd].iocb(fd);
+	}
+
+	/* another thread might have attempted to close this FD in the mean
+	 * time (e.g. timeout task) striking on a previous thread and closing.
+	 * This is detected by both thread_mask and running_mask being 0 after
+	 * we remove ourselves last.
+	 */
+	if ((fdtab[fd].running_mask & tid_bit) &&
+	    fd_clr_running(fd) == 0 && !fdtab[fd].thread_mask) {
+		_fd_delete_orphan(fd);
+	}
+
+	/* we had to stop this FD and it still must be stopped after the I/O
+	 * cb's changes, so let's program an update for this.
+	 */
+	if (must_stop && !(fdtab[fd].update_mask & tid_bit)) {
+		if (((must_stop & FD_POLL_IN)  && !fd_recv_active(fd)) ||
+		    ((must_stop & FD_POLL_OUT) && !fd_send_active(fd)))
+			if (!HA_ATOMIC_BTS(&fdtab[fd].update_mask, tid))
+				fd_updt[fd_nbupdt++] = fd;
+	}
+}
+
 /* Tries to send <npfx> parts from <prefix> followed by <nmsg> parts from <msg>
  * optionally followed by a newline if <nl> is non-null, to file descriptor
  * <fd>. The message is sent atomically using writev(). It may be truncated to