MINOR: fd: Implement fd_takeover().

Implement a new function, fd_takeover(), that lets you become the thread
responsible for the fd. On architectures that do not have a double-width CAS,
use a global rwlock.
fd_set_running() was also changed to be able to compete with fd_takeover(),
either using a dooble-width CAS on both running_mask and thread_mask, or
by claiming a reader on the global rwlock. This extra operation should not
have any measurable impact on modern architectures where threading is
relevant.
diff --git a/include/proto/fd.h b/include/proto/fd.h
index 5ab05d9..8fa7670 100644
--- a/include/proto/fd.h
+++ b/include/proto/fd.h
@@ -60,6 +60,16 @@
  */
 void fd_remove(int fd);
 
+/*
+ * Take over a FD belonging to another thread.
+ * Returns 0 on success, and -1 on failure.
+ */
+int fd_takeover(int fd, void *expected_owner);
+
+#ifndef HA_HAVE_CAS_DW
+__decl_hathreads(HA_RWLOCK_T fd_mig_lock);
+#endif
+
 ssize_t fd_write_frag_line(int fd, size_t maxlen, const struct ist pfx[], size_t npfx, const struct ist msg[], size_t nmsg, int nl);
 
 /* close all FDs starting from <start> */
@@ -299,9 +309,35 @@
 	updt_fd_polling(fd);
 }
 
-static inline void fd_set_running(int fd)
+/* Set the fd as currently running on the current thread.
+ * Retuns 0 if all goes well, or -1 if we no longer own the fd, and should
+ * do nothing with it.
+ */
+static inline int fd_set_running(int fd)
 {
+#ifndef HA_HAVE_CAS_DW
+	HA_RWLOCK_RDLOCK(OTHER_LOCK, &fd_mig_lock);
+	if (!(fdtab[fd].thread_mask & tid_bit)) {
+		HA_RWLOCK_RDUNLOCK(OTHER_LOCK, &fd_mig_lock);
+		return -1;
+	}
 	_HA_ATOMIC_OR(&fdtab[fd].running_mask, tid_bit);
+	HA_RWLOCK_RDUNLOCK(OTHER_LOCK, &fd_mig_lock);
+	return 0;
+#else
+	unsigned long old_masks[2];
+	unsigned long new_masks[2];
+	old_masks[0] = fdtab[fd].running_mask;
+	old_masks[1] = fdtab[fd].thread_mask;
+	do {
+		if (!(old_masks[1] & tid_bit))
+			return -1;
+		new_masks[0] = fdtab[fd].running_mask | tid_bit;
+		new_masks[1] = old_masks[1];
+
+	} while (!(HA_ATOMIC_DWCAS(&fdtab[fd].running_mask, &old_masks, &new_masks)));
+	return 0;
+#endif
 }
 
 static inline void fd_set_running_excl(int fd)
@@ -371,7 +407,8 @@
 		fd_may_send(fd);
 
 	if (fdtab[fd].iocb && fd_active(fd)) {
-		fd_set_running(fd);
+		if (fd_set_running(fd) == -1)
+			return;
 		fdtab[fd].iocb(fd);
 		fd_clr_running(fd);
 	}
diff --git a/src/fd.c b/src/fd.c
index ecc9905..e790f3e 100644
--- a/src/fd.c
+++ b/src/fd.c
@@ -302,6 +302,13 @@
 {
 	int locked = fdtab[fd].running_mask != tid_bit;
 
+	/* We're just trying to protect against a concurrent fd_insert()
+	 * here, not against fd_takeother(), because either we're called
+	 * directly from the iocb(), and we're already locked, or we're
+	 * called from the mux tasklet, but then the mux is responsible for
+	 * making sure the tasklet does nothing, and the connection is never
+	 * destroyed.
+	 */
 	if (locked)
 		fd_set_running_excl(fd);
 
@@ -328,6 +335,66 @@
 		fd_clr_running(fd);
 }
 
+#ifndef HA_HAVE_CAS_DW
+__decl_hathreads(__delc_rwlock(fd_mig_lock));
+#endif
+
+/*
+ * Take over a FD belonging to another thread.
+ * unexpected_conn is the expected owner of the fd.
+ * Returns 0 on success, and -1 on failure.
+ */
+int fd_takeover(int fd, void *expected_owner)
+{
+#ifndef HA_HAVE_CAS_DW
+	int ret;
+
+	HA_RWLOCK_WRLOCK(OTHER_LOCK, &fd_mig_lock);
+	_HA_ATOMIC_OR(&fdtab[fd].running_mask, tid_bit);
+	if (fdtab[fd].running_mask != tid_bit || fdtab[fd].owner != expected_owner) {
+		ret = -1;
+		_HA_ATOMIC_AND(&fdtab[fd].running_mask, ~tid_bit);
+		goto end;
+	}
+	fdtab[fd].thread_mask = tid_bit;
+	_HA_ATOMIC_AND(&fdtab[fd].running_mask, ~tid_bit);
+	ret = 0;
+end:
+	HA_RWLOCK_WRUNLOCK(OTHER_LOCK, &fd_mig_lock);
+	/* Make sure the FD doesn't have the active bit. It is possible that
+	 * the fd is polled by the thread that used to own it, the new thread
+	 * is supposed to call subscribe() later, to activate polling.
+	 */
+	fd_stop_recv(fd);
+	return ret;
+#else
+	unsigned long old_masks[2];
+	unsigned long new_masks[2];
+
+	old_masks[0] = tid_bit;
+	old_masks[1] = fdtab[fd].thread_mask;
+	new_masks[0] = new_masks[1] = tid_bit;
+	/* protect ourself against a delete then an insert for the same fd,
+	 * if it happens, then the owner will no longer be the expected
+	 * connection.
+	 */
+	_HA_ATOMIC_OR(&fdtab[fd].running_mask, tid_bit);
+	if (fdtab[fd].owner != expected_owner) {
+		_HA_ATOMIC_AND(&fdtab[fd].running_mask, ~tid_bit);
+		return -1;
+	}
+	do {
+		if (old_masks[0] != tid_bit || !old_masks[1]) {
+			_HA_ATOMIC_AND(&fdtab[fd].running_mask, ~tid_bit);
+			return -1;
+		}
+	} while (!(_HA_ATOMIC_DWCAS(&fdtab[fd].running_mask, &old_masks,
+		   &new_masks)));
+	_HA_ATOMIC_AND(&fdtab[fd].running_mask, ~tid_bit);
+	return 0;
+#endif /* HW_HAVE_CAS_DW */
+}
+
 /* Deletes an FD from the fdsets.
  * The file descriptor is also closed.
  */