MINOR: pool: introduce pool_item to represent shared pool items

In order to support batch allocation from/to shared pools, we'll have to
support a specific representation for pool objects. The new pool_item
structure will be used for this. For now it only contains a "next"
pointer that matches exactly the current storage model. The few functions
that deal with the shared pool entries were adapted to use the new type.
There is no functionality difference at this point.
diff --git a/include/haproxy/pool-t.h b/include/haproxy/pool-t.h
index c1d2613..29adf65 100644
--- a/include/haproxy/pool-t.h
+++ b/include/haproxy/pool-t.h
@@ -59,6 +59,13 @@
 	struct list by_lru;  /* link to objects by LRU order */
 };
 
+/* This structure is used to represent an element in the pool's shared
+ * free_list.
+ */
+struct pool_item {
+	struct pool_item *next;
+};
+
 /* This describes a complete pool, with its status, usage statistics and the
  * thread-local caches if any. Even if pools are disabled, these descriptors
  * are valid and are used at least to get names and sizes. For small builds
@@ -66,7 +73,7 @@
  * alignment could be removed.
  */
 struct pool_head {
-	void **free_list;
+	struct pool_item *free_list; /* list of free shared objects */
 	unsigned int used;	/* how many chunks are currently in use */
 	unsigned int needed_avg;/* floating indicator between used and allocated */
 	unsigned int allocated;	/* how many chunks have been allocated */
diff --git a/include/haproxy/pool.h b/include/haproxy/pool.h
index ec73cbc..6ee4110 100644
--- a/include/haproxy/pool.h
+++ b/include/haproxy/pool.h
@@ -123,7 +123,7 @@
 	/* ignored without shared pools */
 }
 
-static inline void pool_put_to_shared_cache(struct pool_head *pool, void *item)
+static inline void pool_put_to_shared_cache(struct pool_head *pool, struct pool_item *item)
 {
 	/* ignored without shared pools */
 }
@@ -131,7 +131,7 @@
 #else /* CONFIG_HAP_NO_GLOBAL_POOLS */
 
 void pool_refill_local_from_shared(struct pool_head *pool, struct pool_cache_head *pch);
-void pool_put_to_shared_cache(struct pool_head *pool, void *item);
+void pool_put_to_shared_cache(struct pool_head *pool, struct pool_item *item);
 
 /* returns true if the pool is considered to have too many free objects */
 static inline int pool_is_crowded(const struct pool_head *pool)
diff --git a/src/pool.c b/src/pool.c
index 7b7dd99..7b76802 100644
--- a/src/pool.c
+++ b/src/pool.c
@@ -324,7 +324,7 @@
 		if (unlikely(pool_is_crowded(pool)))
 			pool_free_nocache(pool, item);
 		else
-			pool_put_to_shared_cache(pool, item);
+			pool_put_to_shared_cache(pool, (struct pool_item *)item);
 	}
 }
 
@@ -352,7 +352,7 @@
 		if (unlikely(pool_is_crowded(pool)))
 			pool_free_nocache(pool, item);
 		else
-			pool_put_to_shared_cache(pool, item);
+			pool_put_to_shared_cache(pool, (struct pool_item *)item);
 	} while (pool_cache_bytes > CONFIG_HAP_POOL_CACHE_SIZE * 7 / 8);
 }
 
@@ -404,7 +404,7 @@
 void pool_refill_local_from_shared(struct pool_head *pool, struct pool_cache_head *pch)
 {
 	struct pool_cache_item *item;
-	void *ret;
+	struct pool_item *ret;
 
 	/* we'll need to reference the first element to figure the next one. We
 	 * must temporarily lock it so that nobody allocates then releases it,
@@ -426,14 +426,14 @@
 	}
 
 	/* this releases the lock */
-	HA_ATOMIC_STORE(&pool->free_list, *(void **)ret);
+	HA_ATOMIC_STORE(&pool->free_list, ret->next);
 	HA_ATOMIC_INC(&pool->used);
 
 	/* keep track of where the element was allocated from */
 	POOL_DEBUG_SET_MARK(pool, ret);
 
 	/* now store the retrieved object into the local cache */
-	item = ret;
+	item = (struct pool_cache_item *)ret;
 	LIST_INSERT(&pch->list, &item->by_pool);
 	LIST_INSERT(&th_ctx->pool_lru_head, &item->by_lru);
 	pch->count++;
@@ -446,9 +446,9 @@
  * Both the pool and the item must be valid. Use pool_free() for normal
  * operations.
  */
-void pool_put_to_shared_cache(struct pool_head *pool, void *item)
+void pool_put_to_shared_cache(struct pool_head *pool, struct pool_item *item)
 {
-	void **free_list;
+	struct pool_item *free_list;
 
 	_HA_ATOMIC_DEC(&pool->used);
 	free_list = _HA_ATOMIC_LOAD(&pool->free_list);
@@ -457,7 +457,7 @@
 			__ha_cpu_relax();
 			free_list = _HA_ATOMIC_LOAD(&pool->free_list);
 		}
-		_HA_ATOMIC_STORE((void **)item, (void *)free_list);
+		_HA_ATOMIC_STORE(&item->next, free_list);
 		__ha_barrier_atomic_store();
 	} while (!_HA_ATOMIC_CAS(&pool->free_list, &free_list, item));
 	__ha_barrier_atomic_store();
@@ -469,7 +469,7 @@
  */
 void pool_flush(struct pool_head *pool)
 {
-	void *next, *temp;
+	struct pool_item *next, *temp;
 
 	if (!pool)
 		return;
@@ -491,7 +491,7 @@
 
 	while (next) {
 		temp = next;
-		next = *(void **)temp;
+		next = temp->next;
 		pool_put_to_os(pool, temp);
 	}
 	/* here, we should have pool->allocated == pool->used */
@@ -511,12 +511,12 @@
 		thread_isolate();
 
 	list_for_each_entry(entry, &pools, list) {
-		void *temp;
-		//qfprintf(stderr, "Flushing pool %s\n", entry->name);
+		struct pool_item *temp;
+
 		while (entry->free_list &&
 		       (int)(entry->allocated - entry->used) > (int)entry->minavail) {
 			temp = entry->free_list;
-			entry->free_list = *(void **)temp;
+			entry->free_list = temp->next;
 			pool_put_to_os(entry, temp);
 		}
 	}