blob: 5848bfd2a5b887a6e89be6027d953217ba54806b [file] [log] [blame]
/*
* include/haproxy/task-t.h
* Macros, variables and structures for task management.
*
* Copyright (C) 2000-2010 Willy Tarreau - w@1wt.eu
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation, version 2.1
* exclusively.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with this library; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#ifndef _HAPROXY_TASK_T_H
#define _HAPROXY_TASK_T_H
#include <sys/time.h>
#include <import/eb32sctree.h>
#include <import/eb32tree.h>
#include <haproxy/api-t.h>
#include <haproxy/thread-t.h>
/* values for task->state (32 bits) */
#define TASK_SLEEPING 0x00000000 /* task sleeping */
#define TASK_RUNNING 0x00000001 /* the task is currently running */
#define TASK_GLOBAL 0x00000002 /* The task is currently in the global runqueue */
#define TASK_QUEUED 0x00000004 /* The task has been (re-)added to the run queue */
#define TASK_SHARED_WQ 0x00000008 /* The task's expiration may be updated by other
* threads, must be set before first queue/wakeup */
#define TASK_SELF_WAKING 0x00000010 /* task/tasklet found waking itself */
#define TASK_KILLED 0x00000020 /* task/tasklet killed, may now be freed */
#define TASK_IN_LIST 0x00000040 /* tasklet is in a tasklet list */
#define TASK_HEAVY 0x00000080 /* this task/tasklet is extremely heavy */
#define TASK_WOKEN_INIT 0x00000100 /* woken up for initialisation purposes */
#define TASK_WOKEN_TIMER 0x00000200 /* woken up because of expired timer */
#define TASK_WOKEN_IO 0x00000400 /* woken up because of completed I/O */
#define TASK_WOKEN_SIGNAL 0x00000800 /* woken up by a system signal */
#define TASK_WOKEN_MSG 0x00001000 /* woken up by another task's message */
#define TASK_WOKEN_RES 0x00002000 /* woken up because of available resource */
#define TASK_WOKEN_OTHER 0x00004000 /* woken up for an unspecified reason */
/* use this to check a task state or to clean it up before queueing */
#define TASK_WOKEN_ANY (TASK_WOKEN_OTHER|TASK_WOKEN_INIT|TASK_WOKEN_TIMER| \
TASK_WOKEN_IO|TASK_WOKEN_SIGNAL|TASK_WOKEN_MSG| \
TASK_WOKEN_RES)
#define TASK_F_TASKLET 0x00008000 /* nature of this task: 0=task 1=tasklet */
#define TASK_F_USR1 0x00010000 /* preserved user flag 1, application-specific, def:0 */
/* unused: 0x20000..0x80000000 */
/* These flags are persistent across scheduler calls */
#define TASK_PERSISTENT (TASK_SHARED_WQ | TASK_SELF_WAKING | TASK_KILLED | \
TASK_HEAVY | TASK_F_TASKLET | TASK_F_USR1)
enum {
TL_URGENT = 0, /* urgent tasklets (I/O callbacks) */
TL_NORMAL = 1, /* normal tasks */
TL_BULK = 2, /* bulk task/tasklets, streaming I/Os */
TL_HEAVY = 3, /* heavy computational tasklets (e.g. TLS handshakes) */
TL_CLASSES /* must be last */
};
struct notification {
struct list purge_me; /* Part of the list of signals to be purged in the
case of the LUA execution stack crash. */
struct list wake_me; /* Part of list of signals to be targeted if an
event occurs. */
struct task *task; /* The task to be wake if an event occurs. */
__decl_thread(HA_SPINLOCK_T lock);
};
/* force to split per-thread stuff into separate cache lines */
struct task_per_thread {
// first and second cache lines on 64 bits: thread-local operations only.
struct eb_root timers; /* tree constituting the per-thread wait queue */
struct eb_root rqueue; /* tree constituting the per-thread run queue */
struct task *current; /* current task (not tasklet) */
unsigned int rqueue_ticks; /* Insertion counter for the run queue */
int current_queue; /* points to current tasklet list being run, -1 if none */
unsigned int nb_tasks; /* number of tasks allocated on this thread */
uint8_t tl_class_mask; /* bit mask of non-empty tasklets classes */
// 11 bytes hole here
ALWAYS_ALIGN(2*sizeof(void*));
struct list tasklets[TL_CLASSES]; /* tasklets (and/or tasks) to run, by class */
// third cache line here on 64 bits: accessed mostly using atomic ops
ALWAYS_ALIGN(64);
struct mt_list shared_tasklet_list; /* Tasklet to be run, woken up by other threads */
unsigned int rq_total; /* total size of the run queue, prio_tree + tasklets */
int tasks_in_list; /* Number of tasks in the per-thread tasklets list */
ALWAYS_ALIGN(128);
};
#ifdef DEBUG_TASK
#define TASK_DEBUG_STORAGE \
struct { \
const char *caller_file[2]; \
int caller_line[2]; \
int caller_idx; \
} debug
#else
#define TASK_DEBUG_STORAGE
#endif
/* This part is common between struct task and struct tasklet so that tasks
* can be used as-is as tasklets.
*
* Note that the process() function must ALWAYS return the task/tasklet's
* pointer if the task/tasklet remains valid, and return NULL if it has been
* deleted. The scheduler relies on this to know if it should update its state
* on return.
*/
#define TASK_COMMON \
struct { \
unsigned int state; /* task state : bitfield of TASK_ */ \
/* 16-bit hole here */ \
unsigned int calls; /* number of times process was called */ \
struct task *(*process)(struct task *t, void *ctx, unsigned int state); /* the function which processes the task */ \
void *context; /* the task's context */ \
TASK_DEBUG_STORAGE; \
}
/* The base for all tasks */
struct task {
TASK_COMMON; /* must be at the beginning! */
struct eb32sc_node rq; /* ebtree node used to hold the task in the run queue */
/* WARNING: the struct task is often aliased as a struct tasklet when
* it is NOT in the run queue. The tasklet has its struct list here
* where rq starts and this works because both are exclusive. Never
* ever reorder these fields without taking this into account!
*/
struct eb32_node wq; /* ebtree node used to hold the task in the wait queue */
int expire; /* next expiration date for this task, in ticks */
short nice; /* task prio from -1024 to +1024 */
/* 16-bit hole here */
unsigned long thread_mask; /* mask of thread IDs authorized to process the task */
uint64_t call_date; /* date of the last task wakeup or call */
uint64_t lat_time; /* total latency time experienced */
uint64_t cpu_time; /* total CPU time consumed */
};
/* lightweight tasks, without priority, mainly used for I/Os */
struct tasklet {
TASK_COMMON; /* must be at the beginning! */
struct list list;
/* WARNING: the struct task is often aliased as a struct tasklet when
* it is not in the run queue. The task has its struct rq here where
* list starts and this works because both are exclusive. Never ever
* reorder these fields without taking this into account!
*/
#ifdef DEBUG_TASK
uint64_t call_date; /* date of the last tasklet wakeup or call */
#endif
int tid; /* TID of the tasklet owner, <0 if local */
};
/*
* The task callback (->process) is responsible for updating ->expire. It must
* return a pointer to the task itself, except if the task has been deleted, in
* which case it returns NULL so that the scheduler knows it must not check the
* expire timer. The scheduler will requeue the task at the proper location.
*/
/* A work_list is a thread-safe way to enqueue some work to be run on another
* thread. It consists of a list, a task and a general-purpose argument.
* A work is appended to the list by atomically adding a list element to the
* list and waking up the associated task, which is done using work_add(). The
* caller must be careful about how operations are run as it will definitely
* happen that the element being enqueued is processed by the other thread
* before the call returns. Some locking conventions between the caller and the
* callee might sometimes be necessary. The task is always woken up with reason
* TASK_WOKEN_OTHER and a context pointing to the work_list entry.
*/
struct work_list {
struct mt_list head;
struct task *task;
void *arg;
};
#endif /* _HAPROXY_TASK_T_H */
/*
* Local variables:
* c-indent-level: 8
* c-basic-offset: 8
* End:
*/