sched: s/sched_latency/sched_min_granularity
[usb.git] / include / linux / workqueue.h
blobce6badc98f6dcb3f2b78615cd4aa382d6d956abd
1 /*
2 * workqueue.h --- work queue handling for Linux.
3 */
5 #ifndef _LINUX_WORKQUEUE_H
6 #define _LINUX_WORKQUEUE_H
8 #include <linux/timer.h>
9 #include <linux/linkage.h>
10 #include <linux/bitops.h>
11 #include <asm/atomic.h>
13 struct workqueue_struct;
15 struct work_struct;
16 typedef void (*work_func_t)(struct work_struct *work);
19 * The first word is the work queue pointer and the flags rolled into
20 * one
22 #define work_data_bits(work) ((unsigned long *)(&(work)->data))
24 struct work_struct {
25 atomic_long_t data;
26 #define WORK_STRUCT_PENDING 0 /* T if work item pending execution */
27 #define WORK_STRUCT_FLAG_MASK (3UL)
28 #define WORK_STRUCT_WQ_DATA_MASK (~WORK_STRUCT_FLAG_MASK)
29 struct list_head entry;
30 work_func_t func;
33 #define WORK_DATA_INIT() ATOMIC_LONG_INIT(0)
35 struct delayed_work {
36 struct work_struct work;
37 struct timer_list timer;
40 struct execute_work {
41 struct work_struct work;
44 #define __WORK_INITIALIZER(n, f) { \
45 .data = WORK_DATA_INIT(), \
46 .entry = { &(n).entry, &(n).entry }, \
47 .func = (f), \
50 #define __DELAYED_WORK_INITIALIZER(n, f) { \
51 .work = __WORK_INITIALIZER((n).work, (f)), \
52 .timer = TIMER_INITIALIZER(NULL, 0, 0), \
55 #define DECLARE_WORK(n, f) \
56 struct work_struct n = __WORK_INITIALIZER(n, f)
58 #define DECLARE_DELAYED_WORK(n, f) \
59 struct delayed_work n = __DELAYED_WORK_INITIALIZER(n, f)
62 * initialize a work item's function pointer
64 #define PREPARE_WORK(_work, _func) \
65 do { \
66 (_work)->func = (_func); \
67 } while (0)
69 #define PREPARE_DELAYED_WORK(_work, _func) \
70 PREPARE_WORK(&(_work)->work, (_func))
73 * initialize all of a work item in one go
75 * NOTE! No point in using "atomic_long_set()": useing a direct
76 * assignment of the work data initializer allows the compiler
77 * to generate better code.
79 #define INIT_WORK(_work, _func) \
80 do { \
81 (_work)->data = (atomic_long_t) WORK_DATA_INIT(); \
82 INIT_LIST_HEAD(&(_work)->entry); \
83 PREPARE_WORK((_work), (_func)); \
84 } while (0)
86 #define INIT_DELAYED_WORK(_work, _func) \
87 do { \
88 INIT_WORK(&(_work)->work, (_func)); \
89 init_timer(&(_work)->timer); \
90 } while (0)
92 #define INIT_DELAYED_WORK_DEFERRABLE(_work, _func) \
93 do { \
94 INIT_WORK(&(_work)->work, (_func)); \
95 init_timer_deferrable(&(_work)->timer); \
96 } while (0)
98 /**
99 * work_pending - Find out whether a work item is currently pending
100 * @work: The work item in question
102 #define work_pending(work) \
103 test_bit(WORK_STRUCT_PENDING, work_data_bits(work))
106 * delayed_work_pending - Find out whether a delayable work item is currently
107 * pending
108 * @work: The work item in question
110 #define delayed_work_pending(w) \
111 work_pending(&(w)->work)
114 * work_clear_pending - for internal use only, mark a work item as not pending
115 * @work: The work item in question
117 #define work_clear_pending(work) \
118 clear_bit(WORK_STRUCT_PENDING, work_data_bits(work))
121 extern struct workqueue_struct *__create_workqueue(const char *name,
122 int singlethread,
123 int freezeable);
124 #define create_workqueue(name) __create_workqueue((name), 0, 0)
125 #define create_freezeable_workqueue(name) __create_workqueue((name), 1, 1)
126 #define create_singlethread_workqueue(name) __create_workqueue((name), 1, 0)
128 extern void destroy_workqueue(struct workqueue_struct *wq);
130 extern int FASTCALL(queue_work(struct workqueue_struct *wq, struct work_struct *work));
131 extern int FASTCALL(queue_delayed_work(struct workqueue_struct *wq,
132 struct delayed_work *work, unsigned long delay));
133 extern int queue_delayed_work_on(int cpu, struct workqueue_struct *wq,
134 struct delayed_work *work, unsigned long delay);
136 extern void FASTCALL(flush_workqueue(struct workqueue_struct *wq));
137 extern void flush_scheduled_work(void);
139 extern int FASTCALL(schedule_work(struct work_struct *work));
140 extern int FASTCALL(schedule_delayed_work(struct delayed_work *work,
141 unsigned long delay));
142 extern int schedule_delayed_work_on(int cpu, struct delayed_work *work,
143 unsigned long delay);
144 extern int schedule_on_each_cpu(work_func_t func);
145 extern int current_is_keventd(void);
146 extern int keventd_up(void);
148 extern void init_workqueues(void);
149 int execute_in_process_context(work_func_t fn, struct execute_work *);
151 extern int cancel_work_sync(struct work_struct *work);
154 * Kill off a pending schedule_delayed_work(). Note that the work callback
155 * function may still be running on return from cancel_delayed_work(), unless
156 * it returns 1 and the work doesn't re-arm itself. Run flush_workqueue() or
157 * cancel_work_sync() to wait on it.
159 static inline int cancel_delayed_work(struct delayed_work *work)
161 int ret;
163 ret = del_timer_sync(&work->timer);
164 if (ret)
165 work_clear_pending(&work->work);
166 return ret;
169 extern int cancel_delayed_work_sync(struct delayed_work *work);
171 /* Obsolete. use cancel_delayed_work_sync() */
172 static inline
173 void cancel_rearming_delayed_workqueue(struct workqueue_struct *wq,
174 struct delayed_work *work)
176 cancel_delayed_work_sync(work);
179 /* Obsolete. use cancel_delayed_work_sync() */
180 static inline
181 void cancel_rearming_delayed_work(struct delayed_work *work)
183 cancel_delayed_work_sync(work);
186 #endif