workqueue.h 18.5 KB
Newer Older
L
Linus Torvalds 已提交
1 2 3 4 5 6 7 8 9 10
/*
 * workqueue.h --- work queue handling for Linux.
 */

#ifndef _LINUX_WORKQUEUE_H
#define _LINUX_WORKQUEUE_H

#include <linux/timer.h>
#include <linux/linkage.h>
#include <linux/bitops.h>
11
#include <linux/lockdep.h>
12
#include <linux/threads.h>
A
Arun Sharma 已提交
13
#include <linux/atomic.h>
T
Tejun Heo 已提交
14
#include <linux/cpumask.h>
L
Linus Torvalds 已提交
15 16 17

struct workqueue_struct;

18 19
struct work_struct;
typedef void (*work_func_t)(struct work_struct *work);
20
void delayed_work_timer_fn(unsigned long __data);
21

22 23 24 25 26 27
/*
 * The first word is the work queue pointer and the flags rolled into
 * one
 */
#define work_data_bits(work) ((unsigned long *)(&(work)->data))

28 29
enum {
	WORK_STRUCT_PENDING_BIT	= 0,	/* work item is pending execution */
30
	WORK_STRUCT_DELAYED_BIT	= 1,	/* work item is delayed */
31
	WORK_STRUCT_PWQ_BIT	= 2,	/* data points to pwq */
32
	WORK_STRUCT_LINKED_BIT	= 3,	/* next work is linked to this one */
33
#ifdef CONFIG_DEBUG_OBJECTS_WORK
34 35
	WORK_STRUCT_STATIC_BIT	= 4,	/* static initializer (debugobjects) */
	WORK_STRUCT_COLOR_SHIFT	= 5,	/* color for workqueue flushing */
T
Tejun Heo 已提交
36
#else
37
	WORK_STRUCT_COLOR_SHIFT	= 4,	/* color for workqueue flushing */
38 39
#endif

40 41
	WORK_STRUCT_COLOR_BITS	= 4,

42
	WORK_STRUCT_PENDING	= 1 << WORK_STRUCT_PENDING_BIT,
43
	WORK_STRUCT_DELAYED	= 1 << WORK_STRUCT_DELAYED_BIT,
44
	WORK_STRUCT_PWQ		= 1 << WORK_STRUCT_PWQ_BIT,
45
	WORK_STRUCT_LINKED	= 1 << WORK_STRUCT_LINKED_BIT,
46 47 48 49 50 51
#ifdef CONFIG_DEBUG_OBJECTS_WORK
	WORK_STRUCT_STATIC	= 1 << WORK_STRUCT_STATIC_BIT,
#else
	WORK_STRUCT_STATIC	= 0,
#endif

52 53 54 55 56 57 58
	/*
	 * The last color is no color used for works which don't
	 * participate in workqueue flushing.
	 */
	WORK_NR_COLORS		= (1 << WORK_STRUCT_COLOR_BITS) - 1,
	WORK_NO_COLOR		= WORK_NR_COLORS,

59
	/* special cpu IDs */
60
	WORK_CPU_UNBOUND	= NR_CPUS,
61
	WORK_CPU_END		= NR_CPUS + 1,
62

63
	/*
64 65 66
	 * Reserve 7 bits off of pwq pointer w/ debugobjects turned off.
	 * This makes pwqs aligned to 256 bytes and allows 15 workqueue
	 * flush colors.
67 68 69 70
	 */
	WORK_STRUCT_FLAG_BITS	= WORK_STRUCT_COLOR_SHIFT +
				  WORK_STRUCT_COLOR_BITS,

71
	/* data contains off-queue information when !WORK_STRUCT_PWQ */
72
	WORK_OFFQ_FLAG_BASE	= WORK_STRUCT_COLOR_SHIFT,
73 74 75

	WORK_OFFQ_CANCELING	= (1 << WORK_OFFQ_FLAG_BASE),

76 77
	/*
	 * When a work item is off queue, its high bits point to the last
78 79
	 * pool it was on.  Cap at 31 bits and use the highest number to
	 * indicate that no pool is associated.
80
	 */
81
	WORK_OFFQ_FLAG_BITS	= 1,
82 83 84 85
	WORK_OFFQ_POOL_SHIFT	= WORK_OFFQ_FLAG_BASE + WORK_OFFQ_FLAG_BITS,
	WORK_OFFQ_LEFT		= BITS_PER_LONG - WORK_OFFQ_POOL_SHIFT,
	WORK_OFFQ_POOL_BITS	= WORK_OFFQ_LEFT <= 31 ? WORK_OFFQ_LEFT : 31,
	WORK_OFFQ_POOL_NONE	= (1LU << WORK_OFFQ_POOL_BITS) - 1,
86 87

	/* convenience constants */
T
Tejun Heo 已提交
88
	WORK_STRUCT_FLAG_MASK	= (1UL << WORK_STRUCT_FLAG_BITS) - 1,
89
	WORK_STRUCT_WQ_DATA_MASK = ~WORK_STRUCT_FLAG_MASK,
90
	WORK_STRUCT_NO_POOL	= (unsigned long)WORK_OFFQ_POOL_NONE << WORK_OFFQ_POOL_SHIFT,
91 92 93 94

	/* bit mask for work_busy() return values */
	WORK_BUSY_PENDING	= 1 << 0,
	WORK_BUSY_RUNNING	= 1 << 1,
95 96
};

L
Linus Torvalds 已提交
97
struct work_struct {
98
	atomic_long_t data;
L
Linus Torvalds 已提交
99
	struct list_head entry;
100
	work_func_t func;
101 102 103
#ifdef CONFIG_LOCKDEP
	struct lockdep_map lockdep_map;
#endif
104 105
};

106
#define WORK_DATA_INIT()	ATOMIC_LONG_INIT(WORK_STRUCT_NO_POOL)
107
#define WORK_DATA_STATIC_INIT()	\
108
	ATOMIC_LONG_INIT(WORK_STRUCT_NO_POOL | WORK_STRUCT_STATIC)
109

110 111
struct delayed_work {
	struct work_struct work;
L
Linus Torvalds 已提交
112
	struct timer_list timer;
113 114 115

	/* target workqueue and CPU ->timer uses to queue ->work */
	struct workqueue_struct *wq;
116
	int cpu;
L
Linus Torvalds 已提交
117 118
};

T
Tejun Heo 已提交
119 120 121
/*
 * A struct for workqueue attributes.  This can be used to change
 * attributes of an unbound workqueue.
122 123 124 125
 *
 * Unlike other fields, ->no_numa isn't a property of a worker_pool.  It
 * only modifies how apply_workqueue_attrs() select pools and thus doesn't
 * participate in pool hash calculations or equality comparisons.
T
Tejun Heo 已提交
126 127 128 129
 */
struct workqueue_attrs {
	int			nice;		/* nice level */
	cpumask_var_t		cpumask;	/* allowed CPUs */
130
	bool			no_numa;	/* disable NUMA affinity */
T
Tejun Heo 已提交
131 132
};

133 134 135 136 137
static inline struct delayed_work *to_delayed_work(struct work_struct *work)
{
	return container_of(work, struct delayed_work, work);
}

138 139 140 141
struct execute_work {
	struct work_struct work;
};

142 143 144 145 146 147 148 149 150 151 152 153
#ifdef CONFIG_LOCKDEP
/*
 * NB: because we have to copy the lockdep_map, setting _key
 * here is required, otherwise it could get initialised to the
 * copy of the lockdep_map!
 */
#define __WORK_INIT_LOCKDEP_MAP(n, k) \
	.lockdep_map = STATIC_LOCKDEP_MAP_INIT(n, k),
#else
#define __WORK_INIT_LOCKDEP_MAP(n, k)
#endif

154 155 156 157 158
#define __WORK_INITIALIZER(n, f) {					\
	.data = WORK_DATA_STATIC_INIT(),				\
	.entry	= { &(n).entry, &(n).entry },				\
	.func = (f),							\
	__WORK_INIT_LOCKDEP_MAP(#n, &(n))				\
159 160
	}

161
#define __DELAYED_WORK_INITIALIZER(n, f, tflags) {			\
162
	.work = __WORK_INITIALIZER((n).work, (f)),			\
163
	.timer = __TIMER_INITIALIZER(delayed_work_timer_fn,		\
164 165
				     0, (unsigned long)&(n),		\
				     (tflags) | TIMER_IRQSAFE),		\
166 167
	}

168
#define DECLARE_WORK(n, f)						\
169 170
	struct work_struct n = __WORK_INITIALIZER(n, f)

171
#define DECLARE_DELAYED_WORK(n, f)					\
172
	struct delayed_work n = __DELAYED_WORK_INITIALIZER(n, f, 0)
173

174
#define DECLARE_DEFERRABLE_WORK(n, f)					\
175
	struct delayed_work n = __DELAYED_WORK_INITIALIZER(n, f, TIMER_DEFERRABLE)
176

L
Linus Torvalds 已提交
177
/*
178
 * initialize a work item's function pointer
L
Linus Torvalds 已提交
179
 */
180 181 182
#define PREPARE_WORK(_work, _func)					\
	do {								\
		(_work)->func = (_func);				\
L
Linus Torvalds 已提交
183 184
	} while (0)

185
#define PREPARE_DELAYED_WORK(_work, _func)				\
186
	PREPARE_WORK(&(_work)->work, (_func))
187

188 189 190
#ifdef CONFIG_DEBUG_OBJECTS_WORK
extern void __init_work(struct work_struct *work, int onstack);
extern void destroy_work_on_stack(struct work_struct *work);
T
Tejun Heo 已提交
191 192
static inline unsigned int work_static(struct work_struct *work)
{
193
	return *work_data_bits(work) & WORK_STRUCT_STATIC;
T
Tejun Heo 已提交
194
}
195 196 197
#else
static inline void __init_work(struct work_struct *work, int onstack) { }
static inline void destroy_work_on_stack(struct work_struct *work) { }
T
Tejun Heo 已提交
198
static inline unsigned int work_static(struct work_struct *work) { return 0; }
199 200
#endif

L
Linus Torvalds 已提交
201
/*
202
 * initialize all of a work item in one go
203
 *
D
Dmitri Vorobiev 已提交
204
 * NOTE! No point in using "atomic_long_set()": using a direct
205 206
 * assignment of the work data initializer allows the compiler
 * to generate better code.
L
Linus Torvalds 已提交
207
 */
208
#ifdef CONFIG_LOCKDEP
209
#define __INIT_WORK(_work, _func, _onstack)				\
210
	do {								\
211 212
		static struct lock_class_key __key;			\
									\
213
		__init_work((_work), _onstack);				\
O
Oleg Nesterov 已提交
214
		(_work)->data = (atomic_long_t) WORK_DATA_INIT();	\
215
		lockdep_init_map(&(_work)->lockdep_map, #_work, &__key, 0); \
216 217 218
		INIT_LIST_HEAD(&(_work)->entry);			\
		PREPARE_WORK((_work), (_func));				\
	} while (0)
219
#else
220
#define __INIT_WORK(_work, _func, _onstack)				\
221
	do {								\
222
		__init_work((_work), _onstack);				\
223 224 225 226 227
		(_work)->data = (atomic_long_t) WORK_DATA_INIT();	\
		INIT_LIST_HEAD(&(_work)->entry);			\
		PREPARE_WORK((_work), (_func));				\
	} while (0)
#endif
228

229 230 231
#define INIT_WORK(_work, _func)						\
	do {								\
		__INIT_WORK((_work), (_func), 0);			\
232 233
	} while (0)

234 235 236
#define INIT_WORK_ONSTACK(_work, _func)					\
	do {								\
		__INIT_WORK((_work), (_func), 1);			\
237 238
	} while (0)

239
#define __INIT_DELAYED_WORK(_work, _func, _tflags)			\
240 241
	do {								\
		INIT_WORK(&(_work)->work, (_func));			\
242
		__setup_timer(&(_work)->timer, delayed_work_timer_fn,	\
243 244
			      (unsigned long)(_work),			\
			      (_tflags) | TIMER_IRQSAFE);		\
245 246
	} while (0)

247
#define __INIT_DELAYED_WORK_ONSTACK(_work, _func, _tflags)		\
248 249
	do {								\
		INIT_WORK_ONSTACK(&(_work)->work, (_func));		\
250 251 252
		__setup_timer_on_stack(&(_work)->timer,			\
				       delayed_work_timer_fn,		\
				       (unsigned long)(_work),		\
253
				       (_tflags) | TIMER_IRQSAFE);	\
254 255
	} while (0)

256 257 258 259 260 261
#define INIT_DELAYED_WORK(_work, _func)					\
	__INIT_DELAYED_WORK(_work, _func, 0)

#define INIT_DELAYED_WORK_ONSTACK(_work, _func)				\
	__INIT_DELAYED_WORK_ONSTACK(_work, _func, 0)

262
#define INIT_DEFERRABLE_WORK(_work, _func)				\
263 264 265 266
	__INIT_DELAYED_WORK(_work, _func, TIMER_DEFERRABLE)

#define INIT_DEFERRABLE_WORK_ONSTACK(_work, _func)			\
	__INIT_DELAYED_WORK_ONSTACK(_work, _func, TIMER_DEFERRABLE)
267

268 269 270 271 272
/**
 * work_pending - Find out whether a work item is currently pending
 * @work: The work item in question
 */
#define work_pending(work) \
273
	test_bit(WORK_STRUCT_PENDING_BIT, work_data_bits(work))
274 275 276 277 278 279

/**
 * delayed_work_pending - Find out whether a delayable work item is currently
 * pending
 * @work: The work item in question
 */
280 281
#define delayed_work_pending(w) \
	work_pending(&(w)->work)
282

283
/**
O
Oleg Nesterov 已提交
284 285
 * work_clear_pending - for internal use only, mark a work item as not pending
 * @work: The work item in question
286
 */
O
Oleg Nesterov 已提交
287
#define work_clear_pending(work) \
288
	clear_bit(WORK_STRUCT_PENDING_BIT, work_data_bits(work))
289

T
Tejun Heo 已提交
290 291 292 293
/*
 * Workqueue flags and constants.  For details, please refer to
 * Documentation/workqueue.txt.
 */
294
enum {
295
	WQ_NON_REENTRANT	= 1 << 0, /* guarantee non-reentrance */
296
	WQ_UNBOUND		= 1 << 1, /* not bound to any cpu */
297
	WQ_FREEZABLE		= 1 << 2, /* freeze during suspend */
298
	WQ_MEM_RECLAIM		= 1 << 3, /* may be used for memory reclaim */
299
	WQ_HIGHPRI		= 1 << 4, /* high priority */
300
	WQ_CPU_INTENSIVE	= 1 << 5, /* cpu instensive workqueue */
301
	WQ_SYSFS		= 1 << 6, /* visible in sysfs, see wq_sysfs_register() */
302

303
	__WQ_DRAINING		= 1 << 16, /* internal: workqueue is draining */
304
	__WQ_ORDERED		= 1 << 17, /* internal: workqueue is ordered */
305

306
	WQ_MAX_ACTIVE		= 512,	  /* I like 512, better ideas? */
307
	WQ_MAX_UNBOUND_PER_CPU	= 4,	  /* 4 * #cpus for unbound wq */
308
	WQ_DFL_ACTIVE		= WQ_MAX_ACTIVE / 2,
309
};
310

311 312 313
/* unbound wq's aren't per-cpu, scale max_active according to #cpus */
#define WQ_UNBOUND_MAX_ACTIVE	\
	max_t(int, WQ_MAX_ACTIVE, num_possible_cpus() * WQ_MAX_UNBOUND_PER_CPU)
314

315 316 317 318 319 320 321 322 323 324 325
/*
 * System-wide workqueues which are always present.
 *
 * system_wq is the one used by schedule[_delayed]_work[_on]().
 * Multi-CPU multi-threaded.  There are users which expect relatively
 * short queue flush time.  Don't queue works which can run for too
 * long.
 *
 * system_long_wq is similar to system_wq but may host long running
 * works.  Queue flushing might take relatively long.
 *
326 327 328 329
 * system_unbound_wq is unbound workqueue.  Workers are not bound to
 * any specific CPU, not concurrency managed, and all queued works are
 * executed immediately as long as max_active limit is not reached and
 * resources are available.
T
Tejun Heo 已提交
330
 *
331 332
 * system_freezable_wq is equivalent to system_wq except that it's
 * freezable.
333 334 335
 */
extern struct workqueue_struct *system_wq;
extern struct workqueue_struct *system_long_wq;
336
extern struct workqueue_struct *system_unbound_wq;
337
extern struct workqueue_struct *system_freezable_wq;
338

339
static inline struct workqueue_struct * __deprecated __system_nrt_wq(void)
340 341 342 343
{
	return system_wq;
}

344
static inline struct workqueue_struct * __deprecated __system_nrt_freezable_wq(void)
345 346 347 348 349 350 351
{
	return system_freezable_wq;
}

/* equivlalent to system_wq and system_freezable_wq, deprecated */
#define system_nrt_wq			__system_nrt_wq()
#define system_nrt_freezable_wq		__system_nrt_freezable_wq()
352

353
extern struct workqueue_struct *
354 355
__alloc_workqueue_key(const char *fmt, unsigned int flags, int max_active,
	struct lock_class_key *key, const char *lock_name, ...) __printf(1, 6);
356

357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372
/**
 * alloc_workqueue - allocate a workqueue
 * @fmt: printf format for the name of the workqueue
 * @flags: WQ_* flags
 * @max_active: max in-flight work items, 0 for default
 * @args: args for @fmt
 *
 * Allocate a workqueue with the specified parameters.  For detailed
 * information on WQ_* flags, please refer to Documentation/workqueue.txt.
 *
 * The __lock_name macro dance is to guarantee that single lock_class_key
 * doesn't end up with different namesm, which isn't allowed by lockdep.
 *
 * RETURNS:
 * Pointer to the allocated workqueue on success, %NULL on failure.
 */
373
#ifdef CONFIG_LOCKDEP
374 375 376 377 378 379 380 381 382 383 384 385
#define alloc_workqueue(fmt, flags, max_active, args...)		\
({									\
	static struct lock_class_key __key;				\
	const char *__lock_name;					\
									\
	if (__builtin_constant_p(fmt))					\
		__lock_name = (fmt);					\
	else								\
		__lock_name = #fmt;					\
									\
	__alloc_workqueue_key((fmt), (flags), (max_active),		\
			      &__key, __lock_name, ##args);		\
386 387
})
#else
388 389
#define alloc_workqueue(fmt, flags, max_active, args...)		\
	__alloc_workqueue_key((fmt), (flags), (max_active),		\
390
			      NULL, NULL, ##args)
391 392
#endif

393 394
/**
 * alloc_ordered_workqueue - allocate an ordered workqueue
395
 * @fmt: printf format for the name of the workqueue
396
 * @flags: WQ_* flags (only WQ_FREEZABLE and WQ_MEM_RECLAIM are meaningful)
397
 * @args: args for @fmt
398 399 400 401 402 403 404 405
 *
 * Allocate an ordered workqueue.  An ordered workqueue executes at
 * most one work item at any given time in the queued order.  They are
 * implemented as unbound workqueues with @max_active of one.
 *
 * RETURNS:
 * Pointer to the allocated workqueue on success, %NULL on failure.
 */
406
#define alloc_ordered_workqueue(fmt, flags, args...)			\
407
	alloc_workqueue(fmt, WQ_UNBOUND | __WQ_ORDERED | (flags), 1, ##args)
408

409
#define create_workqueue(name)						\
410
	alloc_workqueue((name), WQ_MEM_RECLAIM, 1)
411
#define create_freezable_workqueue(name)				\
412
	alloc_workqueue((name), WQ_FREEZABLE | WQ_UNBOUND | WQ_MEM_RECLAIM, 1)
413
#define create_singlethread_workqueue(name)				\
414
	alloc_workqueue((name), WQ_UNBOUND | WQ_MEM_RECLAIM, 1)
L
Linus Torvalds 已提交
415 416 417

extern void destroy_workqueue(struct workqueue_struct *wq);

T
Tejun Heo 已提交
418 419
struct workqueue_attrs *alloc_workqueue_attrs(gfp_t gfp_mask);
void free_workqueue_attrs(struct workqueue_attrs *attrs);
420 421
int apply_workqueue_attrs(struct workqueue_struct *wq,
			  const struct workqueue_attrs *attrs);
T
Tejun Heo 已提交
422

423
extern bool queue_work_on(int cpu, struct workqueue_struct *wq,
424
			struct work_struct *work);
425
extern bool queue_delayed_work_on(int cpu, struct workqueue_struct *wq,
426
			struct delayed_work *work, unsigned long delay);
427 428
extern bool mod_delayed_work_on(int cpu, struct workqueue_struct *wq,
			struct delayed_work *dwork, unsigned long delay);
429

430
extern void flush_workqueue(struct workqueue_struct *wq);
431
extern void drain_workqueue(struct workqueue_struct *wq);
432
extern void flush_scheduled_work(void);
L
Linus Torvalds 已提交
433

434
extern int schedule_on_each_cpu(work_func_t func);
L
Linus Torvalds 已提交
435

436
int execute_in_process_context(work_func_t fn, struct execute_work *);
L
Linus Torvalds 已提交
437

438 439 440 441
extern bool flush_work(struct work_struct *work);
extern bool cancel_work_sync(struct work_struct *work);

extern bool flush_delayed_work(struct delayed_work *dwork);
442
extern bool cancel_delayed_work(struct delayed_work *dwork);
443
extern bool cancel_delayed_work_sync(struct delayed_work *dwork);
444

445 446
extern void workqueue_set_max_active(struct workqueue_struct *wq,
				     int max_active);
447
extern bool current_is_workqueue_rescuer(void);
448
extern bool workqueue_congested(int cpu, struct workqueue_struct *wq);
449 450
extern unsigned int work_busy(struct work_struct *work);

T
Tejun Heo 已提交
451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561
/**
 * queue_work - queue work on a workqueue
 * @wq: workqueue to use
 * @work: work to queue
 *
 * Returns %false if @work was already on a queue, %true otherwise.
 *
 * We queue the work to the CPU on which it was submitted, but if the CPU dies
 * it can be processed by another CPU.
 */
static inline bool queue_work(struct workqueue_struct *wq,
			      struct work_struct *work)
{
	return queue_work_on(WORK_CPU_UNBOUND, wq, work);
}

/**
 * queue_delayed_work - queue work on a workqueue after delay
 * @wq: workqueue to use
 * @dwork: delayable work to queue
 * @delay: number of jiffies to wait before queueing
 *
 * Equivalent to queue_delayed_work_on() but tries to use the local CPU.
 */
static inline bool queue_delayed_work(struct workqueue_struct *wq,
				      struct delayed_work *dwork,
				      unsigned long delay)
{
	return queue_delayed_work_on(WORK_CPU_UNBOUND, wq, dwork, delay);
}

/**
 * mod_delayed_work - modify delay of or queue a delayed work
 * @wq: workqueue to use
 * @dwork: work to queue
 * @delay: number of jiffies to wait before queueing
 *
 * mod_delayed_work_on() on local CPU.
 */
static inline bool mod_delayed_work(struct workqueue_struct *wq,
				    struct delayed_work *dwork,
				    unsigned long delay)
{
	return mod_delayed_work_on(WORK_CPU_UNBOUND, wq, dwork, delay);
}

/**
 * schedule_work_on - put work task on a specific cpu
 * @cpu: cpu to put the work task on
 * @work: job to be done
 *
 * This puts a job on a specific cpu
 */
static inline bool schedule_work_on(int cpu, struct work_struct *work)
{
	return queue_work_on(cpu, system_wq, work);
}

/**
 * schedule_work - put work task in global workqueue
 * @work: job to be done
 *
 * Returns %false if @work was already on the kernel-global workqueue and
 * %true otherwise.
 *
 * This puts a job in the kernel-global workqueue if it was not already
 * queued and leaves it in the same position on the kernel-global
 * workqueue otherwise.
 */
static inline bool schedule_work(struct work_struct *work)
{
	return queue_work(system_wq, work);
}

/**
 * schedule_delayed_work_on - queue work in global workqueue on CPU after delay
 * @cpu: cpu to use
 * @dwork: job to be done
 * @delay: number of jiffies to wait
 *
 * After waiting for a given time this puts a job in the kernel-global
 * workqueue on the specified CPU.
 */
static inline bool schedule_delayed_work_on(int cpu, struct delayed_work *dwork,
					    unsigned long delay)
{
	return queue_delayed_work_on(cpu, system_wq, dwork, delay);
}

/**
 * schedule_delayed_work - put work task in global workqueue after delay
 * @dwork: job to be done
 * @delay: number of jiffies to wait or 0 for immediate execution
 *
 * After waiting for a given time this puts a job in the kernel-global
 * workqueue.
 */
static inline bool schedule_delayed_work(struct delayed_work *dwork,
					 unsigned long delay)
{
	return queue_delayed_work(system_wq, dwork, delay);
}

/**
 * keventd_up - is workqueue initialized yet?
 */
static inline bool keventd_up(void)
{
	return system_wq != NULL;
}

562 563 564 565 566
/*
 * Like above, but uses del_timer() instead of del_timer_sync(). This means,
 * if it returns 0 the timer function may be running and the queueing is in
 * progress.
 */
567
static inline bool __deprecated __cancel_delayed_work(struct delayed_work *work)
568
{
569
	bool ret;
570 571 572 573 574 575 576

	ret = del_timer(&work->timer);
	if (ret)
		work_clear_pending(&work->work);
	return ret;
}

577
/* used to be different but now identical to flush_work(), deprecated */
578
static inline bool __deprecated flush_work_sync(struct work_struct *work)
579 580 581 582 583
{
	return flush_work(work);
}

/* used to be different but now identical to flush_delayed_work(), deprecated */
584
static inline bool __deprecated flush_delayed_work_sync(struct delayed_work *dwork)
585 586 587 588
{
	return flush_delayed_work(dwork);
}

589
#ifndef CONFIG_SMP
590
static inline long work_on_cpu(int cpu, long (*fn)(void *), void *arg)
591 592 593 594
{
	return fn(arg);
}
#else
595
long work_on_cpu(int cpu, long (*fn)(void *), void *arg);
596
#endif /* CONFIG_SMP */
597

598 599 600 601 602 603
#ifdef CONFIG_FREEZER
extern void freeze_workqueues_begin(void);
extern bool freeze_workqueues_busy(void);
extern void thaw_workqueues(void);
#endif /* CONFIG_FREEZER */

604 605 606 607 608 609 610
#ifdef CONFIG_SYSFS
int workqueue_sysfs_register(struct workqueue_struct *wq);
#else	/* CONFIG_SYSFS */
static inline int workqueue_sysfs_register(struct workqueue_struct *wq)
{ return 0; }
#endif	/* CONFIG_SYSFS */

L
Linus Torvalds 已提交
611
#endif