mirror of
https://github.com/joel16/android_kernel_sony_msm8994.git
synced 2024-12-01 08:00:42 +00:00
Merge branch 'for-3.9' of git://git.kernel.org/pub/scm/linux/kernel/git/tj/wq
Pull workqueue changes from Tejun Heo:
"A lot of reorganization is going on mostly to prepare for worker pools
with custom attributes so that workqueue can replace custom pool
implementations in places including writeback and btrfs and make CPU
assignment in crypto more flexible.
workqueue evolved from purely per-cpu design and implementation, so
there are a lot of assumptions regarding being bound to CPUs and even
unbound workqueues are implemented as an extension of the model -
workqueues running on the special unbound CPU. Bulk of changes this
round are about promoting worker_pools as the top level abstraction
replacing global_cwq (global cpu workqueue). At this point, I'm
fairly confident about getting custom worker pools working pretty soon
and ready for the next merge window.
Lai's patches are replacing the convoluted mb() dancing workqueue has
been doing with much simpler mechanism which only depends on
assignment atomicity of long. For details, please read the commit
message of 0b3dae68ac
("workqueue: simplify is-work-item-queued-here
test"). While the change ends up adding one pointer to struct
delayed_work, the inflation in percentage is less than five percent
and it decouples delayed_work logic a lot more cleaner from usual work
handling, removes the unusual memory barrier dancing, and allows for
further simplification, so I think the trade-off is acceptable.
There will be two more workqueue related pull requests and there are
some shared commits among them. I'll write further pull requests
assuming this pull request is pulled first."
* 'for-3.9' of git://git.kernel.org/pub/scm/linux/kernel/git/tj/wq: (37 commits)
workqueue: un-GPL function delayed_work_timer_fn()
workqueue: rename cpu_workqueue to pool_workqueue
workqueue: reimplement is_chained_work() using current_wq_worker()
workqueue: fix is_chained_work() regression
workqueue: pick cwq instead of pool in __queue_work()
workqueue: make get_work_pool_id() cheaper
workqueue: move nr_running into worker_pool
workqueue: cosmetic update in try_to_grab_pending()
workqueue: simplify is-work-item-queued-here test
workqueue: make work->data point to pool after try_to_grab_pending()
workqueue: add delayed_work->wq to simplify reentrancy handling
workqueue: make work_busy() test WORK_STRUCT_PENDING first
workqueue: replace WORK_CPU_NONE/LAST with WORK_CPU_END
workqueue: post global_cwq removal cleanups
workqueue: rename nr_running variables
workqueue: remove global_cwq
workqueue: remove worker_pool->gcwq
workqueue: replace for_each_worker_pool() with for_each_std_worker_pool()
workqueue: make freezing/thawing per-pool
workqueue: make hotplug processing per-pool
...
This commit is contained in:
commit
67cb104b4c
@ -52,4 +52,5 @@ extern void async_synchronize_full_domain(struct async_domain *domain);
|
|||||||
extern void async_synchronize_cookie(async_cookie_t cookie);
|
extern void async_synchronize_cookie(async_cookie_t cookie);
|
||||||
extern void async_synchronize_cookie_domain(async_cookie_t cookie,
|
extern void async_synchronize_cookie_domain(async_cookie_t cookie,
|
||||||
struct async_domain *domain);
|
struct async_domain *domain);
|
||||||
|
extern bool current_is_async(void);
|
||||||
#endif
|
#endif
|
||||||
|
@ -27,7 +27,7 @@ void delayed_work_timer_fn(unsigned long __data);
|
|||||||
enum {
|
enum {
|
||||||
WORK_STRUCT_PENDING_BIT = 0, /* work item is pending execution */
|
WORK_STRUCT_PENDING_BIT = 0, /* work item is pending execution */
|
||||||
WORK_STRUCT_DELAYED_BIT = 1, /* work item is delayed */
|
WORK_STRUCT_DELAYED_BIT = 1, /* work item is delayed */
|
||||||
WORK_STRUCT_CWQ_BIT = 2, /* data points to cwq */
|
WORK_STRUCT_PWQ_BIT = 2, /* data points to pwq */
|
||||||
WORK_STRUCT_LINKED_BIT = 3, /* next work is linked to this one */
|
WORK_STRUCT_LINKED_BIT = 3, /* next work is linked to this one */
|
||||||
#ifdef CONFIG_DEBUG_OBJECTS_WORK
|
#ifdef CONFIG_DEBUG_OBJECTS_WORK
|
||||||
WORK_STRUCT_STATIC_BIT = 4, /* static initializer (debugobjects) */
|
WORK_STRUCT_STATIC_BIT = 4, /* static initializer (debugobjects) */
|
||||||
@ -40,7 +40,7 @@ enum {
|
|||||||
|
|
||||||
WORK_STRUCT_PENDING = 1 << WORK_STRUCT_PENDING_BIT,
|
WORK_STRUCT_PENDING = 1 << WORK_STRUCT_PENDING_BIT,
|
||||||
WORK_STRUCT_DELAYED = 1 << WORK_STRUCT_DELAYED_BIT,
|
WORK_STRUCT_DELAYED = 1 << WORK_STRUCT_DELAYED_BIT,
|
||||||
WORK_STRUCT_CWQ = 1 << WORK_STRUCT_CWQ_BIT,
|
WORK_STRUCT_PWQ = 1 << WORK_STRUCT_PWQ_BIT,
|
||||||
WORK_STRUCT_LINKED = 1 << WORK_STRUCT_LINKED_BIT,
|
WORK_STRUCT_LINKED = 1 << WORK_STRUCT_LINKED_BIT,
|
||||||
#ifdef CONFIG_DEBUG_OBJECTS_WORK
|
#ifdef CONFIG_DEBUG_OBJECTS_WORK
|
||||||
WORK_STRUCT_STATIC = 1 << WORK_STRUCT_STATIC_BIT,
|
WORK_STRUCT_STATIC = 1 << WORK_STRUCT_STATIC_BIT,
|
||||||
@ -57,29 +57,36 @@ enum {
|
|||||||
|
|
||||||
/* special cpu IDs */
|
/* special cpu IDs */
|
||||||
WORK_CPU_UNBOUND = NR_CPUS,
|
WORK_CPU_UNBOUND = NR_CPUS,
|
||||||
WORK_CPU_NONE = NR_CPUS + 1,
|
WORK_CPU_END = NR_CPUS + 1,
|
||||||
WORK_CPU_LAST = WORK_CPU_NONE,
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Reserve 7 bits off of cwq pointer w/ debugobjects turned
|
* Reserve 7 bits off of pwq pointer w/ debugobjects turned off.
|
||||||
* off. This makes cwqs aligned to 256 bytes and allows 15
|
* This makes pwqs aligned to 256 bytes and allows 15 workqueue
|
||||||
* workqueue flush colors.
|
* flush colors.
|
||||||
*/
|
*/
|
||||||
WORK_STRUCT_FLAG_BITS = WORK_STRUCT_COLOR_SHIFT +
|
WORK_STRUCT_FLAG_BITS = WORK_STRUCT_COLOR_SHIFT +
|
||||||
WORK_STRUCT_COLOR_BITS,
|
WORK_STRUCT_COLOR_BITS,
|
||||||
|
|
||||||
/* data contains off-queue information when !WORK_STRUCT_CWQ */
|
/* data contains off-queue information when !WORK_STRUCT_PWQ */
|
||||||
WORK_OFFQ_FLAG_BASE = WORK_STRUCT_FLAG_BITS,
|
WORK_OFFQ_FLAG_BASE = WORK_STRUCT_FLAG_BITS,
|
||||||
|
|
||||||
WORK_OFFQ_CANCELING = (1 << WORK_OFFQ_FLAG_BASE),
|
WORK_OFFQ_CANCELING = (1 << WORK_OFFQ_FLAG_BASE),
|
||||||
|
|
||||||
|
/*
|
||||||
|
* When a work item is off queue, its high bits point to the last
|
||||||
|
* pool it was on. Cap at 31 bits and use the highest number to
|
||||||
|
* indicate that no pool is associated.
|
||||||
|
*/
|
||||||
WORK_OFFQ_FLAG_BITS = 1,
|
WORK_OFFQ_FLAG_BITS = 1,
|
||||||
WORK_OFFQ_CPU_SHIFT = WORK_OFFQ_FLAG_BASE + WORK_OFFQ_FLAG_BITS,
|
WORK_OFFQ_POOL_SHIFT = WORK_OFFQ_FLAG_BASE + WORK_OFFQ_FLAG_BITS,
|
||||||
|
WORK_OFFQ_LEFT = BITS_PER_LONG - WORK_OFFQ_POOL_SHIFT,
|
||||||
|
WORK_OFFQ_POOL_BITS = WORK_OFFQ_LEFT <= 31 ? WORK_OFFQ_LEFT : 31,
|
||||||
|
WORK_OFFQ_POOL_NONE = (1LU << WORK_OFFQ_POOL_BITS) - 1,
|
||||||
|
|
||||||
/* convenience constants */
|
/* convenience constants */
|
||||||
WORK_STRUCT_FLAG_MASK = (1UL << WORK_STRUCT_FLAG_BITS) - 1,
|
WORK_STRUCT_FLAG_MASK = (1UL << WORK_STRUCT_FLAG_BITS) - 1,
|
||||||
WORK_STRUCT_WQ_DATA_MASK = ~WORK_STRUCT_FLAG_MASK,
|
WORK_STRUCT_WQ_DATA_MASK = ~WORK_STRUCT_FLAG_MASK,
|
||||||
WORK_STRUCT_NO_CPU = (unsigned long)WORK_CPU_NONE << WORK_OFFQ_CPU_SHIFT,
|
WORK_STRUCT_NO_POOL = (unsigned long)WORK_OFFQ_POOL_NONE << WORK_OFFQ_POOL_SHIFT,
|
||||||
|
|
||||||
/* bit mask for work_busy() return values */
|
/* bit mask for work_busy() return values */
|
||||||
WORK_BUSY_PENDING = 1 << 0,
|
WORK_BUSY_PENDING = 1 << 0,
|
||||||
@ -95,13 +102,16 @@ struct work_struct {
|
|||||||
#endif
|
#endif
|
||||||
};
|
};
|
||||||
|
|
||||||
#define WORK_DATA_INIT() ATOMIC_LONG_INIT(WORK_STRUCT_NO_CPU)
|
#define WORK_DATA_INIT() ATOMIC_LONG_INIT(WORK_STRUCT_NO_POOL)
|
||||||
#define WORK_DATA_STATIC_INIT() \
|
#define WORK_DATA_STATIC_INIT() \
|
||||||
ATOMIC_LONG_INIT(WORK_STRUCT_NO_CPU | WORK_STRUCT_STATIC)
|
ATOMIC_LONG_INIT(WORK_STRUCT_NO_POOL | WORK_STRUCT_STATIC)
|
||||||
|
|
||||||
struct delayed_work {
|
struct delayed_work {
|
||||||
struct work_struct work;
|
struct work_struct work;
|
||||||
struct timer_list timer;
|
struct timer_list timer;
|
||||||
|
|
||||||
|
/* target workqueue and CPU ->timer uses to queue ->work */
|
||||||
|
struct workqueue_struct *wq;
|
||||||
int cpu;
|
int cpu;
|
||||||
};
|
};
|
||||||
|
|
||||||
@ -426,7 +436,6 @@ extern bool cancel_delayed_work_sync(struct delayed_work *dwork);
|
|||||||
extern void workqueue_set_max_active(struct workqueue_struct *wq,
|
extern void workqueue_set_max_active(struct workqueue_struct *wq,
|
||||||
int max_active);
|
int max_active);
|
||||||
extern bool workqueue_congested(unsigned int cpu, struct workqueue_struct *wq);
|
extern bool workqueue_congested(unsigned int cpu, struct workqueue_struct *wq);
|
||||||
extern unsigned int work_cpu(struct work_struct *work);
|
|
||||||
extern unsigned int work_busy(struct work_struct *work);
|
extern unsigned int work_busy(struct work_struct *work);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -27,7 +27,7 @@ DECLARE_EVENT_CLASS(workqueue_work,
|
|||||||
/**
|
/**
|
||||||
* workqueue_queue_work - called when a work gets queued
|
* workqueue_queue_work - called when a work gets queued
|
||||||
* @req_cpu: the requested cpu
|
* @req_cpu: the requested cpu
|
||||||
* @cwq: pointer to struct cpu_workqueue_struct
|
* @pwq: pointer to struct pool_workqueue
|
||||||
* @work: pointer to struct work_struct
|
* @work: pointer to struct work_struct
|
||||||
*
|
*
|
||||||
* This event occurs when a work is queued immediately or once a
|
* This event occurs when a work is queued immediately or once a
|
||||||
@ -36,10 +36,10 @@ DECLARE_EVENT_CLASS(workqueue_work,
|
|||||||
*/
|
*/
|
||||||
TRACE_EVENT(workqueue_queue_work,
|
TRACE_EVENT(workqueue_queue_work,
|
||||||
|
|
||||||
TP_PROTO(unsigned int req_cpu, struct cpu_workqueue_struct *cwq,
|
TP_PROTO(unsigned int req_cpu, struct pool_workqueue *pwq,
|
||||||
struct work_struct *work),
|
struct work_struct *work),
|
||||||
|
|
||||||
TP_ARGS(req_cpu, cwq, work),
|
TP_ARGS(req_cpu, pwq, work),
|
||||||
|
|
||||||
TP_STRUCT__entry(
|
TP_STRUCT__entry(
|
||||||
__field( void *, work )
|
__field( void *, work )
|
||||||
@ -52,9 +52,9 @@ TRACE_EVENT(workqueue_queue_work,
|
|||||||
TP_fast_assign(
|
TP_fast_assign(
|
||||||
__entry->work = work;
|
__entry->work = work;
|
||||||
__entry->function = work->func;
|
__entry->function = work->func;
|
||||||
__entry->workqueue = cwq->wq;
|
__entry->workqueue = pwq->wq;
|
||||||
__entry->req_cpu = req_cpu;
|
__entry->req_cpu = req_cpu;
|
||||||
__entry->cpu = cwq->pool->gcwq->cpu;
|
__entry->cpu = pwq->pool->cpu;
|
||||||
),
|
),
|
||||||
|
|
||||||
TP_printk("work struct=%p function=%pf workqueue=%p req_cpu=%u cpu=%u",
|
TP_printk("work struct=%p function=%pf workqueue=%p req_cpu=%u cpu=%u",
|
||||||
|
@ -57,6 +57,8 @@ asynchronous and synchronous parts of the kernel.
|
|||||||
#include <linux/slab.h>
|
#include <linux/slab.h>
|
||||||
#include <linux/workqueue.h>
|
#include <linux/workqueue.h>
|
||||||
|
|
||||||
|
#include "workqueue_internal.h"
|
||||||
|
|
||||||
static async_cookie_t next_cookie = 1;
|
static async_cookie_t next_cookie = 1;
|
||||||
|
|
||||||
#define MAX_WORK 32768
|
#define MAX_WORK 32768
|
||||||
@ -353,3 +355,15 @@ void async_synchronize_cookie(async_cookie_t cookie)
|
|||||||
async_synchronize_cookie_domain(cookie, &async_running);
|
async_synchronize_cookie_domain(cookie, &async_running);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(async_synchronize_cookie);
|
EXPORT_SYMBOL_GPL(async_synchronize_cookie);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* current_is_async - is %current an async worker task?
|
||||||
|
*
|
||||||
|
* Returns %true if %current is an async worker task.
|
||||||
|
*/
|
||||||
|
bool current_is_async(void)
|
||||||
|
{
|
||||||
|
struct worker *worker = current_wq_worker();
|
||||||
|
|
||||||
|
return worker && worker->current_func == async_run_entry_fn;
|
||||||
|
}
|
||||||
|
@ -83,7 +83,7 @@
|
|||||||
#endif
|
#endif
|
||||||
|
|
||||||
#include "sched.h"
|
#include "sched.h"
|
||||||
#include "../workqueue_sched.h"
|
#include "../workqueue_internal.h"
|
||||||
#include "../smpboot.h"
|
#include "../smpboot.h"
|
||||||
|
|
||||||
#define CREATE_TRACE_POINTS
|
#define CREATE_TRACE_POINTS
|
||||||
|
1538
kernel/workqueue.c
1538
kernel/workqueue.c
File diff suppressed because it is too large
Load Diff
65
kernel/workqueue_internal.h
Normal file
65
kernel/workqueue_internal.h
Normal file
@ -0,0 +1,65 @@
|
|||||||
|
/*
|
||||||
|
* kernel/workqueue_internal.h
|
||||||
|
*
|
||||||
|
* Workqueue internal header file. Only to be included by workqueue and
|
||||||
|
* core kernel subsystems.
|
||||||
|
*/
|
||||||
|
#ifndef _KERNEL_WORKQUEUE_INTERNAL_H
|
||||||
|
#define _KERNEL_WORKQUEUE_INTERNAL_H
|
||||||
|
|
||||||
|
#include <linux/workqueue.h>
|
||||||
|
#include <linux/kthread.h>
|
||||||
|
|
||||||
|
struct worker_pool;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* The poor guys doing the actual heavy lifting. All on-duty workers are
|
||||||
|
* either serving the manager role, on idle list or on busy hash. For
|
||||||
|
* details on the locking annotation (L, I, X...), refer to workqueue.c.
|
||||||
|
*
|
||||||
|
* Only to be used in workqueue and async.
|
||||||
|
*/
|
||||||
|
struct worker {
|
||||||
|
/* on idle list while idle, on busy hash table while busy */
|
||||||
|
union {
|
||||||
|
struct list_head entry; /* L: while idle */
|
||||||
|
struct hlist_node hentry; /* L: while busy */
|
||||||
|
};
|
||||||
|
|
||||||
|
struct work_struct *current_work; /* L: work being processed */
|
||||||
|
work_func_t current_func; /* L: current_work's fn */
|
||||||
|
struct pool_workqueue *current_pwq; /* L: current_work's pwq */
|
||||||
|
struct list_head scheduled; /* L: scheduled works */
|
||||||
|
struct task_struct *task; /* I: worker task */
|
||||||
|
struct worker_pool *pool; /* I: the associated pool */
|
||||||
|
/* 64 bytes boundary on 64bit, 32 on 32bit */
|
||||||
|
unsigned long last_active; /* L: last active timestamp */
|
||||||
|
unsigned int flags; /* X: flags */
|
||||||
|
int id; /* I: worker id */
|
||||||
|
|
||||||
|
/* for rebinding worker to CPU */
|
||||||
|
struct work_struct rebind_work; /* L: for busy worker */
|
||||||
|
|
||||||
|
/* used only by rescuers to point to the target workqueue */
|
||||||
|
struct workqueue_struct *rescue_wq; /* I: the workqueue to rescue */
|
||||||
|
};
|
||||||
|
|
||||||
|
/**
|
||||||
|
* current_wq_worker - return struct worker if %current is a workqueue worker
|
||||||
|
*/
|
||||||
|
static inline struct worker *current_wq_worker(void)
|
||||||
|
{
|
||||||
|
if (current->flags & PF_WQ_WORKER)
|
||||||
|
return kthread_data(current);
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Scheduler hooks for concurrency managed workqueue. Only to be used from
|
||||||
|
* sched.c and workqueue.c.
|
||||||
|
*/
|
||||||
|
void wq_worker_waking_up(struct task_struct *task, unsigned int cpu);
|
||||||
|
struct task_struct *wq_worker_sleeping(struct task_struct *task,
|
||||||
|
unsigned int cpu);
|
||||||
|
|
||||||
|
#endif /* _KERNEL_WORKQUEUE_INTERNAL_H */
|
@ -1,9 +0,0 @@
|
|||||||
/*
|
|
||||||
* kernel/workqueue_sched.h
|
|
||||||
*
|
|
||||||
* Scheduler hooks for concurrency managed workqueue. Only to be
|
|
||||||
* included from sched.c and workqueue.c.
|
|
||||||
*/
|
|
||||||
void wq_worker_waking_up(struct task_struct *task, unsigned int cpu);
|
|
||||||
struct task_struct *wq_worker_sleeping(struct task_struct *task,
|
|
||||||
unsigned int cpu);
|
|
Loading…
Reference in New Issue
Block a user