提交 2d5f0764 编写于 作者: L Lai Jiangshan 提交者: Tejun Heo

workqueue: split apply_workqueue_attrs() into 3 stages

Current apply_workqueue_attrs() includes pwqs-allocation and pwqs-installation,
so when we batch multiple apply_workqueue_attrs()s as a transaction, we can't
ensure the transaction must succeed or fail as a complete unit.

To solve this, we split apply_workqueue_attrs() into three stages.
The first stage does the preparation: allocation memory, pwqs.
The second stage does the attrs-installaion and pwqs-installation.
The third stage frees the allocated memory and (old or unused) pwqs.

As the result, batching multiple apply_workqueue_attrs()s can
succeed or fail as a complete unit:
	1) batch do all the first stage for all the workqueues
	2) only commit all when all the above succeed.

This patch is a preparation for the next patch ("Allow modifying low level
unbound workqueue cpumask") which will do a multiple apply_workqueue_attrs().

The patch doesn't have functionality changed except two minor adjustment:
	1) free_unbound_pwq() for the error path is removed, we use the
	   heavier version put_pwq_unlocked() instead since the error path
	   is rare. this adjustment simplifies the code.
	2) the memory-allocation is also moved into wq_pool_mutex.
	   this is needed to avoid to do the further splitting.

tj: minor updates to comments.
Suggested-by: NTejun Heo <tj@kernel.org>
Cc: Christoph Lameter <cl@linux.com>
Cc: Kevin Hilman <khilman@linaro.org>
Cc: Lai Jiangshan <laijs@cn.fujitsu.com>
Cc: Mike Galbraith <bitbucket@online.de>
Cc: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Cc: Tejun Heo <tj@kernel.org>
Cc: Viresh Kumar <viresh.kumar@linaro.org>
Cc: Frederic Weisbecker <fweisbec@gmail.com>
Signed-off-by: NLai Jiangshan <laijs@cn.fujitsu.com>
Signed-off-by: NTejun Heo <tj@kernel.org>
上级 b787f68c
...@@ -3425,17 +3425,6 @@ static struct pool_workqueue *alloc_unbound_pwq(struct workqueue_struct *wq, ...@@ -3425,17 +3425,6 @@ static struct pool_workqueue *alloc_unbound_pwq(struct workqueue_struct *wq,
return pwq; return pwq;
} }
/* undo alloc_unbound_pwq(), used only in the error path */
static void free_unbound_pwq(struct pool_workqueue *pwq)
{
lockdep_assert_held(&wq_pool_mutex);
if (pwq) {
put_unbound_pool(pwq->pool);
kmem_cache_free(pwq_cache, pwq);
}
}
/** /**
* wq_calc_node_mask - calculate a wq_attrs' cpumask for the specified node * wq_calc_node_mask - calculate a wq_attrs' cpumask for the specified node
* @attrs: the wq_attrs of interest * @attrs: the wq_attrs of interest
...@@ -3498,42 +3487,48 @@ static struct pool_workqueue *numa_pwq_tbl_install(struct workqueue_struct *wq, ...@@ -3498,42 +3487,48 @@ static struct pool_workqueue *numa_pwq_tbl_install(struct workqueue_struct *wq,
return old_pwq; return old_pwq;
} }
/** /* context to store the prepared attrs & pwqs before applying */
* apply_workqueue_attrs - apply new workqueue_attrs to an unbound workqueue struct apply_wqattrs_ctx {
* @wq: the target workqueue struct workqueue_struct *wq; /* target workqueue */
* @attrs: the workqueue_attrs to apply, allocated with alloc_workqueue_attrs() struct workqueue_attrs *attrs; /* attrs to apply */
* struct pool_workqueue *dfl_pwq;
* Apply @attrs to an unbound workqueue @wq. Unless disabled, on NUMA struct pool_workqueue *pwq_tbl[];
* machines, this function maps a separate pwq to each NUMA node with };
* possibles CPUs in @attrs->cpumask so that work items are affine to the
* NUMA node it was issued on. Older pwqs are released as in-flight work /* free the resources after success or abort */
* items finish. Note that a work item which repeatedly requeues itself static void apply_wqattrs_cleanup(struct apply_wqattrs_ctx *ctx)
* back-to-back will stay on its current pwq. {
* if (ctx) {
* Performs GFP_KERNEL allocations. int node;
*
* Return: 0 on success and -errno on failure. for_each_node(node)
*/ put_pwq_unlocked(ctx->pwq_tbl[node]);
int apply_workqueue_attrs(struct workqueue_struct *wq, put_pwq_unlocked(ctx->dfl_pwq);
free_workqueue_attrs(ctx->attrs);
kfree(ctx);
}
}
/* allocate the attrs and pwqs for later installation */
static struct apply_wqattrs_ctx *
apply_wqattrs_prepare(struct workqueue_struct *wq,
const struct workqueue_attrs *attrs) const struct workqueue_attrs *attrs)
{ {
struct apply_wqattrs_ctx *ctx;
struct workqueue_attrs *new_attrs, *tmp_attrs; struct workqueue_attrs *new_attrs, *tmp_attrs;
struct pool_workqueue **pwq_tbl, *dfl_pwq; int node;
int node, ret;
/* only unbound workqueues can change attributes */ lockdep_assert_held(&wq_pool_mutex);
if (WARN_ON(!(wq->flags & WQ_UNBOUND)))
return -EINVAL;
/* creating multiple pwqs breaks ordering guarantee */ ctx = kzalloc(sizeof(*ctx) + nr_node_ids * sizeof(ctx->pwq_tbl[0]),
if (WARN_ON((wq->flags & __WQ_ORDERED) && !list_empty(&wq->pwqs))) GFP_KERNEL);
return -EINVAL;
pwq_tbl = kzalloc(nr_node_ids * sizeof(pwq_tbl[0]), GFP_KERNEL);
new_attrs = alloc_workqueue_attrs(GFP_KERNEL); new_attrs = alloc_workqueue_attrs(GFP_KERNEL);
tmp_attrs = alloc_workqueue_attrs(GFP_KERNEL); tmp_attrs = alloc_workqueue_attrs(GFP_KERNEL);
if (!pwq_tbl || !new_attrs || !tmp_attrs) if (!ctx || !new_attrs || !tmp_attrs)
goto enomem; goto out_free;
/* make a copy of @attrs and sanitize it */ /* make a copy of @attrs and sanitize it */
copy_workqueue_attrs(new_attrs, attrs); copy_workqueue_attrs(new_attrs, attrs);
...@@ -3546,76 +3541,112 @@ int apply_workqueue_attrs(struct workqueue_struct *wq, ...@@ -3546,76 +3541,112 @@ int apply_workqueue_attrs(struct workqueue_struct *wq,
*/ */
copy_workqueue_attrs(tmp_attrs, new_attrs); copy_workqueue_attrs(tmp_attrs, new_attrs);
/*
* CPUs should stay stable across pwq creations and installations.
* Pin CPUs, determine the target cpumask for each node and create
* pwqs accordingly.
*/
get_online_cpus();
mutex_lock(&wq_pool_mutex);
/* /*
* If something goes wrong during CPU up/down, we'll fall back to * If something goes wrong during CPU up/down, we'll fall back to
* the default pwq covering whole @attrs->cpumask. Always create * the default pwq covering whole @attrs->cpumask. Always create
* it even if we don't use it immediately. * it even if we don't use it immediately.
*/ */
dfl_pwq = alloc_unbound_pwq(wq, new_attrs); ctx->dfl_pwq = alloc_unbound_pwq(wq, new_attrs);
if (!dfl_pwq) if (!ctx->dfl_pwq)
goto enomem_pwq; goto out_free;
for_each_node(node) { for_each_node(node) {
if (wq_calc_node_cpumask(attrs, node, -1, tmp_attrs->cpumask)) { if (wq_calc_node_cpumask(attrs, node, -1, tmp_attrs->cpumask)) {
pwq_tbl[node] = alloc_unbound_pwq(wq, tmp_attrs); ctx->pwq_tbl[node] = alloc_unbound_pwq(wq, tmp_attrs);
if (!pwq_tbl[node]) if (!ctx->pwq_tbl[node])
goto enomem_pwq; goto out_free;
} else { } else {
dfl_pwq->refcnt++; ctx->dfl_pwq->refcnt++;
pwq_tbl[node] = dfl_pwq; ctx->pwq_tbl[node] = ctx->dfl_pwq;
} }
} }
mutex_unlock(&wq_pool_mutex); ctx->attrs = new_attrs;
ctx->wq = wq;
free_workqueue_attrs(tmp_attrs);
return ctx;
out_free:
free_workqueue_attrs(tmp_attrs);
free_workqueue_attrs(new_attrs);
apply_wqattrs_cleanup(ctx);
return NULL;
}
/* set attrs and install prepared pwqs, @ctx points to old pwqs on return */
static void apply_wqattrs_commit(struct apply_wqattrs_ctx *ctx)
{
int node;
/* all pwqs have been created successfully, let's install'em */ /* all pwqs have been created successfully, let's install'em */
mutex_lock(&wq->mutex); mutex_lock(&ctx->wq->mutex);
copy_workqueue_attrs(wq->unbound_attrs, new_attrs); copy_workqueue_attrs(ctx->wq->unbound_attrs, ctx->attrs);
/* save the previous pwq and install the new one */ /* save the previous pwq and install the new one */
for_each_node(node) for_each_node(node)
pwq_tbl[node] = numa_pwq_tbl_install(wq, node, pwq_tbl[node]); ctx->pwq_tbl[node] = numa_pwq_tbl_install(ctx->wq, node,
ctx->pwq_tbl[node]);
/* @dfl_pwq might not have been used, ensure it's linked */ /* @dfl_pwq might not have been used, ensure it's linked */
link_pwq(dfl_pwq); link_pwq(ctx->dfl_pwq);
swap(wq->dfl_pwq, dfl_pwq); swap(ctx->wq->dfl_pwq, ctx->dfl_pwq);
mutex_unlock(&wq->mutex); mutex_unlock(&ctx->wq->mutex);
}
/* put the old pwqs */ /**
for_each_node(node) * apply_workqueue_attrs - apply new workqueue_attrs to an unbound workqueue
put_pwq_unlocked(pwq_tbl[node]); * @wq: the target workqueue
put_pwq_unlocked(dfl_pwq); * @attrs: the workqueue_attrs to apply, allocated with alloc_workqueue_attrs()
*
* Apply @attrs to an unbound workqueue @wq. Unless disabled, on NUMA
* machines, this function maps a separate pwq to each NUMA node with
* possibles CPUs in @attrs->cpumask so that work items are affine to the
* NUMA node it was issued on. Older pwqs are released as in-flight work
* items finish. Note that a work item which repeatedly requeues itself
* back-to-back will stay on its current pwq.
*
* Performs GFP_KERNEL allocations.
*
* Return: 0 on success and -errno on failure.
*/
int apply_workqueue_attrs(struct workqueue_struct *wq,
const struct workqueue_attrs *attrs)
{
struct apply_wqattrs_ctx *ctx;
int ret = -ENOMEM;
put_online_cpus(); /* only unbound workqueues can change attributes */
ret = 0; if (WARN_ON(!(wq->flags & WQ_UNBOUND)))
/* fall through */ return -EINVAL;
out_free:
free_workqueue_attrs(tmp_attrs);
free_workqueue_attrs(new_attrs);
kfree(pwq_tbl);
return ret;
enomem_pwq: /* creating multiple pwqs breaks ordering guarantee */
free_unbound_pwq(dfl_pwq); if (WARN_ON((wq->flags & __WQ_ORDERED) && !list_empty(&wq->pwqs)))
for_each_node(node) return -EINVAL;
if (pwq_tbl && pwq_tbl[node] != dfl_pwq)
free_unbound_pwq(pwq_tbl[node]); /*
* CPUs should stay stable across pwq creations and installations.
* Pin CPUs, determine the target cpumask for each node and create
* pwqs accordingly.
*/
get_online_cpus();
mutex_lock(&wq_pool_mutex);
ctx = apply_wqattrs_prepare(wq, attrs);
mutex_unlock(&wq_pool_mutex); mutex_unlock(&wq_pool_mutex);
/* the ctx has been prepared successfully, let's commit it */
if (ctx) {
apply_wqattrs_commit(ctx);
ret = 0;
}
put_online_cpus(); put_online_cpus();
enomem:
ret = -ENOMEM; apply_wqattrs_cleanup(ctx);
goto out_free;
return ret;
} }
/** /**
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册