mirror of
https://github.com/torvalds/linux.git
synced 2024-12-29 06:12:08 +00:00
workqueue: split apply_workqueue_attrs() into 3 stages
Current apply_workqueue_attrs() includes pwqs-allocation and pwqs-installation, so when we batch multiple apply_workqueue_attrs()s as a transaction, we can't ensure the transaction must succeed or fail as a complete unit. To solve this, we split apply_workqueue_attrs() into three stages. The first stage does the preparation: allocation memory, pwqs. The second stage does the attrs-installaion and pwqs-installation. The third stage frees the allocated memory and (old or unused) pwqs. As the result, batching multiple apply_workqueue_attrs()s can succeed or fail as a complete unit: 1) batch do all the first stage for all the workqueues 2) only commit all when all the above succeed. This patch is a preparation for the next patch ("Allow modifying low level unbound workqueue cpumask") which will do a multiple apply_workqueue_attrs(). The patch doesn't have functionality changed except two minor adjustment: 1) free_unbound_pwq() for the error path is removed, we use the heavier version put_pwq_unlocked() instead since the error path is rare. this adjustment simplifies the code. 2) the memory-allocation is also moved into wq_pool_mutex. this is needed to avoid to do the further splitting. tj: minor updates to comments. Suggested-by: Tejun Heo <tj@kernel.org> Cc: Christoph Lameter <cl@linux.com> Cc: Kevin Hilman <khilman@linaro.org> Cc: Lai Jiangshan <laijs@cn.fujitsu.com> Cc: Mike Galbraith <bitbucket@online.de> Cc: Paul E. McKenney <paulmck@linux.vnet.ibm.com> Cc: Tejun Heo <tj@kernel.org> Cc: Viresh Kumar <viresh.kumar@linaro.org> Cc: Frederic Weisbecker <fweisbec@gmail.com> Signed-off-by: Lai Jiangshan <laijs@cn.fujitsu.com> Signed-off-by: Tejun Heo <tj@kernel.org>
This commit is contained in:
parent
b787f68c36
commit
2d5f0764b5
@ -3425,17 +3425,6 @@ static struct pool_workqueue *alloc_unbound_pwq(struct workqueue_struct *wq,
|
||||
return pwq;
|
||||
}
|
||||
|
||||
/* undo alloc_unbound_pwq(), used only in the error path */
|
||||
static void free_unbound_pwq(struct pool_workqueue *pwq)
|
||||
{
|
||||
lockdep_assert_held(&wq_pool_mutex);
|
||||
|
||||
if (pwq) {
|
||||
put_unbound_pool(pwq->pool);
|
||||
kmem_cache_free(pwq_cache, pwq);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* wq_calc_node_mask - calculate a wq_attrs' cpumask for the specified node
|
||||
* @attrs: the wq_attrs of interest
|
||||
@ -3498,6 +3487,114 @@ static struct pool_workqueue *numa_pwq_tbl_install(struct workqueue_struct *wq,
|
||||
return old_pwq;
|
||||
}
|
||||
|
||||
/* context to store the prepared attrs & pwqs before applying */
|
||||
struct apply_wqattrs_ctx {
|
||||
struct workqueue_struct *wq; /* target workqueue */
|
||||
struct workqueue_attrs *attrs; /* attrs to apply */
|
||||
struct pool_workqueue *dfl_pwq;
|
||||
struct pool_workqueue *pwq_tbl[];
|
||||
};
|
||||
|
||||
/* free the resources after success or abort */
|
||||
static void apply_wqattrs_cleanup(struct apply_wqattrs_ctx *ctx)
|
||||
{
|
||||
if (ctx) {
|
||||
int node;
|
||||
|
||||
for_each_node(node)
|
||||
put_pwq_unlocked(ctx->pwq_tbl[node]);
|
||||
put_pwq_unlocked(ctx->dfl_pwq);
|
||||
|
||||
free_workqueue_attrs(ctx->attrs);
|
||||
|
||||
kfree(ctx);
|
||||
}
|
||||
}
|
||||
|
||||
/* allocate the attrs and pwqs for later installation */
|
||||
static struct apply_wqattrs_ctx *
|
||||
apply_wqattrs_prepare(struct workqueue_struct *wq,
|
||||
const struct workqueue_attrs *attrs)
|
||||
{
|
||||
struct apply_wqattrs_ctx *ctx;
|
||||
struct workqueue_attrs *new_attrs, *tmp_attrs;
|
||||
int node;
|
||||
|
||||
lockdep_assert_held(&wq_pool_mutex);
|
||||
|
||||
ctx = kzalloc(sizeof(*ctx) + nr_node_ids * sizeof(ctx->pwq_tbl[0]),
|
||||
GFP_KERNEL);
|
||||
|
||||
new_attrs = alloc_workqueue_attrs(GFP_KERNEL);
|
||||
tmp_attrs = alloc_workqueue_attrs(GFP_KERNEL);
|
||||
if (!ctx || !new_attrs || !tmp_attrs)
|
||||
goto out_free;
|
||||
|
||||
/* make a copy of @attrs and sanitize it */
|
||||
copy_workqueue_attrs(new_attrs, attrs);
|
||||
cpumask_and(new_attrs->cpumask, new_attrs->cpumask, cpu_possible_mask);
|
||||
|
||||
/*
|
||||
* We may create multiple pwqs with differing cpumasks. Make a
|
||||
* copy of @new_attrs which will be modified and used to obtain
|
||||
* pools.
|
||||
*/
|
||||
copy_workqueue_attrs(tmp_attrs, new_attrs);
|
||||
|
||||
/*
|
||||
* If something goes wrong during CPU up/down, we'll fall back to
|
||||
* the default pwq covering whole @attrs->cpumask. Always create
|
||||
* it even if we don't use it immediately.
|
||||
*/
|
||||
ctx->dfl_pwq = alloc_unbound_pwq(wq, new_attrs);
|
||||
if (!ctx->dfl_pwq)
|
||||
goto out_free;
|
||||
|
||||
for_each_node(node) {
|
||||
if (wq_calc_node_cpumask(attrs, node, -1, tmp_attrs->cpumask)) {
|
||||
ctx->pwq_tbl[node] = alloc_unbound_pwq(wq, tmp_attrs);
|
||||
if (!ctx->pwq_tbl[node])
|
||||
goto out_free;
|
||||
} else {
|
||||
ctx->dfl_pwq->refcnt++;
|
||||
ctx->pwq_tbl[node] = ctx->dfl_pwq;
|
||||
}
|
||||
}
|
||||
|
||||
ctx->attrs = new_attrs;
|
||||
ctx->wq = wq;
|
||||
free_workqueue_attrs(tmp_attrs);
|
||||
return ctx;
|
||||
|
||||
out_free:
|
||||
free_workqueue_attrs(tmp_attrs);
|
||||
free_workqueue_attrs(new_attrs);
|
||||
apply_wqattrs_cleanup(ctx);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
/* set attrs and install prepared pwqs, @ctx points to old pwqs on return */
|
||||
static void apply_wqattrs_commit(struct apply_wqattrs_ctx *ctx)
|
||||
{
|
||||
int node;
|
||||
|
||||
/* all pwqs have been created successfully, let's install'em */
|
||||
mutex_lock(&ctx->wq->mutex);
|
||||
|
||||
copy_workqueue_attrs(ctx->wq->unbound_attrs, ctx->attrs);
|
||||
|
||||
/* save the previous pwq and install the new one */
|
||||
for_each_node(node)
|
||||
ctx->pwq_tbl[node] = numa_pwq_tbl_install(ctx->wq, node,
|
||||
ctx->pwq_tbl[node]);
|
||||
|
||||
/* @dfl_pwq might not have been used, ensure it's linked */
|
||||
link_pwq(ctx->dfl_pwq);
|
||||
swap(ctx->wq->dfl_pwq, ctx->dfl_pwq);
|
||||
|
||||
mutex_unlock(&ctx->wq->mutex);
|
||||
}
|
||||
|
||||
/**
|
||||
* apply_workqueue_attrs - apply new workqueue_attrs to an unbound workqueue
|
||||
* @wq: the target workqueue
|
||||
@ -3517,9 +3614,8 @@ static struct pool_workqueue *numa_pwq_tbl_install(struct workqueue_struct *wq,
|
||||
int apply_workqueue_attrs(struct workqueue_struct *wq,
|
||||
const struct workqueue_attrs *attrs)
|
||||
{
|
||||
struct workqueue_attrs *new_attrs, *tmp_attrs;
|
||||
struct pool_workqueue **pwq_tbl, *dfl_pwq;
|
||||
int node, ret;
|
||||
struct apply_wqattrs_ctx *ctx;
|
||||
int ret = -ENOMEM;
|
||||
|
||||
/* only unbound workqueues can change attributes */
|
||||
if (WARN_ON(!(wq->flags & WQ_UNBOUND)))
|
||||
@ -3529,23 +3625,6 @@ int apply_workqueue_attrs(struct workqueue_struct *wq,
|
||||
if (WARN_ON((wq->flags & __WQ_ORDERED) && !list_empty(&wq->pwqs)))
|
||||
return -EINVAL;
|
||||
|
||||
pwq_tbl = kzalloc(nr_node_ids * sizeof(pwq_tbl[0]), GFP_KERNEL);
|
||||
new_attrs = alloc_workqueue_attrs(GFP_KERNEL);
|
||||
tmp_attrs = alloc_workqueue_attrs(GFP_KERNEL);
|
||||
if (!pwq_tbl || !new_attrs || !tmp_attrs)
|
||||
goto enomem;
|
||||
|
||||
/* make a copy of @attrs and sanitize it */
|
||||
copy_workqueue_attrs(new_attrs, attrs);
|
||||
cpumask_and(new_attrs->cpumask, new_attrs->cpumask, cpu_possible_mask);
|
||||
|
||||
/*
|
||||
* We may create multiple pwqs with differing cpumasks. Make a
|
||||
* copy of @new_attrs which will be modified and used to obtain
|
||||
* pools.
|
||||
*/
|
||||
copy_workqueue_attrs(tmp_attrs, new_attrs);
|
||||
|
||||
/*
|
||||
* CPUs should stay stable across pwq creations and installations.
|
||||
* Pin CPUs, determine the target cpumask for each node and create
|
||||
@ -3554,68 +3633,20 @@ int apply_workqueue_attrs(struct workqueue_struct *wq,
|
||||
get_online_cpus();
|
||||
|
||||
mutex_lock(&wq_pool_mutex);
|
||||
ctx = apply_wqattrs_prepare(wq, attrs);
|
||||
mutex_unlock(&wq_pool_mutex);
|
||||
|
||||
/*
|
||||
* If something goes wrong during CPU up/down, we'll fall back to
|
||||
* the default pwq covering whole @attrs->cpumask. Always create
|
||||
* it even if we don't use it immediately.
|
||||
*/
|
||||
dfl_pwq = alloc_unbound_pwq(wq, new_attrs);
|
||||
if (!dfl_pwq)
|
||||
goto enomem_pwq;
|
||||
|
||||
for_each_node(node) {
|
||||
if (wq_calc_node_cpumask(attrs, node, -1, tmp_attrs->cpumask)) {
|
||||
pwq_tbl[node] = alloc_unbound_pwq(wq, tmp_attrs);
|
||||
if (!pwq_tbl[node])
|
||||
goto enomem_pwq;
|
||||
} else {
|
||||
dfl_pwq->refcnt++;
|
||||
pwq_tbl[node] = dfl_pwq;
|
||||
}
|
||||
/* the ctx has been prepared successfully, let's commit it */
|
||||
if (ctx) {
|
||||
apply_wqattrs_commit(ctx);
|
||||
ret = 0;
|
||||
}
|
||||
|
||||
mutex_unlock(&wq_pool_mutex);
|
||||
|
||||
/* all pwqs have been created successfully, let's install'em */
|
||||
mutex_lock(&wq->mutex);
|
||||
|
||||
copy_workqueue_attrs(wq->unbound_attrs, new_attrs);
|
||||
|
||||
/* save the previous pwq and install the new one */
|
||||
for_each_node(node)
|
||||
pwq_tbl[node] = numa_pwq_tbl_install(wq, node, pwq_tbl[node]);
|
||||
|
||||
/* @dfl_pwq might not have been used, ensure it's linked */
|
||||
link_pwq(dfl_pwq);
|
||||
swap(wq->dfl_pwq, dfl_pwq);
|
||||
|
||||
mutex_unlock(&wq->mutex);
|
||||
|
||||
/* put the old pwqs */
|
||||
for_each_node(node)
|
||||
put_pwq_unlocked(pwq_tbl[node]);
|
||||
put_pwq_unlocked(dfl_pwq);
|
||||
|
||||
put_online_cpus();
|
||||
ret = 0;
|
||||
/* fall through */
|
||||
out_free:
|
||||
free_workqueue_attrs(tmp_attrs);
|
||||
free_workqueue_attrs(new_attrs);
|
||||
kfree(pwq_tbl);
|
||||
|
||||
apply_wqattrs_cleanup(ctx);
|
||||
|
||||
return ret;
|
||||
|
||||
enomem_pwq:
|
||||
free_unbound_pwq(dfl_pwq);
|
||||
for_each_node(node)
|
||||
if (pwq_tbl && pwq_tbl[node] != dfl_pwq)
|
||||
free_unbound_pwq(pwq_tbl[node]);
|
||||
mutex_unlock(&wq_pool_mutex);
|
||||
put_online_cpus();
|
||||
enomem:
|
||||
ret = -ENOMEM;
|
||||
goto out_free;
|
||||
}
|
||||
|
||||
/**
|
||||
|
Loading…
Reference in New Issue
Block a user