Commit 2ef269ef authored by Dietmar Eggemann's avatar Dietmar Eggemann Committed by Tejun Heo

cgroup/cpuset: Free DL BW in case can_attach() fails

cpuset_can_attach() can fail. Postpone DL BW allocation until all tasks
have been checked. DL BW is not allocated per-task but as a sum over
all DL tasks migrating.

If multiple controllers are attached to the cgroup next to the cpuset
controller a non-cpuset can_attach() can fail. In this case free DL BW
in cpuset_cancel_attach().

Finally, update cpuset DL task count (nr_deadline_tasks) only in
cpuset_attach().
Suggested-by: default avatarWaiman Long <longman@redhat.com>
Signed-off-by: default avatarDietmar Eggemann <dietmar.eggemann@arm.com>
Signed-off-by: default avatarJuri Lelli <juri.lelli@redhat.com>
Reviewed-by: default avatarWaiman Long <longman@redhat.com>
Signed-off-by: default avatarTejun Heo <tj@kernel.org>
parent 85989106
...@@ -1852,7 +1852,7 @@ current_restore_flags(unsigned long orig_flags, unsigned long flags) ...@@ -1852,7 +1852,7 @@ current_restore_flags(unsigned long orig_flags, unsigned long flags)
} }
extern int cpuset_cpumask_can_shrink(const struct cpumask *cur, const struct cpumask *trial); extern int cpuset_cpumask_can_shrink(const struct cpumask *cur, const struct cpumask *trial);
extern int task_can_attach(struct task_struct *p, const struct cpumask *cs_effective_cpus); extern int task_can_attach(struct task_struct *p);
extern int dl_bw_alloc(int cpu, u64 dl_bw); extern int dl_bw_alloc(int cpu, u64 dl_bw);
extern void dl_bw_free(int cpu, u64 dl_bw); extern void dl_bw_free(int cpu, u64 dl_bw);
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
......
...@@ -198,6 +198,8 @@ struct cpuset { ...@@ -198,6 +198,8 @@ struct cpuset {
* know when to rebuild associated root domain bandwidth information. * know when to rebuild associated root domain bandwidth information.
*/ */
int nr_deadline_tasks; int nr_deadline_tasks;
int nr_migrate_dl_tasks;
u64 sum_migrate_dl_bw;
/* Invalid partition error code, not lock protected */ /* Invalid partition error code, not lock protected */
enum prs_errcode prs_err; enum prs_errcode prs_err;
...@@ -2496,16 +2498,23 @@ static int cpuset_can_attach_check(struct cpuset *cs) ...@@ -2496,16 +2498,23 @@ static int cpuset_can_attach_check(struct cpuset *cs)
return 0; return 0;
} }
static void reset_migrate_dl_data(struct cpuset *cs)
{
cs->nr_migrate_dl_tasks = 0;
cs->sum_migrate_dl_bw = 0;
}
/* Called by cgroups to determine if a cpuset is usable; cpuset_mutex held */ /* Called by cgroups to determine if a cpuset is usable; cpuset_mutex held */
static int cpuset_can_attach(struct cgroup_taskset *tset) static int cpuset_can_attach(struct cgroup_taskset *tset)
{ {
struct cgroup_subsys_state *css; struct cgroup_subsys_state *css;
struct cpuset *cs; struct cpuset *cs, *oldcs;
struct task_struct *task; struct task_struct *task;
int ret; int ret;
/* used later by cpuset_attach() */ /* used later by cpuset_attach() */
cpuset_attach_old_cs = task_cs(cgroup_taskset_first(tset, &css)); cpuset_attach_old_cs = task_cs(cgroup_taskset_first(tset, &css));
oldcs = cpuset_attach_old_cs;
cs = css_cs(css); cs = css_cs(css);
mutex_lock(&cpuset_mutex); mutex_lock(&cpuset_mutex);
...@@ -2516,7 +2525,7 @@ static int cpuset_can_attach(struct cgroup_taskset *tset) ...@@ -2516,7 +2525,7 @@ static int cpuset_can_attach(struct cgroup_taskset *tset)
goto out_unlock; goto out_unlock;
cgroup_taskset_for_each(task, css, tset) { cgroup_taskset_for_each(task, css, tset) {
ret = task_can_attach(task, cs->effective_cpus); ret = task_can_attach(task);
if (ret) if (ret)
goto out_unlock; goto out_unlock;
ret = security_task_setscheduler(task); ret = security_task_setscheduler(task);
...@@ -2524,11 +2533,31 @@ static int cpuset_can_attach(struct cgroup_taskset *tset) ...@@ -2524,11 +2533,31 @@ static int cpuset_can_attach(struct cgroup_taskset *tset)
goto out_unlock; goto out_unlock;
if (dl_task(task)) { if (dl_task(task)) {
cs->nr_deadline_tasks++; cs->nr_migrate_dl_tasks++;
cpuset_attach_old_cs->nr_deadline_tasks--; cs->sum_migrate_dl_bw += task->dl.dl_bw;
} }
} }
if (!cs->nr_migrate_dl_tasks)
goto out_success;
if (!cpumask_intersects(oldcs->effective_cpus, cs->effective_cpus)) {
int cpu = cpumask_any_and(cpu_active_mask, cs->effective_cpus);
if (unlikely(cpu >= nr_cpu_ids)) {
reset_migrate_dl_data(cs);
ret = -EINVAL;
goto out_unlock;
}
ret = dl_bw_alloc(cpu, cs->sum_migrate_dl_bw);
if (ret) {
reset_migrate_dl_data(cs);
goto out_unlock;
}
}
out_success:
/* /*
* Mark attach is in progress. This makes validate_change() fail * Mark attach is in progress. This makes validate_change() fail
* changes which zero cpus/mems_allowed. * changes which zero cpus/mems_allowed.
...@@ -2551,6 +2580,14 @@ static void cpuset_cancel_attach(struct cgroup_taskset *tset) ...@@ -2551,6 +2580,14 @@ static void cpuset_cancel_attach(struct cgroup_taskset *tset)
cs->attach_in_progress--; cs->attach_in_progress--;
if (!cs->attach_in_progress) if (!cs->attach_in_progress)
wake_up(&cpuset_attach_wq); wake_up(&cpuset_attach_wq);
if (cs->nr_migrate_dl_tasks) {
int cpu = cpumask_any(cs->effective_cpus);
dl_bw_free(cpu, cs->sum_migrate_dl_bw);
reset_migrate_dl_data(cs);
}
mutex_unlock(&cpuset_mutex); mutex_unlock(&cpuset_mutex);
} }
...@@ -2651,6 +2688,12 @@ static void cpuset_attach(struct cgroup_taskset *tset) ...@@ -2651,6 +2688,12 @@ static void cpuset_attach(struct cgroup_taskset *tset)
out: out:
cs->old_mems_allowed = cpuset_attach_nodemask_to; cs->old_mems_allowed = cpuset_attach_nodemask_to;
if (cs->nr_migrate_dl_tasks) {
cs->nr_deadline_tasks += cs->nr_migrate_dl_tasks;
oldcs->nr_deadline_tasks -= cs->nr_migrate_dl_tasks;
reset_migrate_dl_data(cs);
}
cs->attach_in_progress--; cs->attach_in_progress--;
if (!cs->attach_in_progress) if (!cs->attach_in_progress)
wake_up(&cpuset_attach_wq); wake_up(&cpuset_attach_wq);
...@@ -3330,7 +3373,7 @@ static int cpuset_can_fork(struct task_struct *task, struct css_set *cset) ...@@ -3330,7 +3373,7 @@ static int cpuset_can_fork(struct task_struct *task, struct css_set *cset)
if (ret) if (ret)
goto out_unlock; goto out_unlock;
ret = task_can_attach(task, cs->effective_cpus); ret = task_can_attach(task);
if (ret) if (ret)
goto out_unlock; goto out_unlock;
......
...@@ -9294,8 +9294,7 @@ int cpuset_cpumask_can_shrink(const struct cpumask *cur, ...@@ -9294,8 +9294,7 @@ int cpuset_cpumask_can_shrink(const struct cpumask *cur,
return ret; return ret;
} }
int task_can_attach(struct task_struct *p, int task_can_attach(struct task_struct *p)
const struct cpumask *cs_effective_cpus)
{ {
int ret = 0; int ret = 0;
...@@ -9308,21 +9307,9 @@ int task_can_attach(struct task_struct *p, ...@@ -9308,21 +9307,9 @@ int task_can_attach(struct task_struct *p,
* success of set_cpus_allowed_ptr() on all attached tasks * success of set_cpus_allowed_ptr() on all attached tasks
* before cpus_mask may be changed. * before cpus_mask may be changed.
*/ */
if (p->flags & PF_NO_SETAFFINITY) { if (p->flags & PF_NO_SETAFFINITY)
ret = -EINVAL; ret = -EINVAL;
goto out;
}
if (dl_task(p) && !cpumask_intersects(task_rq(p)->rd->span,
cs_effective_cpus)) {
int cpu = cpumask_any_and(cpu_active_mask, cs_effective_cpus);
if (unlikely(cpu >= nr_cpu_ids))
return -EINVAL;
ret = dl_bw_alloc(cpu, p->dl.dl_bw);
}
out:
return ret; return ret;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment