On 11/12/2014 11:46 AM, Andrew Cooper wrote:
On 12/11/14 10:40, Juergen Gross wrote:
Commit bac6334b51d9bcfe57ecf4a4cb5288348fcf044a (move domain to
cpupool0 before destroying it) introduced an error in the accounting
of cpupools regarding the number of domains. The number of domains
is nor adjusted when a domain is moved to cpupool0 in kill_domain().
Correct this by introducing a cpupool function doing the move
instead of open coding it by calling sched_move_domain().
Signed-off-by: Juergen Gross <jgr...@suse.com>
Tested-by: Dietmar Hahn <dietmar.h...@ts.fujitsu.com>
---
xen/common/cpupool.c | 47 +++++++++++++++++++++++++++++++++--------------
xen/common/domain.c | 2 +-
xen/include/xen/sched.h | 1 +
3 files changed, 35 insertions(+), 15 deletions(-)
diff --git a/xen/common/cpupool.c b/xen/common/cpupool.c
index 73249d3..c6e3869 100644
--- a/xen/common/cpupool.c
+++ b/xen/common/cpupool.c
@@ -225,6 +225,35 @@ static int cpupool_destroy(struct cpupool *c)
}
/*
+ * Move domain to another cpupool
+ */
+static int cpupool_move_domain_unlocked(struct domain *d, struct cpupool *c)
This isn't an unlocked function. It is strictly called with the
cpupool_lock held. Per prevailing style, it should be named
"__cpupool_move_domain()".
Umpf. Fingers faster than brain. :-)
+{
+ int ret;
+
+ d->cpupool->n_dom--;
+ ret = sched_move_domain(d, c);
+ if ( ret )
+ d->cpupool->n_dom++;
+ else
+ c->n_dom++;
+
+ return ret;
+}
Newline here please.
Once these two issues are fixed, content Reviewed-by: Andrew Cooper
<andrew.coop...@citrix.com>
+int cpupool_move_domain(struct domain *d, struct cpupool *c)
+{
+ int ret;
+
+ spin_lock(&cpupool_lock);
+
+ ret = cpupool_move_domain_unlocked(d, c);
+
+ spin_unlock(&cpupool_lock);
+
+ return ret;
+}
+
+/*
* assign a specific cpu to a cpupool
* cpupool_lock must be held
*/
@@ -338,14 +367,9 @@ static int cpupool_unassign_cpu(struct cpupool *c,
unsigned int cpu)
ret = -EBUSY;
break;
}
- c->n_dom--;
- ret = sched_move_domain(d, cpupool0);
+ ret = cpupool_move_domain_unlocked(d, cpupool0);
if ( ret )
- {
- c->n_dom++;
break;
- }
- cpupool0->n_dom++;
}
rcu_read_unlock(&domlist_read_lock);
if ( ret )
@@ -613,16 +637,11 @@ int cpupool_do_sysctl(struct xen_sysctl_cpupool_op *op)
d->domain_id, op->cpupool_id);
ret = -ENOENT;
spin_lock(&cpupool_lock);
+
c = cpupool_find_by_id(op->cpupool_id);
if ( (c != NULL) && cpumask_weight(c->cpu_valid) )
- {
- d->cpupool->n_dom--;
- ret = sched_move_domain(d, c);
- if ( ret )
- d->cpupool->n_dom++;
- else
- c->n_dom++;
- }
+ ret = cpupool_move_domain_unlocked(d, c);
+
spin_unlock(&cpupool_lock);
cpupool_dprintk("cpupool move_domain(dom=%d)->pool=%d ret %d\n",
d->domain_id, op->cpupool_id, ret);
diff --git a/xen/common/domain.c b/xen/common/domain.c
index a3f51ec..4a62c1d 100644
--- a/xen/common/domain.c
+++ b/xen/common/domain.c
@@ -621,7 +621,7 @@ int domain_kill(struct domain *d)
rc = -EAGAIN;
break;
}
- if ( sched_move_domain(d, cpupool0) )
+ if ( cpupool_move_domain(d, cpupool0) )
return -EAGAIN;
for_each_vcpu ( d, v )
unmap_vcpu_info(v);
diff --git a/xen/include/xen/sched.h b/xen/include/xen/sched.h
index c5157e6..46fc6e3 100644
--- a/xen/include/xen/sched.h
+++ b/xen/include/xen/sched.h
@@ -871,6 +871,7 @@ struct cpupool *cpupool_get_by_id(int poolid);
void cpupool_put(struct cpupool *pool);
int cpupool_add_domain(struct domain *d, int poolid);
void cpupool_rm_domain(struct domain *d);
+int cpupool_move_domain(struct domain *d, struct cpupool *c);
int cpupool_do_sysctl(struct xen_sysctl_cpupool_op *op);
void schedule_dump(struct cpupool *c);
extern void dump_runq(unsigned char key);
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xen.org
http://lists.xen.org/xen-devel
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xen.org
http://lists.xen.org/xen-devel