Re: [PATCH v5 0/9] numa,sched,mm: pseudo-interleaving for automatic NUMA balancing

2014-01-27 Thread Chegu Vinod
had some performance testing, but it could always benefit from more testing, on other systems. Changes since v4: - remove some code that did not help performance - implement all the cleanups suggested by Mel Gorman - lots more testing, by Chegu Vinod and myself - rebase against -tip instead

Re: [PATCH v2 0/7] pseudo-interleaving for automatic NUMA balancing

2014-01-18 Thread Chegu Vinod
had some performance testing, but it could always benefit from more testing, on other systems. Changes since v1: - fix divide by zero found by Chegu Vinod - improve comment, as suggested by Peter Zijlstra - do stats calculations in task_numa_placement in local variables Some performance

Re: [PATCH 0/63] Basic scheduler support for automatic NUMA balancing V8

2013-10-23 Thread Chegu Vinod
Mel Gorman suse.de> writes: > > Weighing in at 63 patches makes the term "basic" in the series title is > a misnomer. > > This series still has roughly the same goals as previous versions. It > reduces overhead of automatic balancing through scan rate reduction > and the avoidance of TLB flushe

Re: kvm_intel: Could not allocate 42 bytes percpu data

2013-07-02 Thread Chegu Vinod
On 7/1/2013 10:49 PM, Rusty Russell wrote: Chegu Vinod writes: On 6/30/2013 11:22 PM, Rusty Russell wrote: Chegu Vinod writes: Hello, Lots (~700+) of the following messages are showing up in the dmesg of a 3.10-rc1 based kernel (Host OS is running on a large socket count box with HT-on

Re: kvm_intel: Could not allocate 42 bytes percpu data

2013-07-01 Thread Chegu Vinod
On 6/30/2013 11:22 PM, Rusty Russell wrote: Chegu Vinod writes: Hello, Lots (~700+) of the following messages are showing up in the dmesg of a 3.10-rc1 based kernel (Host OS is running on a large socket count box with HT-on). [ 82.270682] PERCPU: allocation failed, size=42 align=16, alloc

kvm_intel: Could not allocate 42 bytes percpu data

2013-06-24 Thread Chegu Vinod
Hello, Lots (~700+) of the following messages are showing up in the dmesg of a 3.10-rc1 based kernel (Host OS is running on a large socket count box with HT-on). [ 82.270682] PERCPU: allocation failed, size=42 align=16, alloc from reserved chunk failed [ 82.272633] kvm_intel: Could not

Re: [PATCH 3/2] vfio: Provide module option to disable vfio_iommu_type1 hugepage support

2013-05-30 Thread Chegu Vinod
} + if (unlikely(disable_hugepages)) { + vfio_lock_acct(1); + return 1; + } + /* Lock all the consecutive pages from pfn_base */ for (i = 1, vaddr += PAGE_SIZE; i < npage; i++, vaddr += PAGE_SIZE) { unsigned long pfn = 0; . Te

[PATCH] KVM: x86: Increase the "hard" max VCPU limit

2013-04-27 Thread Chegu Vinod
tapov http://article.gmane.org/gmane.comp.emulators.kvm.devel/99713 ) Signed-off-by: Chegu Vinod --- arch/x86/include/asm/kvm_host.h |2 +- 1 files changed, 1 insertions(+), 1 deletions(-) diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h index 4979778..bc57

Re: Preemptable Ticket Spinlock

2013-04-22 Thread Chegu Vinod
On 4/22/2013 1:50 PM, Jiannan Ouyang wrote: On Mon, Apr 22, 2013 at 4:44 PM, Peter Zijlstra wrote: On Mon, 2013-04-22 at 16:32 -0400, Rik van Riel wrote: IIRC one of the reasons was that the performance improvement wasn't as obvious. Rescheduling VCPUs takes a fair amount of time, quite proba

Re: [PATCH RFC 2/2] kvm: Iterate over only vcpus that are preempted

2013-03-05 Thread Chegu Vinod
continue; if (vcpu == me) continue; if (waitqueue_active(&vcpu->wq)) . Reviewed-by: Chegu Vinod -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to

Re: [PATCH RFC 1/2] kvm: Record the preemption status of vcpus using preempt notifiers

2013-03-05 Thread Chegu Vinod
tate == TASK_RUNNING) + vcpu->preempted = true; kvm_arch_vcpu_put(vcpu); } . Reviewed-by: Chegu Vinod -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kern

Re: [PATCH -v4 0/5] x86,smp: make ticket spinlock proportional backoff w/ auto tuning

2013-02-03 Thread Chegu Vinod
On 1/25/2013 11:05 AM, Rik van Riel wrote: Many spinlocks are embedded in data structures; having many CPUs pounce on the cache line the lock is in will slow down the lock holder, and can cause system performance to fall off a cliff. The paper "Non-scalable locks are dangerous" is a good referen

Re: [PATCH 0/5] x86,smp: make ticket spinlock proportional backoff w/ auto tuning

2013-01-10 Thread Chegu Vinod
On 1/8/2013 2:26 PM, Rik van Riel wrote: <...> Performance is within the margin of error of v2, so the graph has not been update. Please let me know if you manage to break this code in any way, so I can fix it... Attached below is some preliminary data with one of the AIM7 micro-benchmark wor

Re: [PATCH V3 RFC 1/2] sched: Bail out of yield_to when source and target runqueue has one task

2012-11-28 Thread Chegu Vinod
On 11/28/2012 5:09 PM, Chegu Vinod wrote: On 11/27/2012 6:23 AM, Chegu Vinod wrote: On 11/27/2012 2:30 AM, Raghavendra K T wrote: On 11/26/2012 07:05 PM, Andrew Jones wrote: On Mon, Nov 26, 2012 at 05:37:54PM +0530, Raghavendra K T wrote: From: Peter Zijlstra In case of undercomitted

Re: [PATCH V3 RFC 0/2] kvm: Improving undercommit scenarios

2012-11-28 Thread Chegu Vinod
ons(-) . Tested-by: Chegu Vinod -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/

Re: [PATCH V3 RFC 1/2] sched: Bail out of yield_to when source and target runqueue has one task

2012-11-27 Thread Chegu Vinod
On 11/27/2012 2:30 AM, Raghavendra K T wrote: On 11/26/2012 07:05 PM, Andrew Jones wrote: On Mon, Nov 26, 2012 at 05:37:54PM +0530, Raghavendra K T wrote: From: Peter Zijlstra In case of undercomitted scenarios, especially in large guests yield_to overhead is significantly high. when run queu

Re: [PATCH RFC 0/2] kvm: Improving undercommit,overcommit scenarios in PLE handler

2012-09-21 Thread Chegu Vinod
On 9/21/2012 4:59 AM, Raghavendra K T wrote: In some special scenarios like #vcpu <= #pcpu, PLE handler may prove very costly, Yes. because there is no need to iterate over vcpus and do unsuccessful yield_to burning CPU. An idea to solve this is: 1) As Avi had proposed we can modify hardwar