On Mon, Sep 17, 2007 at 09:45:59PM +0200, Oleg Verych wrote:
> The copy list, removed by Ingo is restored. Playing fair game, Willy!
Sorry Oleg, I don't understand why you added me to this thread. And I
don't understand at all what your intent was :-/
I've left others Cc too since they wonder like me.
Regards,
Willy
> Roman, please, find whole thread here:
> http://thread.gmane.org/gmane.linux.kernel/580665
>
> > From: Ingo Molnar <[email protected]>
> > Newsgroups: gmane.linux.kernel,gmane.linux.kernel.ck
> > Subject: Re: Scheduler benchmarks - a follow-up
> > Date: Mon, 17 Sep 2007 13:27:07 +0200
> []
> > In-Reply-To: <[email protected]>
> > User-Agent: Mutt/1.5.14 (2007-02-12)
> > Received-SPF: softfail (mx2: transitioning domain of elte.hu does not designate 157.181.1.14 as permitted sender) client-ip=157.181.1.14; [email protected]; helo=elvis.elte.hu;
> []
> > Archived-At: <http://permalink.gmane.org/gmane.linux.kernel/580689>
>
> >
> > * Rob Hussey <[email protected]> wrote:
> >
> >> Hi all,
> >>
> >> After posting some benchmarks involving cfs
> >> (http://lkml.org/lkml/2007/9/13/385), I got some feedback, so I
> >> decided to do a follow-up that'll hopefully fill in the gaps many
> >> people wanted to see filled.
> >
> > thanks for the update!
> >
> >> I'll start with some selected numbers, which are preceded by the
> >> command used for the benchmark.
> >>
> >> for((i=2; i < 201; i++)); do lat_ctx -s 0 $i; done:
> >> (the left most column is the number of processes ($i))
> >>
> >> 2.6.21 2.6.22-ck1 2.6.23-rc6-cfs-devel
> >>
> >> 15 5.88 4.85 5.14
> >> 16 5.80 4.77 4.76
> >
> > the unbound results are harder to compare because CFS changed SMP
> > balancing to saturate multiple cores better - but this can result in a
> > micro-benchmark slowdown if the other core is idle (and one of the
> > benchmark tasks runs on one core and the other runs on the first core).
> > This affects lat_ctx and pipe-test. (I'll have a look at the hackbench
> > behavior.)
> >
> >> Bound to Single core:
> >
> > these are the more comparable (apples to apples) tests. Usually the most
> > stable of them is pipe-test:
> >
> >> pipe-test:
> >>
> >> 2.6.21 2.6.22-ck1 2.6.23-rc6-cfs-devel
> >>
> >> 1 9.27 8.50 8.55
> >> 2 9.27 8.47 8.55
> >> 3 9.28 8.47 8.54
> >> 4 9.28 8.48 8.54
> >> 5 9.28 8.48 8.54
> >
> > so -ck1 is 0.8% faster in this particular test. (but still, there can be
> > caching effects in either direction - so i usually run the test on both
> > cores/CPUs to see whether there's any systematic spread in the results.
> > The cache-layout related random spread can be as high as 10% on some
> > systems!)
> >
> > many things happened between 2.6.22-ck1 and 2.6.23-cfs-devel that could
> > affect performance of this test. My initial guess would be sched_clock()
> > overhead. Could you send me your system's 'dmesg' output when running a
> > 2.6.22 (or -ck1) kernel? Chances are that your TSC got marked unstable,
> > this turns on a much less precise but also faster sched_clock()
> > implementation. CFS uses the TSC even if the time-of-day code marked it
> > as unstable - going for the more precise but slightly slower variant.
> >
> > To test this theory, could you apply the patch below to cfs-devel (if
> > you are interested in further testing this) - this changes the cfs-devel
> > version of sched_clock() to have a low-resolution fallback like v2.6.22
> > does. Does this result in any measurable increase in performance?
> >
> > (there's also a new sched-devel.git tree out there - if you update to it
> > you'll need to re-pull it against a pristine Linus git head.)
> >
> > Ingo
> >
> > ---
> > arch/i386/kernel/tsc.c | 4 ++--
> > 1 file changed, 2 insertions(+), 2 deletions(-)
> >
> > Index: linux/arch/i386/kernel/tsc.c
> >===================================================================
> > --- linux.orig/arch/i386/kernel/tsc.c
> > +++ linux/arch/i386/kernel/tsc.c
> > @@ -110,9 +110,9 @@ unsigned long long native_sched_clock(vo
> > * very important for it to be as fast as the platform
> > * can achive it. )
> > */
> > - if (unlikely(!tsc_enabled && !tsc_unstable))
> > + if (1 || unlikely(!tsc_enabled && !tsc_unstable))
> > /* No locking but a rare wrong value is not a big deal: */
> > - return (jiffies_64 - INITIAL_JIFFIES) * (1000000000 / HZ);
> > + return jiffies_64 * (1000000000 / HZ);
> >
> > /* read the Time Stamp Counter: */
> > rdtscll(this_offset);
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
[Index of Archives]
[Kernel Newbies]
[Netfilter]
[Bugtraq]
[Photo]
[Stuff]
[Gimp]
[Yosemite News]
[MIPS Linux]
[ARM Linux]
[Linux Security]
[Linux RAID]
[Video 4 Linux]
[Linux for the blind]
[Linux Resources]