Subject: Re: lmbench on Cobalt box
To: Jason R Thorpe <thorpej@zembu.com>
From: David Brownlee <abs@NetBSD.ORG>
List: port-mips
Date: 04/09/2000 12:47:41
On Sat, 8 Apr 2000, Jason R Thorpe wrote:

> On Sat, Apr 08, 2000 at 06:13:47PM +0200, Soren S. Jorvang wrote:
> 
>  > > The numbers don't seem too bad compared with the DEC 60MHz R4400 boxes,
>  > > except the the mmap latency is ~15 times faster.  I've been seeing quite
>  > > shocking figures on both pmax and alpha - that figure is amazingly good.
>  > 
>  > I am rather suspicious of some of those mmap latency figures.
> 
> Me too.  If you look at lat_mmap.c, it doens't just mmap memory, as I
> recall.  It faults the pages in, as well.  Ah yes, just checked, it does
> fault the pages in.
> 
> So, any system that does idle-zeroing of pages on the free list is going
> to basically kick ass on that test, and any system that doesn't is going
> to incur the penalty of ZFOD when those pages are touched.
> 
> So, it is in fact not an "mmap latency" test but rather an "mmap + ZFOD
> latency" test.  And, more than that, it's also testing the latency of
> setting setting up page table mappings for systems that do that lazily.
> 
> And even if the test is labeled correctly, it's not precisely fair.  A
> system with a virtually tagged cache would be foolish to pre-zero pages
> in the idle loop; that cache load would be wasted, and could potentially
> throw other useful data out of the cache (unless the access was done
> un-cached).
> 
> Basically, the test is skewed very much in favor of Linux/i386.  The i386's
> cache is physically tagged (so it is at least not *wasting* the cache load),
> and Linux's VM system sets up all of the page tables when the mmap system
> call occurs.

	Maybe someone could some up with another test that is fair for
	machines with virtually tagged cache, and see if they can get it
	into lmbench? - or if not leave it in a patch in pkgsrc :)

		David/absolute