Subject: Re: "diff" loses on large files?
To: Mike Cheponis <mac@Wireless.Com>
From: Chuck Silvers <chuq@chuq.com>
List: netbsd-help
Date: 07/12/2001 06:44:47
hi,

it's been known for quite some time that gnu diff's algorithm isn't
so great for very large files, I thought there was even a PR about it
(though I can't find it now).

however the panic is something that definitely needs to be fixed,
and I'll look into that in 10 days when I get back from vacation
(if no one else fixes it while I'm gone).

-Chuck


On Thu, Jul 12, 2001 at 12:16:14AM -0700, Mike Cheponis wrote:
> > Date: Thu, 12 Jul 2001 12:30:58 +0200 (CEST)
> > From: Wojciech Puchar <wojtek@wojtek.3miasto.net>
> > To: Mike Cheponis <mac@Wireless.Com>
> > Subject: Re: "diff" loses on large files?
> >
> > > cpu0: AMD Athlon Model 4 (Thunderbird) (686-class), 1100.12 MHz
> > > total memory = 511 MB
> > > avail memory = 468 MB
> > > using 6574 buffers containing 26296 KB of memory
> > >
> > > $ ll *odx
> > > 3065307 -rw-r--r--  1 mac  user  661,969,123 Jul 11 21:45 bad.odx
> > > 3065315 -rw-r--r--  1 mac  user  661,969,123 Jul 11 21:45 ok.odx
> > >
> > > $ diff -H  *odx
> > > diff: memory exhausted
> >
> > ulimit -m / ulimit -d should do
> 
> Well.....
> 
> $ su
> Password:
> # ulimit -a
> cpu time (seconds)         unlimited
> file size (blocks)         unlimited
> data seg size (kbytes)     131072
> stack size (kbytes)        2048
> core file size (blocks)    unlimited
> resident set size (kbytes) 506156
> locked-in-memory size (kb) 168718
> processes                  80
> file descriptors           64
> # ulimit -d 999999
> # ulimit -a
> cpu time (seconds)         unlimited
> file size (blocks)         unlimited
> data seg size (kbytes)     999999
> stack size (kbytes)        2048
> core file size (blocks)    unlimited
> resident set size (kbytes) 506156
> locked-in-memory size (kb) 168718
> processes                  80
> file descriptors           64
> S# diff *odx
> 
> At this point, the kernel panics:
> 
> uvm_fault (0xe3846468, 0x0, 0, 3) -> 1
> kernel: page fault trap, code=0
> stopped in sendmail at uvm_swap_markbad+0xf   addl %ebx, 0x24(%eax)
> 
> Trace produces:
> 
> uvm_swap_markbad+0xf
> uvmfault_anonget+0x262
> uvm_fault+0x943
> trap() at trap+0x413
> -- trap (number 6) --
> 0x480ff95c
> 
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> Now, I did set the data seg size > my available memory.  But should this
> cause a panic?  Seems a rather rude thing to do, panic like that!
> 
> And, to get back to my original issue: diff is braindead.  It should
> -automatically- grab more memory if it can, and, if it can't, it should
> figure out how to use /tmp or some other part of the filesystem if it's
> algorithms are so pathetic that it needs so darn much memory.
> 
> Thanks again -Mike
> 
> 
> p.s. At first I typed in "ulimit -m / ulimit -d should do" at the command
> line (it's late...).
> 
> This had the effect of interpreting "-m /" as "-m 0" and therefore, after
> that command, no other command could be executed.  This would appear to
> be another serious bug.  (fwiw, I'm using "zsh")