NetBSD-Users archive
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]
Re: Networking: Lots of collisions?
On Oct 8, 2010, at 11:12 PM, Fredrik Pettai wrote:
> Hi,
Forgot to mention that its NetBSD/i386.
Looking at vmstat, I can see some things that stand out more than others:
# vmstat
procs memory page disks faults cpu
r b w avm fre flt re pi po fr sr f0 c0 in sy cs us sy id
1 0 0 316088 158332 208 1 0 0 13 53 0 0 836 1763 1678 0 4 96
one of two processes are occasionally in the run queue and during that many
page faults surface:
[...]
0 0 0 316092 158336 0 0 0 0 0 0 0 0 1538 3006 3144 0 6 94
2 0 0 316092 158328 1261 0 0 0 0 0 0 0 1576 3039 3032 8 8 84
0 0 0 316092 158328 0 0 0 0 0 0 0 0 1527 2900 3120 0 3 97
> I just installed a netbsd-5-1-RC4 as a dns server, and I see a lot of
> collisions:
>
> pcn0 in pcn0 out total in total out
> packets errs packets errs colls packets errs packets errs colls
> 39428897 0 22000706 0 5500180 39428897 0 22001100 0 5500180
> 3227 0 1892 0 474 3227 0 1892 0 474
> 3373 0 2060 0 514 3373 0 2060 0 514
> 3168 0 1926 0 482 3168 0 1926 0 482
>
> Now, since it's running in VMware, one could guess that it's a underlying
> problem (in VMware or maybe even in the physical infrastructure).
> But I also have virtualized Linux machines that are quite busy too, and they
> don't show this kind of networking problem.
> (They run in the same VMware hardware)
>
> Trying to do a tcpdump shows that the netbsd system doesn't handle that very
> well either:
>
> # tcpdump -i pcn0
> [...]
> ^C
> 5 packets captured
> 2585 packets received by filter
> 1726 packets dropped by kernel
>
> Doing it on the Linux machine works fine:
>
> # tcpdump -i eth0
> [...]
> ^C
> 2844 packets captured
> 2845 packets received by filter
> 0 packets dropped by kernel
>
> To that I might add that the servers doesn't have any typical CPU load etc.
>
> # top -o cpu
> load averages: 0.59, 0.65, 0.65; up 0+12:32:18
> 23:05:05
> 24 processes: 23 sleeping, 1 on CPU
> CPU states: 0.0% user, 0.0% nice, 2.0% system, 2.0% interrupt, 96.0% idle
> Memory: 306M Act, 2852K Inact, 6040K Wired, 7980K Exec, 117M File, 155M Free
> Swap: 256M Total, 256M Free
> PID USERNAME PRI NICE SIZE RES STATE TIME WCPU CPU COMMAND
> 3929 user 85 0 94M 91M netio 20:49 2.69% 2.69% [dns process]
>
> Anybody else that has seen something similar? (in VMware?)
> Any hints on what to do to make the networking stack more optimized? It's
> currently just the defaults.
>
> /P
Home |
Main Index |
Thread Index |
Old Index