NetBSD-Users archive
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]
Re: Networking: Lots of collisions?
Hi,
NetBSD (and FreeBSD too) seem to have trouble with the default network
adapter presented by VMWare. I'd recommend shutting own the machine and
change the Operating System type to "Other (64-bit)". Then you'll get an
option to install an e1000 adapter. This will show up in NetBSD as wm0
(not pcn0) and works a lot better. On my machine:
PTI:~# netstat -I wm0
Name Mtu Network Address Ipkts Ierrs Opkts Oerrs
Colls
wm0 1500 <Link> 00:0c:29:ab:50:b4 7237520 0 14990168 0
0
Hope this helps.
Jason M.
> On Oct 8, 2010, at 11:12 PM, Fredrik Pettai wrote:
>> Hi,
>
> Forgot to mention that its NetBSD/i386.
>
> Looking at vmstat, I can see some things that stand out more than others:
>
> # vmstat
> procs memory page disks faults cpu
> r b w avm fre flt re pi po fr sr f0 c0 in sy cs us sy
> id
> 1 0 0 316088 158332 208 1 0 0 13 53 0 0 836 1763 1678 0 4
> 96
>
> one of two processes are occasionally in the run queue and during that
> many page faults surface:
>
> [...]
> 0 0 0 316092 158336 0 0 0 0 0 0 0 0 1538 3006 3144 0 6
> 94
> 2 0 0 316092 158328 1261 0 0 0 0 0 0 0 1576 3039 3032 8 8
> 84
> 0 0 0 316092 158328 0 0 0 0 0 0 0 0 1527 2900 3120 0 3
> 97
>
>> I just installed a netbsd-5-1-RC4 as a dns server, and I see a lot of
>> collisions:
>>
>> pcn0 in pcn0 out total in total out
>> packets errs packets errs colls packets errs packets errs colls
>> 39428897 0 22000706 0 5500180 39428897 0 22001100 0
>> 5500180
>> 3227 0 1892 0 474 3227 0 1892 0 474
>> 3373 0 2060 0 514 3373 0 2060 0 514
>> 3168 0 1926 0 482 3168 0 1926 0 482
>>
>> Now, since it's running in VMware, one could guess that it's a
>> underlying problem (in VMware or maybe even in the physical
>> infrastructure).
>> But I also have virtualized Linux machines that are quite busy too, and
>> they don't show this kind of networking problem.
>> (They run in the same VMware hardware)
>>
>> Trying to do a tcpdump shows that the netbsd system doesn't handle that
>> very well either:
>>
>> # tcpdump -i pcn0
>> [...]
>> ^C
>> 5 packets captured
>> 2585 packets received by filter
>> 1726 packets dropped by kernel
>>
>> Doing it on the Linux machine works fine:
>>
>> # tcpdump -i eth0
>> [...]
>> ^C
>> 2844 packets captured
>> 2845 packets received by filter
>> 0 packets dropped by kernel
>>
>> To that I might add that the servers doesn't have any typical CPU load
>> etc.
>>
>> # top -o cpu
>> load averages: 0.59, 0.65, 0.65; up 0+12:32:18
>> 23:05:05
>> 24 processes: 23 sleeping, 1 on CPU
>> CPU states: 0.0% user, 0.0% nice, 2.0% system, 2.0% interrupt, 96.0%
>> idle
>> Memory: 306M Act, 2852K Inact, 6040K Wired, 7980K Exec, 117M File, 155M
>> Free
>> Swap: 256M Total, 256M Free
>> PID USERNAME PRI NICE SIZE RES STATE TIME WCPU CPU
>> COMMAND
>> 3929 user 85 0 94M 91M netio 20:49 2.69% 2.69% [dns
>> process]
>>
>> Anybody else that has seen something similar? (in VMware?)
>> Any hints on what to do to make the networking stack more optimized?
>> It's currently just the defaults.
>>
>> /P
>
>
Home |
Main Index |
Thread Index |
Old Index