NetBSD-Users archive
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]
Re: Networking: Lots of collisions?
Thanks for tip Jason,
I did as you said, and it seems to have removed the collisions.
However, I'm sure the problem/bottleneck is gone. tcpdump still loose most of
the packets.
7 packets captured
2733 packets received by filter
1872 packets dropped by kernel
/P
On Oct 9, 2010, at 12:15 AM, jmitchel%bigjar.com@localhost wrote:
> I forgot to mention that you don't have to reinstall when you change the
> OS type in VMWare, it's just that VMWare only offers the e1000 adapter if
> the OS type is "Other (64-bit)". I don't know why that is.
>
> Jason M.
>
>> On Oct 8, 2010, at 11:12 PM, Fredrik Pettai wrote:
>>> Hi,
>>
>> Forgot to mention that its NetBSD/i386.
>>
>> Looking at vmstat, I can see some things that stand out more than others:
>>
>> # vmstat
>> procs memory page disks faults cpu
>> r b w avm fre flt re pi po fr sr f0 c0 in sy cs us sy
>> id
>> 1 0 0 316088 158332 208 1 0 0 13 53 0 0 836 1763 1678 0 4
>> 96
>>
>> one of two processes are occasionally in the run queue and during that
>> many page faults surface:
>>
>> [...]
>> 0 0 0 316092 158336 0 0 0 0 0 0 0 0 1538 3006 3144 0 6
>> 94
>> 2 0 0 316092 158328 1261 0 0 0 0 0 0 0 1576 3039 3032 8 8
>> 84
>> 0 0 0 316092 158328 0 0 0 0 0 0 0 0 1527 2900 3120 0 3
>> 97
>>
>>> I just installed a netbsd-5-1-RC4 as a dns server, and I see a lot of
>>> collisions:
>>>
>>> pcn0 in pcn0 out total in total out
>>> packets errs packets errs colls packets errs packets errs colls
>>> 39428897 0 22000706 0 5500180 39428897 0 22001100 0
>>> 5500180
>>> 3227 0 1892 0 474 3227 0 1892 0 474
>>> 3373 0 2060 0 514 3373 0 2060 0 514
>>> 3168 0 1926 0 482 3168 0 1926 0 482
>>>
>>> Now, since it's running in VMware, one could guess that it's a
>>> underlying problem (in VMware or maybe even in the physical
>>> infrastructure).
>>> But I also have virtualized Linux machines that are quite busy too, and
>>> they don't show this kind of networking problem.
>>> (They run in the same VMware hardware)
>>>
>>> Trying to do a tcpdump shows that the netbsd system doesn't handle that
>>> very well either:
>>>
>>> # tcpdump -i pcn0
>>> [...]
>>> ^C
>>> 5 packets captured
>>> 2585 packets received by filter
>>> 1726 packets dropped by kernel
>>>
>>> Doing it on the Linux machine works fine:
>>>
>>> # tcpdump -i eth0
>>> [...]
>>> ^C
>>> 2844 packets captured
>>> 2845 packets received by filter
>>> 0 packets dropped by kernel
>>>
>>> To that I might add that the servers doesn't have any typical CPU load
>>> etc.
>>>
>>> # top -o cpu
>>> load averages: 0.59, 0.65, 0.65; up 0+12:32:18
>>> 23:05:05
>>> 24 processes: 23 sleeping, 1 on CPU
>>> CPU states: 0.0% user, 0.0% nice, 2.0% system, 2.0% interrupt, 96.0%
>>> idle
>>> Memory: 306M Act, 2852K Inact, 6040K Wired, 7980K Exec, 117M File, 155M
>>> Free
>>> Swap: 256M Total, 256M Free
>>> PID USERNAME PRI NICE SIZE RES STATE TIME WCPU CPU
>>> COMMAND
>>> 3929 user 85 0 94M 91M netio 20:49 2.69% 2.69% [dns
>>> process]
>>>
>>> Anybody else that has seen something similar? (in VMware?)
>>> Any hints on what to do to make the networking stack more optimized?
>>> It's currently just the defaults.
>>>
>>> /P
>>
>>
>
Home |
Main Index |
Thread Index |
Old Index