[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]
Re: New NetBSD user - ALTQ confusion
On Fri, Dec 17, 2010 at 02:56:51PM +0000, Gerard Lally wrote:
> On Fri, 17 Dec 2010 15:19:30 +0100
> Manuel Bouyer <bouyer%antioche.eu.org@localhost> wrote:
> > On Fri, Dec 17, 2010 at 01:57:48PM +0000, Gerard Lally wrote:
> > > Yes I did increase that substantially.
> > how much ?
> I tried increasing it from the default 2048, first to 16384 and then to
> I also tried tuning TCP in /etc/sysctl.conf:
> But the same thing happened all the time - mbuf usage climbed and
> climbed until the system froze.
if altq is causing latency to increase, it's expected that sendbuf
will be raised up to sendbuf_max. Then a single TCP connection will
eat 8192 mbuf for its send queue
> > > it helped, but eventually the
> > > system still froze.
> > what is your config ?
> I took the queueing section directly
> from /usr/share/examples/pf/faq-example2
thanks. but this doesn't tell me the depth of the altq queues; I don't know
how it's configured with pf nor the default value.
here's what I use on a NetBSD dom0 with altqd:
conditioner vlan3 metroz_cond <tbmeter 131072 16384 <pass> <drop>>
filter vlan3 metroz_cond 0 0 184.108.40.206 0 0
filter6 vlan3 metroz_cond 0 0 2001:660:3302:2828:a800:ff:fe60:f6 0 0
conditioner vlan16 p2pvl_cond <tbmeter 50000000 16384 <pass> <drop>>
filter vlan16 p2pvl_cond 0 0 220.127.116.11 netmask 255.255.255.248 0 0
conditioner vlan20 mlp2pvl_cond <tbmeter 20000000 16384 <pass> <drop>>
filter vlan20 mlp2pvl_cond 0 0 18.104.22.168 netmask 255.255.255.248 0 0
conditioner vlan22 playallvl_cond <tbmeter 20000000 16384 <pass> <drop>>
filter vlan22 playallvl_cond 0 0 22.214.171.124 netmask 255.255.255.252 0 0
and there's no problems with that.
Manuel Bouyer <bouyer%antioche.eu.org@localhost>
NetBSD: 26 ans d'experience feront toujours la difference
Main Index |
Thread Index |