tech-kern archive

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]

Re: RFC: vioif(4) multiqueue support



> Date: Wed, 26 Dec 2018 10:03:15 +0900
> From: Shoichi Yamaguchi <yamaguchi%netbsd.org@localhost>
> 
> > I implemented a patch that make vioif(4) support multi-queue. And I have put
> > the patch on ftp.n.o. I used vioif(4) multiqueue on qemu-kvm on Linux kernel
> > 4.19.5. And It seems to be working fine.
> >
> > https://ftp.netbsd.org/pub/NetBSD/misc/yamaguchi/vioif_mutilq.patch
> 
> Do you have any comments?
> I would like to going to commit the patch if there is no comment until tomorrow.

Hi, Yamaguchi-san!  A lot of Americans and Europeans are on vacation
this time of year, so it might be better to hold off for another week
or two.  Here's a quick review -- I don't know anything about virtio,
so this is just about use of kernel APIs and abstractions.  Someone
who knows something about virtio should take a look too.

> diff --git a/sys/dev/pci/if_vioif.c b/sys/dev/pci/if_vioif.c
> index 3bbd300e88e..769b108e793 100644
> --- a/sys/dev/pci/if_vioif.c
> +++ b/sys/dev/pci/if_vioif.c
>  
>  /* Feature bits */
> -#define VIRTIO_NET_F_CSUM	(1<<0)
> [...]
> +#define VIRTIO_NET_F_MQ			(1<<22)

If you're going to modify all of the lines here, maybe take the
opportunity to convert them to use __BIT?

> @@ -171,73 +184,110 @@ struct virtio_net_ctrl_vlan {
> [...]
>  /*
>   * if_vioifvar.h:
>   */
> +struct vioif_txqueue {
> +	struct virtqueue	*txq_vq;

Why not make this an embedded structure?

struct vioif_txqueue {
	struct virtqueue	txq_vq;
	...
};

struct vioif_softc {
	struct vioif_txqueue	*sc_txq;
	struct vioif_rxqueue	*sc_rxq;
	struct vioif_ctrlqueue	*sc_ctrlq;
	...
};

> +	kmutex_t		*txq_lock;

Why is this a pointer to kmutex_t with mutex_obj_alloc/free and not
just a kmutex_t with mutex_init/destroy?  Is it reused anywhere?  If
it is reused, this needs explanation in the comments.  If it is not,
just use kmutex_t.

Can you write a comment summarizing what locks cover what fields, and,
if more than one lock can be held at once, what the lock order is?

> +struct vioif_rxqueue {
> +	struct virtqueue	*rxq_vq;
> +	kmutex_t		*rxq_lock;

Likewise.

> -#define VIOIF_TX_LOCK(_sc)	mutex_enter(&(_sc)->sc_tx_lock)
> -#define VIOIF_TX_UNLOCK(_sc)	mutex_exit(&(_sc)->sc_tx_lock)
> -#define VIOIF_TX_LOCKED(_sc)	mutex_owned(&(_sc)->sc_tx_lock)
> -#define VIOIF_RX_LOCK(_sc)	mutex_enter(&(_sc)->sc_rx_lock)
> -#define VIOIF_RX_UNLOCK(_sc)	mutex_exit(&(_sc)->sc_rx_lock)
> -#define VIOIF_RX_LOCKED(_sc)	mutex_owned(&(_sc)->sc_rx_lock)
> +#define VIOIF_TXQ_LOCK(_q)	mutex_enter((_q)->txq_lock)
> +#define VIOIF_TXQ_TRYLOCK(_q)	mutex_tryenter((_q)->txq_lock)
> +#define VIOIF_TXQ_UNLOCK(_q)	mutex_exit((_q)->txq_lock)
> +#define VIOIF_TXQ_LOCKED(_q)	mutex_owned((_q)->txq_lock)
> +
> +#define VIOIF_RXQ_LOCK(_q)	mutex_enter((_q)->rxq_lock)
> +#define VIOIF_RXQ_UNLOCK(_q)	mutex_exit((_q)->rxq_lock)
> +#define VIOIF_RXQ_LOCKED(_q)	mutex_owned((_q)->rxq_lock)

Can we just use mutex_enter/exit/&c. without the macros?  Sometimes we
use macros where they are conditional, depending on NET_MPSAFE, but if
there's no need for that, I would prefer to read direct calls to
mutex_enter/exit/&c.

> +static int
> +vioif_alloc_queues(struct vioif_softc *sc)
> +{
> +	int nvq_pairs = sc->sc_max_nvq_pairs;
> +	int nvqs = nvq_pairs * 2;
> +	int i;
> +
> +	sc->sc_rxq = kmem_zalloc(sizeof(sc->sc_rxq[0]) * nvq_pairs,
> +	    KM_NOSLEEP);
> +	if (sc->sc_rxq == NULL)
> +		return -1;
> +
> +	sc->sc_txq = kmem_zalloc(sizeof(sc->sc_txq[0]) * nvq_pairs,
> +	    KM_NOSLEEP);
> +	if (sc->sc_txq == NULL)
> +		return -1;

Check to avoid arithmetic overflow here:

	if (nvq_pairs > INT_MAX/2 - 1 ||
	    nvq_pairs > SIZE_MAX/sizeof(sc->sc_rxq[0]))
		return -1;
	nvqs = nvq_pairs * 2;
	if (...) nvqs++;
	sc->sc_rxq = kmem_zalloc(sizeof(sc->sc_rxq[0]) * nvq_pairs, ...);

Same in all the other allocations like this.  (We should have a
kmem_allocarray -- I have a draft somewhere.)

> @@ -586,69 +759,109 @@ vioif_attach(device_t parent, device_t self, void *aux)
> [...]
> +		/* Limit the number of queue pairs to use */
> +		if (sc->sc_max_nvq_pairs <= ncpu)
> +			sc->sc_req_nvq_pairs = sc->sc_max_nvq_pairs;
> +		else
> +			sc->sc_req_nvq_pairs = ncpu;

How about sc->sc_req_nvq_pairs = MIN(sc->sc_max_nvq_pairs, ncpu)?

> +static void
> +vioif_ctrl_release(struct vioif_softc *sc)
> +{
> +	struct vioif_ctrlqueue *ctrlq = &sc->sc_ctrlq;
> +
> +	mutex_enter(&ctrlq->ctrlq_wait_lock);

KASSERT(ctrlq->ctrlq_inuse != FREE)

Might be helpful to record the lwp that owns this ctrlq, too, for
diagnostics: KASSERT(ctrlq->ctrlq_owner == curlwp).

> diff --git a/sys/dev/pci/virtio_pci.c b/sys/dev/pci/virtio_pci.c
> index 65c5222b774..bb972997be2 100644
> --- a/sys/dev/pci/virtio_pci.c
> +++ b/sys/dev/pci/virtio_pci.c
> @@ -604,8 +677,14 @@ virtio_pci_setup_interrupts(struct virtio_softc *sc)
> [...]
>  	if (pci_intr_type(pc, psc->sc_ihp[0]) == PCI_INTR_TYPE_MSIX) {
> -		psc->sc_ihs = kmem_alloc(sizeof(*psc->sc_ihs) * 2,
> +		psc->sc_ihs = kmem_alloc(sizeof(*psc->sc_ihs) * nmsix,
>  		    KM_SLEEP);

Check for arithmetic overflow here.


Home | Main Index | Thread Index | Old Index