Re: [PATCHv5 3/3] vhost_net: a kernel-level virtio server

From: Avi Kivity
Date: Wed Sep 16 2009 - 09:06:29 EST


On 09/16/2009 02:44 PM, Gregory Haskins wrote:
The problem isn't where to find the models...the problem is how to
aggregate multiple models to the guest.

You mean configuration?

You instantiate multiple vhost-nets. Multiple ethernet NICs is a
supported configuration for kvm.
But this is not KVM.


If kvm can do it, others can.

His slave boards surface themselves as PCI devices to the x86
host. So how do you use that to make multiple vhost-based devices (say
two virtio-nets, and a virtio-console) communicate across the transport?

I don't really see the difference between 1 and N here.
A KVM surfaces N virtio-devices as N pci-devices to the guest. What do
we do in Ira's case where the entire guest represents itself as a PCI
device to the host, and nothing the other way around?

There is no guest and host in this scenario. There's a device side (ppc) and a driver side (x86). The driver side can access configuration information on the device side. How to multiplex multiple devices is an interesting exercise for whoever writes the virtio binding for that setup.

There are multiple ways to do this, but what I am saying is that
whatever is conceived will start to look eerily like a vbus-connector,
since this is one of its primary purposes ;)

I'm not sure if you're talking about the configuration interface or data
path here.
I am talking about how we would tunnel the config space for N devices
across his transport.

Sounds trivial. Write an address containing the device number and register number to on location, read or write data from another. Just like the PCI cf8/cfc interface.

They aren't in the "guest". The best way to look at it is

- a device side, with a dma engine: vhost-net
- a driver side, only accessing its own memory: virtio-net

Given that Ira's config has the dma engine in the ppc boards, that's
where vhost-net would live (the ppc boards acting as NICs to the x86
board, essentially).
That sounds convenient given his hardware, but it has its own set of
problems. For one, the configuration/inventory of these boards is now
driven by the wrong side and has to be addressed.

Why is it the wrong side?

Second, the role
reversal will likely not work for many models other than ethernet (e.g.
virtio-console or virtio-blk drivers running on the x86 board would be
naturally consuming services from the slave boards...virtio-net is an
exception because 802.x is generally symmetrical).

There is no role reversal. The side doing dma is the device, the side accessing its own memory is the driver. Just like that other 1e12 driver/device pairs out there.

I have no idea, that's for Ira to solve.
Bingo. Thus my statement that the vhost proposal is incomplete. You
have the virtio-net and vhost-net pieces covering the fast-path
end-points, but nothing in the middle (transport, aggregation,
config-space), and nothing on the management-side. vbus provides most
of the other pieces, and can even support the same virtio-net protocol
on top. The remaining part would be something like a udev script to
populate the vbus with devices on board-insert events.

Of course vhost is incomplete, in the same sense that Linux is incomplete. Both require userspace.

If he could fake the PCI
config space as seen by the x86 board, he would just show the normal pci
config and use virtio-pci (multiple channels would show up as a
multifunction device). Given he can't, he needs to tunnel the virtio
config space some other way.
Right, and note that vbus was designed to solve this. This tunneling
can, of course, be done without vbus using some other design. However,
whatever solution is created will look incredibly close to what I've
already done, so my point is "why reinvent it"?

virtio requires binding for this tunnelling, so does vbus. Its the same problem with the same solution.

--
error compiling committee.c: too many arguments to function

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/