RE: [RFC net-next 0/8] Introducing subdev bus and devlink extension

From: Parav Pandit
Date: Fri Mar 08 2019 - 12:09:30 EST




> -----Original Message-----
> From: Kirti Wankhede <kwankhede@xxxxxxxxxx>
> Sent: Friday, March 8, 2019 6:19 AM
> To: Parav Pandit <parav@xxxxxxxxxxxx>; Jakub Kicinski
> <jakub.kicinski@xxxxxxxxxxxxx>
> Cc: Or Gerlitz <gerlitz.or@xxxxxxxxx>; netdev@xxxxxxxxxxxxxxx; linux-
> kernel@xxxxxxxxxxxxxxx; michal.lkml@xxxxxxxxxxx; davem@xxxxxxxxxxxxx;
> gregkh@xxxxxxxxxxxxxxxxxxx; Jiri Pirko <jiri@xxxxxxxxxxxx>; Alex
> Williamson <alex.williamson@xxxxxxxxxx>
> Subject: Re: [RFC net-next 0/8] Introducing subdev bus and devlink extension
>
>
>
> >>>>>> <snip>
> >>>>>>
> >>>>>>>>>
> >>>>>>>>> Yes. I got my patches to adapt to mdev way. Will be posting
> >>>>>>>>> RFC
> >>>>>>>>> v2
> >>>> soon.
> >>>>>>>>> Will wait for a day to receive more comments/views from Greg
> >>>>>>>>> and
> >>>>>> others.
> >>>>>>>>>
> >>>>>>>>> As I explained in this cover-letter and discussion, First use
> >>>>>>>>> case is to create and use mdevs in the host (and not in VM).
> >>>>>>>>> Later on, I am sure once we have mdevs available, VM users
> >>>>>>>>> will likely use
> >>>>>>>> it.
> >>>>>>>>>
> >>>>>>>>> So, mlx5_core driver will have two components as starting point.
> >>>>>>>>>
> >>>>>>>>> 1. drivers/net/ethernet/mellanox/mlx5/core/mdev/mdev.c
> >>>>>>>>> This is mdev device life cycle driver which will do,
> >>>>>>>>> mdev_register_device()
> >>>>>>>> and implements mlx5_mdev_ops.
> >>>>>>>>>
> >>>>>>>> Ok. I would suggest not use mdev.c file name, may be add device
> >>>>>>>> name, something like mlx_mdev.c or vfio_mlx.c
> >>>>>>>>
> >>>>>>> mlx5/core is coding convention is not following to prefix mlx to
> >>>>>>> its
> >>>>>>> 40+
> >>>>>> files.
> >>>>>>>
> >>>>>>> it uses actual subsystem or functionality name, such as, sriov.c
> >>>>>>> eswitch.c fw.c en_tc.c (en for Ethernet) lag.c so, mdev.c aligns
> >>>>>>> to rest of the 40+ files.
> >>>>>>>
> >>>>>>>
> >>>>>>>>> 2. drivers/net/ethernet/mellanox/mlx5/core/mdev/mdev_driver.c
> >>>>>>>>> This is mdev device driver which does mdev_register_driver()
> >>>>>>>>> and
> >>>>>>>>> probe() creates netdev by heavily reusing existing code of the
> >>>>>>>>> PF
> >>>> device.
> >>>>>>>>> These drivers will not be placed under drivers/vfio/mdev,
> >>>>>>>>> because this is
> >>>>>>>> not a vfio driver.
> >>>>>>>>> This is fine, right?
> >>>>>>>>>
> >>>>>>>>
> >>>>>>>> I'm not too familiar with netdev, but can you create netdev on
> >>>>>>>> open() call on mlx mdev device? Then you don't have to write
> >>>>>>>> mdev device
> >>>>>> driver.
> >>>>>>>>
> >>>>>>> Who invokes open() and release()?
> >>>>>>> I believe it is the qemu would do open(), release,
> read/write/mmap?
> >>>>>>>
> >>>>>>> Assuming that is the case,
> >>>>>>> I think its incorrect to create netdev in open.
> >>>>>>> Because when we want to map the mdev to VM using above mdev
> >> calls,
> >>>>>>> we
> >>>>>> actually wont be creating netdev in host.
> >>>>>>> Instead, some queues etc will be setup as part of these calls.
> >>>>>>>
> >>>>>>> By default this created mdev is bound to vfio_mdev.
> >>>>>>> And once we unbind the device from this driver, we need to bind
> >>>>>>> to
> >>>>>>> mlx5
> >>>>>> driver so that driver can create the netdev etc.
> >>>>>>>
> >>>>>>> Or did I get open() and friends call wrong?
> >>>>>>>
> >>>>>>
> >>>>>> In 'struct mdev_parent_ops' there are create() and remove(). When
> >>>>>> user creates mdev device by writing UUID to create sysfs, vendor
> >>>>>> driver's
> >>>>>> create() callback gets called. This should be used to
> >>>>>> allocate/commit
> >>>>> Yes. I am already past that stage.
> >>>>>
> >>>>>> resources from parent device and on remove() callback free those
> >>>> resources.
> >>>>>> So there is no need to bind mlx5 driver to that mdev device.
> >>>>>>
> >>>>> If we don't bind mlx5 driver, vfio_mdev driver is bound to it.
> >>>>> Such driver
> >>>> won't create netdev.
> >>>>
> >>>> Doesn't need to.
> >>>>
> >>>> Create netdev from create() callback.
> >>>>
> >>> I strongly believe this is incorrect way to use create() API.
> >>> Because,
> >>> mdev is mediated device from its primary pci device. It is not a
> >>> protocol
> >> device.
> >>>
> >>> It it also incorrect to tell user that vfio_mdev driver is bound to
> >>> this mdev
> >> and mlx5_core driver creating netdev on top of mdev.
> >>>
> >>
> >> vfio_mdev is generic common driver.
> >> Vendor driver who want to partition its device should handle its
> >> child creation and its life cycle. What is wrong in that? Why netdev
> >> has to be created from probe() only and not from create()?
> >>
> > I am not suggesting to invent any new probe() method.
> > create() is generic mdev creation entry point.
> > When create() is implemented by vendor driver, vendor driver doesn't
> know if this mdev will be provisioned for VM or for host.
>
> Vendor driver doesn't need to know if mdev device is going to used for VM
> or host.
>
I explained you the use cases. How can I explain better?
When user wants to use mdev in host, rdma and netdevices needs to be created by the kernel vendor driver.
When user wants to use mdev in VM (passthrough), rdma and netdev won't be created by host kernel driver.
Can you please ack that these two use cases are understood?
Because I get the feeling that it is not from your proposal.
And we should get level set on it first.

These are two different way and obviously it has two different init sequence.
More below.

> > So it must do only generic mdev init sequence.
> > This means, it cannot create netdev here. As simple as that.
> >
> > When user wants to use this mdev in a host, user will first unbind it from
> vfio_mdev driver and binds this mdev to mlx5_driver.
> > probe() of mlx5_core driver is called who did mdev_register_driver.
> > At this point netdev is created.
> >
> > If user wants to use this mdev for VM, than vfio_mdev driver and qemu
> will control it via open/release friend functions.
> >
>
> VFIO interface is generic interface, one example is QEMU, as a user space
> application, uses VFIO interface. But that doesn't mean that if you are using
> VFIO interface that is only be used for VM.
> You can write a user space application for host using VFIO interface.
>
VM is one use case. We picked VM for discussion here.

If mdev are supposed to be consume by single driver = vfio_mdev, there is no need to have it as separate vfio_mdev driver.
Can you please explain why vfio_mdev driver logic was split from the core mdev?

> >>> When we want to map this mdev to VM, what should create() do?
> >>
> >> Mediated device should be created before it is mapped to VM.
> >>
> > Of course.
> > Let me rephrase the question:
> > what shouldn't be done by create() when it wants to map to VM?
> > Answer is: it shouldn't create a netdev, but do necessary initialization so
> that it can be mapped to VM.
> > Because netdev will be created inside the VM not in the host.
> >
> > create() simply doesn't know during creation time, where this mdev will be
> used (VM or host).
> > So it doesn't make any sense to create netdev in create().
> >
>
> As I explained above, I disagree with this comment.
>
But that disagreement doesn't propose a good software solution. :-)
More discussion below.

> > I hope it's clear now.
> >
> >> If you look at the sequence of mdev device creation:
> >> - 'struct device' is created with bus 'mdev_bus_type'
> >> - register the device - device_register(&mdev->dev) -> which calls
> >> vfio_mdev's
> >> probe() -> common code for all vendor drivers
> >> - mdev_device_create_ops() -> calls vendor driver's create() -> this
> >> is for vendor specific allocation and initialization. This is the
> >> callback from where you can do what you want to do for mdev device
> creation and initialization.
> >> Why it has to be named as probe()?
> >
> > I do not intent to create any new probe().
> > I think I explained the flow well above - i.e. role of
> > mdev_driver->probe() vs mdev_device->create().
> >
> >>
> >>> We will have to shift the code from create() to
> >>> mdev_device_driver()-
> >>> probe() to address a use case of selectively mapping a mdev to VM or
> >>> to
> >> host and implement appropriate open/close etc functions for VM case.
> >>>
> >>> So why not start correctly from the beginning?
> >>>
> >>
> >> What is wrong with current implementation which is being used and
> >> tested for multiple devices?
> >>
> > Oh, nothing wrong in current implementation.
> > Which current implementation provisions mdev in host (and not in guest
> VM)?
> >
>
> All mdev vendor driver implementations. Mdev device for host or VM :
> there is no such difference, interface is generic.
>
Can you point to an example driver that creates usable device (not just mdev, actual device for a given mdev)?
I looked at intel_gvt_activate_vgpu() and ioctl(), that doesn't create any gpu in host.
I looked at samples/vfio-mdev/mtty.c that doesn't create any tty device on open() or ioctl() call.

> > I am just using right code split of already available mdev.
> > When user wants to map a device to VM, attach vfio_mdev driver and
> create vfio_device.
> > When user wants to use a device in host, don't attach vfio_mdev driver,
> instead attach, appropriate driver what owns this mdev.
> >
>
> No need to do that. Use VFIO interface in your user space application when
> you want to use device in host.
>
1. Can you explain what do you propose - on which callback should create rdma and netdevice if we should not dedicated vendor bus driver?
2. Can you please also explain how vendor driver should not create these rdma and netdevice, when this device should be used in VM?

So far I heard only two points from you -
i.e. create netdev during create() callback.
Other one was create netdev during open() call.

Both of these modes doesn't fit the need as I explained.
Did I miss something?

3. we certainly do not want to invent new ioctl opcodes on vfio char device to create/delete these devices.
There should be only VFIO_DEVICE_*.
netdev subsystem has gone far ahead to not use such crazy vendor ioctl() extensions.
I hope you do not intent to propose that when you say "Use VFIO interface in your user space application ...".

4. Finally linux kernel core provides a bus model and binding/attaching a specific driver to a device.
I assumed that you are aware of bind/unbind sysfs files of a driver.
Can you please explain why you think that mdev subsystem should not use standard linux kernel driver model?