Re: [PATCH 00/15] Adding GAUDI NIC code to habanalabs driver

From: Oded Gabbay
Date: Thu Sep 10 2020 - 15:15:41 EST

On Thu, Sep 10, 2020 at 6:54 PM Greg KH <gregkh@xxxxxxxxxxxxxxxxxxx> wrote:
> On Thu, Sep 10, 2020 at 06:03:13PM +0300, Oded Gabbay wrote:
> > This patch-set adds support for initializing and using the GAUDI NIC ports,
> > functioning as scale-out interconnect when doing distributed Deep Learning
> > training. The training can be performed over tens of thousands of GAUDIs
> > and it is done using the RDMA-over-converged-Ethernet (RoCE) v2 protocol.
> >
> > Each GAUDI exposes 10x100GbE ports that are designed to scale-out the
> > inter-GAUDI communication by integrating a complete communication engine
> > on-die. This native integration allows users to use the same scaling
> > technology, both inside the server and rack (termed as scale-up), as well
> > as for scaling across racks (scale-out). The racks can be connected
> > directly between GAUDI processors, or through any number of standard
> > Ethernet switches.
> >
> > The driver exposes the NIC ports to the user as standard Ethernet ports by
> > registering each port to the networking subsystem. This allows the user to
> > manage the ports with standard tools such as ifconfig, ethtool, etc. It
> > also enables us to connect to the Linux networking stack and thus support
> > standard networking protocols, such as IPv4, IPv6, TCP, etc. In addition,
> > we can also leverage protocols such as DCB for dynamically configuring
> > priorities to avoid congestion.
> >
> > For each NIC port there is a matching QMAN entity. For RoCE, the user
> > submits workloads to the NIC through the QMAN, same as he does for the
> > compute engines. For regular Ethernet, the user sends and receives packets
> > through the standard Ethernet sockets. Those sockets are used only as a
> > control path. The data path that is used for AI training goes through the
> > RoCE interface.
> >
> > It is important to note that there are some limitations and uniqueness
> > in GAUDI's NIC H/W, compared to other networking adapters that enforced us
> > to use a less-than-common driver design:
> >
> > 1. The NIC functionality is NOT exposed as different PCI Physical
> > Functions. There is a single PF which is used for compute and
> > networking, as the main goal of the NIC ports is to be used as
> > intra-communication and not as standard network interfaces. This
> > implies we can't connect different drivers to handle the networking
> > ports because it is the same device, from the kernel POV, as the
> > compute. Therefore, we must integrate the networking code into the
> > main habanalabs driver.
> That's kind of common, see the long threads on the netdev and IB mailing
> lists about this type of issue on other networking cards today. The
> whole "virtual bus" code should help solve this, if Intel ever gets
> around to posting a new version of that patch series one day...
> But, because you are writing networking driver code here, you really
> should run all of this by the netdev@xxxxxxxxxxxxxxx maintainers and
> developers, as they know how to review this interaction with the network
> stack better than anyone else.
> Care to resend it and cc: them too?
> thanks,
> greg k-h

Sure, np.
Do you have someone specific or should I just send it to that mailing list ?
I don't know anyone there.