[PATCH v1 0/4] GPU Direct RDMA (P2P DMA) for Device Private Pages
From: Yonatan Maman
Date: Tue Oct 15 2024 - 11:24:29 EST
From: Yonatan Maman <Ymaman@xxxxxxxxxx>
This patch series aims to enable Peer-to-Peer (P2P) DMA access in
GPU-centric applications that utilize RDMA and private device pages. This
enhancement is crucial for minimizing data transfer overhead by allowing
the GPU to directly expose device private page data to devices such as
NICs, eliminating the need to traverse system RAM, which is the native
method for exposing device private page data.
To fully support Peer-to-Peer for device private pages, the following
changes are proposed:
`Memory Management (MM)`
* Leverage struct pagemap_ops to support P2P page operations: This
modification ensures that the GPU can directly map device private pages
for P2P DMA.
* Utilize hmm_range_fault to support P2P connections for device private
pages (instead of Page fault)
`IB Drivers`
Add TRY_P2P_REQ flag for the hmm_range_fault call: This flag indicates the
need for P2P mapping, enabling IB drivers to efficiently handle P2P DMA
requests.
`Nouveau driver`
Add support for the Nouveau p2p_page callback function: This update
integrates P2P DMA support into the Nouveau driver, allowing it to handle
P2P page operations seamlessly.
`MLX5 Driver`
Optimize PCI Peer-to-Peer for private device pages, by enabling Address
Translation service(ATS) for ODP memory.
Yonatan Maman (4):
mm/hmm: HMM API for P2P DMA to device zone pages
nouveau/dmem: HMM P2P DMA for private dev pages
IB/core: P2P DMA for device private pages
RDMA/mlx5: Enabling ATS for ODP memory
drivers/gpu/drm/nouveau/nouveau_dmem.c | 117 ++++++++++++++++++++++++-
drivers/infiniband/core/umem_odp.c | 2 +-
drivers/infiniband/hw/mlx5/mlx5_ib.h | 6 +-
include/linux/hmm.h | 2 +
include/linux/memremap.h | 7 ++
mm/hmm.c | 28 ++++++
6 files changed, 156 insertions(+), 6 deletions(-)
--
2.34.1