[RFC v2 00/21] Provide a new two step DMA API mapping API
From: Leon Romanovsky
Date: Thu Sep 12 2024 - 07:16:19 EST
Instead of waiting and waiting when I finally fix NVMe patch which
doesn't work correctly in DMA direct mode (it causes to failures in
module reload), I decided to send the series as is to at least get
a feedback about API and the overall direction.
Thanks
-------------------------------------------------------------------------
Changelog:
v2:
* Remove attr field from dma_iova_attrs
* Added attr parameter to every new API call except alloc_iova
* Embed use_iova boolean into struct dma_iova_state
* Embed struct dma_memory_type into struct dma_iova_state
* Changed function signatures and added new function to initialize state
* Combined dma_set_memory_type() and dma_can_use_iova() to one function, they were used together anyway
* Made dma_start_range/dma_end_range to be NOP for non-iommu case
* Restructured the code to avoid inclusion of linux/pci.h in global header
* Based on DMA static calls series
* Changed iommu/dma functions to get parameters instead of dma_iova_state
* Removed iommu domain pointer from dma_iova_state
* dma_link_range now returns DMA address
* Rewrote NVMe patch
v1: https://lore.kernel.org/all/cover.1719909395.git.leon@xxxxxxxxxx
* Rewrote cover letter
* Changed to API as proposed
https://lore.kernel.org/linux-rdma/20240322184330.GL66976@xxxxxxxx/
* Removed IB DMA wrappers and use DMA API directly
v0: https://lore.kernel.org/all/cover.1709635535.git.leon@xxxxxxxxxx
-------------------------------------------------------------------------
Currently the only efficient way to map a complex memory description through
the DMA API is by using the scatterlist APIs. The SG APIs are unique in that
they efficiently combine the two fundamental operations of sizing and allocating
a large IOVA window from the IOMMU and processing all the per-address
swiotlb/flushing/p2p/map details.
This uniqueness has been a long standing pain point as the scatterlist API
is mandatory, but expensive to use. It prevents any kind of optimization or
feature improvement (such as avoiding struct page for P2P) due to the
impossibility of improving the scatterlist.
Several approaches have been explored to expand the DMA API with additional
scatterlist-like structures (BIO[1], rlist[2]), instead split up the DMAAPI
to allow callers to bring their own data structure.
The API is split up into parts:
- dma_alloc_iova() / dma_free_iova()
To do any pre-allocation required. This is done based on the caller
supplying some details about how much IOMMU address space it would need
in worst case.
- dma_link_range() / dma_unlink_range()
Perform the actual mapping into the pre-allocated IOVA. This is very
similar to dma_map_page().
A driver will extent its mapping size using its own data structure, such as
BIO, to request the required IOVA. Then it will iterate directly over it's
data structure to DMA map each range. The result can then be stored directly
into the HW specific DMA list. No intermediate scatterlist is required.
In this series, examples of three users are converted to the new API to show
the benefits. Each user has a unique flow:
1. RDMA ODP is an example of "SVA mirroring" using HMM that needs to
dynamically map/unmap large numbers of single pages. This becomes
significantly faster in the IOMMU case as the map/unmap is now just
a page table walk, the IOVA allocation is pre-computed once. Significant
amounts of memory are saved as there is no longer a need to store the
dma_addr_t of each page.
2. VFIO PCI live migration code is building a very large "page list"
for the device. Instead of allocating a scatter list entry per allocated
page it can just allocate an array of 'struct page *', saving a large
amount of memory.
3. NVMe PCI demonstrates how a BIO can be converted to a HW scatter
list without having to allocate then populate an intermediate SG table.
This step is first along a path to provide alternatives to scatterlist
and solve some of the abuses and design mistakes, for instance in DMABUF's
P2P support.
The ODP and VFIO versions are complete and fully tested, they can be the
users of the new API to merge it. The NVMe requires more work.
[1] https://lore.kernel.org/all/169772852492.5232.17148564580779995849.stgit@xxxxxxxxxxxxxxxxxxxxx/
[2] https://lore.kernel.org/all/ZD2lMvprVxu23BXZ@xxxxxxxx/
Thanks
Leon Romanovsky (21):
iommu/dma: Provide an interface to allow preallocate IOVA
iommu/dma: Implement link/unlink ranges callbacks
iommu/dma: Add check if IOVA can be used
dma-mapping: initialize IOVA state struct
dma-mapping: provide an interface to allocate IOVA
dma-mapping: set and query DMA IOVA state
dma-mapping: implement link range API
mm/hmm: let users to tag specific PFN with DMA mapped bit
dma-mapping: provide callbacks to link/unlink HMM PFNs to specific
IOVA
RDMA/umem: Preallocate and cache IOVA for UMEM ODP
RDMA/umem: Store ODP access mask information in PFN
RDMA/core: Separate DMA mapping to caching IOVA and page linkage
RDMA/umem: Prevent UMEM ODP creation with SWIOTLB
vfio/mlx5: Explicitly use number of pages instead of allocated length
vfio/mlx5: Rewrite create mkey flow to allow better code reuse
vfio/mlx5: Explicitly store page list
vfio/mlx5: Convert vfio to use DMA link API
nvme-pci: remove optimizations for single DMA entry
nvme-pci: precalculate number of DMA entries for each command
nvme-pci: use new dma API
nvme-pci: don't allow mapping of bvecs with offset
drivers/infiniband/core/umem_odp.c | 216 +++++---------
drivers/infiniband/hw/mlx5/mlx5_ib.h | 1 +
drivers/infiniband/hw/mlx5/odp.c | 44 +--
drivers/iommu/dma-iommu.c | 165 ++++++++++-
drivers/nvme/host/pci.c | 428 ++++++++++++++-------------
drivers/pci/p2pdma.c | 4 +-
drivers/vfio/pci/mlx5/cmd.c | 312 ++++++++++---------
drivers/vfio/pci/mlx5/cmd.h | 23 +-
drivers/vfio/pci/mlx5/main.c | 89 +++---
include/linux/dma-map-ops.h | 7 +
include/linux/dma-mapping.h | 85 ++++++
include/linux/hmm.h | 4 +
include/linux/iommu-dma.h | 43 +++
include/rdma/ib_umem_odp.h | 23 +-
kernel/dma/mapping.c | 232 +++++++++++++++
mm/hmm.c | 34 ++-
16 files changed, 1096 insertions(+), 614 deletions(-)
--
2.46.0