Bounce-buffering makes the system spend more time copying
I/O data. When the I/O transaction take place between
a confidential and a non-confidential endpoints, there is
no other way around.
Introduce a device bitfield to indicate that the device
doesn't need to perform bounce buffering. The capable
device may employ it to save on copying data around.
Signed-off-by: Roman Kisel <romank@xxxxxxxxxxxxxxxxxxx>
---
arch/x86/mm/mem_encrypt.c | 3 +++
include/linux/device.h | 8 ++++++++
include/linux/dma-direct.h | 3 +++
include/linux/swiotlb.h | 3 +++
4 files changed, 17 insertions(+)
diff --git a/arch/x86/mm/mem_encrypt.c b/arch/x86/mm/mem_encrypt.c
index 95bae74fdab2..6349a02a1da3 100644
--- a/arch/x86/mm/mem_encrypt.c
+++ b/arch/x86/mm/mem_encrypt.c
@@ -19,6 +19,9 @@
/* Override for DMA direct allocation check - ARCH_HAS_FORCE_DMA_UNENCRYPTED */
bool force_dma_unencrypted(struct device *dev)
{
+ if (dev->use_priv_pages_for_io)
+ return false;
+
/*
* For SEV, all DMA must be to unencrypted addresses.
*/
diff --git a/include/linux/device.h b/include/linux/device.h
index 80a5b3268986..4aa4a6fd9580 100644
--- a/include/linux/device.h
+++ b/include/linux/device.h
@@ -725,6 +725,8 @@ struct device_physical_location {
* @dma_skip_sync: DMA sync operations can be skipped for coherent buffers.
* @dma_iommu: Device is using default IOMMU implementation for DMA and
* doesn't rely on dma_ops structure.
+ * @use_priv_pages_for_io: Device is using private pages for I/O, no need to
+ * bounce-buffer.
*
* At the lowest level, every device in a Linux system is represented by an
* instance of struct device. The device structure contains the information
@@ -843,6 +845,7 @@ struct device {
#ifdef CONFIG_IOMMU_DMA
bool dma_iommu:1;
#endif
+ bool use_priv_pages_for_io:1;
};
/**
@@ -1079,6 +1082,11 @@ static inline bool dev_removable_is_valid(struct device *dev)
return dev->removable != DEVICE_REMOVABLE_NOT_SUPPORTED;
}
+static inline bool dev_priv_pages_for_io(struct device *dev)
+{
+ return dev->use_priv_pages_for_io;
+}
+
/*
* High level routines for use by the bus drivers
*/
diff --git a/include/linux/dma-direct.h b/include/linux/dma-direct.h
index d7e30d4f7503..b096369f847e 100644
--- a/include/linux/dma-direct.h
+++ b/include/linux/dma-direct.h
@@ -94,6 +94,9 @@ static inline dma_addr_t phys_to_dma_unencrypted(struct device *dev,
*/
static inline dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr)
{
+ if (dev_priv_pages_for_io(dev))
+ return phys_to_dma_unencrypted(dev, paddr);
+
return __sme_set(phys_to_dma_unencrypted(dev, paddr));
}
diff --git a/include/linux/swiotlb.h b/include/linux/swiotlb.h
index 3dae0f592063..35ee10641b42 100644
--- a/include/linux/swiotlb.h
+++ b/include/linux/swiotlb.h
@@ -173,6 +173,9 @@ static inline bool is_swiotlb_force_bounce(struct device *dev)
{
struct io_tlb_mem *mem = dev->dma_io_tlb_mem;
+ if (dev_priv_pages_for_io(dev))
+ return false;
+
return mem && mem->force_bounce;
}