[PATCH 2/2] drm/msm: use drm_cache when available
From: Rob Clark
Date: Mon Aug 05 2019 - 17:21:43 EST
From: Rob Clark <robdclark@xxxxxxxxxxxx>
For a long time drm/msm had been abusing dma_map_* or dma_sync_* to
clean pages for buffers with uncached/writecombine CPU mmap'ings.
But drm/msm is managing it's own iommu domains, and really doesn't want
the additional functionality provided by various DMA API ops.
Let's just cut the abstraction and use drm_cache where possible.
Signed-off-by: Rob Clark <robdclark@xxxxxxxxxxxx>
---
drivers/gpu/drm/msm/msm_gem.c | 8 ++++++++
1 file changed, 8 insertions(+)
diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c
index 8cf6362e64bf..af19ef20d0d5 100644
--- a/drivers/gpu/drm/msm/msm_gem.c
+++ b/drivers/gpu/drm/msm/msm_gem.c
@@ -9,6 +9,8 @@
#include <linux/dma-buf.h>
#include <linux/pfn_t.h>
+#include <drm/drm_cache.h>
+
#include "msm_drv.h"
#include "msm_fence.h"
#include "msm_gem.h"
@@ -48,6 +50,7 @@ static bool use_pages(struct drm_gem_object *obj)
static void sync_for_device(struct msm_gem_object *msm_obj)
{
+#if !defined(HAS_DRM_CACHE)
struct device *dev = msm_obj->base.dev->dev;
if (get_dma_ops(dev)) {
@@ -57,10 +60,14 @@ static void sync_for_device(struct msm_gem_object *msm_obj)
dma_map_sg(dev, msm_obj->sgt->sgl,
msm_obj->sgt->nents, DMA_BIDIRECTIONAL);
}
+#else
+ drm_clflush_sg(msm_obj->sgt);
+#endif
}
static void sync_for_cpu(struct msm_gem_object *msm_obj)
{
+#if !defined(HAS_DRM_CACHE)
struct device *dev = msm_obj->base.dev->dev;
if (get_dma_ops(dev)) {
@@ -70,6 +77,7 @@ static void sync_for_cpu(struct msm_gem_object *msm_obj)
dma_unmap_sg(dev, msm_obj->sgt->sgl,
msm_obj->sgt->nents, DMA_BIDIRECTIONAL);
}
+#endif
}
/* allocate pages from VRAM carveout, used when no IOMMU: */
--
2.21.0