[PATCH] ARC: Allow disabling of prefetch operations for debugging purposes
From: Alexey Brodkin
Date: Thu Jan 18 2018 - 08:48:19 EST
Signed-off-by: Alexey Brodkin <abrodkin@xxxxxxxxxxxx>
---
arch/arc/Kconfig | 5 +++++
arch/arc/include/asm/processor.h | 11 +++++++++++
arch/arc/lib/memcpy-archs.S | 16 ++++++++++++++++
arch/arc/lib/memset-archs.S | 6 ++++++
4 files changed, 38 insertions(+)
diff --git a/arch/arc/Kconfig b/arch/arc/Kconfig
index 9d5fd00d9e91..aa5262e8e43f 100644
--- a/arch/arc/Kconfig
+++ b/arch/arc/Kconfig
@@ -532,6 +532,11 @@ config ARC_DBG_TLB_PARANOIA
bool "Paranoia Checks in Low Level TLB Handlers"
default n
+config ARC_DISABLE_PREFETCH
+ bool "Disable use of prefetch instructions"
+ default n
+ help
+ Suppresses usage of prefetchw and prealloc instructions.
endif
config ARC_UBOOT_SUPPORT
diff --git a/arch/arc/include/asm/processor.h b/arch/arc/include/asm/processor.h
index 8ee41e988169..01011c0515d4 100644
--- a/arch/arc/include/asm/processor.h
+++ b/arch/arc/include/asm/processor.h
@@ -106,6 +106,17 @@ extern unsigned int get_wchan(struct task_struct *p);
*/
#define current_text_addr() ({ __label__ _l; _l: &&_l; })
+#ifdef CONFIG_ARC_DISABLE_PREFETCH
+#define ARCH_HAS_PREFETCH
+#define prefetch(x)
+
+#define ARCH_HAS_PREFETCHW
+#define prefetchw(x)
+
+#define ARCH_HAS_SPINLOCK_PREFETCH
+#define spin_lock_prefetch(x)
+#endif /* CONFIG_ARC_DISABLE_PREFETCH */
+
#endif /* !__ASSEMBLY__ */
/*
diff --git a/arch/arc/lib/memcpy-archs.S b/arch/arc/lib/memcpy-archs.S
index d61044dd8b58..0a29d024be01 100644
--- a/arch/arc/lib/memcpy-archs.S
+++ b/arch/arc/lib/memcpy-archs.S
@@ -41,8 +41,10 @@
#endif
ENTRY_CFI(memcpy)
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
prefetch [r1] ; Prefetch the read location
prefetchw [r0] ; Prefetch the write location
+#endif
mov.f 0, r2
;;; if size is zero
jz.d [blink]
@@ -72,8 +74,10 @@ ENTRY_CFI(memcpy)
lpnz @.Lcopy32_64bytes
;; LOOP START
LOADX (r6, r1)
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
PREFETCH_READ (r1)
PREFETCH_WRITE (r3)
+#endif
LOADX (r8, r1)
LOADX (r10, r1)
LOADX (r4, r1)
@@ -117,9 +121,13 @@ ENTRY_CFI(memcpy)
lpnz @.Lcopy8bytes_1
;; LOOP START
ld.ab r6, [r1, 4]
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
prefetch [r1, 28] ;Prefetch the next read location
+#endif
ld.ab r8, [r1,4]
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
prefetchw [r3, 32] ;Prefetch the next write location
+#endif
SHIFT_1 (r7, r6, 24)
or r7, r7, r5
@@ -162,9 +170,13 @@ ENTRY_CFI(memcpy)
lpnz @.Lcopy8bytes_2
;; LOOP START
ld.ab r6, [r1, 4]
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
prefetch [r1, 28] ;Prefetch the next read location
+#endif
ld.ab r8, [r1,4]
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
prefetchw [r3, 32] ;Prefetch the next write location
+#endif
SHIFT_1 (r7, r6, 16)
or r7, r7, r5
@@ -204,9 +216,13 @@ ENTRY_CFI(memcpy)
lpnz @.Lcopy8bytes_3
;; LOOP START
ld.ab r6, [r1, 4]
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
prefetch [r1, 28] ;Prefetch the next read location
+#endif
ld.ab r8, [r1,4]
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
prefetchw [r3, 32] ;Prefetch the next write location
+#endif
SHIFT_1 (r7, r6, 8)
or r7, r7, r5
diff --git a/arch/arc/lib/memset-archs.S b/arch/arc/lib/memset-archs.S
index 62ad4bcb841a..343f292d92a0 100644
--- a/arch/arc/lib/memset-archs.S
+++ b/arch/arc/lib/memset-archs.S
@@ -11,7 +11,9 @@
#undef PREALLOC_NOT_AVAIL
ENTRY_CFI(memset)
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
prefetchw [r0] ; Prefetch the write location
+#endif
mov.f 0, r2
;;; if size is zero
jz.d [blink]
@@ -48,11 +50,13 @@ ENTRY_CFI(memset)
lpnz @.Lset64bytes
;; LOOP START
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
#ifdef PREALLOC_NOT_AVAIL
prefetchw [r3, 64] ;Prefetch the next write location
#else
prealloc [r3, 64]
#endif
+#endif /* CONFIG_ARC_DISABLE_PREFETCH */
#ifdef CONFIG_ARC_HAS_LL64
std.ab r4, [r3, 8]
std.ab r4, [r3, 8]
@@ -85,7 +89,9 @@ ENTRY_CFI(memset)
lsr.f lp_count, r2, 5 ;Last remaining max 124 bytes
lpnz .Lset32bytes
;; LOOP START
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
prefetchw [r3, 32] ;Prefetch the next write location
+#endif
#ifdef CONFIG_ARC_HAS_LL64
std.ab r4, [r3, 8]
std.ab r4, [r3, 8]
--
2.11.0