Re: [PATCH] microblaze: speedup for word-aligned memcpys

From: Michal Simek
Date: Tue Apr 13 2010 - 04:07:57 EST


Steven J. Magnani wrote:
memcpy performance was measured on a noMMU system having a barrel shifter, 4K caches, and 32-byte write-through cachelines. In this environment, copying word-aligned data in word-sized chunks appears to be about 3% more efficient on packet-sized buffers (1460 bytes) than copying in cacheline-sized chunks.

Skip to word-based copying when buffers are both word-aligned.

Signed-off-by: Steven J. Magnani <steve@xxxxxxxxxxxxxxx>

I added this patch to next branch and I will keep it there for now.

1. I agree that we need several patches like this.
2. The improvement could be there and likely it is but 3% improvement could be caused for different reason.
3. There is necessary to measure it on several hw design and cache configurations to be sure that your expectation is correct.
4. The best will be to monitoring cache behavior but currently there is no any tool which could easily help us with it.

I will talk to xilinx how to monitoring it.

Thanks,
Michal


---
diff -uprN a/arch/microblaze/lib/fastcopy.S b/arch/microblaze/lib/fastcopy.S
--- a/arch/microblaze/lib/fastcopy.S 2010-04-09 21:52:36.000000000 -0500
+++ b/arch/microblaze/lib/fastcopy.S 2010-04-12 15:37:44.000000000 -0500
@@ -69,37 +69,13 @@ a_dalign_done:
blti r4, a_block_done
a_block_xfer:
- andi r4, r7, 0xffffffe0 /* n = c & ~31 */
- rsub r7, r4, r7 /* c = c - n */
-
andi r9, r6, 3 /* t1 = s & 3 */
- /* if temp != 0, unaligned transfers needed */
- bnei r9, a_block_unaligned
-
-a_block_aligned:
- lwi r9, r6, 0 /* t1 = *(s + 0) */
- lwi r10, r6, 4 /* t2 = *(s + 4) */
- lwi r11, r6, 8 /* t3 = *(s + 8) */
- lwi r12, r6, 12 /* t4 = *(s + 12) */
- swi r9, r5, 0 /* *(d + 0) = t1 */
- swi r10, r5, 4 /* *(d + 4) = t2 */
- swi r11, r5, 8 /* *(d + 8) = t3 */
- swi r12, r5, 12 /* *(d + 12) = t4 */
- lwi r9, r6, 16 /* t1 = *(s + 16) */
- lwi r10, r6, 20 /* t2 = *(s + 20) */
- lwi r11, r6, 24 /* t3 = *(s + 24) */
- lwi r12, r6, 28 /* t4 = *(s + 28) */
- swi r9, r5, 16 /* *(d + 16) = t1 */
- swi r10, r5, 20 /* *(d + 20) = t2 */
- swi r11, r5, 24 /* *(d + 24) = t3 */
- swi r12, r5, 28 /* *(d + 28) = t4 */
- addi r6, r6, 32 /* s = s + 32 */
- addi r4, r4, -32 /* n = n - 32 */
- bneid r4, a_block_aligned /* while (n) loop */
- addi r5, r5, 32 /* d = d + 32 (IN DELAY SLOT) */
- bri a_block_done
+ /* if temp == 0, everything is word-aligned */
+ beqi r9, a_word_xfer
a_block_unaligned:
+ andi r4, r7, 0xffffffe0 /* n = c & ~31 */
+ rsub r7, r4, r7 /* c = c - n */
andi r8, r6, 0xfffffffc /* as = s & ~3 */
add r6, r6, r4 /* s = s + n */
lwi r11, r8, 0 /* h = *(as + 0) */



--
Michal Simek, Ing. (M.Eng)
w: www.monstr.eu p: +42-0-721842854
Maintainer of Linux kernel 2.6 Microblaze Linux - http://www.monstr.eu/fdt/
Microblaze U-BOOT custodian
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/