Re: [PATCH] Use correct macros in raid code, not raw asm

From: Neil Brown
Date: Thu Feb 08 2007 - 20:38:43 EST


On Thursday February 8, hpa@xxxxxxxxx wrote:
> Andi Kleen wrote:
> >
> > It should use kernel_fpu_begin() imho. If someone wants to test
> > it in user space again they can add dummy definitions of that
> > to their user space header.
>
> I hadn't seen this thread until now, when Neil pointed me to the thread.
>
> Using kernel_fpu_begin() ... kernel_fpu_end() is probably indeed the
> best option.
>

So does this look right (no, I haven't compiled it yet)

NeilBrown


### Diffstat output
./drivers/md/raid6x86.h | 56 ++++++++++--------------------------------------
1 file changed, 12 insertions(+), 44 deletions(-)

diff .prev/drivers/md/raid6x86.h ./drivers/md/raid6x86.h
--- .prev/drivers/md/raid6x86.h 2007-02-09 12:30:32.000000000 +1100
+++ ./drivers/md/raid6x86.h 2007-02-09 12:36:01.000000000 +1100
@@ -25,20 +25,17 @@

typedef struct {
unsigned int fsave[27];
- unsigned long cr0;
} raid6_mmx_save_t __attribute__((aligned(16)));

/* N.B.: For SSE we only save %xmm0-%xmm7 even for x86-64, since
the code doesn't know about the additional x86-64 registers */
typedef struct {
unsigned int sarea[8*4+2];
- unsigned long cr0;
} raid6_sse_save_t __attribute__((aligned(16)));

/* This is for x86-64-specific code which uses all 16 XMM registers */
typedef struct {
unsigned int sarea[16*4+2];
- unsigned long cr0;
} raid6_sse16_save_t __attribute__((aligned(16)));

/* On x86-64 the stack *SHOULD* be 16-byte aligned, but currently this
@@ -50,7 +47,6 @@ typedef struct {

typedef struct {
unsigned int fsave[27];
- unsigned long cr0;
} raid6_mmx_save_t;

/* On i386, the stack is only 8-byte aligned, but SSE requires 16-byte
@@ -58,7 +54,6 @@ typedef struct {
a properly-sized area correctly. */
typedef struct {
unsigned int sarea[8*4+3];
- unsigned long cr0;
} raid6_sse_save_t;

/* Find the 16-byte aligned save area */
@@ -66,56 +61,29 @@ typedef struct {

#endif

-#ifdef __KERNEL__ /* Real code */
-
-/* Note: %cr0 is 32 bits on i386 and 64 bits on x86-64 */
-
-static inline unsigned long raid6_get_fpu(void)
-{
- unsigned long cr0;
-
- preempt_disable();
- asm volatile("mov %%cr0,%0 ; clts" : "=r" (cr0));
- return cr0;
-}
-
-static inline void raid6_put_fpu(unsigned long cr0)
-{
- asm volatile("mov %0,%%cr0" : : "r" (cr0));
- preempt_enable();
-}
-
-#else /* Dummy code for user space testing */
-
-static inline unsigned long raid6_get_fpu(void)
-{
- return 0xf00ba6;
-}
-
-static inline void raid6_put_fpu(unsigned long cr0)
-{
- (void)cr0;
-}
-
+#ifndef __KERNEL__
+/* for user-space testing */
+#define kernel_fpu_begin()
+#define kernel_fpu_end();
#endif

static inline void raid6_before_mmx(raid6_mmx_save_t *s)
{
- s->cr0 = raid6_get_fpu();
+ kernel_fpu_begin();
asm volatile("fsave %0 ; fwait" : "=m" (s->fsave[0]));
}

static inline void raid6_after_mmx(raid6_mmx_save_t *s)
{
asm volatile("frstor %0" : : "m" (s->fsave[0]));
- raid6_put_fpu(s->cr0);
+ kernel_fpu_end();
}

static inline void raid6_before_sse(raid6_sse_save_t *s)
{
unsigned int *rsa = SAREA(s);

- s->cr0 = raid6_get_fpu();
+ kernel_fpu_begin();

asm volatile("movaps %%xmm0,%0" : "=m" (rsa[0]));
asm volatile("movaps %%xmm1,%0" : "=m" (rsa[4]));
@@ -140,14 +108,14 @@ static inline void raid6_after_sse(raid6
asm volatile("movaps %0,%%xmm6" : : "m" (rsa[24]));
asm volatile("movaps %0,%%xmm7" : : "m" (rsa[28]));

- raid6_put_fpu(s->cr0);
+ kernel_fpu_end();
}

static inline void raid6_before_sse2(raid6_sse_save_t *s)
{
unsigned int *rsa = SAREA(s);

- s->cr0 = raid6_get_fpu();
+ kernel_fpu_begin();

asm volatile("movdqa %%xmm0,%0" : "=m" (rsa[0]));
asm volatile("movdqa %%xmm1,%0" : "=m" (rsa[4]));
@@ -172,7 +140,7 @@ static inline void raid6_after_sse2(raid
asm volatile("movdqa %0,%%xmm6" : : "m" (rsa[24]));
asm volatile("movdqa %0,%%xmm7" : : "m" (rsa[28]));

- raid6_put_fpu(s->cr0);
+ kernel_fpu_end();
}

#ifdef __x86_64__
@@ -181,7 +149,7 @@ static inline void raid6_before_sse16(ra
{
unsigned int *rsa = SAREA(s);

- s->cr0 = raid6_get_fpu();
+ kernel_fpu_begin();

asm volatile("movdqa %%xmm0,%0" : "=m" (rsa[0]));
asm volatile("movdqa %%xmm1,%0" : "=m" (rsa[4]));
@@ -222,7 +190,7 @@ static inline void raid6_after_sse16(rai
asm volatile("movdqa %0,%%xmm14" : : "m" (rsa[56]));
asm volatile("movdqa %0,%%xmm15" : : "m" (rsa[60]));

- raid6_put_fpu(s->cr0);
+ kernel_fpu_end();
}

#endif /* __x86_64__ */
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/