Re: [PATCH] ARC: Fix __udelay parentheses

From: Joe Perches
Date: Wed Aug 28 2013 - 15:01:52 EST


On Wed, 2013-08-28 at 18:53 +0000, Mischa Jonker wrote:
> > > Make sure that usecs is casted to long long, to ensure that the (usecs
> > > * 4295 * HZ) multiplication is 64 bit.
> > >
> > > Initially, the (usecs * 4295 * HZ) part was done as a 32 bit
> > > multiplication, with the result casted to 64 bit. This led to some
> > > bits falling off.
> > >
> > > Signed-off-by: Mischa Jonker <mjonker@xxxxxxxxxxxx>
> > > ---
> > > arch/arc/include/asm/delay.h | 4 ++--
> > > 1 file changed, 2 insertions(+), 2 deletions(-)
> > >
> > > diff --git a/arch/arc/include/asm/delay.h
> > > b/arch/arc/include/asm/delay.h index 442ce5d..8d35fe1 100644
> > > --- a/arch/arc/include/asm/delay.h
> > > +++ b/arch/arc/include/asm/delay.h
> > > @@ -56,8 +56,8 @@ static inline void __udelay(unsigned long usecs)
> > > /* (long long) cast ensures 64 bit MPY - real or emulated
> > > * HZ * 4295 is pre-evaluated by gcc - hence only 2 mpy ops
> > > */
> > > - loops = ((long long)(usecs * 4295 * HZ) *
> > > - (long long)(loops_per_jiffy)) >> 32;
> > > + loops = (((long long) usecs) * 4295 * HZ *
> > > + (long long) loops_per_jiffy) >> 32;
> >
> > Shouldn't this be unsigned long long or u64?
> Yes, it should, but that is not directly related to the issue:)
>
> > Why is it >> 32 again?
> >
> > The comment above it doesn't seem to match the code.
> >
> The original code explains about the >> 32:
>
> /*
> * Normal Math for computing loops in "N" usecs
> * -we have precomputed @loops_per_jiffy
> * -1 sec has HZ jiffies
> * loops per "N" usecs = ((loops_per_jiffy * HZ / 1000000) * N)
> *
> * Approximate Division by multiplication:
> * -Mathematically if we multiply and divide a number by same value the
> * result remains unchanged: In this case, we use 2^32
> * -> (loops_per_N_usec * 2^32 ) / 2^32
> * -> (((loops_per_jiffy * HZ / 1000000) * N) * 2^32) / 2^32
> * -> (loops_per_jiffy * HZ * N * 4295) / 2^32
> *
> * -Divide by 2^32 is very simply right shift by 32
> * -We simply need to ensure that the multiply per above eqn happens in
> * 64-bit precision (if CPU doesn't support it - gcc can emaulate it)
> */

I don't see the loops_per_jiffy initial shift << 32.

> The problem is that the original code already _tried_ to cast to 64 bits (to ensure a 64 bit multiply), but only did so after the multiply (so 32 bit multiply).

I know that. It's the use of a signed long long
vs the unsigned long long that I think wrong.

Why cast a unsigned to a signed?


--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/