On 2015-09-25 07:41, Austin S Hemmelgarn wrote:Apologies for not replying yesterday like I said I would.
On 2015-09-24 16:14, Theodore Ts'o wrote:OK, just started a couple of runs in parallel using different generators
On Thu, Sep 24, 2015 at 03:11:23PM -0400, Austin S Hemmelgarn wrote:I sincerely apologize about this, I should have been more specific right
That is a startling result. Please say what architecture, kernel
version, dieharder version and commandline arguments you are using to
get 10% WEAK or FAILED assessments from dieharder on /dev/urandom.
I do not remember what exact dieharder version or command-line
(this was almost a decade ago), except that I compiled it from source
myself, I do remember it was a 32-bit x86 processor (as that was
sadly all I
had to run Linux on at the time), and an early 2.6 series kernel
(which if I
remember correctly was already EOL by the time I was using it).
It might have been nice if you had said this from the beginning
instead of making an unqualified statement with the assumption that it
was applicable to kernels likely to be used today in non-obsolete
systems. Otherwise it risks generating a click-bait article on
Phoronix that would get people really worried for no good reason...
from the beginning (I need to get better about that when talking to
people, I'm so used to dealing with some of my friends who couldn't
event tell you the difference between RAM and a hard drive, think a bus
is only something you use for transportation, and get confused when I
try to properly explain even relatively simple CS and statistics
I don't think this was what I hit, I'm pretty sure I had serialized the
There was a bug a long, long time ago (which where we weren't doing
sufficient locking and if two processes raced reading from
/dev/urandom at the same time, it was possible that the two processes
would get the same value read out from /dev/urandom). This was fixed
a long time ago, though, and in fact the scalability problem which
Andi is trying to fix was caused by that extra locking that was
It's possible that is what you saw. I don't know, since there was no
reproduction information to back up your rather startling claim.
If you can reproduce consistent Dieharder failures, please do let us
know with detailed reproduction instructures.
using the following command line:
dieharder -a -m 32 -k 1 -Y 1 -g XXX
with one each for:
glibc random() (039)
The above command line will run all dieharder tests with 12800 psamples,
using a higher than default precision, and re-running tests that return
WEAK until it gets a PASS or FAIL. Even on the relatively fast (at
least, fast for a desktop) system I'm running them on, I expect it will
take quite some time to finish (although regardless of that I'm probably
not going to be getting back to it until Monday).
Interestingly, based on what dieharder is already saying about
performance, /dev/urandom is slower than AES_OFB (at least, on this
particular system, happy to provide hardware specs if someone wants).
Description: S/MIME Cryptographic Signature