Myths about /dev/urandom (2014)

(2uo.de)

51 points | by signa11 4 hours ago

10 comments

  • xiphmont 3 hours ago
    Half the entropy is trying to figure out which pieces of this article's text are supposed to be the silly falsehoods being corrected, and which pieces are just the second or third paragraph of a preceding 'Fact'. Deadpool is easier to follow.
    • Tomte 1 hour ago
      Hey, someone submitted my old article. On my birthday!

      Oh, people hate it… and even someone I definitely look up to.

      You‘re absolutely right, though, I don‘t remember it being that bad, and probably I just read over it when resurrecting the article, because I‘m so familiar with every word.

      I‘ll slap some <hr> tags on it when I‘m back home from my holiday.

      • Sophira 34 minutes ago
        If it helps, it used to look like this: https://web.archive.org/web/20140309183752/http://www.2uo.de...

        Definitely a lot more readable! Something must have changed in the meantime.

        It looks like some links have gone too, like the one in the sentence "how does /dev/random know how much entropy there is available to give out?"

      • gblargg 1 hour ago
        Label the myths rather than leaving them as plain statements.
      • flyinprogrammer 1 hour ago
        Most importantly -- Happy birthday!!!
    • roter 2 hours ago
      I saw a note from an earlier year's discussion saying the css has been changed over the years. Perhaps it was easier then to discern fact or myth, truth or fiction.
    • notnmeyer 2 hours ago
      glad i’m not the only one. i’m more or less baffled reading that.
    • cyanydeez 1 hour ago
      the article is why you need to tell your LLM to 'make noistakes'
  • sph 3 hours ago
    This is a good place as any to ask, last time I didn't get any answer: has there ever been a serious Linux exploit from manipulating/predicting bad PRNG? Apart from the Debian SSH key generation fiasco from years ago, of course.

    Having a good entropy source makes mathematical sense, and you want something a bit more "random" than a dice roll, but I wonder at which point it becomes security theatre.

    Of all the possible avenues for exploiting a modern OS might have, I figure kernel PRNG prediction to be very, very far down the list of things to try.

    • vlovich123 2 hours ago
      It’s both hard to attack but also a hugely audited system with a lot of attention paid.

      That being said, [1] from 2012. The challenge with security is that structural weaknesses can take a long time to be discovered but once they are it’s catastrophic. Modern Linux finally switched to CSPRNG and proper construction and relies less on the numerology of entropy estimation it had been using (ie real security instead of theater). RDRAND has also been there for a long time on the x86 side which is useful because even if it’s insecure it gets mixed with other entropy sources like instruction execution time and scheduling jitter to protect standalone servers and iot devices.

      Of course you hit the nail on the head in terms of the challenge of distinguishing security theater because you won’t know if the hardening is useful until there’s a problem, but there’s enough knowledgeable people on it that it’s less security theater than it might seem if you know what’s going on.

      [1] https://www.usenix.org/system/files/conference/usenixsecurit...

      • jonhohle 1 hour ago
        30 years ago BSDs already had non-blocking /dev/random (there was no difference to /dev/urandom). OpenBSD especially wouldn’t have shipped something known insecure. Blocking random probably caused more issues (DOS, random hangs, etc.) than a no blocking CSPRNG would have.
    • hannob 2 hours ago
      I think this one is among the most significant findings: https://factorable.net/

      I also believe there were some android ASLR issues based on the same weakness (i.e., low early boot-time entropy).

      But this is all quite old, and there've been massive improvements. Basically, "don't use a very old linux kernel" is your mitigation for these issues.

    • wat10000 25 minutes ago
      You can analyze it much like you'd analyze a password. If you construct a password from four words taken from a list of 1024 words, that's 40 bits of entropy. On average, a brute force attacker would have to try 2^39 (half the possibilities) random passwords before cracking your account. You can then apply that number to the time/money required for one attempt, and see if it's sufficiently secure for your tastes. If the answer comes back as 10 minutes, maybe it's not good enough. If it's 10 quadrillion years, you're probably OK.

      If you have bad PRNG, you should be able to quantify it in terms of bits. The Debian bug resulted in 15 bits of randomness, since all inputs to the PRNG were erased except for the pid, which was 15 bits at the time.

      Another real-world example, albeit not Linux. I once worked on a program that had the option of encrypting save files. The encryption was custom (not done by me!) and had a bit of an issue. The encryption itself was not bad, but the save file's master encryption key was generated from the current time. This reduced the number of bits of randomness to well within brute-force range, especially if you could guess at roughly when the key was created. This was convenient for users who had lost their passwords, but somewhat less convenient for users who wanted to actually protect their data.

      An attacker isn't going to spontaneously try breaking your PRNG, but if you do have an issue, it's a real concern. It'll be far down the list of things to try just because any modern system will hopefully have very good randomness.

  • throw0101c 3 hours ago
  • jonhohle 1 hour ago
    Back in the dinosaur days (around 2005) I was working on a PHP CMS used by a big registrar. Occasionally page loads would block for seconds. It appeared randomly (natch) and was relatively unreproducible.

    I couldn’t find any good way to debug it and a friend suggested GDB. I had never thought of using such a low level debugger on a scripting language, but what choice did I have? Fired it up, found a blocked process and sure enough it was blocked on reads to /dev/random.

    I leaned two things that day: the decision to make and keep /dev/random blocking was dumb and GDB (or lldb, or valgrind, etc.) is useful for debugging just about anything.

  • chaboud 2 hours ago
    I woke up around 4am, read this, and wondered if I was still in a dream state given the meandering nature of it.

    Were the man page musings written in response to the (alleged, but... uh... NSA) kleptographic backdoor in Dual_EC_DRBG? It requires multiple successive outputs to compromise and derive internal PRNG state, if memory serves.

    In that one construction, /dev/random blocking on seeding would have a mild state-hiding advantage over /dev/urandom, I imagine... but, sheesh. Nobody use that generator.

  • iamtedd 2 hours ago
    That was hard to tell where the additional commentary on the fact ended and the next myth started.
  • iamtedd 2 hours ago
    Twelve years later, if there's still so much misconception about /dev/(u)random, has the man page been fixed?

    Edit: can't count.

  • NooneAtAll3 3 hours ago
    (2014)
    • ape4 3 hours ago
      Ah, I wonder what's change since then.
      • vbezhenar 3 hours ago
        Here's quote from the article:

        > Note from 2024: This article was published on March 16th, 2014. It is still correct in its discussion of entropy and randomness, but the Linux kernel random number generator has been reworked several times since then and does not look like this anymore. Good news: the separation between /dev/urandom and /dev/random is practically gone.

        My understanding is that on modern Linux system:

        At early boot phases, /dev/random can still block, because not enough entropy has been seeded yet. /dev/urandom will not block, but the random data might be of poor quality and not suitable for crypto purposes. This happens very early in the boot, so probably it's not even possible to run user stuff at this time. At least on my laptop, the message "random: crng init done" gets logged almost instantly after boot and long before even initrd starts. Might be different for exotic platforms, I guess.

        Once there was enough entropy seeded, both /dev/random and /dev/urandom works identically, they don't block and they return high quality random data. So for most userspace purposes, these files can be used interchangeably, one is not better than another.

        • aomix 2 hours ago
          It started looking a whole lot like OpenBSD’s random number system. Private entropy pool from good system entropy seeds a ChaCha20 stream with random reseeds for forward secrecy in case of compromise. I think Linux is even more paranoid in the early boot environment where even in the presence of a seed file it prefers to get system entropy mixed in before confidently saying it can do crypto activities.
        • mananaysiempre 3 hours ago
          > Might be different for exotic platforms, I guess.

          Short-lived isolated VMs (like might be used for CI) are one place where entropy can be a problem. The relevant definition of “platform” here is less about the CPU architecture and more about the environment.

          • ChrisSD 2 hours ago
            VMs should have VirtIO RNG set up so there is a high quality entropy source at boot time, provided by the host system.
            • mananaysiempre 2 hours ago
              Should, yes. Will, perhaps, but better be aware of the potential problem and check.

              Just yesterday I encountered people complaining about a VM not connecting to a cloud service when they neglected to put their DNS server’s address in the config for the DHCP server used by that particular host. And a dysfunctional RNG is much more difficult to detect.

  • EMPTYCONTOUR 3 hours ago
    [flagged]