• irotsoma
    link
    English
    286 months ago

    Because computers have come even close to needing more than 16 exabytes of memory for anything. And how many applications need to do basic mathematical operations on numbers greater than 2^64. Most applications haven’t even exceeded the need for 32 bit operations, so really the push to 64bit was primarily to appease more than 4GB of memory without slow workarounds.

    • @[email protected]
      link
      fedilink
      English
      56 months ago

      I know a google engineer who was saying they’re having to update their code bases to handle > 16 exabytes of storage, if you can imagine. But yeah, that’s storage, not RAM.

      • @[email protected]
        link
        fedilink
        English
        46 months ago

        I would kind of enjoy the trouble of needing to store and owning the place for 16 exabytes…

    • Justin
      link
      fedilink
      English
      46 months ago

      Tons of computing is done on x86 these days with 256 bit numbers, and even 512-bit numbers.

      • @pivot_root
        link
        English
        156 months ago

        Being pedantic, but…

        The amd64 ISA doesn’t have native 256-bit integer operations, let alone 512-bit. Those numbers you mention are for SIMD instructions, which is just 8x 32-bit integer operations running at the same time.

        • @[email protected]
          link
          fedilink
          English
          36 months ago

          The ISA does include sse2 though which is 128 bit, already more than the pointer width. They also doubled the number of xmm registers compared to 32-bit sse2.

          Back in the days using those instructions often gained you nothing as the CPUs didn’t come with enough APUs to actually do operations on the whole vector in parallel.

        • Justin
          link
          fedilink
          English
          1
          edit-2
          6 months ago

          Ah fair enough, I figured that since the registers are 512 bit, that they’d support 512 bit math.

          It does look like you can load/store and do binary operations on 512-bit numbers, at least.

          Not much difference between 8x64 and 512 when it comes to integer math, anyways. Add and subtract are completely identical.

      • @[email protected]
        link
        fedilink
        English
        6
        edit-2
        6 months ago

        You can always combine integer operations in smaller chunks to simulate something that’s too big to fit in a register. Python even does this transparently for you, so your integers can be as big as you want.

        The fundamental problem that led to requiring 64-bit was when we needed to start addressing more than 4 GB of RAM. It’s kind of similar to the problem of the Internet, where 4 billion unique IP addresses falls rather short of what we need. IPv6 has a host of improvements, but the massively improved address space is what gets talked about the most since that’s what is desperately needed.

        Going back to RAM though, it’s sort of interesting that at the lowest levels of accessing memory, it is done in chunks that are larger than 8 bits, and that’s been the case for a long time now. CPUs have to provide the illusion that an 8-bit byte is the smallest addressible unit of memory since software would break badly were this not the case, but it’s somewhat amusing to me that we still shouldn’t really need more than 32 bits to address RAM at the lowest levels even with the 16 GB I have in my laptop right now. I’ve worked with 32-bit microcontrollers where the byte size is > 8 bits, and yeah, you can have plenty of addressible memory in there if you wanted.