The fact that this has been replicated is amazing!

  • @Adeptfuckup
    link
    English
    61 year ago

    Hitch your tits and pucker up. We’re entering a new age of industry. Much like the original Industrial Revolution, technology is going to advance at an extremely rapid pace. Fusion, quantum computing supremacy. Just… wow. How far off is general AI with this new room temperature superconductor?

    • Dr. Dabbles
      link
      English
      131 year ago

      Fusion is no closer than ever before, and AGI is hilariously over hyped. Also no closer than ever before.

      • @[email protected]
        link
        fedilink
        English
        21 year ago

        And Fusion is pretty close to begin with. Commonwealth Fusion is well within their purpose time table so far. They don’t need any new superconductors for their project.

    • @[email protected]
      link
      fedilink
      English
      31 year ago

      Stupid question probably - is computing power what is holding back general AI? I’ve not heard that.

      • Dr. Dabbles
        link
        English
        141 year ago

        What’s holding back AGI is a complete lack of progress toward anything like intelligence. What we have now isn’t intelligent, it’s multi-variable probability.

        • @JGrffn
          link
          English
          41 year ago

          It’s not that it’s not intelligent, it’s that predictive language models are obviously just one piece of the puzzle, and we’re going to need all the pieces to get to AGI. It’s looking incredibly doable if we figured out how to make something that’s dumb but sounds smarter than most of us already. We just need to connect it to other models that handle other things better.

          • Dr. Dabbles
            link
            English
            11 year ago

            You don’t speak predictively. It’s not one of the pieces, it’s a parlor trick.

          • @Ogmios
            link
            English
            1
            edit-2
            1 year ago

            deleted by creator

        • @MuThyme
          link
          English
          3
          edit-2
          1 year ago

          deleted by creator

            • @MuThyme
              link
              English
              2
              edit-2
              1 year ago

              deleted by creator

                • @MuThyme
                  link
                  English
                  2
                  edit-2
                  1 year ago

                  deleted by creator

        • @[email protected]
          link
          fedilink
          English
          11 year ago

          Really appreciate the write up! I didn’t know the computing power required!

          Another stupid question (if you don’t mind) - adding superconductors to GPUs doesn’t really se like it would make a huge difference on the heat generation. Sure, some of the heat generated is through trace resistance, but the overwhelming majority is the switching losses of the transistors which will not be effected by superconductor technology. Are we assuming these superconductors will be able to replace semiconductors too? Where are these CPU/GPU efficiencies coming from?

            • @[email protected]
              link
              fedilink
              English
              21 year ago

              Semiconductors are used for transistors because they give us the ability to electrically control whether they conduct or resist electrical current. I don’t know what mechanism you’d use to do that with superconductors. I agree you don’t ‘have’ to have resistance in order to achieve this functionality, but at this time semiconductors or mechanical relays are the only ways we have to do that. My focus is not in semiconductor / IC design either so I may by way off base, but I don’t know of a mechanism that would allow superconductors to function as transistors (or “electrically controlled electrical connections”), but I really hope I’m wrong!

      • @knotthatone
        link
        English
        41 year ago

        Simply throwing computing power at the existing models won’t get us general AI. It will let us develop bigger and more complex models, but there’s no guarantee that’ll get us closer to the real thing.