The 200 year-old company may soon go public on the back of AI-powered education products.

  • @[email protected]
    link
    fedilink
    English
    182 days ago

    Wikipedia is just like valve lol
    do literally nothing and your opponents will kill themselves

  • @[email protected]
    link
    fedilink
    English
    342 days ago

    Encyclopaedia Britannica—now known as just Britannica— is all in on artificial intelligence, and may soon go public at a valuation of nearly $1 billion

    All in on enshittification, like everyone else.

  • Diplomjodler
    link
    English
    49
    edit-2
    1 day ago

    I miss the days when you could just do a thing and as long as you made more money than you spent, just keep doing it. Now the line must always go up, no matter what. And if your business model isn’t profitable enough they’ll just shut down 200 years of tradition without a second thought.

    • @ikidd
      link
      English
      41 day ago

      I think that profit ship sailed for EB long ago.

    • @then_three_more
      link
      English
      72 days ago

      Surely companies have always adapted and changed products. They’re not exactly able to make much of encyclopedias anymore when they’re main competitor is free to use and has volunteers writing and editing it.

  • @[email protected]
    link
    fedilink
    English
    392 days ago

    More general-purpose models like ChatGPT suffer from hallucinations because they have hoovered up the entire internet, including all the junk and misinformation.

    Incorrect. ChatGPT hallucinates because that’s how LLMs work. Hoovering up misinformation is a separate problem.

    A company in the space of selling educational books that has seen its fortunes go the opposite direction is Chegg. The company has seen its stock price plummet almost in lock-step with the rise of OpenAI’s ChatGPT, as students canceled their subscriptions to its online knowledge platform.

    Incorrect. Chegg is a cheating platform. It is the opposite of a knowledge platform.

    Why is Gizmodo paying people to write articles who apparently know pretty much nothing about the subject they are writing about?

    • @Alexstarfire
      link
      English
      62 days ago

      Having bad information in your dataset surely has to increase the odds of hallucinations though.

    • FaceDeer
      link
      fedilink
      -22 days ago

      How dare you say something insufficiently negative about the stuff everyone hates.

      • Lemminary
        link
        English
        62 days ago

        The downvotes are for the naïveté of the statement. Many people here use LLMs every day and have stated so in other threads. We just don’t think this is necessarily a proper use case given that you’re dealing with factual information. You can see as much in other comments on this thread pointing out the hallucinations.

        • FaceDeer
          link
          fedilink
          31 day ago

          Whereas I use LLMs every day, have actually written code that uses them, and I understand that they’re perfectly fine dealing with factual information when used in the proper framework. You’d be using retrieval-augmented generation (RAG) in an application like this.

          The “but hallucinations!” Objection goes in the same bin as “they can’t do fingers.” It’s an old concern that’s had a lot of work done to resolve it but that the general public haven’t bothered to keep up with.

          • Lemminary
            link
            English
            11 day ago

            “they can’t do fingers.” It’s an old concern

            Have you seen those gorilla hands, though? Yes, there are five fingers there but everyone got fucking man hands. lmao

            It seems RAG helps mitigate but doesn’t eliminate hallucinations yet. Not to mention it’s quite expensive and has trouble extracting information based on abstract concepts. It sounds promising but it’s not the silver bullet I’m being sold on.