The 200 year-old company may soon go public on the back of AI-powered education products.
Wikipedia is just like valve lol
do literally nothing and your opponents will kill themselvesEncyclopaedia Britannica—now known as just Britannica— is all in on artificial intelligence, and may soon go public at a valuation of nearly $1 billion
All in on enshittification, like everyone else.
I miss the days when you could just do a thing and as long as you made more money than you spent, just keep doing it. Now the line must always go up, no matter what. And if your business model isn’t profitable enough they’ll just shut down 200 years of tradition without a second thought.
I think that profit ship sailed for EB long ago.
Surely companies have always adapted and changed products. They’re not exactly able to make much of encyclopedias anymore when they’re main competitor is free to use and has volunteers writing and editing it.
More general-purpose models like ChatGPT suffer from hallucinations because they have hoovered up the entire internet, including all the junk and misinformation.
Incorrect. ChatGPT hallucinates because that’s how LLMs work. Hoovering up misinformation is a separate problem.
A company in the space of selling educational books that has seen its fortunes go the opposite direction is Chegg. The company has seen its stock price plummet almost in lock-step with the rise of OpenAI’s ChatGPT, as students canceled their subscriptions to its online knowledge platform.
Incorrect. Chegg is a cheating platform. It is the opposite of a knowledge platform.
Why is Gizmodo paying people to write articles who apparently know pretty much nothing about the subject they are writing about?
Having bad information in your dataset surely has to increase the odds of hallucinations though.
Because they know their audience.
I mean, this makes sense and could actually be a positive thing to look stuff up
How dare you say something insufficiently negative about the stuff everyone hates.
The downvotes are for the naïveté of the statement. Many people here use LLMs every day and have stated so in other threads. We just don’t think this is necessarily a proper use case given that you’re dealing with factual information. You can see as much in other comments on this thread pointing out the hallucinations.
Whereas I use LLMs every day, have actually written code that uses them, and I understand that they’re perfectly fine dealing with factual information when used in the proper framework. You’d be using retrieval-augmented generation (RAG) in an application like this.
The “but hallucinations!” Objection goes in the same bin as “they can’t do fingers.” It’s an old concern that’s had a lot of work done to resolve it but that the general public haven’t bothered to keep up with.
“they can’t do fingers.” It’s an old concern
Have you seen those gorilla hands, though? Yes, there are five fingers there but everyone got fucking man hands. lmao
It seems RAG helps mitigate but doesn’t eliminate hallucinations yet. Not to mention it’s quite expensive and has trouble extracting information based on abstract concepts. It sounds promising but it’s not the silver bullet I’m being sold on.