News: 0175295833

  ARM Give a man a fire and he's warm for a day, but set fire to him and he's warm for the rest of his life (Terry Pratchett, Jingo)

AI 'Bubble' Will Burst 99% of Players, Says Baidu CEO (theregister.com)

(Monday October 21, 2024 @05:22PM (msmash) from the shape-of-things-to-come dept.)


Baidu CEO Robin Li has proclaimed that hallucinations produced by large language models are no longer a problem, and predicted [1]a massive wipeout of AI startups when the "bubble" bursts . From a report:

> "The most significant change we're seeing over the past 18 to 20 months is the accuracy of those answers from the large language models," gushed the CEO at last week's Harvard Business Review Future of Business Conference. "I think over the past 18 months, that problem has pretty much been solved â" meaning when you talk to a chatbot, a frontier model-based chatbot, you can basically trust the answer," he added.

>

> Li also described the AI sector as in an "inevitable bubble," similar to the dot-com bubble in the '90s. "Probably one percent of the companies will stand out and become huge and will create a lot of value or will create tremendous value for the people, for the society. And I think we are just going through this kind of process," stated Li. The CEO also guesstimated it will be another 10 to 30 years before human jobs are displaced by the technology. "Companies, organizations, governments and ordinary people all need to prepare for that kind of paradigm shift," he warned.



[1] https://www.theregister.com/2024/10/20/asia_tech_news_roundup/



I mean, accurate enough. (Score:3)

by HBI ( 10338492 )

If you constrain the answers from the LLM, you can avoid the hallucinations. Of course, you're creating a jury-rigged system of input control which requires constant maintenance. Whether you've also constrained the usefulness of the answers is debatable.

As far as 99% of the players going bust, that is easy to predict. Even Gartner essentially says that with their hype cycle.

Re: (Score:2)

by Hadlock ( 143607 )

The current solution seems to be to have a second AI evaluate something like, "is this a valid request?" and then if true, pass it along

Re: (Score:3)

by SchroedingersCat ( 583063 )

In other words, to understand the Ultimate Answer, you need to produce the Ultimate Question which requires much more sophisticated AI.

Re: (Score:2)

by ravenshrike ( 808508 )

Why not have 3 AIs that must be in agreement? We can call them Melchior, Balthasar, and Caspar.

Re: (Score:2)

by GoTeam ( 5042081 )

> Why not have 3 AIs that must be in agreement? We can call them Melchior, Balthasar, and Caspar.

Because they rarely ever report back to the king... I mean requestor...

Re: (Score:2)

by taustin ( 171655 )

That amounts to "training an AI on AI generated sources," which has been pretty conclusively proven to provide garbage output.

Re: I mean, accurate enough. (Score:2)

by Hadlock ( 143607 )

I'm not sure how filtering input to receive good output has anything to do with the completely separate training step, that happens wholly before the "using the model" step

Re: (Score:2)

by Geoffrey.landis ( 926948 )

> If you constrain the answers from the LLM, you can avoid the hallucinations.

What does that mean, "constrain the answers"? Constrain them to what?

Re: (Score:2)

by HBI ( 10338492 )

There have been some interesting articles about this topic. Cross-referencing multiple LLMs is one concept. Another, more common one is based on pure input control, disallowing certain prompts. All have the net effect of limiting the answers given, aka output.

Your opinion of this and its efficacy for your use case is intensely personal.

Re: (Score:2)

by Geoffrey.landis ( 926948 )

> There have been some interesting articles about this topic. Cross-referencing multiple LLMs is one concept.

i.e., you're saying to have one LLM filter the results of the first one, throwing out the bad answers? Not sure if I would have labeled that "constraint," but ok, sounds like an approach, as long as the two models have different training.

> Another, more common one is based on pure input control, disallowing certain prompts.

How do you decide what inputs to disallow?

> All have the net effect of limiting the answers given, aka output. Your opinion of this and its efficacy for your use case is intensely personal.

not sure what this means, but yes, I expect everybody's use case is going to be different.

Re: I mean, accurate enough. (Score:2)

by Brownstar ( 139242 )

Look up retrieval augmented generation.

Basically you add specific context to your system and it saves it off. That is then used to constrain what is generated.

Obviously if you provide crap context, youâ(TM)ll still get crap out put.

Notebooklm from Google is a good example of this that you can use, without needing to roll your own solution

Re: (Score:2)

by Thelasko ( 1196535 )

> As far as 99% of the players going bust, that is easy to predict. Even Gartner essentially says that with their hype cycle.

Just remember, during a gold rush, invest in shovels!

Re: (Score:2)

by AvitarX ( 172628 )

I suspect it's already too late to invest in Nvidia.

Uh-huh, I've Got a Bridge to Sell Ya (Score:2)

by crunchy_one ( 1047426 )

> meaning when you talk to a chatbot, a frontier model-based chatbot, you can basically trust the answer

Re: (Score:3)

by SchroedingersCat ( 583063 )

You can trust the answer as much as you can trust the first hit on Google search results. In other words - YMMV.

gushed the CEO... (Score:2)

by narcc ( 412956 )

> "I think over the past 18 months, that problem has pretty much been solved" meaning when you talk to a chatbot, a frontier model-based chatbot, you can basically trust the answer,"

This is very silly, but it doesn't look like he actually believes that:

> The CEO also guesstimated it will be another 10 to 30 years before human jobs are displaced by the technology.

It looks like the CEO has already been replaced by a chatbot...

Difference between boom and bubble (Score:3)

by Targon ( 17348 )

People don't seem to have an understanding between a boom and a bubble. A boom will be when there is a surge in companies getting into any given area. A bubble is when you have USELESS companies jumping into that sector, they don't have product, they don't have expertise, but they have venture capital money. When there is a large number of useless companies jumping into a booming sector, that becomes a bubble that will eventually pop.

Hype...yea, let's take an area that has the potential for growth, and generate so much hype that useless companies start to pop up just to take advantage of the boom, and then, 2-3 years later, they collapse.

Not a great comparison (Score:3)

by RogueWarrior65 ( 678876 )

In the dotcom era, anyone could build a website and startups blew their funding on expensive offices, Herman-Miller chairs, Silicon Graphics workstations, and a powered paragliders (yes I saw all of this at a bankruptcy auction in Santa Monica). None of that stuff was required to build a product. LLMs require a lot of expensive infrastructure and you need the proverbial killer-app to pitch to investors who are a little bit smarter than they were 20+ years ago. That's going to increase the signal-to-noise ratio.

Re: (Score:2)

by toxonix ( 1793960 )

I wanted an SGI but all they'd give me was an Indy. At least it had the Nintendo 64 dev board installed in it.

What problems do LLMs solve? (Score:2, Insightful)

by RobinH ( 124750 )

When faced with a so-called solution, you really need to focus on what problem it's solving. I've seen generative AI that can create all kinds of clipart or thumbnail images which are good enough for a business presentation or a low-production-value YouTube video. That's solving a real problem that people have, and I see evidence of people using it (and digital artists looking to change careers). On the other hand, where are the legions of people who are using LLMs to solve real-world problems? High sch

Re: (Score:3)

by serviscope_minor ( 664417 )

The next generation will do all the useful things. We are on the next, you say? I mean the next next one of course. Now hand me that briefcase of VC cash.

Re: (Score:3)

by Whateverthisis ( 7004192 )

I'm a huge AI skeptic. However...

I have found uses for it. My family and I go on relatively long vacations to places we've never been, and the research to plan for it can sometimes outstrip the time we actually are on the vacation. ChatGPT does a great job of condensing a rather large amount of information into a condensed, reasonably well organized narrative that cuts down that research tremendously.

I've also used it for creating marketing material at work and designing entire D&D campaigns.

Re: (Score:2)

by BadDreamer ( 196188 )

These examples are good, and they have one thing in common. You either do not worry about the accuracy of the output, or you have enough expertise to easily correct and make use of it despite the errors.

This is the huge issue with text generators. They're really good at assembling something which will look correct (if often a bit verbose) but which has subtle, and sometimes gross, faults.

And this is the main issue with it being used to create actionable tasks, especially when that is automated and the tasks

Re: (Score:2)

by Thelasko ( 1196535 )

You know when your boss asks you to put together a bullshit presentation that is a waste of everyone's time? LLMs are great for that!

Re: (Score:2)

by Lije Baley ( 88936 )

Bingo. The problem is that there are way too many people employed creating B.S. presentations and doing other things that glorify their managers' existence. Among them are plenty who can't just switch to a trade or healthcare job due to physical limitations. Though I may not like it, I'd much rather see companies continue paying people to do B.S. jobs than to see them on welfare.

Re: (Score:2)

by RobinH ( 124750 )

I'm well aware of the fact that there's a middle-management layer in corporations where this kind of waste happens, but it's important to realize that the money to pay these people comes out of real productivity (somewhere) in the organization. In the case of something like an automobile factory it's the people (or robots) attaching components onto the frame, or at the foundry where we need to use massive amounts of energy to heat a batch of steel to the right temperature, or in the case of Starbuck's it's

Re: (Score:2)

by neilo_1701D ( 2765337 )

> Seriously, if LLMs were as valuable a the hype claimed, you'd already be seeing a profound impact, and we're simply not seeing it.

It depends where you are looking.

If you have a Microsoft Copilot enterprise license, Microsoft Dynamics 365 (their flagship ERP system) and the extensive setup needed, you can do some very nice natural language queries of the database that would be a pain otherwise:

> How many jeans in style 501 that are black and size 35 can I have delivered to 123 Anystreet, SomeTown, SomeState by Friday?

We have customers using that right now; their customer support call centers can use these natural language queries whils

Re: (Score:2)

by gweihir ( 88907 )

I found two uses so far: Better lies ("how do I say xyz in a positive way") and better search, but with very strong limitations as asking for references often fails or produces very little. I had one instance of something I was searching but did not know the term for. ChatCPT provided several candidates for the search term and conventional DuckDuckGo got me what I was looking for. That is basically it. Not that impressive, more an incremental improvement. Useful, yes, but not a game-changer.

To be sure, this

"the society" (Score:3)

by VeryFluffyBunny ( 5037285 )

He read about it in a magazine once. Apparently, people react positively to the term so he's going to use it.

So... (Score:3)

by zkiwi34 ( 974563 )

Was that his prediction, or an AI prediction?

the usual (Score:2)

by CoderFool ( 1366191 )

What usually happens is something new and hot comes out and everyone things it will solve all society's ill and and bring world peace. So everyone jumps on board.

After a while, when it becomes clear to most what this new thing is actually capable of and where it is actually useful, then most people jump off it onto the next new shiny thing.

examples include bluetooth, agile, self-driving cars, EVs

AI Telephone? (Score:2)

by gtall ( 79522 )

I wonder what happens when you string together AI-bots to play telephone? Do the responses converge to a single response or do they gyrate wildly?

Bullshit (Score:2)

by gweihir ( 88907 )

Hallucinations are alive and well. So is overlooking of critical details. You just need to ask something that you could not have readily looked up in Wikipedia.

I do agree to 10-30 years before significant job loss. I do expect specialized models will make that likely. It will mostly affect "bullshit" jobs, but there are many of those.

Looks nice to me but about the only way you are likely to get Linus to take
in kernel debugging patches is to turn them into hex and disguise them as USB
firmware ;)

- Alan Cox's guide on submitting Linux patches, today:
chapter #3, kernel debuggers