Bonum Certa Men Certa

The Great LLM Delusion - Part IV: Academic Papers as Microsoft Marketing for LLMs

posted by Roy Schestowitz on Jan 31, 2024

Photo capturing two very different architectural styles for student buildings at the UCLA Campus, California

"Strange and misleading article about LLMs," as explained by an anonymous contributor

THIS series coincides with Microsoft hype and vapourware (much-needed distractions). This part is a guest post of sorts, an unedited version of something a reader sent us. The supporting material is in there.

To be clear, you can talk to a chatbot. The chatbot won't talk back to you. It'll just spew out words, some of them merely plagiarised based on words similar to what you said (which the chatbot does not grasp, it's more like a Web search, except there's no attribution/link to source). Chatbots might lower the bar for journalism - to the point where Web as a whole will lose legitimacy, trust, value etc. Then what? Going back to physical libraries? Saying you can compose a physical book using a chatbot is like saying you can make a very large meal by assembling trash and cooking parts of it (LLMs are "digital pagpag"). Given reports like "Scammy AI-Generated Book Rewrites Are Flooding Amazon", this is already a real problem. "These 'AI' stock increases based on fake increases in revenue," an associate has remarked, "appear funded by mass firings to appease the LARPers in the financial community, That can only go on so long before they run out of people to take care of the core income-generating activities, a line which I suspect they have already crossed."

Will Microsoft also start spewing out "papers" or "publication" made by its chatbots, in order to generate hype about chatbots? That probably would not work, as the quality would not meet basic criteria.

Without further ado, here is the contributor's message:


I stumbled upon a recent article you may find curious.

While reading comments on a post at Bruce Schneier's blog, I saw a user who posted the following link as a kind of "proof" that conversations with LLM-based chatbots can be "useful" and "interesting."

Of course, it sparked my interest at first, but as I started reading it, red flags started to pop up here and there.

I do not know much about Quanta Magazine's credibility. At first, I thought that it was some semi-crackpot pop science news site, but after a shallow search, I saw a good rank from a fact-checking site.

The article was published on January 22, 2024, and the research it discussed was released on October 26, 2023. May be it does not mean much—just a few months—but it is a bit suspicious that the research paper is apparently not peer reviewed (just published on arXiv and cited in ~2–3 sources), and the article about it came out in parallel with "AI" swindle failure unraveling.

It seems like the article is desperately trying to spark new interest in readers regarding LLMs and chatbots, saying that there is some evidence that there is "much more than just autocomplete."

Following are some dubious parts.

1. The article talks about the "understanding" of something by LLMs but presents no clear definition of it.

The thing that can pass as a semi-definition (from the research paper)—"combinations that were unlikely to exist in the training data"—is, in my opinion, misleading for ordinary people. Much like other misnomers in the field (e.g., "hallucinations").

I guess it may be suitable to talk about "competence" instead, as in the "competence without comprehension" phrase from Dennet's writings.

2. The paper described in the article seems to support (or go in the direction of) the vague idea that if you shovel a lot of data and complexity into "AI" (LLM in this case), then "something" will emerge ("skills" and "ability to generalize" in this case, as stated in the paper and researcher's comments in the article). I find it concerning.

3. "Research scientist at Google DeepMind" among the authors of the paper, so it is probably not clearly independent (from corporate influence) research.

4. “[They] cannot be just mimicking what has been seen in the training data,” said Sébastien Bubeck, a mathematician and computer scientist at Microsoft Research who was not part of the work. “That’s the basic insight.”

Wait, what? Why is this part inserted in the article at all? Some guy from Microsoft is eager to tell us that LLMs are "something more." No bullshit. What a surprise!

5. The paper starts with this passage: "With LLMs shifting their role from statistical modeling of language to serving as general-purpose AI agents..."

I mean, what the fuck?! LLMs are not "shifting" anywhere; they are poorly shoehorned into use cases where a "general-purpose AI agent" is required (whatever it is, it does not exist in our reality anyway) by people who want to reap profits from selling half-assed "products" based almost entirely on lies! LLMs are definitely not suitable for general-purpose tasks other than text manipulation or some kinds of entertainment where facts, preciseness, and responsibilities do not matter at all.

One of the researchers acknowledges that it is not about accuracy.

"Arora adds that the work doesn’t say anything about the accuracy of what LLMs write. “In fact, it’s arguing for originality,” he said. “These things have never existed in the world’s training corpus. Nobody has ever written this. It has to hallucinate.”

I need to make it clear: I have no competence to review the actual paper; this task requires actual experts in the field.

As far as I understand the paper, the researchers devised some abstractions to describe observations they already made and try to construct a method that would be useful to work with their definitions and hypotheses that have a little in common with laymen's definitions (e.g., for terms like "understanding" and "creativity") and perceptions of the matter.

I tried to read the paper with an open mind to avoid at least some obvious biases. I have no problems with the paper; maybe it is actual useful research that will serve to advance the field (and not the companies of con artists)—I cannot say for sure.

What bothers me are the misnomers, misleading, and vague terms and descriptions in the paper (less) and the article (a great deal) based on it. In my opinion, the article commits the crime of severely misinforming the reader.

Other Recent Techrights' Posts

Links 07/10/2024:China’s 'Deflation' (Price Decreases), Brazil Still Bars Twitter ("X")
Links for the day
Links 07/10/2024: "Creative Computing" Turns 50, Long War in Middle East Turns 1
Links for the day
Gemini Links 07/10/2024: Luck and Dishonesty, Gaming Getting Worse
Links for the day
Over at Tux Machines...
GNU/Linux news for the past day
IRC Proceedings: Sunday, October 06, 2024
IRC logs for Sunday, October 06, 2024
EPO: We Give Recognition to Frauds
Good to see some frank recognition right there in the EPO's own Web site
Even Though We Don't Focus on statCounter for Now (Not Our Top Priority) GNU/Linux Reaches New Highs This Month:
We caught GNU/Linux at 4.86% before, but only temporarily
Links 06/10/2024: Ham Radio for Recovery, Health Problems Worldwide
Links for the day
Gemini Links 06/10/2024: Special Interest Galore and Religion
Links for the day
Keeping Control Out of Dictators' Hands
When people are just "numbers"...
Links 06/10/2024: Misinformation Growing on the Web, "Hey Hi" Hype Waning for Lack of RoI
Links for the day
[Meme] Years Have Passed and EPO Management Still Isn't Obeying a Ruling From a Court Regarding Communications Between Staff
Representatives talking to their staff is "privacy violation"?
Presentations of the Staff Union of the European Patent Office in Its Headquarters Tomorrow After Work
Annual General Meeting and reports
Gemini Links 06/10/2024: SSH Keys and Hobby Game Development
Links for the day
Over at Tux Machines...
GNU/Linux news for the past day
IRC Proceedings: Saturday, October 05, 2024
IRC logs for Saturday, October 05, 2024
[Meme] How to Keep Granting Hundreds of Thousands of Fake Patents (Without Upsetting Anybody in Politics and Media)
This is very Kremlin-like
EPO Examiners to Adopt Resolution Condemning EPO Management for Breaking the Law in Order to Grant Many Illegal Software Patents
Europe's second-largest institution (EPO) is a law-breaking institution hiding behind the veil of "law"
[Meme] Sup, Nazi?
"Come back, one year"
Calling "Nazi" and "Right Wing" Everyone Who Does Not Agree With You (Even Leftists Whose Views on Some Issues Slightly Differ From Yours)
Oil money has become exceptionally notorious for takeover of online platforms and institutions/NGOs (using them to incite society inwards, not upwards)
EFF Losing the Plot
Like the Linux Foundation and OSI, the EFF has succumbed to corporate influence and is derailing itself (along with its original mission)
Links 05/10/2024: Patents Being Squashed, EFF Insists on Children's Access to Porn
Links for the day
Gemini Links 05/10/2024: Multitudinous Agreeable Futures and Misfin Mail
Links for the day
EFF Celebrates Microsoft Windows and Microsoft Office as "Digital Inclusion", Mocks GNU/Linux-Based ChromeOS
Yet another example/evidence that EFF has become a rotten pile of junk
Links 05/10/2024: Amazon Culling 14,000 Managers, About 160 People Resign From Automattic
Links for the day
Microsoft Moles in Nerdearla, Openwashing and Whitewashing Microsoft With Its Latest Ponzi Scheme and Storytelling
Also GPL violations en masse
The Danger of Outsourcing Your Platform to Social Control Media and Getting "Information" There
Stella is probably not aware of what she has just done
Over at Tux Machines...
GNU/Linux news for the past day
IRC Proceedings: Friday, October 04, 2024
IRC logs for Friday, October 04, 2024
Links 05/10/2024: Shift to ARM, Microsoft XBox Crisis
Links for the day