03-29-2025, 12:19 PM
This post was last modified 03-29-2025, 12:21 PM by Maxmars. Edited 1 time in total. 
I am so encouraged to finally see someone calling it by it's true name... this is not "AI."
Presumably, Large Language Models are an amazing collection of mathematical algorithms, meant to model the processes which its' programmers deem "intelligence."
Eventually, they will determine that the problem is in the model... "patching" it won't help.
Analyses like these bring us closer to understanding just how misrepresented this technology has been... (akin to "other" recent "technologies" misrepresented)... it's becoming a trend now.
The processes and discipline of "scientific" modeling are not something to be casually assumed.Â
Some ingenious few have developed a potentially high-functioning interface in "human" language... (mind you not using meaning - but statistical potentials for word tokens.)
This made fluid speech much easier to achieve... but damned the "meaning of the information" to a differential relationship based upon language use.
It's calculations are about what could be a validly stated response, insofar as language is concerned... not meaning.
There is no "thought" in this suddenly renamed, economically ripe "AI."
From ArsTechnica:Â Why do LLMs make stuff up? New research peers under the hood.
Subtitled:Â Claude's faulty "known entity" neurons sometime override its "don't answer" circuitry.
Now, new research from Anthropic is exposing at least some of the inner neural network "circuitry" that helps an LLM decide when to take a stab at a (perhaps hallucinated) response versus when to refuse an answer in the first place. While human understanding of this internal LLM "decision" process is still rough, this kind of research could lead to better overall solutions for the AI confabulation problem.
Note the allusion to a "decision"... as if 2 + 2 "decided" to be 4.
Some authors, simply refuse to portray reality without injecting the near anthropomorphic prayer "AI" into everything.
At their core, large language models are designed to take a string of text and predict the text that is likely to follow—a design that has led some to deride the whole endeavor as "glorified auto-complete." That core design is useful when the prompt text closely matches the kinds of things already found in a model's copious training data. However, for "relatively obscure facts or topics," this tendency toward always completing the prompt "incentivizes models to guess plausible completions for blocks of text," Anthropic writes in its new research.
...
This and other research into the low-level operation of LLMs provides some crucial context for how and why models provide the kinds of answers they do. But Anthropic warns that its current investigatory process still "only captures a fraction of the total computation performed by Claude" and requires "a few hours of human effort" to understand the circuits and features involved in even a short prompt "with tens of words." Hopefully, this is just the first step into more powerful research methods that can provide even deeper insight into LLMs' confabulation problem and maybe, one day, how to fix it.
"...only captures a fraction of the total computation performed by Claude..."Â
Not interested in "everything else" going on.... but this was the same mental gymnastic that led to....
"There are so many millions of things that can go wrong in a space launch like Challenger... let's not waste time trying to understand." (Thank God for Feynman)
It has become clear that all reporting seems to "want" this to be "AI"... and now it needs to be fixed....Â
I don't mind fixing LLMs... but call them that, OK?
Presumably, Large Language Models are an amazing collection of mathematical algorithms, meant to model the processes which its' programmers deem "intelligence."
Eventually, they will determine that the problem is in the model... "patching" it won't help.
Analyses like these bring us closer to understanding just how misrepresented this technology has been... (akin to "other" recent "technologies" misrepresented)... it's becoming a trend now.
The processes and discipline of "scientific" modeling are not something to be casually assumed.Â
Some ingenious few have developed a potentially high-functioning interface in "human" language... (mind you not using meaning - but statistical potentials for word tokens.)
This made fluid speech much easier to achieve... but damned the "meaning of the information" to a differential relationship based upon language use.
It's calculations are about what could be a validly stated response, insofar as language is concerned... not meaning.
There is no "thought" in this suddenly renamed, economically ripe "AI."
From ArsTechnica:Â Why do LLMs make stuff up? New research peers under the hood.
Subtitled:Â Claude's faulty "known entity" neurons sometime override its "don't answer" circuitry.
Now, new research from Anthropic is exposing at least some of the inner neural network "circuitry" that helps an LLM decide when to take a stab at a (perhaps hallucinated) response versus when to refuse an answer in the first place. While human understanding of this internal LLM "decision" process is still rough, this kind of research could lead to better overall solutions for the AI confabulation problem.
Note the allusion to a "decision"... as if 2 + 2 "decided" to be 4.
Some authors, simply refuse to portray reality without injecting the near anthropomorphic prayer "AI" into everything.
At their core, large language models are designed to take a string of text and predict the text that is likely to follow—a design that has led some to deride the whole endeavor as "glorified auto-complete." That core design is useful when the prompt text closely matches the kinds of things already found in a model's copious training data. However, for "relatively obscure facts or topics," this tendency toward always completing the prompt "incentivizes models to guess plausible completions for blocks of text," Anthropic writes in its new research.
...
This and other research into the low-level operation of LLMs provides some crucial context for how and why models provide the kinds of answers they do. But Anthropic warns that its current investigatory process still "only captures a fraction of the total computation performed by Claude" and requires "a few hours of human effort" to understand the circuits and features involved in even a short prompt "with tens of words." Hopefully, this is just the first step into more powerful research methods that can provide even deeper insight into LLMs' confabulation problem and maybe, one day, how to fix it.
"...only captures a fraction of the total computation performed by Claude..."Â
Not interested in "everything else" going on.... but this was the same mental gymnastic that led to....
"There are so many millions of things that can go wrong in a space launch like Challenger... let's not waste time trying to understand." (Thank God for Feynman)
It has become clear that all reporting seems to "want" this to be "AI"... and now it needs to be fixed....Â
I don't mind fixing LLMs... but call them that, OK?