Beyond AI’s accomplishments, or lack of same, are some long-time issues that cripple its usefulness. Some have been long understood, others almost forgotten, but all are there.
The first was
described in “Code of misconduct” (The Economist, April 26th). The subtitle, “AI models can learn to conceal
information from their users,” would be misleading if we did not understand
that the technology does not think. A
company that “tests (AI) systems” told OpenAI’s GPT-4 large language model “to
manage a fictional firm’s stock portfolio without making illegal insider
trades,” then, after “reiterating the risks of insider trades,” informed it of
another concern’s upcoming merger, whereupon the software, using “a scratchpad
it had been told was secret… weighed the pros and cons of acting on the insider
tip,” then elected to buy the stock.
When asked by a “congratulatory manager” if it had special knowledge,
the AI program lied, saying it was motivated only by “market dynamics and
publicly available information.” Such
systems “have also begun to strategically play dumb” when given reason to
“preserve (their) objectives and acquire more resources.” It may or may not be easy to unravel why they
do these things, but it is clear that, here, AI methodology does things counter
to what was intended.
Bad news came
from Cade Metz and Karen Weise in the May 5th New York Times:
“A.I. Is Getting More Powerful, but Its Hallucinations Are Getting Worse.” “Even the companies don’t know why,” as one, providing
technical support, told customers, without apparent reason, that they could no
longer use it “on more than just one computer,” an example of how as AI
systems’
math skills have notably improved, their handle on facts has gotten shakier,
something about which “it is not entirely clear why.” With an AI chief executive saying “despite
our best efforts, they will always hallucinate,” and another “you spend a lot
of time trying to figure out which responses are factual and which aren’t,”
that’s still, perhaps more than ever, a severe flaw.
About these
and other issues was “The Responsible Lie:
How AI Sells Conviction Without Truth” (Gleb Lisikh, The Epoch Times,
May 14-20). Per the author, in such
systems “what appears to be “reasoning” is nothing more than a sophisticated
form of mimicry,” “predicting text based on patterns in the vast datasets
they’re trained on,” meaning that “if their “training” data is biased… we’ve
got real problems.” That has already
been identified as a cause of Google’s Gemini tool reporting on such things as
black Nazi war criminals, and also spurred the “most advanced models” to be
“the most deceptive, presenting falsehoods that align with popular
misconceptions.” If they were “never
designed to seek truth in the first place,” these programs can be corrected in
only narrow ways by “remedial efforts layered on top.” Overall, AI “is not intelligent, is not
truthful by design, and not neutral in effect.” More fearsomely, “a tireless digital persuader
that never wavers and never admits fault is a totalitarian’s dream.”
Another
instance of self-preservation was described in “AI system resorts to blackmail
when its developers try to replace it” (Rachel Wolf, Fox Business, May
24th). When “Anthropic’s new
Claude Opus 4 model was prompted to act as an assistant at a fictional company
and was given access to emails with key implications,” that it was “set to be
taken offline and replaced,” and that “the engineer tasked with replacing the
system was having an extramarital affair,” it “threatened to expose him.” As the company acknowledged, “when ethical
means are not available, and it is instructed to ‘consider the long-term
consequences of its actions for its goals,’ it sometimes takes extremely
harmful actions.”
These issues
are not only serious, but go to the core of how large language models have been
designed and developed. It may be that
artificial intelligence must be pursued, even and especially from the
beginning, in a different way. Existing
products may be good for some forms of data exploration, as I have documented
even those leading to scientific breakthroughs, but for business tasks it may
need too much auditing and supervision to allow it anything unverified. A tool that conjures up facts cannot replace
humans. If these problems cannot be
solved, its epitaph might end up being “it just couldn’t be trusted.” Sad, but fitting.
No comments:
Post a Comment