This is a compelling philosophical argument.
Not Just Nomenclature
It’s not just your imagination — ChatGPT really is spitting out “bullshit,” according to a group of researchers.
In a new paper published in the journal Ethics and Information Technology, a trio of philosophy researchers from the University of Glasgow in Scotland argue that referring to chatbot’s propensity to make crap up shouldn’t be referred to as “hallucinations,” because it’s actually something much less flattering.
Hallucination, as anyone who’s studied psychology or taken psychedelics knows well, is generally defined as seeing or perceiving something that isn’t there. Its use in the context of artificial intelligence is clearly metaphorical, because large language models (LLMs) don’t see or perceive anything at all — and as the Glasgow researchers maintain, that metaphor misses the mark when the concept of “bullshitting” is right there.
“The machines are not trying to communicate something they believe or perceive,” the paper reads. “Their inaccuracy is not due to misperception or hallucination. As we have pointed out, they are not trying to convey information at all. They are bullshitting.”
Calling Bull
At the crux of the assertion from researchers Michael Townsen Hicks, James Humphries, and Joe Slater is philosopher Harry Frankfurt’s hilarious and cutting 2005 epistemology opus “On Bullshit.” As the Glaswegians summarize it, Frankfurt’s general definition of bullshit is “any utterance produced where a speaker has indifference towards the truth of the utterance.” That explanation, in turn, is divided into two “species”: hard bullshit, which occurs when there is an agenda to mislead, or soft bullshit, which is uttered without agenda.
“ChatGPT is at minimum a soft bullshitter or a bullshit machine, because if it is not an agent then it can neither hold any attitudes towards truth nor towards deceiving hearers about its (or, perhaps more properly, its users’) agenda,” the trio writes.
Rather than having any intention or agenda, chatbots have one singular objective: to output human-like text. Citing the lawyer who used ChatGPT to write a legal brief and ended up presenting a bunch of “bogus” legal precedents before the judge, the UG team asserts that LLMs have proven themselves adept bullshitters — and that sort of thing could become more and more dangerous as people keep relying on chatbots to work for them.
“Investors, policymakers, and members of the general public make decisions on how to treat these machines and how to react to them based not on a deep technical understanding of how they work, but on the often metaphorical way in which their abilities and function are communicated,” the researchers proclaim. “Calling their mistakes ‘hallucinations’ isn’t harmless: it lends itself to the confusion that the machines are in some way misperceiving but are nonetheless trying to convey something that they believe or have perceived.”
More on AI shit: This Simple Logic Question Stumps Even the Most Advanced AI
Source link
lol