87

Recently, there has been considerable interest in large language models: machine learning systems which produce human-like text and dialogue. Applications of these systems have been plagued by persistent inaccuracies in their output; these are often called “AI hallucinations”. We argue that these falsehoods, and the overall activity of large language models, is better understood as bullshit in the sense explored by Frankfurt (On Bullshit, Princeton, 2005): the models are in an important way indifferent to the truth of their outputs. We distinguish two ways in which the models can be said to be bullshitters, and argue that they clearly meet at least one of these definitions. We further argue that describing AI misrepresentations as bullshit is both a more useful and more accurate way of predicting and discussing the behaviour of these systems.

you are viewing a single comment's thread
view the rest of the comments
[-] xhieron@lemmy.world 29 points 6 months ago

Control the language and you control the thought. I pitched a fit when "hallucinate" was put forward by the tech giants to describe their LLMs' falsehoods, and it mostly fell on deaf ears in my circles. Hallucinating isn't what these things do. They bullshit.

[-] AcausalRobotGod@awful.systems 19 points 6 months ago

Hallucination also hid that literally everything they produce is a 'hallucination' because that's how they work. "Bullshit" is much more apt, as a bullshitter is sometimes and even often right.

load more comments (4 replies)
this post was submitted on 15 Jun 2024
87 points (100.0% liked)

SneerClub

1012 readers
24 users here now

Hurling ordure at the TREACLES, especially those closely related to LessWrong.

AI-Industrial-Complex grift is fine as long as it sufficiently relates to the AI doom from the TREACLES. (Though TechTakes may be more suitable.)

This is sneer club, not debate club. Unless it's amusing debate.

[Especially don't debate the race scientists, if any sneak in - we ban and delete them as unsuitable for the server.]

founded 2 years ago
MODERATORS