jeffjarvis,
@jeffjarvis@mastodon.social avatar

Sigh. LLMs don't hallucinate. They assemble words. They know nothing, understand nothing. They are being misused by tech platforms and misrepresented by media.
When A.I. Chatbots Hallucinate https://www.nytimes.com/2023/05/01/business/ai-chatbots-hallucinatation.html?smid=tw-share

simon,
@simon@simonwillison.net avatar

@jeffjarvis An interesting thing about that article is that "When did The New York Times first report on AI?" is a question that's almost the perfect case of something a language model will be unable to answer correctly

Explaining why that is to people is so hard though!

One of the big skills involved in using this tech productively is developing robust instincts in terms of what questions it will likely answer well and what questions it's going to blatantly mess up

simon,
@simon@simonwillison.net avatar

@jeffjarvis A key message I've been trying to get across to people is that, while this stuff looks trivially easy to use, it really isn't - it takes a great deal of experience and knowledge to use it in a way that avoids the many, many traps and pitfalls

Once you climb that steep and invisible learning curve it's one of the most powerful new tools I've ever encountered - but figuring out how to help guide people there is proving really difficult

22,

@simon @jeffjarvis this is the message of the classic xkcd 1425: “In CS, it can be hard to explain the difference between the easy and the virtually impossible”

As a developer who works daily with very highly skilled and technical people (financial asset traders), the opacity for them of what I do continually surprises me. The gentle “I’m not sure how long this will take but…” is just as often followed by a request that’s a one-liner change or a request that needs three teams to coordinate work on.

larsmb,
@larsmb@mastodon.online avatar

@22 @simon @jeffjarvis Looks like someone paid for the research team during the last 9 years then ...

jakob,
@jakob@pxi.social avatar

@simon @jeffjarvis I don't think it takes extraordinary skill - or at least it would not need to, if the mental models of people who drive public discourse were better informed.

Of course you can't understand or explain a thing well, if the metaphor you are using to describe it is ill equipped and leads you down unhelpful paths.

https://pxi.social/@jakob/110283919520935594

jakob,
@jakob@pxi.social avatar

@simon @jeffjarvis incidentally, "answering" is already an unhelpful frame.

To answer a question, as an act of human-like communication, an interlocutor needs to access a knowledge model that binds truth values to propositions. An LLM cannot answer a question in that sense.

Some prompts can generate useful outputs, though. That's how I would try to approach defining sensible constraints.

simon,
@simon@simonwillison.net avatar

@jakob @jeffjarvis that's an illustration of the problem: despite lacking a knowledge model, they CAN answer many questions really effectively - they do that all the time

I'm not convinced that trying to explain to people that what they see LLMs do every day isn't technically correct or possible is a useful path

jakob,
@jakob@pxi.social avatar

@simon @jeffjarvis it is epistemically false to purport there is a schema of question and answer that LLMs can partake in. LLMs generate outputs from large N data pools, stochastic training algorithms and prompts. That you frame the input prompt as a question is a choice.

People see things all the time where an initial model of what they see is incongruent with the underlying causes. That's where journalists and Benoit Blanc come in. Hopefully only one of them is fictional.

simon,
@simon@simonwillison.net avatar

@jakob @jeffjarvis I understand how they work - the thing I find so interesting about them is how many things they manage to be useful for despite the very clear flaws and limitations in the way they are built

sheldonrampton,

@simon @jakob @jeffjarvis I find the same thing is the case with human beings. Despite the very clear flaws and limitations in the way WE are built, we still often manage to be useful and occasionally even inspired.

jsit,
@jsit@social.coop avatar

@jeffjarvis Calling them “AI” in the first place is a big part of the problem.

profcarroll,
@profcarroll@federate.social avatar

@jeffjarvis Classic both-sides trick from the grey lady. Use the bad but industrial term but then include the sentence that “critics” oppose using the term because it’s bad and misleading.

robertklaschka,
@robertklaschka@mastodon.online avatar

@jeffjarvis I've come to conclusion that using words like hallucinate is part of a deliberate effort to make people think of these technologies as similar to human Intelligence. Similar to the vast amount of time spent making robots dance so we anthropomorphise them.

Kraemer_HB,
@Kraemer_HB@mastodon.social avatar

@robertklaschka It is in the best interest of the publisher that journalists paint LLMs as fully capable of taking their positions as journalists. And some journalists seem to hallucinate about LLMs, ironically. @jeffjarvis

KatEmm,

@robertklaschka @jeffjarvis it also makes what the ai is doing seem more deep and interesting...it's having a mental adventure! Instead of more negative anthro language like: it's messing up, it's bullshitting..m

_L1vY_,
@_L1vY_@mstdn.social avatar

@KatEmm @robertklaschka @jeffjarvis

Right? It's not a hallucination. The device is generating a statistically likely word from its wide samples of content.

StarkRG,

@KatEmm @robertklaschka @jeffjarvis Or, more realistically, "It was trained by a group of idiots."

jeffjarvis,
@jeffjarvis@mastodon.social avatar

@StarkRG @KatEmm @robertklaschka
Yes, namely all of us.

_L1vY_,
@_L1vY_@mstdn.social avatar

@robertklaschka @jeffjarvis
Those dancing robots got people to accept ultraviolent police machinery out on our streets in broad daylight 😤

whynothugo,
@whynothugo@fosstodon.org avatar

@jeffjarvis I also know nothing, understand nothing. I just assemble words into patterns that I’ve observed from other humans. It’s not that I think that AI is super smart, it’s that I think we’re overestimating how smart humans actually are.

morri,

@jeffjarvis one article called them "elaborate text predict"

atatassault,

@jeffjarvis

Yeah, the word hallucinate is terrible. Luke Lafreniere (that's probably wrong, I'm terrible with french words) from LinusTechTips called it "LLMs are Confidently Wrong", aka, bullshitting.

costrike,

@jeffjarvis When the NYT Aids and Abets

seth,

@jeffjarvis Spot on

ids1024,
@ids1024@fosstodon.org avatar

@jeffjarvis "Figuring out why chatbots make things up and how to solve the problem has become one of the most pressing issues facing researchers as the tech industry races toward the development of new A.I. systems."

Yeah, I don't know if "why chatbots make things up" is that big of a question. Of course language models will do that. It's really a bigger question why they give an acceptable answer more often than one might expect.

UlrikeHahn,
@UlrikeHahn@fediscience.org avatar
tanepiper,
@tanepiper@tane.codes avatar

@jeffjarvis Yep - I built @StochasticEntropy and asked OpenAI to confirm it's not returning anyone elses responses from my empty prompts - they confirmed it's just stochastic chains

martinvermeer,
@martinvermeer@fediscience.org avatar

@jeffjarvis Not untrue, but only meaningful to the extent that it not also applies to lots of human thinking, or 'thinking'.

grumpasaurus,
@grumpasaurus@fosstodon.org avatar

@jeffjarvis anthropomorphism at its finest

Aradayn,
@Aradayn@mastodon.social avatar

@jeffjarvis
I've seen people use the word "confabulate" as an alternative.

I prefer "fabricate," and I think it's better to apply it to all of their output, not just what we classify as "mistakes."

azizhp,

@jeffjarvis fair enough, but the term was coined by the researchers in LLMs themselves and predates the (unexpectedly) public release of ChatGPT. Like "temperature", the term has a specific and unique meaning in this context.

alec,
@alec@perkins.pub avatar

@jeffjarvis I agree generally, but I like @simon’s position that it’s more productive to focus on straightforward communication of the harms instead of getting caught up in lower priority problem of anthropomorphism. https://simonwillison.net/2023/Apr/7/chatgpt-lies/

ankitpati,
@ankitpati@mastodon.social avatar

@jeffjarvis I understand your sentiment, but I also agree with the label of “hallucination” as commonly applied to LLMs.

We’ve been anthropomorphising technology long before LLMs.

Computer “viruses,” “worms,” and “Trojan horses” come to mind. None of these are “real” in the physical sense, but they convey the concept accurately to someone familiar with the physical objects these words describe.

Jfillian,
@Jfillian@twit.social avatar

@jeffjarvis the busy folk over at snopes.com are going to be overwhelmed.

ravigupta,
@ravigupta@mastodon.social avatar

@jeffjarvis But they are acing medical tests somehow?

simon,
@simon@simonwillison.net avatar

@ravigupta @jeffjarvis Most of those "it got an A on this test!" things are multiple choice questions

Multiple choice questions are the kind of thing a language model can be incredibly effective at - much, much easier than "when did the NY Times first mention AI?"

Enema_Cowboy,
@Enema_Cowboy@dotnet.social avatar

@jeffjarvis So they're basically the Sergeant Shultz of technology?

coffe,

@jeffjarvis exactly.

And it's something we absolutely need to understand.

https://fosstodon.org/@coffe/110292695353936488

jeber,
@jeber@mastodon.social avatar

@jeffjarvis
Sometimes we need to not anthropomorphize everything.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • random
  • kavyap
  • thenastyranch
  • GTA5RPClips
  • tester
  • InstantRegret
  • DreamBathrooms
  • ngwrru68w68
  • magazineikmin
  • everett
  • Youngstown
  • mdbf
  • slotface
  • rosin
  • cisconetworking
  • megavids
  • khanakhh
  • normalnudes
  • osvaldo12
  • cubers
  • tacticalgear
  • Durango
  • ethstaker
  • modclub
  • anitta
  • provamag3
  • Leos
  • JUstTest
  • lostlight
  • All magazines