simon,
@simon@simonwillison.net avatar

I wonder how much people's opinions of LLMs are shaped by the first application of them that they encounter

If your first ever mental model of LLMs is that they're for plagiarism and cheating on homework, I can see how that would cloud your overall opinion compared to if you start out by using them to help debug a weird error message

I'd love to see more research around how people think about and understand these things

brianstorms,
@brianstorms@mastodon.social avatar

@simon

For me it was not the first impression of an LLM application, or even any early first impressions of LLMs in action, that formed my opinion.

It is a long-in-development impression based on years of observation of the executives at, and funders of, companies developing and rolling out LLM applications, that have formed, and hardened, my opinions.

And those opinions all swirl around a profound and ironclad sense of distrust.

3psboyd,
@3psboyd@mastodon.social avatar

@simon My first interaction was probably AI Dungeon and was very cool, and my second interaction was pundits saying "We should replace search engines with these."

JMMaok,
@JMMaok@mastodon.online avatar

@simon

For so many of us, our first experience is with their output of almost-plausible garbage getting in the way of information we are trying to evaluate, whether that is internet search, scammy GitHub or other comments trying to build up an identity, or application materials.

I think a big divide is people whose jobs involve producing things, especially where others review their work and/or bear the risk, and people whose work mostly involves evaluation of various types.

shram86,
@shram86@mastodon.gamedev.place avatar

@simon if it can't reliably Google an answer for you without lying about it, why do you think it would be good for something so infamously obtuse as programming errors

simon,
@simon@simonwillison.net avatar

@shram86 because I've been using it very successfully for exactly that for over a year now

The things they are good and bad at are extremely unintuitive!

You have to spend a bunch of time with these models to develop intuition as to what they are useful for

shram86,
@shram86@mastodon.gamedev.place avatar

@simon interesting!

beamflash,
@beamflash@hachyderm.io avatar

@simon @shram86 And yet most people are going to blindly use them and accept whatever output they're given

simon,
@simon@simonwillison.net avatar

@beamflash @shram86 that's a big claim!

This is the research I'm most interested in seeing right now: how do people actually use these things, and how quickly do they learn when and when not to trust them?

tehstu,
@tehstu@hachyderm.io avatar

@simon honestly, I started out by understanding the source of training material.

simon,
@simon@simonwillison.net avatar

@tehstu I'm looking forward to the first useful LLM that's trained entirely on public domain text - I think it's going to happen at some point, there's clearly demand for it

luis_in_brief,
@luis_in_brief@social.coop avatar

@simon @tehstu the eleuther.ai folks are working on a pure PD+openly-licensed corpus.

simon,
@simon@simonwillison.net avatar

@luis_in_brief @tehstu oh amazing! I hadn't heard about that

simon,
@simon@simonwillison.net avatar

@luis_in_brief @tehstu got a link to more information? I couldn't find it on their website

luis_in_brief,
@luis_in_brief@social.coop avatar

@simon @tehstu I think maybe only in their discord of yet?

  • All
  • Subscribed
  • Moderated
  • Favorites
  • random
  • DreamBathrooms
  • everett
  • ngwrru68w68
  • magazineikmin
  • rosin
  • Youngstown
  • slotface
  • InstantRegret
  • Durango
  • ethstaker
  • kavyap
  • cisconetworking
  • thenastyranch
  • osvaldo12
  • JUstTest
  • tacticalgear
  • khanakhh
  • mdbf
  • Leos
  • normalnudes
  • modclub
  • GTA5RPClips
  • cubers
  • tester
  • megavids
  • provamag3
  • anitta
  • lostlight
  • All magazines