Kaldo,
Kaldo avatar

Did we really need an LLM summary of an otherwise already short article? Why do you assume it's even able to correctly transcribe the point behind the article in the first place? For example, it says:

The article claims that these changes are harmful to the Fediverse for several reasons:
They violate the Fediverse’s ethos of user autonomy and privacy, by forcing users to give up their data and follow Threads’ rules.

The article never said this. If anything, the author of the article even acquiesces "Granted, these sound like basic table stakes for federation to work well within the Fediverse. Most Mastodon servers collect roughly about the same amount of data for basic features to work correctly. ".

So how can this then be "violating the fediverse's ethos" when it is something the fediverse already does? The issue is not trusting facebook with this data, not the principle of data collection itself. Because of subtle nuance like this I'd say the summary is just misrepresenting the original point and just generating incorrect clickbait. There's other stuff in it that just seems made up since it's not mentioned in the article at all.

TL;DR Fuck LLMs, stop thinking they understand context. They are just glorified autocomplete algorithms.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • fediverse
  • DreamBathrooms
  • mdbf
  • ethstaker
  • magazineikmin
  • GTA5RPClips
  • rosin
  • thenastyranch
  • Youngstown
  • InstantRegret
  • slotface
  • osvaldo12
  • kavyap
  • khanakhh
  • Durango
  • megavids
  • everett
  • cisconetworking
  • normalnudes
  • tester
  • ngwrru68w68
  • cubers
  • modclub
  • tacticalgear
  • provamag3
  • Leos
  • anitta
  • JUstTest
  • lostlight
  • All magazines