Large Language Models

sebsauvage, French
@sebsauvage@framapiaf.org avatar


Essayons de résumer où on en est sur ces IA de type LLM (+ une nouvelle faille) : https://sebsauvage.net/links/?0aif1Q

tradjincal,
@tradjincal@ludosphere.fr avatar

@sebsauvage j'ai l'impression aussi qu'economiquement parlant, il y a que Nvidia qui fait sont beurre et que les autres boîtes sur les investissements des banques.

sebsauvage,
@sebsauvage@framapiaf.org avatar

@tradjincal
tout à fait !
J'avais oublié, je l'ajoute.

drahardja,
@drahardja@sfba.social avatar

There was a paper shared recently about the exponential amount of training data to get incremental performance gains in , but I seem to have misplaced it. Do you know what I’m referring to? Mind sharing the link if you have it?

anmey,
@anmey@social.anoxinon.de avatar

I think one of the biggest fears people have about AI is that it isn't perfect as assumed, but that, like us humans, it takes the given information, assumes the most likely outcome, and presents it plausibly.

kellogh,
@kellogh@hachyderm.io avatar

@anmey yeah, there’s this paradox — we kinda want computers to think like humans, but when they get plausibly good at it, we complain that they don’t think like computers anymore

grumpybozo,
@grumpybozo@toad.social avatar

I’d like to trust this story, but it fails to link to its supposed source or provide enough info to find it elsewise. A few clicks around the site makes me think that it may well be nothing but a -composed content farm. https://cosocial.ca/@kgw/112498693958537559

feld,
@feld@bikeshed.party avatar
dvshkn,
@dvshkn@fosstodon.org avatar

I gave the pizza question to Golden Gate Claude. This is truly the people's LLM. Google btfo.

tripplehelix,
@tripplehelix@fosstodon.org avatar

@dvshkn What bridge?

chikim,
@chikim@mastodon.social avatar

Llama.cpp now supports the distributed inference, meaning you can use multiple computers to speed up the response time! Network is the main bottleneck, so all machines need to be hard wired, not connected through wifi. # https://github.com/ggerganov/llama.cpp/tree/master/examples/rpc

cheukting_ho,
@cheukting_ho@fosstodon.org avatar

opening keynote by @t_redactyl - and illusions

pauleveritt,
@pauleveritt@fosstodon.org avatar

@cheukting_ho @t_redactyl Sunday night: saying bye to Jodie, thinking "whew, glad I’m not getting on a plane tomorrow direct to next conference.”

Today: sad I wasn't there to see Jodie.

chikim,
@chikim@mastodon.social avatar
kellogh,
@kellogh@hachyderm.io avatar

@chikim i love what they’ve been doing with phi!

hypolite,

How would anyone trust the products these people put worth? They aren’t working on making LLMs more accurate (spoiler alert: they can’t, by design), they’re working to make them more appealing to companies targeting unsuspecting consumers. By any means necessary.


RE: mastodon.social/users/nixCraft…

heluecht,

@hypolite @nixCraft I just watched some demos. They are very irritating. I really don't appreciate this cheeky behaviour of that system.

troed,
@troed@ioc.exchange avatar

I'm worried for my tech friends.

The vitriol, and - honestly - ignorance around LLM-based "AI" is starting to fill my feeds from normally sane and technologically literate people.

You should be able to see through the hype and misuse. LLMs aren't encyclopedias - they're tools that are able to manipulate data of various sorts in ways that are very similar to how humans do it.

Yes, I compare LLMs to human brains. It's not the same as saying they're conscious (yet) - but the way LLMs work is apparently in many ways similar to how our brains work.

One fascinating insight into that comes from research done on what happens to the ability of LLMs to recall information as they are exposed to large and larger corpuses. Apparently they're better at recalling the early and late information, whilst starting to lose some in the middle.

In human psychology we call that the primacy and recency effect - because our brains do the same.

LLMs are absolutely awesome for a wide variety of tasks (and we have by no means found them all). Every second you spend not understanding this is a second on the way to your own irrelevance (if these tools would aid someone in your chosen area of work) or to become a grumpy old person yelling at clouds.

angiebaby,
@angiebaby@mas.to avatar

@troed

Snake oil in tech form is still snake oil.

smurthys,
@smurthys@hachyderm.io avatar

I just finished a productive Copilot session on a complex programming task. I came up with much of the algorithms, and wrote a lot of the code, and had to guide it a lot throughout, but credit where due, Copilot did make small but meaningful contributions along the way.

Overall, not a pair programmer but someone useful to talk to when WFH alone on complex tasks.

Enough for Copilot to earn a ✋🏽. And I like how it responded to that. It has got that part down. 😉

finestructure,
@finestructure@mastodon.social avatar

•This• is the compelling use case for me. If I use a translator to write messages in French I'm not forced to come up with an initial attempt and I lose the learning aspect of that.

If instead I put something into ChatGPT and it not only corrects but explains what my mistakes were that's a huge win in terms of learning from your mistakes.

(I still don't trust the thing 100% but it's also not a high stakes situation – I'm not engaging in a nuclear arms treaty after all 😅)

finestructure,
@finestructure@mastodon.social avatar

@groue Yeah, my partner pointed that and a few other things out as well and now I’m less convinced this is working as well as I thought it was.

Yet another example where the answer sounds good but only because I don’t have the expertise to verify.

finestructure,
@finestructure@mastodon.social avatar

@rene I’ve actually noticed that about French punctuation, too, at least when it comes to ! and ? 🤷‍♂️

m0bi13, Polish
@m0bi13@pol.social avatar

Ciekawa musi być praca nad bezpieczeństwem "sztucznej inteligencji" 🤔

#LLM #AI #SztucznaInteligencja

rotnroll666,
@rotnroll666@mastodon.social avatar

Absolutely unbelievable but here we are. by default using messages, files etc for building and training models, enabled by default and opting out requires a manual email from the workspace owner.

https://slack.com/intl/en-gb/trust/data-management/privacy-principles

What a time to be alive in IT. 🤦‍♂️

tanepiper,
@tanepiper@tane.codes avatar
LChoshen,
@LChoshen@sigmoid.social avatar

Do LLMs learn foundational concepts required to build world models? (less than expected)

We address this question with 🌐🐨EWoK (Elements of World Knowledge)🐨🌐

a flexible cognition-inspired framework to test knowledge across physical and social domains

https://ewok-core.github.io

luis_in_brief,
@luis_in_brief@social.coop avatar

@LChoshen I was just talking about this problem with a friend the other day. Really interesting data, thank you for sharing!

  • All
  • Subscribed
  • Moderated
  • Favorites
  • llm
  • Durango
  • DreamBathrooms
  • InstantRegret
  • tacticalgear
  • magazineikmin
  • Youngstown
  • thenastyranch
  • mdbf
  • slotface
  • rosin
  • Leos
  • kavyap
  • modclub
  • ethstaker
  • JUstTest
  • everett
  • GTA5RPClips
  • cubers
  • khanakhh
  • ngwrru68w68
  • anitta
  • provamag3
  • cisconetworking
  • tester
  • osvaldo12
  • megavids
  • normalnudes
  • lostlight
  • All magazines