louis,
@louis@emacs.ch avatar

deleted_by_author

  • Loading...
    not_null_p,
    @not_null_p@emacs.ch avatar

    @louis Nice! Thanks for a nice showcase of another obvious limitation of LLMs.

    Hallucinating is bad enough, but to completely fabricate new explanations on non-existing stuffs really take the cake here (but not too surprising; since fundamentally there is no concept of "truth" in LLMs anyway).

    This reminded me of another case where a user asked an LLM a technical question, and it give a very plausible answer; only for the user to realized later that the correct answer is the complete "reverse" of the one that the LLM gives.

    This is especially hilarious because I know a couple of companies that is building their product around Mistral AI.

    Can't wait for the bubble to burst

    EDIT: And no, RAG won't save LLMs; people already have to downplay the hype of what RAG can do to mitigate issues with LLMs.

    amoroso,
    @amoroso@fosstodon.org avatar

    @louis Huzzaz.

    olav,
    @olav@emacs.ch avatar

    @louis Have you genuinely tried to make use of AI? If so I'd be very surprised if you don't find it useful.

    The AI discourse on Mastodon is all just "AI hallucinated -> AI is completely useless"

    In any case, it works on my machine:

    louis,
    @louis@emacs.ch avatar

    @olav And this is what ChatGPT has to tell about HUJU Common Lisp on my machine.

    Here you can see that LLMs actually hallucinate depending and what it predicts you want to know.

    Apparently "Does it exist?" is different from "Tell me about property N of X". Because in the latter case it assumes that it already exists.

    To be more precise, when I refer to AI in the Mastodon discourse, I mean LLMs. There are legitimate use cases for Machine Learning.

    The problem is how LLMs are sold to the world as actually useful while they are utterly broken. And a society that relies on broken software ultimately becomes a broken society.

    olav,
    @olav@emacs.ch avatar

    @louis "ChatGPT 3.5"

    louis,
    @louis@emacs.ch avatar

    @olav I wonder, do LLMs actually have knowledge of the real world, or is it just a giant text transformation engine?

    Because I'd feel completely different about an actual knowledge database that I can query in human language, which has access to real knowledge.

    olav,
    @olav@emacs.ch avatar

    @louis I don't really know. That discussion quickly becomes a philosophical discussion about what knowledge even is.

    In practice I find ChatGPT4 very useful for questions in a broad range of areas. As opposed to reading articles, books, docs etc, it can tailor the info for you, so that if you ask the right question you get just the right info back instead of having to scavenge and piece together what you want to know. Especially useful if you don't know exactly what you're looking for.

    fwiw my feeling from talking to it is that it has a lot of knowledge.

    I wonder, do LLMs actually have knowledge of the real world, or is it just a giant text transformation engine?

    Idk

    using LLM along with knowledgebases is quite popular. I think it's called RAG?

    Basic example would be passing a whole wikipedia article about the subject to the llm...

    zyd,
    @zyd@emacs.ch avatar

    @louis

    And a society that relies on broken software ultimately becomes a broken society.

    I think of it as the inverse: a broken society produces broken software. Why would a society such as ours be so driven for this crap? Your point still stands however.

    @olav

    louis,
    @louis@emacs.ch avatar

    @olav I see know that you use ChatGPT 4, while I used ChatGPT 3.5. However in my former examples I was using Mistral AI's LLM - which is praised specifically in the Fediverse as being "superior" to ChatGPT.

    olav,
    @olav@emacs.ch avatar

    @louis Which version of Mistral LLM? I agree that hallucination is a problem, encyclopaedias still serve a purpose for sure.

    For me ChatGPT 4 only rly hallucinates when I ask it Guix questions heh.

    4 and 3.5 are miles apart.

    louis,
    @louis@emacs.ch avatar

    @olav I used the "Large" model of Mistral AI with "superior capabilities".

    olav,
    @olav@emacs.ch avatar

    @louis well of course they're gonna say that :p

    offset___cyan,
    @offset___cyan@emacs.ch avatar

    @louis excited to see your LOUIS mastodon clone coming soon

    louis,
    @louis@emacs.ch avatar

    @offset___cyan it already exists.

    mitchmarq42xyz,
    @mitchmarq42xyz@emacs.ch avatar

    @louis wait you asked an AI and this is what it said? or are those real?

    louis,
    @louis@emacs.ch avatar

    @mitchmarq42xyz yes they are real. That's the Mistral AI with the "Large" model turned on.

    I could ask it about any XYZ implementation and it would elaborate on it and invent all kinds of stuff.

    mitchmarq42xyz,
    @mitchmarq42xyz@emacs.ch avatar

    @louis so there is no HCL or Louis common lisp, those are just hallucinated? so confused

    louis,
    @louis@emacs.ch avatar

    @mitchmarq42xyz pure hallucination. this piece of software doesn't have the slightest clue of anything.

    mitchmarq42xyz,
    @mitchmarq42xyz@emacs.ch avatar

    @louis ah, makes sense now got it

  • All
  • Subscribed
  • Moderated
  • Favorites
  • ai
  • kavyap
  • thenastyranch
  • mdbf
  • DreamBathrooms
  • ngwrru68w68
  • magazineikmin
  • InstantRegret
  • Youngstown
  • vwfavf
  • slotface
  • everett
  • osvaldo12
  • rosin
  • khanakhh
  • megavids
  • tester
  • Durango
  • tacticalgear
  • GTA5RPClips
  • cisconetworking
  • ethstaker
  • cubers
  • normalnudes
  • modclub
  • provamag3
  • Leos
  • anitta
  • JUstTest
  • All magazines