I just issued a data deletion request to #StackOverflow to erase all of the associations between my name and the questions, answers and comments I have on the platform.
One of the key ways in which #RAG works to supplement #LLMs is based on proven associations. Higher ranked Stack Overflow members' answers will carry more weight in any #LLM that is produced.
By asking for my name to be disassociated from the textual data, it removes a semantic relationship that is helpful for determining which tokens of text to use in an #LLM.
If you sell out your user base without consultation, expect a backlash.
there’s a big need for something stronger than #RAG, but more flexible and cheaper than a giant all-knowing #LLM.
A great part about RAG is that it’s just a database. You just issue INSERT/UPDATE and yeah, that’s how you maintain knowledge. No million dollar training process
i low key don't want to see a big jump in #LLM or #AI capabilities anytime soon. rn they're capable enough that my mom wants to use them, but bad enough that even she has an intuitive sense for when they're wrong
that's how you build "AIQ", the skill of using it. Lots of people toying with them, to feel out their capabilities and limitations
They use #OpenAI, which means my GitHub OSS has almost certainly been used in training data.
They rely on OpenAI's promise to not ingest any code that is used for "context".
They specifically do not disclaim that their tool could result in me violating someone else's copyright, and they could suggest the same code to someone else, too.
Uninstall this crap, now. It's dangerous and irresponsible
I am trying to build a RAG with LLAMA 3 and... getting really crazy with the strange formats I get in the response....
Not only the response, but additional text, XML tags... #Llama3#LLM#RAG
Back in 2018, Dario Amodei worked at OpenAI. And looking at one of its first A.I. models, he wondered: What would happen as you fed an artificial intelligence more and more data? He and his colleagues decided to study it, and they found that the A.I. didn’t just get better with more data; it got better exponentially.
•This• is the compelling #LLM use case for me. If I use a translator to write messages in French I'm not forced to come up with an initial attempt and I lose the learning aspect of that.
If instead I put something into ChatGPT and it not only corrects but explains what my mistakes were that's a huge win in terms of learning from your mistakes.
(I still don't trust the thing 100% but it's also not a high stakes situation – I'm not engaging in a nuclear arms treaty after all 😅)
I just came across this (h/t to Peter Krupa), and it blew my mind. It highlights the problem with LLMs in general with pinpoint accuracy, and wraps it in a well known metaphorical idiom that everyone understands — which instantly becomes a meta reference. …
Ofcourse results needs to be verified and confirmed in practice but after reading the
MedGemini paper from Google there is no doubt in my mind AI will change the world of medicines. Not replacing people but augmenting them during diagnosis, operations and treatment of patients. https://arxiv.org/abs/2404.18416 #AI#medicines#generativeAI#LLM#GoogleGemini#MedGemini
I've had occasion to ask an AI about a thing twice lately (a recent online phenomenon, and a book recommendation). Both times I asked both Gemini and ChatGPT, and both times one gave a reasonable if bland answer, and the other (a different one each time) gave a plausible but completely fictional ("hallucinated") answer.
When do we acknowledge that LLMs, and "AI" in general, aren't quite ready to revolutionize the world?
Ce matin... deux BOT de scrapping pour alimenter des modèles d'IA/#LLM ont abusé du forum d'@osm_fr
C'est pas la première fois et ça devient vraiment une plaie, surtout quand #ClaudeBot requête les URL de notre ancien #phpBB, remplacé il y a plusieurs années par #discourse
Malgrès plus de 130 000 erreurs 404 rien que ce matin, il continuait à un rythme effréné...
Autre bot albert-bot... de albertai.com (rien avoir avec l'Albert cocorico), bloqué lui aussi.
In #homeassistant, using #nodered to make an API call to a #llamacpp server running #mistral 7B model. I create a prompt that asks it to summarize all the data in my house from the sensors. The results are pretty impressive for such a little model. Now I get a customized rundown, Jarvis style.
Useful? Probably not. But cool as hell. :cool_skelly:
Yay, I too got my 7-day suspension badge from Stack Overflow from adding an #LLM#AI disclaimer back after it was first reverted to my four (4) answers!
"The output from an LLM is a derivative work of the data used to train the LLM.
If we fail to recognise this, or are unable to uphold this in law, copyright (and copyleft on which it depends) is dead. Copyright will still be used against us by corporations, but its utility to FOSS to preserve freedom is gone."
So… Big Tech is allowed to blatantly steal the work, styles and therewith the job opportunities of thousands of artists and writers without being reprimanded, but it takes similarity to the voice of a famous actor to spark public outrage about AI. 🤔
Watching #GoogleIO and there are some cool demonstrations of data center cloud computing, but there's also this fog of dystopia surrounding these demos.
The announcements for search are horrifying. Google is full mask off.
Phrases like "search for something, and we'll collect all this data for you" basically equates to:
"We sucked up ALL the data from people who really did the work, and we're going to give you the results of their hard work, but we wont take you to the site that generated the data. You can stay on the search page, and the site's traffic will plummet."
"The biggest question raised by a future populated by unexceptional A.I., however, is existential. Should we as a society be investing tens of billions of dollars, our precious electricity that could be used toward moving away from fossil fuels, and a generation of the brightest math and science minds on incremental improvements in mediocre email writing?" (From an NYT article. See original thread.)
Do you REALLY want to get a feel for how GPT-4o does what it does? Just complete this poem — by doing so, you’ll have performed a computation similar to the one it does when you feed it a text-plus-image prompt.
Fix your shitty autocorrect! There’s no such thing at “there’re” so quit putting it into my content.
And how come I get a word suggestion as I type, I click on it, and an entirely different word is inserted that wasn’t even one of the options offered - sometimes not even an English word?!