Large Language Models

PieterPeach, (edited )
@PieterPeach@mastodon.social avatar

While driving, my iPhone not only read aloud a text message, but also accurately described an attached photo.
That was unexpected

Richard_Ipsum,

@PieterPeach It feels random which photos it does it to though. I think the model is currently only trained on certain items and if you photo doesn’t contain it there’s no response.

PieterPeach,
@PieterPeach@mastodon.social avatar

@Richard_Ipsum interesting, will keep an eye on it. have only heard it happen once.

chikim,
@chikim@mastodon.social avatar

Apparently Meta is planning to release two small varients of Llama-3 next week "as a precursor to the launch of the biggest version of Llama 3, expected this summer." Command-r-plus, mixtral 8x22b, Google CodeGemma... All of sudden companies are releasing LLMS like crazy! Where's Apple? Maybe In WWDC 2024? lol https://www.theinformation.com/articles/meta-platforms-to-launch-small-versions-of-llama-3-next-week

chikim,
@chikim@mastodon.social avatar

@ppatel Yeah I don't think they'll go open source, but they'll probably make a way for developers to take adantage of it.

ppatel,
@ppatel@mstdn.social avatar

@chikim I'm very curious about how they'll improve their CreateML and other ML libraries/APIs. XCode improvements will be the most interesting to watch if they manage to get code authoring.

troed,
@troed@ioc.exchange avatar

I see lots of posts here on Mastodon where people state that today's "AI" (LLMs) have no use, waste energy and are just doing copyright infringement on a vast scale.

I don't get it.

I just put together "summarize.sh" - a bit of glue between some open source and self-hosted LLMs. It takes a Youtube URL as its only parameter, and outputs a summary in text of the important parts of the spoken words in the video.

That is, I run yt-dlp, Whisper and finally Mixtral 8x7b. And I no longer need to sit through someone yapping about for a few minutes to tell me what should've been a short blog post.

Example output from a 4 minute video:

"The text describes a video tutorial on how to reset a Corsair keyboard when it's not working properly. The keyboard in question has three white flashing lights at the top and is experiencing issues with its RGB lighting and key input. To reset the keyboard, the user should unplug the USB cables from the computer, hold down the escape key, and then plug the USB cables back into the computer while still holding down the escape key. After releasing the escape key, the keyboard's lights should flash, indicating that it has been reset. The tutorial notes that this method has worked for other Corsair keyboards as well."

How is this not a great thing to have?

gray17,
@gray17@mastodon.social avatar

@troed
and I think it misses a couple important points about the Schoen story.

these are kind of nits, but some of them might be important in some contexts.

also, I suspect that LLM summarization works better for scripted info-dumps like Sabine's videos. I've seen much more hallucination when LLMs try to summarize unscripted conversations

gray17,
@gray17@mastodon.social avatar

@troed the wrong years, I'm guessing the LLM interpreted "last year" based on a wrong idea of the current year, but this is bad. there's no way to know if a year in the summary is based on Sabine saying an absolute year or a relative year.

(also, the summary elides the distinction between publication date and retraction date)

forteller,
@forteller@tutoteket.no avatar

Either everyone has to be bound by copyright law, or noone.

If those alternatives are not acceptable, we need to reform copyright to clearly state who is bound and who is not. And if so it's not the poor and the individual that should be bound, while the billion dollar companies gets a free pass.

forteller,
@forteller@tutoteket.no avatar

Noone is above the law. That is the first principle of rule of law. Why should anyone care about any law if we blatantly disregard that?

The rich have always been able to pay their way out of the laws applying to them, and been less scrutinized. But this kind of total lack of concern for the law that we see today from the LLM corporations and the politicians seems like a new level.

How can we abide with content creators getting their livelihoods ruined with copyright strikes under this regime?

algorights, Spanish
@algorights@mastodon.social avatar

¿Los grandes modelos de lenguaje () son de izquierdas o de derechas? Según un paper recién publicado, de centro-izquierda.
El gráfico viene de un estudio en el que se les preguntó a y similares por opiniones políticas. Aquí puedes leerlo al completo: https://arxiv.org/abs/2402.01789

harriorrihar,
@harriorrihar@mas.to avatar

@algorights

Parecer no es ser. Simplemente es su forma de impactar de forma global. Esto lo hacen todos los canales desde Disney hasta Planeta. Adaptan el discurso y las narrativas a cada nicho de mercado, y en el caso de la IA, a cada individuo, porque lo único importante es el mercado.

maxleibman,
@maxleibman@mastodon.social avatar

An doesn’t know anything. It doesn’t reason about anything. It doesn’t understand anything.

It doesn’t think; it simulates thinking.

“Why does that matter?” you might ask. “What’s the difference between simulations of thinking and actual thinking?”

And for some applications, you’d be right. Simulations have value.

But here’s the thing: the best flight simulator in the world—or even the best flight simulator that will exist in your lifetime—won’t get you from New York to Barcelona.

mwyman,
@mwyman@mastodon.social avatar
dpflug,
@dpflug@hachyderm.io avatar

There appears to be exactly one LLM that is attempting ethical data sourcing.

https://huggingface.co/kernelmachine/silo-pdswby-1.3b

I don't have a GPU that'll run it, so I have no idea what it's like, but it deserves more attention for the effort. Boost for visibility if that's your thing?

bornach,
@bornach@masto.ai avatar

Asked (formerly ) a familiar riddle but with numbers changed to make it impossible. It generated the same solution but substituting the numbers so that it ends up with the nonsense claim:

10 + 5 = 23

bornach,
@bornach@masto.ai avatar

I tried the prompt:

"I have an empty opaque bag. I put two apples and one banana in the bag. I either remove the banana or I remove one apple. I then remove all remaining fruits from the bag. Is it possible to tell what is in the bag now?"

with #BingCopilot which got the right answer, but it confused Perplexity.ai which also cited a website on how to build a disaster survival kit.
#LLM #AI #PerplexityAI #chatbot #GenerativeAI

glyph,
@glyph@mastodon.social avatar

@bornach I was going to post something like "I guess programmers' jobs are safe" but as I was looking at it I realized that for most companies, 15 is close enough to 23 that they'll just use the AI and call it a win

lupyuen,
@lupyuen@qoto.org avatar

"a Large Language Model (#LLM) can be convinced to tell you how to build a bomb if you prime it with a few dozen less-harmful questions first"

https://techcrunch.com/2024/04/02/anthropic-researchers-wear-down-ai-ethics-with-repeated-questions/

freemo,
@freemo@qoto.org avatar

@lupyuen I actually have no issue with AI or anything else instructing people on how to make bombs. Knowledge should never be illegal.

AmpBenzScientist,
@AmpBenzScientist@qoto.org avatar

@freemo @lupyuen I don't see the problem except that it didn't specify if it was fission, fission-fusion or pure fusion.

Conventional energetic devices are just containers that fail to hold a chemical reaction.

There's even an argument that not knowing how to make a bomb is worse. For example a young agent finding a rental van with a lot of fertilizer and saying that it's fine exactly a year after setting a residence on fire and massacring a religious community.

Or making a funny tiktok where a glitter prank goes in an unexpected direction because they used aluminum powder.

"A little learning is a dangerous thing; drink deep, or taste not the Pierian spring." Alexander Pope

cigitalgem,
@cigitalgem@sigmoid.social avatar

I am giving a talk @indianauniv in Bloomington THIS FRIDAY. I will cover security risks (that is ) identified by BIML. Open to the public.

https://spice.luddy.indiana.edu/garymcgrawtalk/

picture

cigitalgem,
@cigitalgem@sigmoid.social avatar

My talk @indianauniv this Friday will be directly based on this study that BIML published 1.24.24

https://berryvilleiml.com/results/BIML-LLM24.pdf

joshstrange,
@joshstrange@mastodon.social avatar

I always find it funny to hear people compare AI (aka LLMs) to Crypto. One is incredibly useful already and the other is full of scams after well over a decade of trying to prove its usefulness...

Even if LLMs don't lead to AGI (my guess is they won't) they provide value TODAY and have a clear runway for improvement that will bring further value (like running them faster, which we know is possible).

jdiller,
@jdiller@hachyderm.io avatar

@danhulton @joshstrange it’s a typical hype cycle, there are some fundamentally useful things here and in the fullness of time some will shake out and become part of the fabric of life. Same as web apps and mobile before it.

I’m still convinced there’s no there there with crypto. It serves no useful purpose and will eventually be forgotten.

danhulton,
@danhulton@hachyderm.io avatar

@jdiller @joshstrange Oh yeah, they're not THE SAME. But you CAN compare them.

I think LLMs are in roughly the same place as AR/VR right now. They both need a lot more work before they can fulfill the claims being made for them. And it's likely that by the time they get there, we'll understand better how a lot of the proposed use cases were nonsense anyway, and the actual market/uses for the tech are much smaller and more focused.

winterschon,
@winterschon@hachyderm.io avatar
ramikrispin,
@ramikrispin@mstdn.social avatar

(1/2) Generative AI for Beginners Course 🚀

The Generative AI for Beginners course by Microsoft provides an introduction to the foundations of GenAI 👇🏼

https://github.com/microsoft/generative-ai-for-beginners

The course code examples are with both Python 🐍 and TypeScrip.

ramikrispin,
@ramikrispin@mstdn.social avatar

(2/2) This 18 lessons course covers a variety of topics, such as:
✅ Prompt engineering
✅ Text generation applications
✅ Image generation applications
✅ Retrieval augmented generation (RAG) and vector databases
✅ Open source models and Hugging Face 🤗
✅ Fine-tuning LLMs

https://microsoft.github.io/generative-ai-for-beginners/#/

publicvoit, German
@publicvoit@graz.social avatar

startet eine Umfrage via E-Mail, wo man mit und ohne tiefer gehende Kenntnis von mitteilen soll, wo man glaubt, dass diese Technologie Positives für die IT bringen kann/wird.

Gute Argumente sollen fefe überzeugen, dass diese Technologie kein Scam ist.

https://blog.fefe.de/?ts=98f66681

anmey,
@anmey@social.anoxinon.de avatar

@publicvoit E-Mail? 😀 steht da nicht.. aber gut. Ki & Security .. wenn wir E-Mails nur noch durch Llm zusammengefasst betrachten ist das ein security Pluspunkt. Und Texte um Menschen Risiken zu erklären lassen sich auch gut erstellen. 😁

publicvoit,
@publicvoit@graz.social avatar

@anmey AFAIK ist E-Mail die einzige Möglichkeit fefe Feedback zukommen zu lassen. Wenn du andere findest, lass es mich wissen.

ajsadauskas,
@ajsadauskas@aus.social avatar

New York City's new LLM-powered chatbot (chat.nyc.gov) is happy to help you with all sorts of questions.

For example, how to go about opening a butcher shop for cannibals on the Upper East Side.

No, really:

5am,
@5am@fosstodon.org avatar

If you keep hearing about AI, ML, GPT and LLMs, and you’re wondering what all the fuss is about, I wrote a post about locally hosting your own LLMs using Ollama for added privacy and control. I provide examples of getting help with coding, generating image descriptions and (of course) generating stories about a dog named Fido...

https://www.samhowell.uk/posts/2024/03/local-llms-for-productivity-and-privacy/

trimtab,

@5am

Nice post. Thanks.

What performance per model do you observe with that not-so-recent GeForce GTX 1070 Ti from the example ?

5am,
@5am@fosstodon.org avatar

@trimtab I haven't done any proper benchmarking with the GPU, but generally speaking, with the smaller 7B models (mistral, llama2) the responses were pretty quick, with barely any delay. With the larger models (dolphincoder etc.), I noticed the increased delay, but nothing to complain about.

underdarkGIS,
@underdarkGIS@fosstodon.org avatar

🤔 Where do we stand re ? Still no magic bullet to solve our problems?

I must be missing something.

Any pointers, ?

rustprooflabs,

@djh @underdarkGIS There was a post published yesterday about setting up a RAG with Postgres. This might be a start in a useful direction... haven't played with it myself yet but it's on my list!

https://bonesmoses.org/2024/pg-phriday-dirty-postgres-rag/

underdarkGIS,
@underdarkGIS@fosstodon.org avatar

@rustprooflabs awesome post. Thanks so much for sharing

kellogh,
@kellogh@hachyderm.io avatar

Automatic refutation of misinformation.

A new paper offers a system to correct misinformation using an . The approach seems solid, and the results seem strong. I haven’t dug in deep yet, but I’m hopeful about this one

https://arxiv.org/abs/2403.11169

kellogh,
@kellogh@hachyderm.io avatar

The biggest hole, assuming this actually works, is identifying potentially misleading remarks

This is typically done with a layered approach — one or more “pre-filter” steps have an extremely high TP (true positive) rate of identification. Each step along the way has a progressively better FP (false positive) rate and also progressively more expensive to run

So the end answer might involve LLMs for identification, but only in the final steps

zzzeek,
@zzzeek@hachyderm.io avatar

@kellogh ah the LLM was able to correct misinformation by quoting Ron DeSantis, perfect. he's of course totally above bar

troed,
@troed@ioc.exchange avatar

I was looking at integrating a local creating descriptions for photos into our locally hosted yesterday. I found SalesForce/blip, but after testing I find the descriptions not as detailed as I had hoped. Is there a better "state of the art" available?

(I know about the Nextcloud app "Recognize" but I'm under the impression it's not as detailed either)

persagen,
@persagen@mastodon.social avatar

Bioinformatics and Biomedical Informatics with ChatGPT: Year One Review
https://arxiv.org/abs/2403.15274

... covering omics, genetics, biomedical text mining, drug discovery, biomedical image understanding, bioinformatics programming, and bioinformatics education ...

#LLM #ChatGPT #NLP #AI #bioinformatics #biomedical

HxxxKxxx,
@HxxxKxxx@det.social avatar

Learn about the inner workings of Large Language Models like ChatGPT in this workshop. Discover architectural fundamentals, training methods, application areas, and how to apply LLMs to your projects. Join us for presentations, quizzes, and hands-on Python exercises.

https://www.bremen-research.de/data-train/courses/course-details?event_id=65

ian,
@ian@phpc.social avatar

How to cause an to hallucinate: ask it which phones support n70 (the band).

"The Galaxy S22 supports n70"

lol no it very much does not, 7B.

bobmagicii,
@bobmagicii@phpc.social avatar

@ian if i ask my ai to render a picture of the s22 supporting n70 would you believe it more

ian,
@ian@phpc.social avatar
chikim,
@chikim@mastodon.social avatar

Grok is a LLM from Elon Musk's xAI, and it's 638GB in fp16! Running on a consumer hardware will be pretty impossible anytime soon even with quantized. Maybe Mac Studio with 192GB. https://huggingface.co/hpcai-tech/grok-1

chikim,
@chikim@mastodon.social avatar

@ppatel It's just a base model which is pretty useless for chat. We need to wait for a fine tuned model. It's going to take a lot of GPU power, so open source teams with small budget won't be able to fine tune it.

ppatel,
@ppatel@mstdn.social avatar

@chikim Yes. But, even as a base model, it doesn't have the type of performance researchers are looking for. This is some of the chatter I hear from couple of discord groups I'm monitoring. Let's forget hardware performance, I'm talking ratings.

ChristosArgyrop,

Show me an illustration about the future of the C programming language

nschont, French
@nschont@mastodon.mim-libre.fr avatar

Un corpus LLM basé sur les textes avant 1884, avantage des textes longs et non pas les posts des réseaux non-sociaux ou des forums

https://www.zdnet.fr/actualites/common-corpus-un-corpus-de-textes-libres-de-droit-pour-nourrir-les-llm-39965094.htm#xtor=RSS-1

nschont,
@nschont@mastodon.mim-libre.fr avatar

@dad avec des réponses en vieux français

formidableinc,
@formidableinc@framapiaf.org avatar

@nschont @dad oui j'allais le dire.

Et j'attends avec impatience les réponses de cette IA sur des sujets comme.. les droits des femmes, ou l'éducation des jeunes enfants... :)

  • All
  • Subscribed
  • Moderated
  • Favorites
  • llm
  • DreamBathrooms
  • mdbf
  • ethstaker
  • magazineikmin
  • cubers
  • rosin
  • thenastyranch
  • Youngstown
  • InstantRegret
  • slotface
  • osvaldo12
  • kavyap
  • khanakhh
  • Durango
  • megavids
  • everett
  • tacticalgear
  • modclub
  • normalnudes
  • ngwrru68w68
  • cisconetworking
  • tester
  • GTA5RPClips
  • Leos
  • anitta
  • provamag3
  • JUstTest
  • lostlight
  • All magazines