metin, to ai
@metin@graphics.social avatar

𝚆𝚑𝚎𝚗 𝚆𝚒𝚕𝚕 𝚝𝚑𝚎 𝙶𝚎𝚗𝙰𝙸 𝙱𝚞𝚋𝚋𝚕𝚎 𝙱𝚞𝚛𝚜𝚝?

https://garymarcus.substack.com/p/when-will-the-genai-bubble-burst

#AI #ArtificialIntelligence #ML #MachineLearning #DeepLearning #LLM #LLMs #tech #technology #science #IT

ami, to ai
@ami@mastodon.world avatar

I keep seeing posts about , and s... and the majority are vastly incorrect and full of fear.

There's a lack of understanding of what they are, how they work, what they are capable now and in the future.

Governments are looking to legislate on AI and they lack understanding as well.

AI is a tool that will improve other tools from traffic lights to your oven and climate control.

Banning AI is like banning pens because people write nasty things that make you scared.

cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

New podcast from TechTarget discusses BIML's LLM Risk Analysis in great detail. Have a listen.

https://berryvilleiml.com/2024/04/01/tech-target-podcast-biml-discusses-23-black-box-llm-foundation-model-risks/

remixtures, to ai Portuguese
@remixtures@tldr.nettime.org avatar

: "Machine learning and algorithmic systems are useful tools whose potential we are only just beginning to grapple with—but we have to understand what these technologies are and what they are not. They are neither “artificial” or “intelligent”—they do not represent an alternate and spontaneously-occurring way of knowing independent of the human mind. People build these systems and train them to get a desired outcome. Even when outcomes from AI are unexpected, usually one can find their origins somewhere in the data systems they were trained on. Understanding this will go a long way toward responsibly shaping how and when AI is deployed, especially in a defense contract, and will hopefully alleviate some of our collective sci-fi panic.

This doesn’t mean that people won’t weaponize AI—and already are in the form of political disinformation or realistic impersonation. But the solution to that is not to outlaw AI entirely, nor is it handing over the keys to a nuclear arsenal to computers. We need a common sense system that respects innovation, regulates uses rather than the technology itself, and does not let panic, AI boosters, or military tacticians dictate how and when important systems are put under autonomous control." https://www.eff.org/deeplinks/2024/03/how-avoid-ai-apocalypse-one-easy-step

arnicas, to ai
@arnicas@mstdn.social avatar

btw, this poster is fab on Threads with ML papers. Seems you can follow here now! https://www.threads.net/@sung.kim.mw/post/C5Ji17TPJYj

cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

I am giving a talk in Bloomington 4/5 (the Friday before the eclipse). Come hear me present the new BIML LLM work in person!

https://spice.luddy.indiana.edu/garymcgrawtalk/

cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

"Tech companies and venture capitalists have been throwing money at AI startups following the success of Open AI’s ChatGPT. More than $29 billion was invested in generative AI companies last year, according to research firm PitchBook. " https://www.wsj.com/tech/ai/amazon-invests-2-75-billion-in-ai-startup-anthropic-87bb869e

hermeticlibrary, to ai
@hermeticlibrary@mastodon.social avatar
tero, to LLMs

have really created a paradigm shift in machine learning. It used to be so that you would train an model to perform a task by collecting a dataset reflecting the task, with task output labels, and then using supervised learning to learn this task by doing.

Now a new paradigm has emerged: Train by reading about the task. We have such generalist models that we can let them learn about the domain by reading all the books and other content about it, and then utilize that learned knowledge to perform the task. Note that task labels are missing. You might need those to measure the performance but you don't need those for training.

Of course if you have both example performances as task labels and lots of general material about the topic, you can actually use both to get even better performance.

Here is a good example of training the model not by example performances, but by general written knowledge about the topic. surpasses the quality levels of previous state-of-the-art despite not having been trained for this task.

This is the power of generalist models; they unlock new ways to train them, which for example allow us to surpass human-level by side-stepping imitative objectives. This isn't the only way to train skills these models enable, there are countless other ways, but this is an uncharted territory.

The classic triad of supervised learning, unsupervised learning and reinforcement learning are going to have an explosion of new training methodologies to become their peers because of this.

https://www.nature.com/articles/s41592-024-02235-4

datasciencejobsusa, to datascience
@datasciencejobsusa@mastodon.social avatar
KathyReid, (edited ) to ML
@KathyReid@aus.social avatar

Delighted to be able to publicise a paper that was presented at the @ALTAnlp 2023 Workshop at the end of last year, co-authored with my supervisor, Associate Professor @eltwilliams, and written as part of my research at School of Cybernetics.

Titled "Right the docs: Characterising voice dataset documentation practices used in machine learning", it combines both exploratory interviews and documentation analysis to characterise how large voice datasets - e.g. , @mozilla's , and several others, document their .

Unsurprisingly, it finds that the practices seen currently do not meet the needs of the practitioners who use these datasets.

We show, once again, in the words of Nithya Sambasivan - "everyone wants to do the model work, but nobody wants to do the data work" ...

https://aclanthology.org/2023.alta-1.6/

Citation:

Reid, K., Williams, E.T., 2023. Right the docs: Characterising voice dataset documentation practices used in machine learning, in: Muresan, S., Chen, V., Casey, K., David, V., Nina, D., Koji, I., Erik, E., Stefan, U. (Eds.), Proceedings of the 21st Annual Workshop of the Australasian Language Technology Association. Association for Computational Linguistics, Melbourne, Australia, pp. 51–66.

chikim, to ai
@chikim@mastodon.social avatar

Maybe we have an open source competitor for ElevenLabs? Check out their demo which they switch between original and synthesized. I can't tell. lol Apparently they're going to fully open source codebase and model weights. https://jasonppy.github.io/VoiceCraft_web/

chikim, to llm
@chikim@mastodon.social avatar

Grok is a LLM from Elon Musk's xAI, and it's 638GB in fp16! Running on a consumer hardware will be pretty impossible anytime soon even with quantized. Maybe Mac Studio with 192GB. https://huggingface.co/hpcai-tech/grok-1

cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

Nice to see data lakes released...but what we need are data oceans. This new dataset is off by many orders of magnitude. Humans have a hard time with trillions...
https://huggingface.co/blog/Pclanglais/common-corpus

everythingopen, to AWS
@everythingopen@fosstodon.org avatar

Continuing our #EverythingOpen Schedule Highlights, we present Faisal Masood of #AWS who will talk about the #ML life-cycle of #data preparation, model #training, testing and deployment, and the role that #automation and #monitoring tools play.

Faisal shows you how to build a model workflow where all team members can collaborate to create a #CI and delivery pipeline for ML models.

🗓️ Schedule: https://2024.everythingopen.au/schedule/

🗓️ Schedule: https://2024.everythingopen.au/attend/tickets/

maugendre, to mathematics
@maugendre@hachyderm.io avatar

Concentration of measures:
Talagrand's "work illustrates the idea that the interplay of many random events can, counter-intuitively, lead to outcomes that are more predictable, and gives estimates for the extent to which the uncertainty is reigned in."

Marianne Freiberger: https://plus.maths.org/content/abel-prize-2024 @data @mathematics

maugendre,
@maugendre@hachyderm.io avatar

"Majorizing measures provide bounds for the supremum of stochastic processes. They represent the most general possible form of the chaining argument".

Michel Talagrand, 1996, https://projecteuclid.org/journals/annals-of-probability/volume-24/issue-3/Majorizing-measures-the-generic-chaining/10.1214/aop/1065725175.full @data @mathematics

maugendre,
@maugendre@hachyderm.io avatar

"Majorizing measures provide bounds for the supremum of stochastic processes. They represent the most general possible form of the chaining argument".

Michel Talagrand, 1996, https://projecteuclid.org/journals/annals-of-probability/volume-24/issue-3/Majorizing-measures-the-generic-chaining/10.1214/aop/1065725175.full @mathematics

mempko, to ML
@mempko@fosstodon.org avatar

I don't think the tech nerds out there understand how upsetting generative AI is to artists. Not because it will replace them, but because there will be a generation of soulless creation devoid of humanity.

Also, how many children are looking at the progress and thinking 'what's the point of becoming an artist?'. Or how many school directors are thinking 'what's the point of a fine art budget'.

cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

I don't believe we can filter our way out of drinking a polluted ocean of training data. https://www.techtarget.com/searchEnterpriseAI/news/366574580/Microsoft-hires-DeepMind-co-founder-amid-Google-Apple-news

cigitalgem,
@cigitalgem@sigmoid.social avatar

"The issue is that [Google] trained up the [Gemini] foundation model on the polluted ocean and now they're trying to stop the pollution from getting out with a filter, and that doesn't work," he said. "These models were built by drinking a data ocean without cleaning it first. And we have to do better than that." And Microsoft has the same problem, he added.

https://www.techtarget.com/searchEnterpriseAI/news/366574580/Microsoft-hires-DeepMind-co-founder-amid-Google-Apple-news

hermeticlibrary, to ai
@hermeticlibrary@mastodon.social avatar
vij, to ai
@vij@sfba.social avatar

Why would I want to listen to a story from a brain nobody designed?

Why would I want to enjoy a seashore nobody authored?

Learn to create planets, nerd.

‘Why would I want to listen to dialog nobody wrote?’ https://x.com/brkeogh/status/1770339997775786441

cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

Insurance companies have a deep understanding of statistics and ML methods, but they don't yet understand how BIG big datasets are. Immense.


https://berryvilleiml.com/2024/03/19/biml-featured-on-theinsurertv/

Lobrien, to ML

Prompt "engineering" boils my blood. Can you imagine if you were working on a stream prediction system and the quality of the output depended on prepending a stream of magic numbers? You'd disdain anyone claiming that was a sustainable solution for a business. (I mean, I can imagine it, because that's exactly the kind of crap you see in consulting.)

kisharrington, to ML
@kisharrington@mastodon.social avatar

a 🔥 minihackathon for album + DL4MicEverywhere across 🇺🇸 🇵🇹 🇩🇪 🇫🇮

will follow up here when the DL4Mic album catalog drops

  • All
  • Subscribed
  • Moderated
  • Favorites
  • anitta
  • kavyap
  • cisconetworking
  • thenastyranch
  • magazineikmin
  • hgfsjryuu7
  • DreamBathrooms
  • InstantRegret
  • Youngstown
  • slotface
  • PowerRangers
  • Durango
  • everett
  • rosin
  • normalnudes
  • vwfavf
  • modclub
  • ethstaker
  • khanakhh
  • tacticalgear
  • ngwrru68w68
  • osvaldo12
  • mdbf
  • tester
  • cubers
  • Leos
  • GTA5RPClips
  • provamag3
  • All magazines