lampinen, to ArtificialIntelligence
@lampinen@sigmoid.social avatar

How well can we understand an LLM by interpreting its representations? What can we learn by comparing brain and model representations? Our new paper (https://arxiv.org/abs/2405.05847) highlights intriguing biases in learned feature representations that make interpreting them more challenging! 1/9
#intrepretability #deeplearning #representation #transformers

lampinen,
@lampinen@sigmoid.social avatar

For example, if we train a model to compute a simple, linear feature and a hard, highly non-linear one, the easy feature is naturally learned first, but both are generalized perfectly by the end of training. However, the easy feature dominates the representations! 3/9

lampinen,
@lampinen@sigmoid.social avatar

This paper is really just us finally following up on a weird finding about RSA (figure on the here) from a paper Katherine Hermann & I had at NeurIPS back in the dark ages (2020): https://x.com/khermann_/status/1323353860283326464
Thanks to my coauthors @scychan_brains & Katherine! 9/9

ramikrispin, to llm
@ramikrispin@mstdn.social avatar

Fine Tuning LLM Models – Generative AI Course 👇🏼

FreeCodeCamp released today a new course for fine tuning LLM models. The course, by Krish Naik, focuses on different tuning methods such as QLORA, LORA, and Quantization using different models such as Llama2, Gradient, and Google Gemma model.

📽️: https://www.youtube.com/watch?v=iOdFUJiB0Zc

HxxxKxxx, to ArtificialIntelligence German
@HxxxKxxx@det.social avatar

Vom 16.9.-19.9.2024 richten wir an der Universität zu Köln wieder eine Sommerschule zum Thema
"Deep Learning for Language Analysis“ aus,

Weitere Informationen: http://ml-school.uni-koeln.de/

#DeepLearning #LanguageAnalysis #SummerSchool #UniKöln #MLSchool

SIB, to ArtificialIntelligence
@SIB@mstdn.science avatar

“The Protein Universe Atlas is a groundbreaking resource for exploring the diversity of proteins. Its user-friendly web interface empowers researchers, biocurators and, students in navigating the “dark matter” to explore proteins of unknown function.”

🥁 That’s what the committee said about this work, one of the 2023 👏

👉 Find out more about this and the other outputs: https://tinyurl.com/ye2yrpxx

video/mp4

koen, to ArtificialIntelligence
@koen@procolix.social avatar

Paul Gerke presents on infrastructure for at @nluug

metin, (edited ) to ai
@metin@graphics.social avatar

So… Big Tech is allowed to blatantly steal the work, styles and therewith the job opportunities of thousands of artists and writers without being reprimanded, but it takes similarity to the voice of a famous actor to spark public outrage about AI. 🤔

https://www.theregister.com/2024/05/21/scarlett_johansson_openai_accusation/

rubinjoni,
@rubinjoni@mastodon.social avatar

@metin Better late, than never.

metin,
@metin@graphics.social avatar

@rubinjoni Definitely. 👍

ramikrispin, to machinelearning
@ramikrispin@mstdn.social avatar

MLX Examples 🚀

The MLX is Apple's framework for machine learning applications on Apple silicon. The MLX examples repository provides a set of examples for using the MLX framework. This includes examples of:
✅ Text models such as transformer, Llama, Mistral, and Phi-2 models
✅ Image models such as Stable Diffusion
✅ Audio and speech recognition with OpenAI's Whisper
✅ Support for some Hugging Face models

🔗 https://github.com/ml-explore/mlx-examples

Lobrien,

@ramikrispin @BenjaminHan How do this and corenet (https://github.com/apple/corenet) fit together? The corenet repo has examples for inference with MLX for models trained with corenet; is that it, does MLX not have, e.g., activation and loss fns, optimizers, etc.?

ramikrispin,
@ramikrispin@mstdn.social avatar

@Lobrien @BenjaminHan The corenet is deep learning application where the MLX is array framework for high performance on Apple silicon. This mean that if you are using mac with M1-3 CPU it should perform better when using MLX on the backend (did not test it myself)

metin, (edited ) to ai
@metin@graphics.social avatar
metin, to ai
@metin@graphics.social avatar
ramikrispin, to ArtificialIntelligence
@ramikrispin@mstdn.social avatar

(1/2) MIT Introduction to Deep Learning 🚀🚀🚀

MIT launched the 2024 edition of the Introduction to Deep Learning course by Prof. Alexander Amini and Prof.Ava Amini. The course started at the end of April and will run until June. The course lectures are published weekly. The course syllabus keeps changing from year to year, reflecting the rapid changes in this field.

ramikrispin,
@ramikrispin@mstdn.social avatar

(2/2) The course covers the following topics:
✅ Deep learning foundation
✅ Computer vision
✅ Deep generative modeling
✅ Reinforcement learning
✅ Robot learning
✅ Text to image

Resources 📚
Course website 🔗: http://introtodeeplearning.com/
Video lectures 📽️: https://www.youtube.com/playlist?list=PLtBw6njQRU-rwp5__7C0oIVt26ZgjG9NI

ramikrispin, to datascience
@ramikrispin@mstdn.social avatar

(1/2) Happy Tuesday! ☀️

Deep Generative Models - New Stanford Course 🚀👇🏼

Stanford University released a new course last week focusing on Deep Generative Models. The course, by Prof. Stefano Ermon, focuses on the models beyond GenAI models.

#genai #DataScience #MachineLearning #deeplearning

metin, to ai
@metin@graphics.social avatar
ramikrispin, to OpenAI
@ramikrispin@mstdn.social avatar

The new OpenAI model is out - GPT 4 Omni, supporting video, audio, and vision 🤯

https://openai.com/index/hello-gpt-4o/

#openai #datascience #llm #deeplearning #genai

ramikrispin, to datascience
@ramikrispin@mstdn.social avatar

(1/2) Google released a new foundation model for time series forecasting 🚀

The TimeFM (Time Series Foundation Model) is a foundation model for time series forecasting applications. This pre-trained model was developed by the Google Research team. It joins the recent trend of leveraging foundation models for time series forecasting, which includes Salesforce's Moirai and Amazon's Chronos.

image/png

ramikrispin,
@ramikrispin@mstdn.social avatar

(2/2) License: Apache-2.0 license 🦄

Release notes 📖: https://research.google/blog/a-decoder-only-foundation-model-for-time-series-forecasting/
Code 🔗: https://github.com/google-research/timesfm
Hugging Face 🤗: https://huggingface.co/google/timesfm-1.0-200m

Image credit: Model's paper

neuromatch, to ai
@neuromatch@neuromatch.social avatar

Passionate about nurturing neuroscience or AI talent? Join as a professional development mentor for Neuromatch Academy. Spend just one hour a week for 2-3 weeks guiding students through their academic and professional paths. No prep needed! We are accepting on a rolling basis! Apply here: https://airtable.com/appd4DSKbwTVkCWAS/pagXTqRh1IeMqN3sE/form
Learn more about it: https://neuromatch.io/mentoring/
#Neuromatch #Mentorship #AI #DeepLearning

  • All
  • Subscribed
  • Moderated
  • Favorites
  • megavids
  • thenastyranch
  • DreamBathrooms
  • ethstaker
  • InstantRegret
  • Youngstown
  • magazineikmin
  • osvaldo12
  • slotface
  • khanakhh
  • rosin
  • kavyap
  • everett
  • tacticalgear
  • JUstTest
  • ngwrru68w68
  • GTA5RPClips
  • cisconetworking
  • mdbf
  • cubers
  • Durango
  • anitta
  • modclub
  • normalnudes
  • tester
  • Leos
  • provamag3
  • lostlight
  • All magazines