simondueckert, to windows German
@simondueckert@colearn.social avatar

ist eine App für , und , mit der ihr Open-Source-Sprachmodelle (LLMs) wie , , & Co. lokal auf eurem Rechner verwenden könnt: https://lmstudio.ai

Das ist Datenschutz-freundlich und spart Energie - eine GPT-Anfrage verbraucht 15x mehr Energie, als eine Google-Suche).

Der KI MOOC ist eine gute Gelegenheit, um neben "klassischen" KI-Tools auch mal offene Varianten auszuprobieren: https://www.meetup.com/de-DE/cogneon/events/297769514/

kellogh, to llm
@kellogh@hachyderm.io avatar

i low key don't want to see a big jump in or capabilities anytime soon. rn they're capable enough that my mom wants to use them, but bad enough that even she has an intuitive sense for when they're wrong

that's how you build "AIQ", the skill of using it. Lots of people toying with them, to feel out their capabilities and limitations

kellogh,
@kellogh@hachyderm.io avatar

the last week or two there's been some big jumps, in terms of sustainability.

  1. with apple's new models, they bought and paid for all the written works used for training

  2. microsoft's and also apple's new models are both tiny and capable. in apple's case, tiny enough to fit on a phone

that's what happens when you slow down, things become more sustainable

jamesravey, to ai
@jamesravey@fosstodon.org avatar

I've been investigating whether small models like and generalise well to bio-medical Q&A use cases. Small models that don't require data centres full of GPUs to run are starting to become competitive with big commercial LLMs. https://brainsteam.co.uk/2024/04/26/can-phi-and-llama-do-biology/

remixtures, to ai Portuguese
@remixtures@tldr.nettime.org avatar

: "How did Microsoft cram a capability potentially similar to GPT-3.5, which has at least 175 billion parameters, into such a small model? Its researchers found the answer by using carefully curated, high-quality training data they initially pulled from textbooks. "The innovation lies entirely in our dataset for training, a scaled-up version of the one used for phi-2, composed of heavily filtered web data and synthetic data," writes Microsoft. "The model is also further aligned for robustness, safety, and chat format."

Much has been written about the potential environmental impact of AI models and datacenters themselves, including on Ars. With new techniques and research, it's possible that machine learning experts may continue to increase the capability of smaller AI models, replacing the need for larger ones—at least for everyday tasks. That would theoretically not only save money in the long run but also require far less energy in aggregate, dramatically decreasing AI's environmental footprint. AI models like Phi-3 may be a step toward that future if the benchmark results hold up to scrutiny.

Phi-3 is immediately available on Microsoft's cloud service platform Azure, as well as through partnerships with machine learning model platform Hugging Face and Ollama, a framework that allows models to run locally on Macs and PCs."

https://arstechnica.com/information-technology/2024/04/microsofts-phi-3-shows-the-surprising-power-of-small-locally-run-ai-language-models/

adr, to llm
@adr@mastodon.social avatar

Aha, is up now. GGUF models, even. Good deal. https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-gguf

  • All
  • Subscribed
  • Moderated
  • Favorites
  • Leos
  • Durango
  • magazineikmin
  • InstantRegret
  • hgfsjryuu7
  • vwfavf
  • Youngstown
  • slotface
  • thenastyranch
  • ngwrru68w68
  • rosin
  • kavyap
  • PowerRangers
  • DreamBathrooms
  • anitta
  • khanakhh
  • mdbf
  • tacticalgear
  • cubers
  • ethstaker
  • osvaldo12
  • everett
  • cisconetworking
  • GTA5RPClips
  • modclub
  • tester
  • normalnudes
  • provamag3
  • All magazines