falken, to ai
@falken@qoto.org avatar

I open sourced a thing !

A super simple web GUI for @mozilla 's

Should pass the SO test.

https://codeberg.org/falken/simple-gpt

stlhood, to random
@stlhood@fosstodon.org avatar

New blog post! : four months of progress towards democratizing AI: https://future.mozilla.org/news/llamafile-four-months-of-progress-towards-democratizing-ai/

governa, to random
@governa@fosstodon.org avatar

0.7 Brings AVX-512 Support: 10x Faster Prompt Eval Times For AMD Zen 4

https://www.phoronix.com/news/Llamafile-0.7

simon, to random
@simon@simonwillison.net avatar

I put together some detailed notes showing how I use Claude and ChatGPT as part of my daily workflow - in this case describing how I used them for a 6 minute side quest to create myself a GeoJSON map of the boundary of the Adirondack Park in upstate New York
https://simonwillison.net/2024/Mar/22/claude-and-chatgpt-case-study/

simon,
@simon@simonwillison.net avatar
itnewsbot, to ArtificialIntelligence
@itnewsbot@schleuss.online avatar

A Straightforward AI Voice Assistant, On a Pi - With AI being all the rage at the moment it’s been somewhat annoying that using a ... - https://hackaday.com/2024/02/20/a-straightforward-ai-voice-assistant-on-a-pi/

ErikJonker, to llm
@ErikJonker@mastodon.social avatar

Playing around with small LLM's on my very old laptop using Llamafile , all those models are totally unreliable, hallucinate a lot. Not surprising if you think about how relatively small they are, my resources etc. This is fun to play around with and you can easily change the settings use another model etc. As a learning experience great, for practical use i will stick with ChatGPT4 .
https://github.com/Mozilla-Ocho/llamafile

ErikJonker, to llm
@ErikJonker@mastodon.social avatar

Nog nooit echt plezier gehad van glasvezel met Gigabit snelheid maar nu wel fijn met het downloaden van Mixtral-8x7B-Instruct , 30 GB 😆
https://github.com/Mozilla-Ocho/llamafile

ErikJonker,
@ErikJonker@mastodon.social avatar

@cbontenbal ...het was me al gelukt het kleinste model direct te draaien (zie beschrijving op genoemde Github), nu het grootste model met de gewichten apart omdat windows geen executable groter dan 4GB aan kan 😀 . Betreffende opensource modellen zijn nog een stuk slechter als GPT4 maar het experimenteren vind ik interessant in mijn vrije tijd.

ErikJonker, (edited ) to ai
@ErikJonker@mastodon.social avatar

Amazing, with everybody can experiment with a large language model on their own computer, it's maybe not up to the level of GPT4 but still amazing, running on local resources. My laptop is four years old 🙂
https://github.com/Mozilla-Ocho/llamafile

simon, to random
@simon@simonwillison.net avatar

Many options for running Mistral models in your terminal using LLM

I wrote about a whole bunch of different ways you can use my LLM tool to run prompts through Mistral 7B, Mixtral 8x7B and the new Mistral-medium from the terminal:

https://simonwillison.net/2023/Dec/18/mistral/

simon,
@simon@simonwillison.net avatar

Added another option: you can run Mixtral as a llamafile and then configure my LLM tool to talk to it via its OpenAI-compatible localhost API endpoint https://simonwillison.net/2023/Dec/18/mistral/#llamafile-openai

joe, (edited ) to llm
@joe@toot.works avatar

I know that LLM and AI are kind of dirty words in the app dev community and I get the reasons why. Even if you just keep living your life, folks like Dropbox (https://boingboing.net/2023/12/14/dropbox-is-sharing-users-files-with-openai-heres-how-to-opt-out.html) and GitHub are using your data to train their LLMs. I think that there is an opportunity in ethically sourced AI, though. Even if you disagree with that statement, there is a "know your enemy" argument.

That's the reason why I have been trying to learn this shit, lately. ...

#LLM #AI

joe,
@joe@toot.works avatar

... I have tried running #Llama 2 (https://ai.meta.com/llama/) locally and I haven't been able to figure it out yet but it is the reason why I decided to replace my 2017 MBP with a 2021 MBP (with double the RAM and an M1 Pro instead of an i7).

I did have some success with #llamafile (https://github.com/Mozilla-Ocho/llamafile), though. Mozilla released something that lets you package a set of weights as a single multi-gigabyte executable that can run on macOS, Windows, Linux, FreeBSD, OpenBSD, or NetBSD. ...

ben, to random

@simon is there a llamafile for llava 1.5 13B? Your blog post is awesome, and now I'm hoping to try the larger model.

simon,
@simon@simonwillison.net avatar

@ben @freakazoid you should be need to make a new llamafile to try it out if you download the GGUF and run it like this: https://simonwillison.net/2023/Nov/29/llamafile/#llamafile-trying-other-models

happyborg, to llm
@happyborg@fosstodon.org avatar

Try a local using 's :

  1. Get the 4.3GB file:
    curl -LO https://huggingface.co/jartine/llava-v1.5-7B-GGUF/resolve/main/llava-v1.5-7b-q4-server.llamafile

  2. Make executable

  3. Run it

  4. Visit http://127.0.0.1:8080/

You're welcome!

This is a text and image capable LLM, so ~ 4

h/t @simon

More in Simon's blog: https://simonwillison.net/2023/Nov/29/llamafile/

devinprater, to ai

Well, I'm sure an NVDA addon with this will be here soon. Just amazing how far things have come. I mean, I hope there's a Windows version of the Llava part of this anyway.

llamafile is the new best way to run a LLM on your own computer:

https://simonwillison.net/2023/Nov/29/llamafile/

  • All
  • Subscribed
  • Moderated
  • Favorites
  • JUstTest
  • magazineikmin
  • Youngstown
  • osvaldo12
  • khanakhh
  • slotface
  • tacticalgear
  • mdbf
  • InstantRegret
  • kavyap
  • DreamBathrooms
  • thenastyranch
  • everett
  • rosin
  • anitta
  • Durango
  • GTA5RPClips
  • ethstaker
  • modclub
  • cisconetworking
  • ngwrru68w68
  • tester
  • normalnudes
  • cubers
  • Leos
  • megavids
  • provamag3
  • lostlight
  • All magazines