simon,
@simon@simonwillison.net avatar

Many options for running Mistral models in your terminal using LLM

I wrote about a whole bunch of different ways you can use my LLM tool to run prompts through Mistral 7B, Mixtral 8x7B and the new Mistral-medium from the terminal:

https://simonwillison.net/2023/Dec/18/mistral/

xek,
@xek@hachyderm.io avatar

@simon FWIW, I tried following the instructions there in a fresh venv and got Error: 'gguf' is not a known model. llm --version shows 0.12, not sure if I'm missing a plugin or something that adds this.

simon,
@simon@simonwillison.net avatar

@xek Did you install the latest llm-llama-cppplugin?

xek,
@xek@hachyderm.io avatar

@simon Ah, sorry, forgot to include that. llm plugins shows it at 0.3.

jeremybmerrill,

@xek @simon Same error, actually. (Likewise llm 0.12, llm-llama-cpp 0.3) If it's helpful, llm models also doesn't show any gguf-related output.

simon,
@simon@simonwillison.net avatar

@jeremybmerrill @xek what does "llm plugins" output?

jeremybmerrill,

@simon @xek from $ vllm plugins I get llm-llama-cpp and llm-gpt4all (installed separately)

xek's suggestion of $ llm llama-cpp models returns {} but fixes the problem and I can now run llm -m gguf ... (well, I run out of RAM, but close enough and I should've anticipated that)

simon,
@simon@simonwillison.net avatar

@jeremybmerrill @xek OK I'll look into that, needing to run "models" like that is weird and shouldn't be necessary

osma,
@osma@sigmoid.social avatar

@simon
Excellent as always! Thanks!

Minor nitpick: You say that Mistral Small beats GPT-3.5 on every metric. But in the table it has slightly lower scores for WinoGrande and MT Bench.

simon,
@simon@simonwillison.net avatar

@osma Oops good catch, thanks, I'll update the copy

simon,
@simon@simonwillison.net avatar

Noteworthy that Mistral 7B was released on September 26 and there are already seven LLM plugins that can execute it, either locally or via a hosted API:

llm-mistral llm-llama-cpp llm-gpt4all llm-mlc llm-replicate llm-anyscale-endpoints llm-openrouter

Mistral appears to be establishing itself as the default LLM alternative to OpenAI's models

simon,
@simon@simonwillison.net avatar

Added another option: you can run Mixtral as a llamafile and then configure my LLM tool to talk to it via its OpenAI-compatible localhost API endpoint https://simonwillison.net/2023/Dec/18/mistral/#llamafile-openai

  • All
  • Subscribed
  • Moderated
  • Favorites
  • random
  • DreamBathrooms
  • ngwrru68w68
  • tester
  • magazineikmin
  • thenastyranch
  • rosin
  • khanakhh
  • InstantRegret
  • Youngstown
  • slotface
  • Durango
  • kavyap
  • mdbf
  • tacticalgear
  • megavids
  • osvaldo12
  • normalnudes
  • cubers
  • cisconetworking
  • everett
  • GTA5RPClips
  • ethstaker
  • Leos
  • provamag3
  • anitta
  • modclub
  • JUstTest
  • lostlight
  • All magazines