docbibi,
@docbibi@freiburg.social avatar

@simon I wonder what to make of an llm chat (local model) answer that abruptly stops. I got relatively consistent results by following it with "continue", but it stops again and there's an ensuing ping pong match.

Do you know what's happening? Maybe it's related to the token limit?
(In this case it's the Llama 3 8B Instruct.)

simon,
@simon@simonwillison.net avatar

@docbibi might be that there's a max_token setting that defaults to a shorter value too, which plugin are you using?

docbibi,
@docbibi@freiburg.social avatar

@simon I'm afraid only llm-gpt4all (v0.4), llm is also up-to-date (V0.13.1).

simon,
@simon@simonwillison.net avatar

@docbibi here's the problem: I have that defaulting to 200 max tokens! That's a bad default https://github.com/simonw/llm-gpt4all/blob/363559a3accd49c5c0757a1bc843e0376c902bf2/llm_gpt4all.py#L76

simon, (edited )
@simon@simonwillison.net avatar
  • All
  • Subscribed
  • Moderated
  • Favorites
  • random
  • DreamBathrooms
  • everett
  • tacticalgear
  • magazineikmin
  • thenastyranch
  • rosin
  • tester
  • Youngstown
  • khanakhh
  • slotface
  • ngwrru68w68
  • kavyap
  • mdbf
  • InstantRegret
  • megavids
  • osvaldo12
  • GTA5RPClips
  • ethstaker
  • normalnudes
  • Durango
  • cisconetworking
  • anitta
  • modclub
  • cubers
  • Leos
  • provamag3
  • JUstTest
  • lostlight
  • All magazines