stsquad,
@stsquad@mastodon.org.uk avatar

I've finally finished a post I started about 2 weeks ago on my experiences with 's like . Unlike the last post it was 100% written by me which explains why its taken so long to finish 😄 It's very much from the perspective of someone who's a novice still learning about how these things work: https://www.bennee.com/~alex/blog/2023/12/10/a-systems-programmers-perspectives-on-generative-ai/

(replies to this comment appear on the blog)

penguin42,
@penguin42@mastodon.org.uk avatar

@stsquad AI stuff is interesting; I've played with the free ChatGPT's & Bard; they seem to be pretty good at explaining or searching for stuff;and yes I've seen others suggest use for review. That arm version conversion you mention is impressive; you can't blame it getting unstuck on an MMU! I'm tempted to try a local run; I'm being told that the llama.cpp runs OK on a 32G RAM host even without GPU. I find the way the parameter types shrink from f32->f16->f8->i4 fascinating architecturally.

stsquad,
@stsquad@mastodon.org.uk avatar

@penguin42 the quantised 7B models run pretty well in my experience if you have enough RAM and cores. Certainly enough to experiment with.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • Blog
  • DreamBathrooms
  • mdbf
  • ngwrru68w68
  • magazineikmin
  • thenastyranch
  • rosin
  • khanakhh
  • osvaldo12
  • Youngstown
  • slotface
  • Durango
  • kavyap
  • InstantRegret
  • tacticalgear
  • provamag3
  • ethstaker
  • cisconetworking
  • modclub
  • tester
  • GTA5RPClips
  • cubers
  • everett
  • normalnudes
  • megavids
  • Leos
  • anitta
  • JUstTest
  • lostlight
  • All magazines