@Suiseiseki There are various model files in different sizes. They need to fit inside your VRAM. So, you can get normal-sized models and run. them on Linux. I do in fact have a Linux box with 24GB of VRAM but that is still not enough for the largest models. A Mac, if you spend the money, can run a model in over 190GB.
@mischievoustomato@Suiseiseki@Moon the highest end alchemist card currently available is still pretty weak for ai stuff afaik, whether battlemage changes the situation remains to be seen
@Moon@Suiseiseki So I really can't just go get a MEGA ULTRA GAMER ELITE PC with a 4090 (24 GB VRAM) and run the biggest AI stuff on it by having it share/offload some of the model to system RAM? I've never really delved into the big AI/GPU stuff.
@feld@Suiseiseki@vic I haven't kept up with the state of the art with games. It used to not really be a problem that VRAM wasn't shared because you just pushed never pulled. I assume shared is better because you don't even have to copy. But for AI stuff you're pushing AND pulling.
Add comment