RE: Run even larger AI models locally with LM Studio

You are viewing a single comment's thread:

Mac Studio and even Mac Minis are very popular option for LLM due to how unified memory works. Nowhere can you get ~188 VRAM for less than the cost of even a single A100 40G.



0
0
0.000
2 comments
avatar

I'm getting 23 tokens per second using the 5 bit Mixtal 2.7 model.

0
0
0.000
avatar

macs have a big edge for this.
I would recommend the 4 bit, the 5 bit isn't much better and takes a lot more ram. I'd stick with 4 bit, or something like 8 bit if you can get there.

0
0
0.000