Hacker News new | past | comments | ask | show | jobs | submit login

Koboldcpp (a nice frontend for llama.cpp) is The Way.

You really want to run OSX though, as its not very fast without Metal (or Vulkan). Also, you need a relatively high memory M1 model to run the better llama variants.




I'll take a look into Koboldcpp, a frontend is always nice, thanks! I do have the max specs on this M1.




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: