Hacker News new | past | comments | ask | show | jobs | submit login

Has anyone run this on a Mac with M1 chip with GPU acceleration enabled?

I did my first hello world fine-tuning of Llama2 today, using Google Colab and various pieces of code. So no resume from checkpoint, I had to specify the number of epochs manually and no tracking of the loss and I didn't get the lora weights extracted, with llama2 7B I ended up with a 13gb gguf file, that I have successfully tried out locally with llama.cpp and its webui.

The axolotl seems like a nice project that hopefully makes the fine tuning easier.




An M1 does not have the umph for training :( Just rent or buy a GPU. You can train a 7B model easily on a consumer GPU.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: