Hacker News new | past | comments | ask | show | jobs | submit login
A poor man's guide to fine-tuning Llama 2 (duarteocarmo.com)
62 points by duarteoc on Sept 26, 2023 | hide | past | favorite | 5 comments



Hey man, thanks for the article. I like it that it is concise and simple. One thing that's not clear to me is about the inference stage: where does this inference process runs? Do you need to run it in a GPU powered instance or could it be runned in a consumer laptop?


Has anyone run this on a Mac with M1 chip with GPU acceleration enabled?

I did my first hello world fine-tuning of Llama2 today, using Google Colab and various pieces of code. So no resume from checkpoint, I had to specify the number of epochs manually and no tracking of the loss and I didn't get the lora weights extracted, with llama2 7B I ended up with a 13gb gguf file, that I have successfully tried out locally with llama.cpp and its webui.

The axolotl seems like a nice project that hopefully makes the fine tuning easier.


An M1 does not have the umph for training :( Just rent or buy a GPU. You can train a 7B model easily on a consumer GPU.


I'm curious to see the results, although as an english-only speaker, it may be hard for me to assess. Interesting write up nonetheless


really appreciate the simple explanation




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: