Hacker News new | past | comments | ask | show | jobs | submit login

3 billion parameters. Does that mean I will be able to run on a 8gb consumer GPU?



Probably not out of the box but if some of the local deep learning wizards get a quantized version working well and optimize it a bit, definitely.


Means that once it's incorporated into llama.cpp, you can run it on your laptop.


Hopefully on phones too


No, I could only get 2.7B to run on 8GB VRam unfortunatly.


it is 2.7B


Loading seems to have worked on my laptop's RTX 3070, `nvidia-smi` shows `5188MiB / 8192MiB` in memory usage.


their pytorch_model.bin is 10.4GB


I just loaded this on my laptop's RTX 3070 GPU by following the instructions here: https://huggingface.co/replit/replit-code-v1-3b

I don't know how I can test the model, but it seem loading worked. When I run `nvidia-smi` on another terminal, I see `5188MiB / 8192MiB` in the memory-usage column.


you can load it but you cant run inference? whats the issue?


No issue, I'm simply unfamiliar with python machine learning APIs.

I managed to run inference locally by installing the requirements and running app.py from the demo: https://huggingface.co/spaces/replit/replit-code-v1-3b-demo/...

It is very fast on my RTX 3070, VRAM usage goes to ~= 6.3GB during inference.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: