/s
We already have something we're testing that's gotten 70b training down to dual 16GB GPUs, and we're working it making it even smaller...
/s