Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Honestly you can run this on a 16GB VRAM GPU with llama.cpp. Just try it!
 help



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: