Hacker News new | past | comments | ask | show | jobs | submit login

That may well be true. I know that earlier models like Llama 1 65B could tolerate more aggressive quantization, which supports that idea.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: