Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

My understanding is that models are currently undertrained and not very "dense", so Q4 doesn't hurt very much now but it may in future denser models.


That may well be true. I know that earlier models like Llama 1 65B could tolerate more aggressive quantization, which supports that idea.




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: