Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Larger llms do pretty well with this.

Smaller ones don't.



Large ones do better than small ones but still do worse than I would have expected before I tested them. E.g. `o1` doesn't know things which are repeated several times on wikipedia.


o1 is not too large, and the emphasis is on reasoning rather than memorization.

Try the largest llama models, and phrase your prompt like a sentence to be completed instead of you asking a question.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: