Hacker News new | past | comments | ask | show | jobs | submit login

It could very well be legit, but if you "have not tested its answer yet" the fact that it can generate something that looks plausible doesn't really tell you much. Generating plausible-sounding but incorrect answers is like the #1 most common failure mode for LLMs.





in recent usage, that only happens 10% of the time for me. Usually the results are grounded and so work usually fine.

Could you imagine if any other software failed silently and plausibly 10% of the time? It would never get off the ground. VC money is a hell of a drug



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: