Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

There are millions of websites, and a local LLM cannot scrape all of them to make sense of them. Think about it. OpenAI can do it because they spend millions to train its systems.

Many sites have hidden sitemaps that cannot be found unless submitted to google directly. (Not even listed in robots txt most of the time). There is no way a local LLM can keep up with up to date internet.





No, the AI will just use Google, DDG, Bing, etc on behalf of the user (behind the scenes). The ads will be shown to the AI which will ignore them.



Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: