Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I wonder if this can be solved by adding a disconnected, 2nd layer which can answer meta-questions about the bot’s answers, such as “does this answer follow the rules?”, and this can be used to tweak the answers.

Perhaps it can also be used to answer questions that require some sort of chain-of-thought, like “what would be the first step to answering the question? What is the next one?”, etc, then compile these “thoughts” to a final answer.

Basically you have the LLM that answers questions, and a “supervisor” LLM which can help guide it.

Does something like this exist?



Lots of people have tried solving prompt injection like this. I don't think it's a responsible strategy:

- https://simonwillison.net/2023/May/2/prompt-injection-explai...

- https://simonwillison.net/2022/Sep/17/prompt-injection-more-...




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: