Wait, how does rewriting a prompt until it gives you the output you expect help the LLM learn? Are you suggesting better prompting gets fed back into the training process in some helpful way? This feels confused.
You think OpenAI isn’t using your prompts and results to train better models? Think of yourself as a large RLHF experiment LOL
Kinda like Netflix did with people watching movies 10 years ago. The data’s there, and abundant. People are massaging their chatbot to get better results. You can measure when people are satisfied. So… obviously…