I wouldn't call a step in a history-aware conversation a completely new universe. By that logic, every single time a token is generated is a new universe even though the token is largely dependent on the prompt, which includes custom instructions, chat history, and all tokens generated in the response so far.
Well, I would have also said or thought that each token actually is a new universe in a sense. You could rotate between different LLMs for each token for example or instances of the same LLM or branch into different possibilities. Input gets cycled again as a whole.