> Chain of thought has nothing to do with “tapping into better answers”. It’s simply asking the model to break up the output into smaller tasks and gives it more time and space to reason.
It doesn't give it more time and space to reason, though. Time isn't usually bounded on a single turn and space is context-window limited (and every turn in Chain of Thought is done within the context window, so it doesn't add any space.)
What it does is push the output toward a shape that resembles a particular idealization of (the explanation of) human reasoning, producing results that look more like an explanation of reasoning and sometimes producing more satisfying conclusions.
It does give more space in the sense that the model generates more tokens of steps/etc that it can then base its actual answer on, rather than being forced into generating the answer right away.
> It does give more space in the sense that the model generates more tokens of steps/etc that it can then base its actual answer on,
It doesn't give more space in the sense of increasing the upper bound on space used; it may bias the space used higher than a single naive prompt aiming to respond to the same question, but it doesn't alter the constraints.
It doesn't give it more time and space to reason, though. Time isn't usually bounded on a single turn and space is context-window limited (and every turn in Chain of Thought is done within the context window, so it doesn't add any space.)
What it does is push the output toward a shape that resembles a particular idealization of (the explanation of) human reasoning, producing results that look more like an explanation of reasoning and sometimes producing more satisfying conclusions.