Posing a technical question to a reasoning LLM may elicit a series of "thinking-like" sentences. For example,
Suppose this model is already pre-fossilized into its 600 billion weight values.
Facts like "donut-shaped coil" could simply be the highest ranked transformer continuation to the prompt "A toroidal solenoid is a ...", right?
However, I am not clear how do reasoning phrases like "Let me start by recalling ..., right?" manage to come out of the transformer?
The training corpus will likely not contain any phrases like "let me start by recalling" on web pages related to toroidal solenoids, correct?
Is this particular sentence added during model fine-tuning, so that this particular transformer always answers any question with the phrase "Let me start by recalling ..., right?"
How much reasoning is actually true reasoning in reasoning models?
