Great observation! You nailed it with the comment. 💗To clarify, when I mentioned this as an 'alternative,' I was referring to the implementation method (using a seamless pipeline without enforced tags), not necessarily a breakthrough in arithmetic capability at the 1B scale. What you're seeing here is a classic example of hallucination in small-parameter models. The model is faithfully following the instruction to 'reason step-by-step' (CoT), but due to its limited size (1B), it hallucinates the intermediate calculations while maintaining a confident tone. Maintaining logic while ensuring factual accuracy in such compact models is indeed one of the biggest challenges we are currently facing and working to optimize.