LingVo.club
Level
New training method helps models do long multiplication — Level A2 — brown wooden blocks on white surface

New training method helps models do long multiplicationCEFR A2

29 Dec 2025

Level A2 – High beginner / Elementary
2 min
118 words

Researchers compared standard fine-tuning with a different method called Implicit Chain of Thought (ICoT). They looked at how models handle long calculations and whether the models can keep intermediate results during many steps.

Under standard fine-tuning, models with two to 12 layers had less than 1% accuracy on four-digit multiplication. By contrast, the ICoT-trained model reached 100% accuracy. The team found they could read running sums from the ICoT model’s internal states, which shows the model learned to remember useful intermediate values.

The researchers also added a training objective that teaches a model to track running sums. Adding this objective raised a two-layer model’s accuracy to 99% without explicit chain-of-thought supervision.

Difficult words

  • fine-tuningsmall training to improve a model
  • implicitnot directly shown or written
  • intermediatebetween the first and last steps
  • accuracyhow often answers are correct
  • internalinside a system or model
  • objectivea goal used during training

Tip: hover, focus or tap highlighted words in the article to see quick definitions while you read or listen.

Discussion questions

  • Do you write intermediate results when you do long calculations? Why?
  • Would you like a model that remembers intermediate values? Why or why not?
  • How can remembering intermediate values help solve a problem?

Related articles

New training method helps models do long multiplication — English Level A2 | LingVo.club