{"id":1141870,"date":"2025-06-17T09:00:00","date_gmt":"2025-06-17T16:00:00","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?p=1141870"},"modified":"2025-06-17T06:32:47","modified_gmt":"2025-06-17T13:32:47","slug":"new-methods-boost-reasoning-in-small-and-large-language-models","status":"publish","type":"post","link":"https:\/\/www.microsoft.com\/en-us\/research\/blog\/new-methods-boost-reasoning-in-small-and-large-language-models\/","title":{"rendered":"New methods boost reasoning in small and large language models"},"content":{"rendered":"\n
\"The<\/figure>\n\n\n\n

Artificial intelligence is advancing across a wide range of fields, with one of the most important developments being its growing capacity for reasoning. This capability could help AI becomes a reliable partner in critical domains like scientific research and healthcare.<\/p>\n\n\n\n

To support this progress, we\u2019ve identified three primary strategies to strengthen reasoning capabilities in both small and large language models: improve architectural design to boost performance in smaller models; incorporate mathematical reasoning techniques to increase reliability; and build stronger generalization capabilities to enable reasoning across a variety of fields.<\/p>\n\n\n\n

Smarter reasoning in smaller models<\/h2>\n\n\n\n

While language models trained on broad world knowledge hold great potential, they lack the ability to learn continuously and refine their understanding. This limitation becomes especially pronounced in smaller models, where limited capacity makes strong reasoning even harder.<\/p>\n\n\n\n

The problem stems from how current language models operate. They rely on fast, pattern recognition-based responses that break down in complex scenarios. In contrast, people use deliberate, step-by-step reasoning, test different approaches, and evaluate outcomes. To address this gap, we\u2019re building methods to enable stronger reasoning in smaller systems.<\/p>\n\n\n\n

rStar-Math<\/a> is a method that uses Monte Carlo Tree Search (MCTS) to simulate deeper, more methodical reasoning in smaller models. It uses a three-step, self-improving cycle: <\/p>\n\n\n\n