You are currently viewing MIT’s Smarter Path for LLMs on Tough Problems
Featured image for MIT's Smarter Path for LLMs on Tough Problems

MIT’s Smarter Path for LLMs on Tough Problems

  • Post author:
  • Post category:News
  • Post comments:0 Comments

MIT’s Smarter Path for LLMs on Tough Problems

Image sourced from news.mit.edu
Image sourced from news.mit.edu

Large language models struggle with hard questions because they often get stuck with a fixed amount of thinking time. MIT researchers fixed that with a method called instance-adaptive scaling. It lets LLMs adjust how much compute they use on the fly, focusing effort on promising ideas while skipping dead ends. Their tests showed it cuts compute in half while matching accuracy on math problems of different difficulties, according to the MIT News report.

How It Works, Step by Step

Standard “inference-time scaling” gives LLMs extra steps to reason—like generating multiple paths and picking the best with a process reward model (PRM). But it uses the same budget for every problem, wasting time on easy ones.

MIT’s version watches the action as it happens:

  • The PRM checks the question and partial answers at each step, scoring how likely a path leads to success.
  • Hard questions get more paths explored; confident spots trim them down.
  • Humans do this naturally: sketch ideas, chase the good ones, ditch the rest. Navid Azizan, a mechanical engineering professor at MIT, compared it to knowing what you don’t know.

One issue: PRMs tend to act too sure. The team calibrated them to output a range of probabilities instead of one number. This gives honest uncertainty, so the LLM doesn’t cut corners too soon. Lead author Young-Jin Park noted it makes scaling reliable without slashing accuracy.

Why It Matters Now

Reasoning setups boost LLMs on complex tasks, but they guzzle energy. A study by Hugging Face and Salesforce found reasoning models use 100 times more power than plain ones on 1,000 prompts. DeepSeek’s R1 jumped from 50 watt-hours to over 308,000 with reasoning on (Mint; Nature on DeepSeek math). MIT’s adaptive method helps by saving compute where possible.

It also lets smaller LLMs beat bigger ones on hard stuff, cutting energy overall. See VentureBeat on multimodal reasoning and Quanta on hard problems. The paper hits NeurIPS this week, with authors from MIT and IBM.

Implementation Guide

No public code yet, but here’s how to build it based on the MIT description. Use libraries like Hugging Face Transformers for the LLM and PRM.

  1. Set up your base LLM and PRM: Pick an LLM (e.g., Llama or Mistral). Train or fine-tune a PRM on reasoning traces—reward high for correct paths, low for failures. OpenAI’s o1-style setups offer PRM examples.
  2. Calibrate the PRM: Run it on held-out data. Train it to output probability distributions (e.g., via temperature scaling or Platt scaling) that match real success rates. Aim for proper uncertainty: wide ranges on tough spots.
  3. Dynamic scaling loop:
    • Start with question. Generate N initial paths (N based on quick PRM difficulty estimate).
    • At each step: PRM scores current partials. Sample budget for next tokens/paths proportional to top scores (e.g., keep top-k where k drops if probs > 0.8).
    • Stop early if overall success prob > threshold or max steps hit.
  4. Test and tune: Run on math benchmarks like GSM8K. Track tokens used vs. accuracy. Adjust PRM weights if over/under-confident.

For code inspo, check inference-time scaling repos on GitHub (search “test-time compute LLM”). Kristjan Greenewald from MIT-IBM said the adaptation runs live, problem-by-problem.

Separate OpenAI work trains LLMs to “confess” reasoning slips, per MIT Technology Review. Pairs well with MIT’s scaling for trustworthy hard-problem solving.

More stories at letsjustdoai.com

Seb

I love AI and automations, I enjoy seeing how it can make my life easier. I have a background in computational sciences and worked in academia, industry and as consultant. This is my journey about how I learn and use AI.

Leave a Reply