×
MIT’s CodeSteer boosts LLM accuracy 30% by coaching code use
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

MIT researchers have developed CodeSteer, a “smart coach” system that guides large language models to switch between text and code generation to solve complex problems more accurately. The system boosted LLM accuracy on symbolic tasks like math problems and Sudoku by more than 30 percent, addressing a key weakness where models often default to less effective textual reasoning even when code would be more appropriate.

How it works: CodeSteer operates as a smaller, specialized LLM that iteratively guides larger models through problem-solving processes.

  • The system first analyzes a query to determine whether text or code would be more effective, then generates prompts directing the larger LLM accordingly.
  • After receiving an answer, CodeSteer reviews the response and continues prompting the model to refine its approach until reaching a correct solution.
  • A symbolic checker evaluates code complexity to prevent the larger LLM from using overly simple or inefficient solutions, while a self-answer checker verifies correctness.

Key performance gains: Testing across 37 complex symbolic tasks showed significant improvements in LLM capabilities.

  • Average accuracy increased from 53.3 percent to 86.4 percent when CodeSteer was added to existing models.
  • The system enabled less sophisticated models to outperform more advanced models with enhanced reasoning skills.
  • Performance remained consistent across different LLMs and on previously unseen tasks.

Why this matters: The approach addresses a fundamental limitation in how LLMs handle computational versus linguistic tasks.

  • Initially trained to understand and predict human language, LLMs are more likely to answer queries using text, even when code would be more effective for problems like comparing numbers or solving math equations.
  • This could improve LLM performance on complex real-world applications like robot path planning or supply chain scheduling.

The bigger picture: Rather than developing entirely new models, the MIT team focused on enhancing existing capabilities through strategic guidance.

  • “There is a race to develop better and better models that are capable of doing everything, but we’ve taken a complementary approach,” says Chuchu Fan, an associate professor at MIT and the study’s senior author.
  • Fine-tuning the smaller CodeSteer model doesn’t alter the larger LLM, eliminating risks to its other capabilities.

What they’re saying: External experts praised the approach for its practical impact on LLM performance.

  • “This simple yet impactful method enables state-of-the-art LLMs to achieve significant performance improvements without requiring direct fine-tuning,” said Jinsung Yoon, a staff research scientist at Google Cloud AI.
  • Chi Wang from Google DeepMind highlighted how “this intelligent collaboration among diverse AI ‘agents’ paves the way for more robust and versatile applications in complex real-world scenarios.”

What’s next: The researchers plan to streamline CodeSteer’s iterative prompting process for faster performance and explore developing unified models that can switch between reasoning modes without requiring a separate assistant.

This “smart coach” helps LLMs switch between text and code

Recent News

Roblox launches AI age verification for teens amid safety concerns

Video selfies unlock unfiltered chat for verified teens, but experts warn predators adapt quickly.

Scammers use AI deepfakes of Indian chief minister to promote fake investment scheme

Cybercriminals weaponized trusted public figures to lend credibility to sophisticated fraud schemes.

Google adds visual song search history to Circle to Search on Android

A clock icon now reveals your personal soundtrack archaeology in a visual grid format.