Skip to content

Meta & UCSD Develop DeepConf: Boosting Language Model Efficiency & Accuracy

DeepConf boosts language models' efficiency and accuracy. It filters out low-quality paths, reducing computational costs without compromising performance.

In this image a woman is speaking with the help of microphone, besides to her a man is standing,...
In this image a woman is speaking with the help of microphone, besides to her a man is standing, and he is holding a paper in his hand, in the background we can see a bottle, Cup, glass, papers, laptop on the table and also we can find couple of chairs, speaker and a projector screen.

Meta & UCSD Develop DeepConf: Boosting Language Model Efficiency & Accuracy

Researchers from Meta and the University of California, San Diego, have developed DeepConf, a method that enhances language models' efficiency and accuracy. It operates in offline and online modes, with the latter stopping low-confidence paths early, leading to significant token consumption reductions.

DeepConf achieved remarkable results on the gpt-oss-120B model for AIME 2025. In offline mode, it reached 99.9% accuracy, and in online mode, it scored 97.9% with an 84.7% reduction in token consumption. This method filters out low-quality solution paths by analyzing a model's confidence in its predictions.

The aggressive variant of DeepConf reduced token consumption by up to 84.7% in mathematical tasks, maintaining accuracy. The conservative variant saved up to 59% without compromising performance. Notably, DeepConf does not require additional training and can be integrated into existing systems with minimal code changes.

DeepConf's ability to improve mathematical reasoning in language models, reduce computational cost, and increase accuracy positions it as a crucial player in the evolution of language models. However, concerns about long-term viability due to rising energy costs and the need to prove economic viability have been raised.

Read also:

Latest