Back to Search Start Over

Solving Quantitative Reasoning Problems with Language Models

Authors :
Lewkowycz, Aitor
Andreassen, Anders
Dohan, David
Dyer, Ethan
Michalewski, Henryk
Ramasesh, Vinay
Slone, Ambrose
Anil, Cem
Schlag, Imanol
Gutman-Solo, Theo
Wu, Yuhuai
Neyshabur, Behnam
Gur-Ari, Guy
Misra, Vedant
Publication Year :
2022

Abstract

Language models have achieved remarkable performance on a wide range of tasks that require natural language understanding. Nevertheless, state-of-the-art models have generally struggled with tasks that require quantitative reasoning, such as solving mathematics, science, and engineering problems at the college level. To help close this gap, we introduce Minerva, a large language model pretrained on general natural language data and further trained on technical content. The model achieves state-of-the-art performance on technical benchmarks without the use of external tools. We also evaluate our model on over two hundred undergraduate-level problems in physics, biology, chemistry, economics, and other sciences that require quantitative reasoning, and find that the model can correctly answer nearly a third of them.<br />Comment: 12 pages, 5 figures + references and appendices

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2206.14858
Document Type :
Working Paper