AstroMLab 1: Who wins astronomy jeopardy!?

dc.contributor.authorTing, Y. S.en
dc.contributor.authorNguyen, T. D.en
dc.contributor.authorGhosal, T.en
dc.contributor.authorPan, R.en
dc.contributor.authorArora, H.en
dc.contributor.authorSun, Z.en
dc.contributor.authorde Haan, T.en
dc.contributor.authorRamachandra, N.en
dc.contributor.authorWells, A.en
dc.contributor.authorMadireddy, S.en
dc.contributor.authorAccomazzi, A.en
dc.date.accessioned2025-05-23T09:25:40Z
dc.date.available2025-05-23T09:25:40Z
dc.date.issued2025en
dc.description.abstractWe present a comprehensive evaluation of proprietary and open-weights large language models using the first astronomy-specific benchmarking dataset. This dataset comprises 4,425 multiple-choice questions curated from the Annual Review of Astronomy and Astrophysics, covering a broad range of astrophysical topics.1 Our analysis examines model performance across various astronomical subfields and assesses response calibration, crucial for potential deployment in research environments. Claude-3.5-Sonnet outperforms competitors by up to 4.6 percentage points, achieving 85.0% accuracy. For proprietary models, we observed a universal reduction in cost every 3-to-12 months to achieve similar score in this particular astronomy benchmark. open-weights models have rapidly improved, with LLaMA-3-70b (80.6%) and Qwen-2-72b (77.7%) now competing with some of the best proprietary models. We identify performance variations across topics, with non-English-focused models generally struggling more in exoplanet-related fields, stellar astrophysics, and instrumentation related questions. These challenges likely stem from less abundant training data, limited historical context, and rapid recent developments in these areas. This pattern is observed across both open-weights and proprietary models, with regional dependencies evident, highlighting the impact of training data diversity on model performance in specialized scientific domains. Top-performing models demonstrate well-calibrated confidence, with correlations above 0.9 between confidence and correctness, though they tend to be slightly underconfident. The development for fast, low-cost inference of open-weights models presents new opportunities for affordable deployment in astronomy. The rapid progress observed suggests that LLM-driven research in astronomy may become feasible in the near future.en
dc.description.sponsorshipThis research was conducted using resources and services provided by the National Computational Infrastructure (NCI) , which receives support from the Australian Government, and the Oak Ridge Leadership Computing Facility Frontier Nodes . We are also grateful for support from Microsoft\u2019s Accelerating Foundation Models Research (AFMR) program , which played a crucial role in enabling this benchmarking work. The work at Argonne National Laboratory was supported by the U.S. Department of Energy, Office of High Energy Physics and Advanced Scientific Computing Research , through the SciDAC-RAPIDS2 institute. Argonne National Laboratory is a U.S. Department of Energy Office of Science Laboratory operated by UChicago Argonne LLC under contract no. DE-AC02-06CH11357 . The views expressed herein do not necessarily represent the views of the U.S. Department of Energy or the United States Government. This research was conducted using resources and services provided by the National Computational Infrastructure (NCI), Australia, which receives support from the Australian Government, and the Oak Ridge Leadership Computing Facility Frontier Nodes, United States, which is a DOE Office of Science User Facility at the Oak Ridge National Laboratory supported by the U.S. Department of Energy under Contract No. DE-AC05-00OR22725. We are also grateful for support from Microsoft's Accelerating Foundation Models Research (AFMR) program, United States, which played a crucial role in enabling this benchmarking work. The work at Argonne National Laboratory was supported by the U.S. Department of Energy, Office of High Energy Physics and Advanced Scientific Computing Research, through the SciDAC-RAPIDS2 institute. Argonne National Laboratory is a U.S. Department of Energy Office of Science Laboratory operated by UChicago Argonne LLC, United States under contract no. DE-AC02-06CH11357. The views expressed herein do not necessarily represent the views of the U.S. Department of Energy or the United States Government.en
dc.description.statusPeer-revieweden
dc.format.extent29en
dc.identifier.issn2213-1337en
dc.identifier.scopus85210381998en
dc.identifier.urihttp://www.scopus.com/inward/record.url?scp=85210381998&partnerID=8YFLogxKen
dc.identifier.urihttps://hdl.handle.net/1885/733751962
dc.language.isoenen
dc.rights © 2024 The Author(s) en
dc.sourceAstronomy and Computingen
dc.subjectAstronomyen
dc.subjectBenchmarkingen
dc.subjectLarge Language Modelsen
dc.subjectQuestion Answeringen
dc.subjectScientific Knowledge Assessmenten
dc.titleAstroMLab 1: Who wins astronomy jeopardy!?en
dc.typeJournal articleen
dspace.entity.typePublicationen
local.contributor.affiliationTing, Y. S.; School of Computing, ANU College of Systems and Society, The Australian National Universityen
local.contributor.affiliationNguyen, T. D.; University of Pennsylvaniaen
local.contributor.affiliationGhosal, T.; Oak Ridge National Laboratoryen
local.contributor.affiliationPan, R.; Hong Kong University of Science and Technologyen
local.contributor.affiliationArora, H.; Indian Institute of Technology Patnaen
local.contributor.affiliationSun, Z.; Tsinghua Universityen
local.contributor.affiliationde Haan, T.; High Energy Accelerator Research Organization, Institute of Particle and Nuclear Physicsen
local.contributor.affiliationRamachandra, N.; Argonne National Laboratoryen
local.contributor.affiliationWells, A.; Argonne National Laboratoryen
local.contributor.affiliationMadireddy, S.; Argonne National Laboratoryen
local.contributor.affiliationAccomazzi, A.; Harvard-Smithsonian Center for Astrophysicsen
local.identifier.citationvolume51en
local.identifier.doi10.1016/j.ascom.2024.100893en
local.identifier.puref881fb9f-ea5e-4091-a995-a0b097ef116ben
local.identifier.urlhttps://www.scopus.com/pages/publications/85210381998en
local.type.statusPublisheden

Downloads