ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2403.03230
27
45

Large language models surpass human experts in predicting neuroscience results

4 March 2024
Xiaoliang Luo
Akilles Rechardt
Guangzhi Sun
Kevin K. Nejad
Felipe Y´a˜nez
Bati Yilmaz
Kangjoo Lee
Alexandra O. Cohen
Valentina Borghesani
Anton Pashkov
Daniele Marinazzo
Jonathan Nicholas
Alessandro Salatiello
Ilia Sucholutsky
Pasquale Minervini
Sepehr Razavi
Roberta Rocca
Elkhan Yusifov
Tereza Okalova
Nianlong Gu
Martin Ferianc
Mikail Khona
Kaustubh R. Patil
Pui-Shee Lee
Rui Mata
Nicholas E. Myers
Jennifer K Bizley
Sebastian Musslick
I. Bilgin
Guiomar Niso
Justin M. Ales
Michael Gaebler
Apurva Ratan Murty
Leyla Loued-Khenissi
Anna Behler
Chloe M. Hall
J. Dafflon
Sherry Dongqi Bao
Bradley C. Love
ArXivPDFHTML
Abstract

Scientific discoveries often hinge on synthesizing decades of research, a task that potentially outstrips human information processing capacities. Large language models (LLMs) offer a solution. LLMs trained on the vast scientific literature could potentially integrate noisy yet interrelated findings to forecast novel results better than human experts. To evaluate this possibility, we created BrainBench, a forward-looking benchmark for predicting neuroscience results. We find that LLMs surpass experts in predicting experimental outcomes. BrainGPT, an LLM we tuned on the neuroscience literature, performed better yet. Like human experts, when LLMs were confident in their predictions, they were more likely to be correct, which presages a future where humans and LLMs team together to make discoveries. Our approach is not neuroscience-specific and is transferable to other knowledge-intensive endeavors.

View on arXiv
Comments on this paper