Large language models surpass human experts in predicting neuroscience results.
Détails
ID Serval
serval:BIB_FF43875B71A5
Type
Article: article d'un périodique ou d'un magazine.
Collection
Publications
Institution
Titre
Large language models surpass human experts in predicting neuroscience results.
Périodique
Nature human behaviour
ISSN
2397-3374 (Electronic)
ISSN-L
2397-3374
Statut éditorial
Publié
Date de publication
02/2025
Peer-reviewed
Oui
Volume
9
Numéro
2
Pages
305-315
Langue
anglais
Notes
Publication types: Journal Article
Publication Status: ppublish
Publication Status: ppublish
Résumé
Scientific discoveries often hinge on synthesizing decades of research, a task that potentially outstrips human information processing capacities. Large language models (LLMs) offer a solution. LLMs trained on the vast scientific literature could potentially integrate noisy yet interrelated findings to forecast novel results better than human experts. Here, to evaluate this possibility, we created BrainBench, a forward-looking benchmark for predicting neuroscience results. We find that LLMs surpass experts in predicting experimental outcomes. BrainGPT, an LLM we tuned on the neuroscience literature, performed better yet. Like human experts, when LLMs indicated high confidence in their predictions, their responses were more likely to be correct, which presages a future where LLMs assist humans in making discoveries. Our approach is not neuroscience specific and is transferable to other knowledge-intensive endeavours.
Mots-clé
Humans, Neurosciences, Language, Brain/physiology, Brain/diagnostic imaging
Pubmed
Web of science
Open Access
Oui
Création de la notice
02/12/2024 16:55
Dernière modification de la notice
04/03/2025 8:56