任务(项目管理)
水准点(测量)
心理学
认知科学
神经科学
地理
经济
大地测量学
管理
作者
Xiaoliang Luo,Akilles Rechardt,Guangzhi Sun,Kevin Kermani Nejad,Felipe Yáñez,Bati Yilmaz,Kangjoo Lee,Alexandra O. Cohen,Valentina Borghesani,Anton Pashkov,Daniele Marinazzo,Jonathan Nicholas,Alessandro Salatiello,Ilia Sucholutsky,Pasquale Minervini,S. Morteza Razavi,Roberta Rocca,Elkhan Yusifov,Tereza Okalova,Nianlong Gu
标识
DOI:10.1038/s41562-024-02046-9
摘要
Abstract Scientific discoveries often hinge on synthesizing decades of research, a task that potentially outstrips human information processing capacities. Large language models (LLMs) offer a solution. LLMs trained on the vast scientific literature could potentially integrate noisy yet interrelated findings to forecast novel results better than human experts. Here, to evaluate this possibility, we created BrainBench, a forward-looking benchmark for predicting neuroscience results. We find that LLMs surpass experts in predicting experimental outcomes. BrainGPT, an LLM we tuned on the neuroscience literature, performed better yet. Like human experts, when LLMs indicated high confidence in their predictions, their responses were more likely to be correct, which presages a future where LLMs assist humans in making discoveries. Our approach is not neuroscience specific and is transferable to other knowledge-intensive endeavours.
科研通智能强力驱动
Strongly Powered by AbleSci AI