放大倍数
缩放比例
病理
计算机科学
人工智能
医学
数学
几何学
作者
Eric Zimmermann,Eugene Vorontsov,Julian Viret,Adam Casson,Michal Zelechowski,George Shaikovski,Neil Tenenholtz,James M. Hall,Thomas J. Fuchs,Nicolò Fusi,Siqi Liu,Kristen Severson
出处
期刊:Cornell University - arXiv
日期:2024-08-01
被引量:10
标识
DOI:10.48550/arxiv.2408.00738
摘要
Foundation models are rapidly being developed for computational pathology applications. However, it remains an open question which factors are most important for downstream performance with data scale and diversity, model size, and training algorithm all playing a role. In this work, we present the result of scaling both data and model size, surpassing previous studies in both dimensions, and introduce two new models: Virchow 2, a 632M parameter vision transformer, and Virchow 2G, a 1.85B parameter vision transformer, each trained with 3.1M histopathology whole slide images. To support this scale, we propose domain-inspired adaptations to the DINOv2 training algorithm, which is quickly becoming the default method in self-supervised learning for computational pathology. We achieve state of the art performance on twelve tile-level tasks, as compared to the top performing competing models. Our results suggest that data diversity and domain-specific training can outperform models that only scale in the number of parameters, but, on average, performance benefits from domain-tailoring, data scale, and model scale.
科研通智能强力驱动
Strongly Powered by AbleSci AI