TY - JOUR
T1 - A whole-slide foundation model for digital pathology from real-world data
AU - Xu, Hanwen
AU - Usuyama, Naoto
AU - Bagga, Jaspreet
AU - Zhang, Sheng
AU - Rao, Rajesh
AU - Naumann, Tristan
AU - Wong, Cliff
AU - Gero, Zelalem
AU - González, Javier
AU - Gu, Yu
AU - Xu, Yanbo
AU - Wei, Mu
AU - Wang, Wenhui
AU - Ma, Shuming
AU - Wei, Furu
AU - Yang, Jianwei
AU - Li, Chunyuan
AU - Gao, Jianfeng
AU - Rosemon, Jaylen
AU - Bower, Tucker
AU - Lee, Soohee
AU - Weerasinghe, Roshanthi
AU - Wright, Bill J.
AU - Robicsek, Ari
AU - Piening, Brian
AU - Bifulco, Carlo
AU - Wang, Sheng
AU - Poon, Hoifung
N1 - Publisher Copyright:
© The Author(s) 2024.
PY - 2024
Y1 - 2024
N2 - Digital pathology poses unique computational challenges, as a standard gigapixel slide may comprise tens of thousands of image tiles1–3. Prior models have often resorted to subsampling a small portion of tiles for each slide, thus missing the important slide-level context4. Here we present Prov-GigaPath, a whole-slide pathology foundation model pretrained on 1.3 billion 256 × 256 pathology image tiles in 171,189 whole slides from Providence, a large US health network comprising 28 cancer centres. The slides originated from more than 30,000 patients covering 31 major tissue types. To pretrain Prov-GigaPath, we propose GigaPath, a novel vision transformer architecture for pretraining gigapixel pathology slides. To scale GigaPath for slide-level learning with tens of thousands of image tiles, GigaPath adapts the newly developed LongNet5 method to digital pathology. To evaluate Prov-GigaPath, we construct a digital pathology benchmark comprising 9 cancer subtyping tasks and 17 pathomics tasks, using both Providence and TCGA data6. With large-scale pretraining and ultra-large-context modelling, Prov-GigaPath attains state-of-the-art performance on 25 out of 26 tasks, with significant improvement over the second-best method on 18 tasks. We further demonstrate the potential of Prov-GigaPath on vision–language pretraining for pathology7,8 by incorporating the pathology reports. In sum, Prov-GigaPath is an open-weight foundation model that achieves state-of-the-art performance on various digital pathology tasks, demonstrating the importance of real-world data and whole-slide modelling.
AB - Digital pathology poses unique computational challenges, as a standard gigapixel slide may comprise tens of thousands of image tiles1–3. Prior models have often resorted to subsampling a small portion of tiles for each slide, thus missing the important slide-level context4. Here we present Prov-GigaPath, a whole-slide pathology foundation model pretrained on 1.3 billion 256 × 256 pathology image tiles in 171,189 whole slides from Providence, a large US health network comprising 28 cancer centres. The slides originated from more than 30,000 patients covering 31 major tissue types. To pretrain Prov-GigaPath, we propose GigaPath, a novel vision transformer architecture for pretraining gigapixel pathology slides. To scale GigaPath for slide-level learning with tens of thousands of image tiles, GigaPath adapts the newly developed LongNet5 method to digital pathology. To evaluate Prov-GigaPath, we construct a digital pathology benchmark comprising 9 cancer subtyping tasks and 17 pathomics tasks, using both Providence and TCGA data6. With large-scale pretraining and ultra-large-context modelling, Prov-GigaPath attains state-of-the-art performance on 25 out of 26 tasks, with significant improvement over the second-best method on 18 tasks. We further demonstrate the potential of Prov-GigaPath on vision–language pretraining for pathology7,8 by incorporating the pathology reports. In sum, Prov-GigaPath is an open-weight foundation model that achieves state-of-the-art performance on various digital pathology tasks, demonstrating the importance of real-world data and whole-slide modelling.
UR - http://www.scopus.com/inward/record.url?scp=85193847602&partnerID=8YFLogxK
U2 - 10.1038/s41586-024-07441-w
DO - 10.1038/s41586-024-07441-w
M3 - Article
C2 - 38778098
AN - SCOPUS:85193847602
SN - 0028-0836
JO - Nature
JF - Nature
ER -