Vorontsov, Eugene
PRISM: A Multi-Modal Generative Foundation Model for Slide-Level Histopathology
Shaikovski, George, Casson, Adam, Severson, Kristen, Zimmermann, Eric, Wang, Yi Kan, Kunz, Jeremy D., Retamero, Juan A., Oakley, Gerard, Klimstra, David, Kanan, Christopher, Hanna, Matthew, Zelechowski, Michal, Viret, Julian, Tenenholtz, Neil, Hall, James, Fusi, Nicolo, Yousfi, Razik, Hamilton, Peter, Moye, William A., Vorontsov, Eugene, Liu, Siqi, Fuchs, Thomas J.
Foundation models in computational pathology promise to unlock the development of new clinical decision support systems and models for precision medicine. However, there is a mismatch between most clinical analysis, which is defined at the level of one or more whole slide images, and foundation models to date, which process the thousands of image tiles contained in a whole slide image separately. The requirement to train a network to aggregate information across a large number of tiles in multiple whole slide images limits these models' impact. In this work, we present a slide-level foundation model for H&E-stained histopathology, PRISM, that builds on Virchow tile embeddings and leverages clinical report text for pre-training. Using the tile embeddings, PRISM produces slide-level embeddings with the ability to generate clinical reports, resulting in several modes of use. Using text prompts, PRISM achieves zero-shot cancer detection and sub-typing performance approaching and surpassing that of a supervised aggregator model. Using the slide embeddings with linear classifiers, PRISM surpasses supervised aggregator models. Furthermore, we demonstrate that fine-tuning of the PRISM slide encoder yields label-efficient training for biomarker prediction, a task that typically suffers from low availability of training data; an aggregator initialized with PRISM and trained on as little as 10% of the training data can outperform a supervised baseline that uses all of the data.
Virchow: A Million-Slide Digital Pathology Foundation Model
Vorontsov, Eugene, Bozkurt, Alican, Casson, Adam, Shaikovski, George, Zelechowski, Michal, Liu, Siqi, Severson, Kristen, Zimmermann, Eric, Hall, James, Tenenholtz, Neil, Fusi, Nicolo, Mathieu, Philippe, van Eck, Alexander, Lee, Donghun, Viret, Julian, Robert, Eric, Wang, Yi Kan, Kunz, Jeremy D., Lee, Matthew C. H., Bernhard, Jan, Godrich, Ran A., Oakley, Gerard, Millar, Ewan, Hanna, Matthew, Retamero, Juan, Moye, William A., Yousfi, Razik, Kanan, Christopher, Klimstra, David, Rothrock, Brandon, Fuchs, Thomas J.
The use of artificial intelligence to enable precision medicine and decision support systems through the analysis of pathology images has the potential to revolutionize the diagnosis and treatment of cancer. Such applications will depend on models' abilities to capture the diverse patterns observed in pathology images. To address this challenge, we present Virchow, a foundation model for computational pathology. Using self-supervised learning empowered by the DINOv2 algorithm, Virchow is a vision transformer model with 632 million parameters trained on 1.5 million hematoxylin and eosin stained whole slide images from diverse tissue and specimen types, which is orders of magnitude more data than previous works. The Virchow model enables the development of a pan-cancer detection system with 0.949 overall specimen-level AUC across 17 different cancer types, while also achieving 0.937 AUC on 7 rare cancer types. The Virchow model sets the state-of-the-art on the internal and external image tile level benchmarks and slide level biomarker prediction tasks. The gains in performance highlight the importance of training on massive pathology image datasets, suggesting scaling up the data and network architecture can improve the accuracy for many high-impact computational pathology applications where limited amounts of training data are available.
Non-normal Recurrent Neural Network (nnRNN): learning long time dependencies while improving expressivity with transient dynamics
Kerg, Giancarlo, Goyette, Kyle, Touzel, Maximilian Puelma, Gidel, Gauthier, Vorontsov, Eugene, Bengio, Yoshua, Lajoie, Guillaume
A recent strategy to circumvent the exploding and vanishing gradient problem in RNNs, and to allow the stable propagation of signals over long time scales, is to constrain recurrent connectivity matrices to be orthogonal or unitary. This ensures eigenvalues with unit norm and thus stable dynamics and training. However this comes at the cost of reduced expressivity due to the limited variety of orthogonal transformations. We propose a novel connectivity structure based on the Schur decomposition and a splitting of the Schur form into normal and non-normal parts. This allows to parametrize matrices with unit-norm eigenspectra without orthogonality constraints on eigenbases. The resulting architecture ensures access to a larger space of spectrally constrained matrices, of which orthogonal matrices are a subset. This crucial difference retains the stability advantages and training speed of orthogonal RNNs while enhancing expressivity, especially on tasks that require computations over ongoing input sequences.
Towards Non-saturating Recurrent Units for Modelling Long-term Dependencies
Chandar, Sarath, Sankar, Chinnadhurai, Vorontsov, Eugene, Kahou, Samira Ebrahimi, Bengio, Yoshua
Modelling long-term dependencies is a challenge for recurrent neural networks. This is primarily due to the fact that gradients vanish during training, as the sequence length increases. Gradients can be attenuated by transition operators and are attenuated or dropped by activation functions. Canonical architectures like LSTM alleviate this issue by skipping information through a memory mechanism. We propose a new recurrent architecture (Non-saturating Recurrent Unit; NRU) that relies on a memory mechanism but forgoes both saturating activation functions and saturating gates, in order to further alleviate vanishing gradients. In a series of synthetic and real world tasks, we demonstrate that the proposed model is the only model that performs among the top 2 models across all tasks with and without long-term dependencies, when compared against a range of other architectures.