Distilling Knowledge from Ensembles of Cluster-Constrained-Attention Multiple-Instance Learners for Whole Slide Image Classification
Abstract
The peculiar nature of whole slide imaging (WSI), digitizing conventional glass slides to obtain multiple high resolution images which capture microscopic details of a patient’s histopathological features, has garnered increased interest from the computer vision research community over the last two decades. Given the unique computational space and time complexity inherent to gigapixel-size whole slide image data, researchers have proposed novel machine learning algorithms to aid in the performance of diagnostic tasks in clinical pathology. One effective algorithm represents a Whole slide image as a bag of smaller image patches, which can be represented as low-dimension image patch embeddings. Weakly supervised deep-learning methods, such as cluster-constrained-attention multiple instance learning (CLAM), have shown promising results when combined with image patch embeddings. While traditional ensemble classifiers yield improved task performance, such methods come with a steep cost in model complexity. Through knowledge distillation, it is possible to retain some performance improvements from an ensemble, while minimizing costs to model complexity. In this work, we implement a weakly supervised ensemble using clustering-constrained-attention multiple-instance learners (CLAM), which uses attention and instance-level clustering to identify task salient regions and feature extraction in whole slides. By applying logit-based and attention-based knowledge distillation, we show it ismore »
- Authors:
-
- ORNL
- Publication Date:
- Research Org.:
- Oak Ridge National Laboratory (ORNL), Oak Ridge, TN (United States)
- Sponsoring Org.:
- USDOE Office of Science (SC)
- OSTI Identifier:
- 1965255
- DOE Contract Number:
- AC05-00OR22725
- Resource Type:
- Conference
- Resource Relation:
- Conference: The 4th International Workshop on Big Data Tools, Methods, and Use Cases for Innovative Scientific Discovery - Osaka, , Japan - 12/17/2022 5:00:00 AM-
- Country of Publication:
- United States
- Language:
- English
Citation Formats
Alamudun, Folami, Dash, Sajal, Hinkle, Jacob, Hernandez Arreguin, Benjamin, Tsaris, Aristeidis, and Yoon, Hong-Jun. Distilling Knowledge from Ensembles of Cluster-Constrained-Attention Multiple-Instance Learners for Whole Slide Image Classification. United States: N. p., 2022.
Web. doi:10.1109/BigData55660.2022.10020938.
Alamudun, Folami, Dash, Sajal, Hinkle, Jacob, Hernandez Arreguin, Benjamin, Tsaris, Aristeidis, & Yoon, Hong-Jun. Distilling Knowledge from Ensembles of Cluster-Constrained-Attention Multiple-Instance Learners for Whole Slide Image Classification. United States. https://doi.org/10.1109/BigData55660.2022.10020938
Alamudun, Folami, Dash, Sajal, Hinkle, Jacob, Hernandez Arreguin, Benjamin, Tsaris, Aristeidis, and Yoon, Hong-Jun. 2022.
"Distilling Knowledge from Ensembles of Cluster-Constrained-Attention Multiple-Instance Learners for Whole Slide Image Classification". United States. https://doi.org/10.1109/BigData55660.2022.10020938. https://www.osti.gov/servlets/purl/1965255.
@article{osti_1965255,
title = {Distilling Knowledge from Ensembles of Cluster-Constrained-Attention Multiple-Instance Learners for Whole Slide Image Classification},
author = {Alamudun, Folami and Dash, Sajal and Hinkle, Jacob and Hernandez Arreguin, Benjamin and Tsaris, Aristeidis and Yoon, Hong-Jun},
abstractNote = {The peculiar nature of whole slide imaging (WSI), digitizing conventional glass slides to obtain multiple high resolution images which capture microscopic details of a patient’s histopathological features, has garnered increased interest from the computer vision research community over the last two decades. Given the unique computational space and time complexity inherent to gigapixel-size whole slide image data, researchers have proposed novel machine learning algorithms to aid in the performance of diagnostic tasks in clinical pathology. One effective algorithm represents a Whole slide image as a bag of smaller image patches, which can be represented as low-dimension image patch embeddings. Weakly supervised deep-learning methods, such as cluster-constrained-attention multiple instance learning (CLAM), have shown promising results when combined with image patch embeddings. While traditional ensemble classifiers yield improved task performance, such methods come with a steep cost in model complexity. Through knowledge distillation, it is possible to retain some performance improvements from an ensemble, while minimizing costs to model complexity. In this work, we implement a weakly supervised ensemble using clustering-constrained-attention multiple-instance learners (CLAM), which uses attention and instance-level clustering to identify task salient regions and feature extraction in whole slides. By applying logit-based and attention-based knowledge distillation, we show it is possible to retain some performance improvements resulting from the ensemble at zero cost to model complexity.},
doi = {10.1109/BigData55660.2022.10020938},
url = {https://www.osti.gov/biblio/1965255},
journal = {},
number = ,
volume = ,
place = {United States},
year = {2022},
month = {12}
}