skip to main content
OSTI.GOV title logo U.S. Department of Energy
Office of Scientific and Technical Information

Title: Distilling Knowledge from Ensembles of Cluster-Constrained-Attention Multiple-Instance Learners for Whole Slide Image Classification

Abstract

The peculiar nature of whole slide imaging (WSI), digitizing conventional glass slides to obtain multiple high resolution images which capture microscopic details of a patient’s histopathological features, has garnered increased interest from the computer vision research community over the last two decades. Given the unique computational space and time complexity inherent to gigapixel-size whole slide image data, researchers have proposed novel machine learning algorithms to aid in the performance of diagnostic tasks in clinical pathology. One effective algorithm represents a Whole slide image as a bag of smaller image patches, which can be represented as low-dimension image patch embeddings. Weakly supervised deep-learning methods, such as cluster-constrained-attention multiple instance learning (CLAM), have shown promising results when combined with image patch embeddings. While traditional ensemble classifiers yield improved task performance, such methods come with a steep cost in model complexity. Through knowledge distillation, it is possible to retain some performance improvements from an ensemble, while minimizing costs to model complexity. In this work, we implement a weakly supervised ensemble using clustering-constrained-attention multiple-instance learners (CLAM), which uses attention and instance-level clustering to identify task salient regions and feature extraction in whole slides. By applying logit-based and attention-based knowledge distillation, we show it ismore » possible to retain some performance improvements resulting from the ensemble at zero cost to model complexity.« less

Authors:
ORCiD logo [1];  [1]; ORCiD logo [1]; ORCiD logo [1];  [1]; ORCiD logo [1]
  1. ORNL
Publication Date:
Research Org.:
Oak Ridge National Laboratory (ORNL), Oak Ridge, TN (United States)
Sponsoring Org.:
USDOE Office of Science (SC)
OSTI Identifier:
1965255
DOE Contract Number:  
AC05-00OR22725
Resource Type:
Conference
Resource Relation:
Conference: The 4th International Workshop on Big Data Tools, Methods, and Use Cases for Innovative Scientific Discovery - Osaka, , Japan - 12/17/2022 5:00:00 AM-
Country of Publication:
United States
Language:
English

Citation Formats

Alamudun, Folami, Dash, Sajal, Hinkle, Jacob, Hernandez Arreguin, Benjamin, Tsaris, Aristeidis, and Yoon, Hong-Jun. Distilling Knowledge from Ensembles of Cluster-Constrained-Attention Multiple-Instance Learners for Whole Slide Image Classification. United States: N. p., 2022. Web. doi:10.1109/BigData55660.2022.10020938.
Alamudun, Folami, Dash, Sajal, Hinkle, Jacob, Hernandez Arreguin, Benjamin, Tsaris, Aristeidis, & Yoon, Hong-Jun. Distilling Knowledge from Ensembles of Cluster-Constrained-Attention Multiple-Instance Learners for Whole Slide Image Classification. United States. https://doi.org/10.1109/BigData55660.2022.10020938
Alamudun, Folami, Dash, Sajal, Hinkle, Jacob, Hernandez Arreguin, Benjamin, Tsaris, Aristeidis, and Yoon, Hong-Jun. 2022. "Distilling Knowledge from Ensembles of Cluster-Constrained-Attention Multiple-Instance Learners for Whole Slide Image Classification". United States. https://doi.org/10.1109/BigData55660.2022.10020938. https://www.osti.gov/servlets/purl/1965255.
@article{osti_1965255,
title = {Distilling Knowledge from Ensembles of Cluster-Constrained-Attention Multiple-Instance Learners for Whole Slide Image Classification},
author = {Alamudun, Folami and Dash, Sajal and Hinkle, Jacob and Hernandez Arreguin, Benjamin and Tsaris, Aristeidis and Yoon, Hong-Jun},
abstractNote = {The peculiar nature of whole slide imaging (WSI), digitizing conventional glass slides to obtain multiple high resolution images which capture microscopic details of a patient’s histopathological features, has garnered increased interest from the computer vision research community over the last two decades. Given the unique computational space and time complexity inherent to gigapixel-size whole slide image data, researchers have proposed novel machine learning algorithms to aid in the performance of diagnostic tasks in clinical pathology. One effective algorithm represents a Whole slide image as a bag of smaller image patches, which can be represented as low-dimension image patch embeddings. Weakly supervised deep-learning methods, such as cluster-constrained-attention multiple instance learning (CLAM), have shown promising results when combined with image patch embeddings. While traditional ensemble classifiers yield improved task performance, such methods come with a steep cost in model complexity. Through knowledge distillation, it is possible to retain some performance improvements from an ensemble, while minimizing costs to model complexity. In this work, we implement a weakly supervised ensemble using clustering-constrained-attention multiple-instance learners (CLAM), which uses attention and instance-level clustering to identify task salient regions and feature extraction in whole slides. By applying logit-based and attention-based knowledge distillation, we show it is possible to retain some performance improvements resulting from the ensemble at zero cost to model complexity.},
doi = {10.1109/BigData55660.2022.10020938},
url = {https://www.osti.gov/biblio/1965255}, journal = {},
number = ,
volume = ,
place = {United States},
year = {2022},
month = {12}
}

Conference:
Other availability
Please see Document Availability for additional information on obtaining the full-text document. Library patrons may search WorldCat to identify libraries that hold this conference proceeding.

Save / Share: