Masked autoencoders with handcrafted feature predictions: Transformer for weakly supervised esophageal cancer classification.

Esophageal cancer Multi-instance learning Self-supervised learning Transformer Whole slide image analysis

Journal

Computer methods and programs in biomedicine
ISSN: 1872-7565
Titre abrégé: Comput Methods Programs Biomed
Pays: Ireland
ID NLM: 8506513

Informations de publication

Date de publication:
22 Nov 2023
Historique:
received: 06 12 2022
revised: 28 10 2023
accepted: 19 11 2023
medline: 29 11 2023
pubmed: 29 11 2023
entrez: 28 11 2023
Statut: aheadofprint

Résumé

Esophageal cancer is a serious disease with a high prevalence in Eastern Asia. Histopathology tissue analysis stands as the gold standard in diagnosing esophageal cancer. In recent years, there has been a shift towards digitizing histopathological images into whole slide images (WSIs), progressively integrating them into cancer diagnostics. However, the gigapixel sizes of WSIs present significant storage and processing challenges, and they often lack localized annotations. To address this issue, multi-instance learning (MIL) has been introduced for WSI classification, utilizing weakly supervised learning for diagnosis analysis. By applying the principles of MIL to WSI analysis, it is possible to reduce the workload of pathologists by facilitating the generation of localized annotations. Nevertheless, the approach's effectiveness is hindered by the traditional simple aggregation operation and the domain shift resulting from the prevalent use of convolutional feature extractors pretrained on ImageNet. We propose a MIL-based framework for WSI analysis and cancer classification. Concurrently, we introduce employing self-supervised learning, which obviates the need for manual annotation and demonstrates versatility in various tasks, to pretrain feature extractors. This method enhances the extraction of representative features from esophageal WSI for MIL, ensuring more robust and accurate performance. We build a comprehensive dataset of whole esophageal slide images and conduct extensive experiments utilizing this dataset. The performance on our dataset demonstrates the efficiency of our proposed MIL framework and the pretraining process, with our framework outperforming existing methods, achieving an accuracy of 93.07% and AUC (area under the curve) of 95.31%. This work proposes an effective MIL method to classify WSI of esophageal cancer. The promising results indicate that our cancer classification framework holds great potential in promoting the automatic whole esophageal slide image analysis.

Sections du résumé

BACKGROUND AND OBJECTIVE OBJECTIVE
Esophageal cancer is a serious disease with a high prevalence in Eastern Asia. Histopathology tissue analysis stands as the gold standard in diagnosing esophageal cancer. In recent years, there has been a shift towards digitizing histopathological images into whole slide images (WSIs), progressively integrating them into cancer diagnostics. However, the gigapixel sizes of WSIs present significant storage and processing challenges, and they often lack localized annotations. To address this issue, multi-instance learning (MIL) has been introduced for WSI classification, utilizing weakly supervised learning for diagnosis analysis. By applying the principles of MIL to WSI analysis, it is possible to reduce the workload of pathologists by facilitating the generation of localized annotations. Nevertheless, the approach's effectiveness is hindered by the traditional simple aggregation operation and the domain shift resulting from the prevalent use of convolutional feature extractors pretrained on ImageNet.
METHODS METHODS
We propose a MIL-based framework for WSI analysis and cancer classification. Concurrently, we introduce employing self-supervised learning, which obviates the need for manual annotation and demonstrates versatility in various tasks, to pretrain feature extractors. This method enhances the extraction of representative features from esophageal WSI for MIL, ensuring more robust and accurate performance.
RESULTS RESULTS
We build a comprehensive dataset of whole esophageal slide images and conduct extensive experiments utilizing this dataset. The performance on our dataset demonstrates the efficiency of our proposed MIL framework and the pretraining process, with our framework outperforming existing methods, achieving an accuracy of 93.07% and AUC (area under the curve) of 95.31%.
CONCLUSION CONCLUSIONS
This work proposes an effective MIL method to classify WSI of esophageal cancer. The promising results indicate that our cancer classification framework holds great potential in promoting the automatic whole esophageal slide image analysis.

Identifiants

pubmed: 38016392
pii: S0169-2607(23)00602-8
doi: 10.1016/j.cmpb.2023.107936
pii:
doi:

Types de publication

Journal Article

Langues

eng

Sous-ensembles de citation

IM

Pagination

107936

Informations de copyright

Copyright © 2023. Published by Elsevier B.V.

Déclaration de conflit d'intérêts

Declaration of Competing Interest There are no known conflicts of interest associated with this publication entitled “Masked Autoencoders with handcrafted feature predictions: Transformer for weakly supervised esophageal cancer classification”.

Auteurs

Yunhao Bai (Y)

the School of Electrical and Information Engineering, Tianjin University, Tianjin, China.

Wenqi Li (W)

Department of Pathology, Key Laboratory of Cancer Prevention and Therapy, National Clinical Research Center for Cancer, Tianjin Medical University Cancer Institute and Hospital, Tianjin, China.

Jianpeng An (J)

the School of Electrical and Information Engineering, Tianjin University, Tianjin, China.

Lili Xia (L)

the School of Electrical and Information Engineering, Tianjin University, Tianjin, China.

Huazhen Chen (H)

the School of Electrical and Information Engineering, Tianjin University, Tianjin, China.

Gang Zhao (G)

Department of Pathology, Key Laboratory of Cancer Prevention and Therapy, National Clinical Research Center for Cancer, Tianjin Medical University Cancer Institute and Hospital, Tianjin, China.

Zhongke Gao (Z)

the School of Electrical and Information Engineering, Tianjin University, Tianjin, China. Electronic address: zhongkegao@tju.edu.cn.

Classifications MeSH