A convolutional neural network with self-attention for fully automated metabolic tumor volume delineation of head and neck cancer in [Formula: see text]F]FDG PET/CT.
Convolutional neural network
FDG PET
HNC
Head and neck cancer
MTV
Metabolic tumor volume
Journal
European journal of nuclear medicine and molecular imaging
ISSN: 1619-7089
Titre abrégé: Eur J Nucl Med Mol Imaging
Pays: Germany
ID NLM: 101140988
Informations de publication
Date de publication:
07 2023
07 2023
Historique:
received:
14
11
2022
accepted:
14
03
2023
medline:
5
7
2023
pubmed:
20
4
2023
entrez:
20
04
2023
Statut:
ppublish
Résumé
PET-derived metabolic tumor volume (MTV) and total lesion glycolysis of the primary tumor are known to be prognostic of clinical outcome in head and neck cancer (HNC). Including evaluation of lymph node metastases can further increase the prognostic value of PET but accurate manual delineation and classification of all lesions is time-consuming and prone to interobserver variability. Our goal, therefore, was development and evaluation of an automated tool for MTV delineation/classification of primary tumor and lymph node metastases in PET/CT investigations of HNC patients. Automated lesion delineation was performed with a residual 3D U-Net convolutional neural network (CNN) incorporating a multi-head self-attention block. 698 [Formula: see text]F]FDG PET/CT scans from 3 different sites and 5 public databases were used for network training and testing. An external dataset of 181 [Formula: see text]F]FDG PET/CT scans from 2 additional sites was employed to assess the generalizability of the network. In these data, primary tumor and lymph node (LN) metastases were interactively delineated and labeled by two experienced physicians. Performance of the trained network models was assessed by 5-fold cross-validation in the main dataset and by pooling results from the 5 developed models in the external dataset. The Dice similarity coefficient (DSC) for individual delineation tasks and the primary tumor/metastasis classification accuracy were used as evaluation metrics. Additionally, a survival analysis using univariate Cox regression was performed comparing achieved group separation for manual and automated delineation, respectively. In the cross-validation experiment, delineation of all malignant lesions with the trained U-Net models achieves DSC of 0.885, 0.805, and 0.870 for primary tumor, LN metastases, and the union of both, respectively. In external testing, the DSC reaches 0.850, 0.724, and 0.823 for primary tumor, LN metastases, and the union of both, respectively. The voxel classification accuracy was 98.0% and 97.9% in cross-validation and external data, respectively. Univariate Cox analysis in the cross-validation and the external testing reveals that manually and automatically derived total MTVs are both highly prognostic with respect to overall survival, yielding essentially identical hazard ratios (HR) ([Formula: see text]; [Formula: see text] vs. [Formula: see text]; [Formula: see text] in cross-validation and [Formula: see text]; [Formula: see text] vs. [Formula: see text]; [Formula: see text] in external testing). To the best of our knowledge, this work presents the first CNN model for successful MTV delineation and lesion classification in HNC. In the vast majority of patients, the network performs satisfactory delineation and classification of primary tumor and lymph node metastases and only rarely requires more than minimal manual correction. It is thus able to massively facilitate study data evaluation in large patient groups and also does have clear potential for supervised clinical application.
Identifiants
pubmed: 37079128
doi: 10.1007/s00259-023-06197-1
pii: 10.1007/s00259-023-06197-1
pmc: PMC10317885
doi:
Substances chimiques
Fluorodeoxyglucose F18
0Z5B2CJX4D
Types de publication
Journal Article
Langues
eng
Sous-ensembles de citation
IM
Pagination
2751-2766Subventions
Organisme : Berliner Krebsgesellschaft
ID : ZSF201720
Organisme : Bundesministerium für Bildung und Forschung
ID : 03ZIK42
Informations de copyright
© 2023. The Author(s).
Références
Nichols AC, Theurer J, Prisman E, Read N, Berthelet E, Tran E, et al. Radiotherapy versus transoral robotic surgery and neck dissection for oropharyngeal squamous cell carcinoma (ORATOR): an open-label, phase 2, randomised trial. Lancet Oncol. 2019;20(10):1349–59.
doi: 10.1016/S1470-2045(19)30410-3
pubmed: 31416685
Wang L, Bai J, Duan P. Prognostic value of 18F-FDG PET/CT functional parameters in patients with head and neck cancer. Nucl Med Commun. 2019;40(4):361–9. https://doi.org/10.1097/mnm.0000000000000974 .
doi: 10.1097/mnm.0000000000000974
pubmed: 30664599
Zschaeck S, Li Y, Lin Q, Beck M, Amthauer H, Bauersachs L, et al. Prognostic value of baseline [18F]-fluorodeoxyglucose positron emission tomography parameters MTV, TLG and asphericity in an international multicenter cohort of nasopharyngeal carcinoma patients. PLoS ONE. 2020;15(7): e0236841. https://doi.org/10.1371/journal.pone.0236841 .
doi: 10.1371/journal.pone.0236841
pubmed: 32730364
pmcid: 7392321
Zschaeck S, Weingärtner J, Lombardo E, Marschner S, Hajiyianni M, Beck M, et al. 18F-Fluorodeoxyglucose positron emission tomography of head and neck cancer: location and HPV specific parameters for potential treatment individualization. Front Oncol. 2022;12. https://doi.org/10.3389/fonc.2022.870319 .
Marschner S, Lombardo E, Minibek L, Holzgreve A, Kaiser L, Albert N, et al. Risk stratification using 18F-FDG PET/CT and artificial neural networks in head and neck cancer patients undergoing radiotherapy. Diagnostics. 2021;11(9):1581. https://doi.org/10.3390/diagnostics11091581 .
doi: 10.3390/diagnostics11091581
pubmed: 34573924
pmcid: 8468242
Wang Y, Lombardo E, Avanzo M, Zschaek S, Weingärtner J, Holzgreve A, et al. Deep learning based time-to-event analysis with PET, CT and joint PET/CT for head and neck cancer prognosis. Comput Methods Programs Biomed. 2022;222: 106948. https://doi.org/10.1016/j.cmpb.2022.106948 .
doi: 10.1016/j.cmpb.2022.106948
pubmed: 35752119
Castelli J, Depeursinge A, Devillers A, Campillo-Gimenez B, Dicente Y, Prior JO, et al. PET-based prognostic survival model after radiotherapy for head and neck cancer. Eur J Nucl Med Mol Imaging. 2018;46(3):638–49. https://doi.org/10.1007/s00259-018-4134-9 .
doi: 10.1007/s00259-018-4134-9
pubmed: 30132054
Foster B, Bagci U, Mansoor A, Xu Z, Mollura DJ. A review on segmentation of positron emission tomography images. Comput Biol Med. 2014;50(1):76–96. https://linkinghub.elsevier.com/retrieve/pii/S0010482514001000 .
Sadaghiani MS, Rowe SP, Sheikhbahaei S. Applications of artificial intelligence in oncologic 18F-FDG PET/CT imaging: a systematic review. Ann Transl Med. 2021, 9(9), pp. 823–823. https://doi.org/10.21037/atm-20-6162
Aggarwal R, Sounderajah V, Martin G, Ting DSW, Karthikesalingam A, King D, et al. Diagnostic accuracy of deep learning in medical imaging: a systematic review and meta-analysis. npj Digit Med. 2021;4(1). https://doi.org/10.1038/s41746-021-00438-z .
Bera K, Braman N, Gupta A, Velcheti V, Madabhushi A. Predicting cancer outcomes with radiomics and artificial intelligence in radiology. Nat Rev Clin Oncol. 2021;19(2):132–46. https://doi.org/10.1038/s41571-021-00560-7 .
doi: 10.1038/s41571-021-00560-7
pubmed: 34663898
pmcid: 9034765
Chen X, Wang X, Zhang K, Fung KM, Thai TC, Moore K, et al. Recent advances and clinical applications of deep learning in medical image analysis. Med Image Anal. 2022;79: 102444. https://doi.org/10.1016/j.media.2022.102444 .
doi: 10.1016/j.media.2022.102444
pubmed: 35472844
pmcid: 9156578
Li S, Liu J, Wang Z, Cao Z, Yang Y, Wang B, et al. Application of PET/CT-based deep learning radiomics in head and neck cancer prognosis: a systematic review. Radiology Science. 2022;1(1). https://doi.org/10.15212/radsci-2022-0006 .
Ha R, Chin C, Karcich J, Liu MZ, Chang P, Mutasa S, et al. Prior to initiation of chemotherapy, can we predict breast tumor response? Deep learning convolutional neural networks approach using a breast MRI tumor dataset. J Digit Imaging. 2018;32(5):693–701. https://doi.org/10.1007/s10278-018-0144-1 .
doi: 10.1007/s10278-018-0144-1
pmcid: 6737125
Xu Y, Hosny A, Zeleznik R, Parmar C, Coroller T, Franco I, et al. Deep learning predicts lung cancer treatment response from serial medical imaging. Clin Cancer Res. 2019;25(11):3266–75. https://doi.org/10.1158/1078-0432.ccr-18-2495 .
doi: 10.1158/1078-0432.ccr-18-2495
pubmed: 31010833
pmcid: 6548658
Braman N, Adoui ME, Vulchi M, Turk P, Etesami M, Fu P, et al. Deep learning-based prediction of response to HER2-targeted neoadjuvant chemotherapy from pre-treatment dynamic breast MRI: a multi-institutional validation study. 2020.
Jiang Y, Jin C, Yu H, Wu J, Chen C, Yuan Q, et al. Development and validation of a deep learning CT signature to predict survival and chemotherapy benefit in gastric cancer. Ann Surg. 2020;274(6):e1153-61. https://doi.org/10.1097/sla.0000000000003778 .
doi: 10.1097/sla.0000000000003778
Yang J, Chen J, Kuang K, Lin T, He J, Ni B. MIA-Prognosis: a deep learning framework to predict therapy response. In: Medical Image Computing and Computer Assisted Intervention–MICCAI 2020. Springer International Publishing; 2020. p. 211–20.
Mu W, Jiang L, Zhang J, Shi Y, Gray JE, Tunali I, et al. Non-invasive decision support for NSCLC treatment using PET/CT radiomics. Nat Commun. 2020;11(1). https://doi.org/10.1038/s41467-020-19116-x .
Lao J, Chen Y, Li ZC, Li Q, Zhang J, Liu J, et al. A deep learning-based radiomics model for prediction of survival in Glioblastoma Multiforme. Sci Rep. 2017;7(1). https://doi.org/10.1038/s41598-017-10649-8 .
Hosny A, Parmar C, Coroller TP, Grossmann P, Zeleznik R, Kumar A, et al. Deep learning for lung cancer prognostication: a retrospective multi-cohort radiomics study. PLoS Med. 2018;15(11): e1002711. https://doi.org/10.1371/journal.pmed.1002711 .
doi: 10.1371/journal.pmed.1002711
pubmed: 30500819
pmcid: 6269088
Kickingereder P, Isensee F, Tursunova I, Petersen J, Neuberger U, Bonekamp D, et al. Automated quantitative tumour response assessment of MRI in neuro-oncology with artificial neural networks: a multicentre, retrospective study. Lancet Oncol. 2019;20(5):728–40. https://doi.org/10.1016/s1470-2045(19)30098-1 .
doi: 10.1016/s1470-2045(19)30098-1
pubmed: 30952559
Wang S, Liu Z, Rong Y, Zhou B, Bai Y, Wei W, et al. Deep learning provides a new computed tomography-based prognostic biomarker for recurrence prediction in high-grade serous ovarian cancer. Radiother Oncol. 2019;132:171–7. https://doi.org/10.1016/j.radonc.2018.10.019 .
doi: 10.1016/j.radonc.2018.10.019
pubmed: 30392780
Peng H, Dong D, Fang MJ, Li L, Tang LL, Chen L, et al. Prognostic value of deep learning PET/CT-Based radiomics: potential role for future individual induction chemotherapy in advanced nasopharyngeal carcinoma. Clin Cancer Res. 2019;25(14):4271–9. https://doi.org/10.1158/1078-0432.ccr-18-3065 .
doi: 10.1158/1078-0432.ccr-18-3065
pubmed: 30975664
Drukker K, Edwards A, Papaioannou J, Giger M. Deep learning predicts breast cancer recurrence in analysis of consecutive MRIs acquired during the course of neoadjuvant chemotherapy. In: Medical Imaging 2020: Computer-Aided Diagnosis, Hahn HK, Mazurowski MA, editors, vol. 11314. International Society for Optics and Photonics. SPIE. p. 1131410. https://doi.org/10.1117/12.2549044 .
Zhou T, Fu H, Zhang Y, Zhang C, Lu X, Shen J, et al. M[Formula: see text]-Net: multi-modal multi-channel network for overall survival time prediction of brain tumor patients. In: Medical Image Computing and Computer Assisted Intervention–MICCAI 2020. Springer International Publishing; 2020. p. 221–31.
Starke S, Leger S, Zwanenburg A, Leger K, Lohaus F, Linge A, et al. 2D and 3D convolutional neural networks for outcome modelling of locally advanced head and neck squamous cell carcinoma. Sci Rep. 2020;10(1). https://doi.org/10.1038/s41598-020-70542-9 .
Zhang Y, Lobo-Mueller EM, Karanicolas P, Gallinger S, Haider MA, Khalvati F. Improving prognostic performance in resectable pancreatic ductal adenocarcinoma using radiomics and deep learning features fusion in CT images. Sci Rep. 2021;11(1). https://doi.org/10.1038/s41598-021-80998-y .
Sokooti H, de Vos B, Berendsen F, Lelieveldt BPF, Išgum I, Staring M. Nonrigid image registration using multi-scale 3D convolutional neural networks. In: Medical Image Computing and Computer Assisted Intervention - MICCAI 2017. Springer International Publishing. 2017;232–239.
Balakrishnan G, Zhao A, Sabuncu MR, Dalca AV, Guttag J. An unsupervised learning model for deformable medical image registration. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. IEEE. p. 9252–60. https://doi.org/10.1109/cvpr.2018.00964 .
de Vos BD, Berendsen FF, Viergever MA, Sokooti H, Staring M, Išgum I. A deep learning framework for unsupervised affine and deformable image registration. Med Image Anal. 2019;52:128–43. https://doi.org/10.1016/j.media.2018.11.010 .
doi: 10.1016/j.media.2018.11.010
pubmed: 30579222
Wang J, Zhang M. DeepFLASH: an efficient network for learning-based medical image registration. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE. p. 4443–51. https://doi.org/10.1109/cvpr42600.2020.00450 .
Fu Y, Lei Y, Wang T, Curran WJ, Liu T, Yang X. Deep learning in medical image registration: a review. Phys Med Biol. 2020;65(20):20TR01. https://doi.org/10.1088/1361-6560/ab843e .
Haskins G, Kruger U, Yan P. Deep learning in medical image registration: a survey. Mach Vis Appl. 2020;31(1-2). https://doi.org/10.1007/s00138-020-01060-x .
Kumar A, Kim J, Lyndon D, Fulham M, Feng D. an ensemble of fine-tuned convolutional neural networks for medical image classification. IEEE J Biomed Health Inform. 2017;21(1):31–40. https://doi.org/10.1109/jbhi.2016.2635663 .
doi: 10.1109/jbhi.2016.2635663
pubmed: 28114041
Wang SH, Phillips P, Sui Y, Liu B, Yang M, Cheng H. Classification of Alzheimer’s Disease based on eight-layer convolutional neural network with leaky rectified linear unit and max pooling. J Med Syst. 2018;42(5). https://doi.org/10.1007/s10916-018-0932-7 .
Frid-Adar M, Diamant I, Klang E, Amitai M, Goldberger J, Greenspan H. GAN-based synthetic medical image augmentation for increased CNN performance in liver lesion classification. Neurocomputing. 2018;321:321–31. https://doi.org/10.1016/j.neucom.2018.09.013 .
doi: 10.1016/j.neucom.2018.09.013
Schlemper J, Oktay O, Schaap M, Heinrich M, Kainz B, Glocker B, et al. Attention gated networks: learning to leverage salient regions in medical images. Med Image Anal. 2019;53:197–207. https://doi.org/10.1016/j.media.2019.01.012 .
doi: 10.1016/j.media.2019.01.012
pubmed: 30802813
pmcid: 7610718
Xie Y, Zhang J, Xia Y. Semi-supervised adversarial model for benign-malignant lung nodule classification on chest CT. Med Image Anal. 2019;57:237–48. https://doi.org/10.1016/j.media.2019.07.004 .
doi: 10.1016/j.media.2019.07.004
pubmed: 31352126
Sharma H, Jain JS, Bansal P, Gupta S. Feature extraction and classification of Chest X-Ray images using CNN to detect pneumonia. In: 2020 10th International Conference on Cloud Computing, Data Science & Engineering (Confluence). IEEE. p. 227–31. https://doi.org/10.1109/confluence47617.2020.9057809 .
Abbas A, Abdelsamea MM, Gaber MM. Classification of COVID-19 in chest X-ray images using DeTraC deep convolutional neural network. Appl Intell. 2020;51(2):854–64. https://doi.org/10.1007/s10489-020-01829-7 .
doi: 10.1007/s10489-020-01829-7
Li X, Jia M, Islam MT, Yu L, Xing L. Self-supervised feature learning via exploiting multi-modal data for retinal disease diagnosis. IEEE Trans Med Imaging. 2020;39(12):4023–33. https://doi.org/10.1109/tmi.2020.3008871 .
doi: 10.1109/tmi.2020.3008871
pubmed: 32746140
Guan Q, Huang Y, Zhong Z, Zheng Z, Zheng L, Yang Y. Thorax disease classification with attention guided convolutional neural network. Pattern Recogn Lett. 2020;131:38–45. https://doi.org/10.1016/j.patrec.2019.11.040 .
doi: 10.1016/j.patrec.2019.11.040
Shorfuzzaman M, Hossain MS. MetaCOVID: A Siamese neural network framework with contrastive loss for n-shot diagnosis of COVID-19 patients. Pattern Recogn. 2021;113: 107700. https://doi.org/10.1016/j.patcog.2020.107700 .
doi: 10.1016/j.patcog.2020.107700
Azizi S, Mustafa B, Ryan F, Beaver Z, Freyberg J, Deaton J, et al. Big self-supervised models advance medical image classification. In: 2021 IEEE/CVF International Conference on Computer Vision (ICCV). IEEE. p, 3458–68. https://doi.org/10.1109/iccv48922.2021.00346 .
Chen L, Dohopolski M, Zhou Z, Wang K, Wang R, Sher D, et al. Attention guided lymph node malignancy prediction in head and neck cancer. Int J Radiat Oncol Biol Phys. 2021;110(4):1171–9. https://doi.org/10.1016/j.ijrobp.2021.02.004 .
doi: 10.1016/j.ijrobp.2021.02.004
pubmed: 33561508
Zheng Y, Liu D, Georgescu B, Nguyen H, Comaniciu D. 3D deep learning for efficient and robust landmark detection in volumetric data. In: Lecture Notes in Computer Science. Springer International Publishing; 2015. p. 565–72.
Ding J, Li A, Hu Z, Wang L. Accurate pulmonary nodule detection in computed tomography images using deep convolutional neural networks. In: Medical Image Computing and Computer Assisted Intervention - MICCAI 2017. Springer International Publishing; 2017. p. 559–67.
Yan K, Tang Y, Peng Y, Sandfort V, Bagheri M, Lu Z, et al. MULAN: multitask universal lesion analysis network for joint lesion detection, tagging, and segmentation. In: Lecture Notes in Computer Science. Springer International Publishing; 2019. p. 194–202.
Nair T, Precup D, Arnold DL, Arbel T. Exploring uncertainty measures in deep networks for Multiple sclerosis lesion detection and segmentation. Med Image Anal. 2020;59: 101557. https://doi.org/10.1016/j.media.2019.101557 .
doi: 10.1016/j.media.2019.101557
pubmed: 31677438
Wang D, Zhang Y, Zhang K, Wang L. FocalMix: Semi-supervised learning for 3D medical image detection. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE. p. 3950–9. https://doi.org/10.1109/cvpr42600.2020.00401 .
Liu Y, Zhang F, Zhang Q, Wang S, Wang Y, Yu Y. Cross-view correspondence reasoning based on bipartite graph convolutional network for mammogram mass detection. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE. p. 3811–21. https://doi.org/10.1109/cvpr42600.2020.00387 .
Mei J, Cheng MM, Xu G, Wan LR, Zhang H. SANet: A slice-aware network for pulmonary nodule detection. IEEE Trans Pattern Anal Mach Intell. 2021;1–1. https://doi.org/10.1109/tpami.2021.3065086 .
Baur C, Denner S, Wiestler B, Navab N, Albarqouni S. Autoencoders for unsupervised anomaly segmentation in brain MR images: a comparative study. Med Image Anal. 2021;69: 101952. https://doi.org/10.1016/j.media.2020.101952 .
doi: 10.1016/j.media.2020.101952
pubmed: 33454602
Ronneberger O, Fischer P, Brox T. U-Net: Convolutional networks for biomedical image segmentation. 2015.
Milletari F, Navab N, Ahmadi SA. V-Net: fully convolutional neural networks for volumetric medical image segmentation. In: 2016 Fourth International Conference on 3D Vision (3DV). IEEE. p. 565–71. https://doi.org/10.1109/3dv.2016.79 .
Zhou Z, Siddiquee MMR, Tajbakhsh N, Liang J. UNet++: a nested U-net architecture for medical image segmentation. In: Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support. Springer International Publishing; 2018. p. 3–11.
Gibson E, Giganti F, Hu Y, Bonmati E, Bandula S, Gurusamy K, et al. Automatic multi-organ segmentation on abdominal CT with dense V-networks. IEEE Trans Med Imaging. 2018;37(8):1822–34. https://doi.org/10.1109/tmi.2018.2806309 .
doi: 10.1109/tmi.2018.2806309
pubmed: 29994628
pmcid: 6076994
Xue Y, Xu T, Zhang H, Long LR, Huang X. SegAN: adversarial network with multi-scale L1 loss for medical image segmentation. Neuroinformatics. 2018;16(3–4):383–92. https://doi.org/10.1007/s12021-018-9377-x .
doi: 10.1007/s12021-018-9377-x
pubmed: 29725916
Yu L, Wang S, Li X, Fu CW, Heng PA. Uncertainty-aware self-ensembling model for semi-supervised 3D left atrium segmentation. In: Lecture Notes in Computer Science. Springer International Publishing; 2019. p. 605–13.
Bai W, Chen C, Tarroni G, Duan J, Guitton F, Petersen SE, et al. Self-supervised learning for cardiac MR image segmentation by anatomical position prediction. In: Lecture Notes in Computer Science. Springer International Publishing; 2019. p. 541–9.
Baumgartner CF, Tezcan KC, Chaitanya K, Hötker AM, Muehlematter UJ, Schawkat K, et al. PHiSeg: capturing uncertainty in medical image segmentation. In: Lecture Notes in Computer Science. Springer International Publishing; 2019. p. 119–27.
Wang X, Han S, Chen Y, Gao D, Vasconcelos N. Volumetric attention for 3D medical image segmentation and detection. In: Lecture Notes in Computer Science. Springer International Publishing; 2019. p. 175–84.
Fan DP, Zhou T, Ji GP, Zhou Y, Chen G, Fu H, et al. Inf-Net: Automatic COVID-19 lung infection segmentation from CT images. IEEE Trans Med Imaging. 2020;39(8):2626–37. https://doi.org/10.1109/tmi.2020.2996645 .
doi: 10.1109/tmi.2020.2996645
pubmed: 32730213
Chen J, Lu Y, Yu Q, Luo X, Adeli E, Wang Y, et al. TransUNet: transformers make strong encoders for medical image segmentation. 2021.
Zuo Q, Chen S, Wang Z. R2AU-Net: attention recurrent residual convolutional neural network for multimodal medical image segmentation. Secur Commun Netw. 2021;2021:1–10. https://doi.org/10.1155/2021/6625688 .
doi: 10.1155/2021/6625688
Samarasinghe G, Jameson M, Vinod S, Field M, Dowling J, Sowmya A, et al. Deep learning for segmentation in radiation therapy planning: a review. J Med Imaging Radiat Oncol. 2021;65(5):578–95.
doi: 10.1111/1754-9485.13286
pubmed: 34313006
Hatamizadeh A, Tang Y, Nath V, Yang D, Myronenko A, Landman B, et al. UNETR: Transformers for 3D medical image segmentation. In: 2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV). IEEE. p. 1748–58. https://doi.org/10.1109/wacv51458.2022.00181 .
Li S, Xiao J, He L, Peng X, Yuan X. The tumor target segmentation of Nasopharyngeal Cancer in CT images based on deep learning methods. Technol Cancer Res Treat. 2019;18:153303381988456. http://journals.sagepub.com/doi/10.1177/1533033819884561 .
Guo Z, Guo N, Gong K, Zhong S, Li Q. Gross tumor volume segmentation for head and neck cancer radiotherapy using deep dense multi-modality network. Phys Med Biol. 2019;64(20):205015. https://iopscience.iop.org/article/10.1088/1361-6560/ab440d .
Moe YM, Groendahl AR, Mulstad M, Tomic O, Indahl U, Dale E, et al. Deep learning for automatic tumour segmentation in PET/CT images of patients with head and neck cancers. 2019.
Andrearczyk V, Oreiller V, Vallières M, Castelli J, Elhalawani H, Jreige M, et al. Automatic segmentation of head and neck tumors and nodal metastases in PET-CT scans. In: Proceedings of Machine Learning Research. 2020. p. 33–43 https://tinyurl.com/yc5sq5jy .
Groendahl AR, Knudtsen IS, Huynh BN, Mulstad M, Moe YM, Knuth F, et al. A comparison of methods for fully automatic segmentation of tumors and involved nodes in PET/CT of head and neck cancers. Phys Med Biol. 2021;66(6): 065012. https://doi.org/10.1088/1361-6560/abe553 .
doi: 10.1088/1361-6560/abe553
pubmed: 33666176
Moe YM, Groendahl AR, Tomic O, Dale E, Malinen E, Futsaether CM. Deep learning-based auto-delineation of gross tumour volumes and involved nodes in PET/CT images of head and neck cancer patients. Eur J Nucl Med Mol Imaging. 2021;48(9):2782–92. https://doi.org/10.1007/s00259-020-05125-x .
doi: 10.1007/s00259-020-05125-x
pubmed: 33559711
pmcid: 8263429
Zhao B, Soraghan J, Caterina GD, Grose D. Segmentation of Head and Neck Tumours Using Modified U-net. In: 2019 27th European Signal Processing Conference (EUSIPCO). IEEE; p. 1–4. https://doi.org/10.23919/eusipco.2019.8902637 .
Bielak L, Wiedenmann N, Berlin A, Nicolay NH, Gunashekar DD, Hägele L, et al. Convolutional neural networks for head and neck tumor segmentation on 7-channel multiparametric MRI: a leave-one-out analysis. Radiat Oncol. 2020;15(1). https://doi.org/10.1186/s13014-020-01618-z .
Outeiral RR, Bos P, Al-Mamgani A, Jasperse B, Simões R, van der Heide UA. Oropharyngeal primary tumor segmentation for radiotherapy planning on magnetic resonance imaging using deep learning. Phys Imaging Radiat Oncol. 2021;19:39–44. https://doi.org/10.1016/j.phro.2021.06.005 .
doi: 10.1016/j.phro.2021.06.005
Badrigilan S, Nabavi S, Abin AA, Rostampour N, Abedi I, Shirvani A, et al. Deep learning approaches for automated classification and segmentation of head and neck cancers and brain tumors in magnetic resonance images: a meta-analysis study. Int J Comput Assist Radiol Surg. 2021;16(4):529–42. https://doi.org/10.1007/s11548-021-02326-z .
doi: 10.1007/s11548-021-02326-z
pubmed: 33666859
Schouten JP, Noteboom S, Martens RM, Mes SW, Leemans CR, de Graaf P, et al. Automatic segmentation of head and neck primary tumors on MRI using a multi-view CNN. Cancer Imaging. 2022;22(1). https://doi.org/10.1186/s40644-022-00445-7 .
Ren J, Eriksen JG, Nijkamp J, Korreman SS. Comparing different CT, PET and MRI multi-modality image combinations for deep learning-based head and neck tumor segmentation. Acta Oncol. 2021;60(11):1399–406. https://doi.org/10.1080/0284186x.2021.1949034 .
doi: 10.1080/0284186x.2021.1949034
pubmed: 34264157
Guo Z, Guo N, Gong K, Li Q. Automatic multi-modality segmentation of gross tumor volume for head and neck cancer radiotherapy using 3D U-Net. In: Mori K, Hahn HK, editors. Medical Imaging 2019: Computer-Aided Diagnosis, vol. 10950. International Society for Optics and Photonics. SPIE; p. 1095009. https://doi.org/10.1117/12.2513229 .
Andrearczyk V, Fontaine P, Oreiller V, Castelli J, Jreige M, Prior JO, et al. Multi-task deep segmentation and radiomics for automatic prognosis in head and neck cancer. In: Rekik I, Adeli E, Park SH, Schnabel J, editors. Predictive Intelligence in Medicine. Springer International Publishing, Cham; p. 147–156. ISBN 978-3-030-87602-9.
Sobirov I, Nazarov O, Alasmawi H, Yaqub M. Automatic segmentation of head and neck tumor: how powerful transformers are? 2022.
Andrearczyk V, Oreiller V, Boughdad S, Rest CCL, Elhalawani H, Jreige M, et al. Overview of the HECKTOR Challenge at MICCAI 2021:automatic head and neck tumor segmentation and outcome prediction in PET/CT Images. In: Lecture Notes in Computer Science. Springer International Publishing. 2022. p. 1–37.
Xie J, Peng Y. The Head and Neck Tumor Segmentation Based on 3D U-Net. In: Lecture Notes in Computer Science. Springer International Publishing. 2022. p. 92–98.
Ha SC, Oh JS, Roh JL, Moon H, Kim JS, Cho KJ, et al. Pretreatment tumor SUV max predicts disease-specific and overall survival in patients with head and neck soft tissue sarcoma. Eur J Nucl Med Mol Imaging. 2017;44:33–40.
doi: 10.1007/s00259-016-3456-8
pubmed: 27448574
Shiri I, Arabi H, Sanaat A, Jenabi E, Becker M, Zaidi H. Fully automated gross tumor volume delineation from pet in head and neck cancer using deep learning algorithms. Clin Nucl Med. 2021;46(11):872–83. https://doi.org/10.1097/rlu.0000000000003789 .
doi: 10.1097/rlu.0000000000003789
pubmed: 34238799
Vallières M, Kay-Rivest E, Perrin L, Liem X, Furstoss C, Khaouam N, et al. Data from Head-Neck-PET-CT. 2017.
Wee L, Dekker A. Data from Head-Neck-Radiomics-HN1. 2019.
Grossberg AJ, Mohamed ASR, Elhalawani H, Bennett WC, Smith KE, Nolan TS, et al. Imaging and clinical data archive for head and neck squamous cell carcinoma patients treated with radiotherapy. Sci Data. 2018;5(1). https://doi.org/10.1038/sdata.2018.173 .
Zuley ML, Jarosz R, Kirk S, Lee Y, Colen R, Garcia K, et al. Radiology data from The Cancer Genome Atlas Head-Neck Squamous Cell Carcinoma [TCGA-HNSC] collection. 2016.
Beichel RR, Ulrich EJ, Bauer C, Wahle A, Brown B, Chang T, et al. Data from QIN-HEADNECK. 2015.
Hofheinz F, Pötzsch C, Oehme L, Beuthien-Baumann B, Steinbach J, Kotzerke J, et al. Automatic volume delineation in oncological PET. Evaluation of a dedicated software tool and comparison with manual delineation in clinical data sets. Nuklearmedizin. 2012;51:9–16.
Hofheinz F, Langner J, Petr J, Beuthien-Baumann B, Steinbach J, Kotzerke J, et al. An automatic method for accurate volume delineation of heterogeneous tumors in PET. Med Phys. 2013;40(8): 082503.
doi: 10.1118/1.4812892
pubmed: 23927348
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need. In: Guyon I, Luxburg UV, Bengio S, Wallach H, Fergus R, Vishwanathan S, et al., editors. Advances in Neural Information Processing Systems, vol. 30. Curran Associates, Inc.; p. 5999–6009. https://proceedings.neurips.cc/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf .
R Core Team. R: A Language and Environment for Statistical Computing. R Foundation for statistical computing, Vienna, Austria. 2022.
Hofheinz F, Langner J, Petr J, Beuthien-Baumann B, Steinbach J, Kotzerke J, et al. An automatic method for accurate volume delineation of heterogeneous tumors in PET. Med Phys. 2013;40(8): 082503.
doi: 10.1118/1.4812892
pubmed: 23927348
Han S, Oh JS, Kim Yi, Seo SY, Lee GD, Park MJ, et al. Fully automatic quantitative measurement of 18F-FDG PET/CT in Thymic Epithelial tumors using a convolutional neural network. Clin Nucl Med. 2022;47(7):590–598.
Gudi S, Ghosh-Laskar S, Agarwal JP, Chaudhari S, Rangarajan V, Paul SN, et al. Interobserver Variability in the delineation of gross tumour volume and specified organs-at-risk during IMRT for head and neck cancers and the impact of FDG-PET/CT on such variability at the primary site. J Med Imaging Radiat Sci. 2017;48(2):184–92. https://doi.org/10.1016/j.jmir.2016.11.003 .
doi: 10.1016/j.jmir.2016.11.003
pubmed: 31047367
Miyabe J, Hanamoto A, Tatsumi M, Hamasaki T, Takenaka Y, Nakahara S, et al. Metabolic tumor volume of primary tumor predicts survival better than T classification in the larynx preservation approach. Cancer Sci. 2017;108(10):2030–8. https://doi.org/10.1111/cas.13345 .
doi: 10.1111/cas.13345
pubmed: 28787757
pmcid: 5623730
Lin L, Dou Q, Jin YM, Zhou GQ, Tang YQ, Chen WL, et al. Deep learning for automated contouring of primary tumor volumes by MRI for Nasopharyngeal Carcinoma. Radiology. 2019;291(3):677–86. https://doi.org/10.1148/radiol.2019182012 .
doi: 10.1148/radiol.2019182012
pubmed: 30912722
Cho JK, Hyun SH, Choi N, Kim MJ, Padera TP, Choi JY, et al. Significance of lymph node metastasis in cancer dissemination of head and neck cancer. Translational Oncology. 2015;8(2):119–25. https://doi.org/10.1016/j.tranon.2015.03.001 .
doi: 10.1016/j.tranon.2015.03.001
pubmed: 25926078
pmcid: 4415144
Huang Y, Feng M, He Q, Yin J, Xu P, Jiang Q, et al. Prognostic value of pretreatment 18F-FDG PET-CT for nasopharyngeal carcinoma patients. Medicine. 2017;96(17): e6721. https://doi.org/10.1097/md.0000000000006721 .
doi: 10.1097/md.0000000000006721
pubmed: 28445287
pmcid: 5413252
Ahmed KB, Goldgof GM, Paul R, Goldgof DB, Hall LO. Discovery of a generalization gap of convolutional neural networks on COVID-19 X-Rays Classification. IEEE Access. 2021;9:72970–9. https://doi.org/10.1109/access.2021.3079716 .
doi: 10.1109/access.2021.3079716
pubmed: 34178559
Aide N, Lasnon C, Veit-Haibach P, Sera T, Sattler B, Boellaard R. EANM/EARL harmonization strategies in PET quantification: from daily practice to multicentre oncological studies. Eur J Nucl Med Mol Imaging. 2017;44(S1):17–31. https://doi.org/10.1007/s00259-017-3740-2 .
doi: 10.1007/s00259-017-3740-2
pubmed: 28623376
pmcid: 5541084
Mali SA, Ibrahim A, Woodruff HC, Andrearczyk V, Müller H, Primakov S, et al. Making radiomics more reproducible across scanner and imaging protocol variations: a review of harmonization methods. J Perinat Med. 2021;11(9):842. https://doi.org/10.3390/jpm11090842 .
doi: 10.3390/jpm11090842