[Evaluation of Radiograph Accuracy in Skull X-ray Images Using Deep Learning].

X-ray image artificial intelligence (AI) classification deep convolutional neural network (DCNN) radiograph accuracy

Journal

Nihon Hoshasen Gijutsu Gakkai zasshi
ISSN: 1881-4883
Titre abrégé: Nihon Hoshasen Gijutsu Gakkai Zasshi
Pays: Japan
ID NLM: 7505722

Informations de publication

Date de publication:
2022
Historique:
entrez: 20 1 2022
pubmed: 21 1 2022
medline: 22 1 2022
Statut: ppublish

Résumé

Accurate positioning is essential for radiography, and it is especially important to maintain image reproducibility in follow-up observations. The decision on re-taking radiographs is entrusting to the individual radiological technologist. The evaluation is a visual and qualitative evaluation and there are individual variations in the acceptance criteria. In this study, we propose a method of image evaluation using a deep convolutional neural network (DCNN) for skull X-ray images. The radiographs were obtained from 5 skull phantoms and were classified by simple network and VGG16. The discrimination ability of DCNN was verified by recognizing the X-ray projection angle and the retake of the radiograph. DCNN architectures were used with the different input image sizes and were evaluated by 5-fold cross-validation and leave-one-out cross-validation. Using the 5-fold cross-validation, the classification accuracy was 99.75% for the simple network and 80.00% for the VGG16 in small input image sizes, and when the input image size was general image size, simple network and VGG16 showed 79.58% and 80.00%, respectively. The experimental results showed that the combination between the small input image size, and the shallow DCNN architecture was suitable for the four-category classification in X-ray projection angles. The classification accuracy was up to 99.75%. The proposed method has the potential to automatically recognize the slight projection angles and the re-taking images to the acceptance criteria. It is considered that our proposed method can contribute to feedback for re-taking images and to reduce radiation dose due to individual subjectivity.

Identifiants

pubmed: 35046219
doi: 10.6009/jjrt.780104
doi:

Types de publication

Journal Article

Langues

jpn

Sous-ensembles de citation

IM

Pagination

23-32

Auteurs

Hideyoshi Mitsutake (H)

Department of Radiological Technology, Teikyo University Hospital.

Haruyuki Watanabe (H)

School of Radiological Technology, Gunma Prefectural College of Health Sciences.

Aya Sakaguchi (A)

School of Radiological Technology, Gunma Prefectural College of Health Sciences (Current address: Department of Radiological Technology, Seikei-kai Chiba Medical Center).

Kiyoshi Uchiyama (K)

Department of Radiological Technology, Teikyo University Hospital.

Yongbum Lee (Y)

School of Health Sciences, Faculty of Medicine, Niigata University.

Norio Hayashi (N)

School of Radiological Technology, Gunma Prefectural College of Health Sciences.

Masayuki Shimosegawa (M)

School of Radiological Technology, Gunma Prefectural College of Health Sciences.

Toshihiro Ogura (T)

School of Radiological Technology, Gunma Prefectural College of Health Sciences.

Articles similaires

Databases, Protein Protein Domains Protein Folding Proteins Deep Learning

Failed radial head arthroplasty treated by removal of the implant.

Juan Ameztoy Gallego, Blanca Diez Sanchez, Afonso Vaquero-Picado et al.
1.00
Humans Male Female Middle Aged Range of Motion, Articular
Humans Middle Aged Female Male Surveys and Questionnaires

Classifications MeSH