Application of Improved CycleGAN in Laser-Visible Face Image Translation.

CycleGAN RRDB module identity loss least squares method

Journal

Sensors (Basel, Switzerland)
ISSN: 1424-8220
Titre abrégé: Sensors (Basel)
Pays: Switzerland
ID NLM: 101204366

Informations de publication

Date de publication:
27 May 2022
Historique:
received: 12 04 2022
revised: 23 05 2022
accepted: 25 05 2022
entrez: 10 6 2022
pubmed: 11 6 2022
medline: 14 6 2022
Statut: epublish

Résumé

CycleGAN is widely used in various image translations, such as thermal-infrared-visible-image translation, near-infrared-visible-image translation, and shortwave-infrared-visible-image translation. However, most image translations are used for infrared-to-visible translation, and the wide application of laser imaging has an increasingly strong demand for laser-visible image translation. In addition, the current image translation is mainly aimed at frontal face images, which cannot be effectively utilized to translate faces at a certain angle. In this paper, we construct a laser-visible face mapping dataset; in case of the gradient dispersion of the objective function of the original adversarial loss, the least squares loss function is used to replace the cross-entropy loss function and an identity loss function is added to strengthen the network constraints on the generator. The experimental results indicate that the SSIM value of the improved model increases by 1.25%, 8%, 0, 8%, the PSNR value is not much different, and the FID value decreases by 11.22, 12.85, 43.37 and 72.19, respectively, compared with the CycleGAN, Pix2pix, U-GAN-IT and StarGAN models. In the profile image translation, in view of the poor extraction effect of CycleGAN's original feature extraction module ResNet, the RRDB module is used to replace it based on the first improvement. The experimental results show that, compared with the CycleGAN, Pix2pix, U-GAN-IT, StarGAN and the first improved model, the SSIM value of the improved model increased by 3.75%, 10.67%, 2.47%, 10.67% and 2.47%, respectively; the PSNR value increased by 1.02, 2.74, 0.32, 0.66 and 0.02, respectively; the FID value reduced by 26.32, 27.95, 58.47, 87.29 and 15.1, respectively. Subjectively, the contour features and facial features were better conserved.

Identifiants

pubmed: 35684676
pii: s22114057
doi: 10.3390/s22114057
pmc: PMC9185648
pii:
doi:

Types de publication

Journal Article

Langues

eng

Sous-ensembles de citation

IM

Subventions

Organisme : Key Basic Research Projects of the Basic Strengthening Pro-gram
ID : 2020-JCJQ-ZD-071

Références

IEEE Trans Image Process. 2004 Apr;13(4):600-12
pubmed: 15376593
Sensors (Basel). 2022 Feb 23;22(5):
pubmed: 35270898
Sensors (Basel). 2022 Mar 09;22(6):
pubmed: 35336291

Auteurs

Mingyu Qin (M)

Graduate School, Department of Electronic and Optical Engineering, Space Engineering University, Beijing 101416, China.

Youchen Fan (Y)

School of Space Information, Space Engineering University, Beijing 101416, China.

Huichao Guo (H)

Department of Electronic and Optical Engineering, Space Engineering University, Beijing 101416, China.

Mingqian Wang (M)

Graduate School, Department of Electronic and Optical Engineering, Space Engineering University, Beijing 101416, China.

Articles similaires

1.00
Humans Magnetic Resonance Imaging Brain Infant, Newborn Infant, Premature
Cephalometry Humans Anatomic Landmarks Software Internet
Humans Artificial Intelligence Neoplasms Prognosis Image Processing, Computer-Assisted
Humans Breast Neoplasms Female Deep Learning Ultrasonography, Mammary

Classifications MeSH