LOANet: a lightweight network using object attention for extracting buildings and roads from UAV aerial remote sensing images.
Context features
Lightweight network
Object attention
Remote sensing image
Semantic segmentation
Journal
PeerJ. Computer science
ISSN: 2376-5992
Titre abrégé: PeerJ Comput Sci
Pays: United States
ID NLM: 101660598
Informations de publication
Date de publication:
2023
2023
Historique:
received:
11
04
2023
accepted:
08
06
2023
medline:
7
8
2023
pubmed:
7
8
2023
entrez:
7
8
2023
Statut:
epublish
Résumé
Semantic segmentation for extracting buildings and roads from uncrewed aerial vehicle (UAV) remote sensing images by deep learning becomes a more efficient and convenient method than traditional manual segmentation in surveying and mapping fields. In order to make the model lightweight and improve the model accuracy, a lightweight network using object attention (LOANet) for buildings and roads from UAV aerial remote sensing images is proposed. The proposed network adopts an encoder-decoder architecture in which a lightweight densely connected network (LDCNet) is developed as the encoder. In the decoder part, the dual multi-scale context modules which consist of the atrous spatial pyramid pooling module (ASPP) and the object attention module (OAM) are designed to capture more context information from feature maps of UAV remote sensing images. Between ASPP and OAM, a feature pyramid network (FPN) module is used to fuse multi-scale features extracted from ASPP. A private dataset of remote sensing images taken by UAV which contains 2431 training sets, 945 validation sets, and 475 test sets is constructed. The proposed basic model performs well on this dataset, with only 1.4M parameters and 5.48G floating point operations (FLOPs), achieving excellent mean Intersection-over-Union (mIoU). Further experiments on the publicly available LoveDA and CITY-OSM datasets have been conducted to further validate the effectiveness of the proposed basic and large model, and outstanding mIoU results have been achieved. All codes are available on https://github.com/GtLinyer/LOANet.
Identifiants
pubmed: 37547422
doi: 10.7717/peerj-cs.1467
pii: cs-1467
pmc: PMC10403170
doi:
Types de publication
Journal Article
Langues
eng
Pagination
e1467Informations de copyright
©2023 Han et al.
Déclaration de conflit d'intérêts
The authors declare there are no competing interests.
Références
IEEE Trans Pattern Anal Mach Intell. 2017 Apr;39(4):640-651
pubmed: 27244717
Sensors (Basel). 2019 Sep 23;19(19):
pubmed: 31547609
Sensors (Basel). 2020 Jan 10;20(2):
pubmed: 31936791
IEEE Trans Pattern Anal Mach Intell. 2017 Dec;39(12):2481-2495
pubmed: 28060704
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):834-848
pubmed: 28463186