The use of remote sensing images for land cover analysis has broad prospects. At present, the resolution of aerial remote sensing images is getting higher and higher, and the span of time and space is getting larger and larger, therefore segmenting target objects enconter great difficulties. Convolutional neural networks are widely used in many image semantic segmentation tasks, but existing models often use simple accumulation of various convolutional layers or the direct stacking of interfeature reuse of up- and downsampling, the network very heavy. To improve the accuracy of land cover segmentation, we propose a multichannel feature fusion lozenge network. The multichannel feature fusion lozenge network (MLNet) is a three-sided network composed of three branches: one branch uses different levels of feature indexes to sample to maintain the integrity of high-frequency information; one branch focuses on contextual information and strengthens the compatibility of information within and between classes; and the last branch uses feature integration to filter redundant information based on multiresolution segmentation to extract key features. Compared with FCN, UNet, PSP, and other serial single road computing models, the MLNet, which performs feature fusion after three-way parallelism structure, can significantly improve the accuracy with only small increase in complexity. Experimental results show that the average accuracy of 85.30% is obtained on the land cover data set, which is much higher than that of 82.98% of FCN, 81.87% of UNet, 77.52% of SegNet, and 83.09% of EspNet, which proves the effectiveness of the model. |
ACCESS THE FULL ARTICLE
No SPIE Account? Create one
CITATIONS
Cited by 47 scholarly publications.
Image segmentation
Convolution
Buildings
Remote sensing
Clouds
Image fusion
Data modeling