Back to Search Start Over

Portrait Semantic Segmentation Method Based on Dual Modal Information Complementarity

Authors :
Guang Feng
Chong Tang
Source :
Applied Sciences, Vol 14, Iss 4, p 1439 (2024)
Publication Year :
2024
Publisher :
MDPI AG, 2024.

Abstract

Semantic segmentation of human images is a research hotspot in the field of computer vision. At present, the semantic segmentation models based on U-net generally lack the ability to capture the spatial information of images. At the same time, semantic incompatibility exists because the feature maps of encoder and decoder are directly connected in the skip connection stage. In addition, in low light scenes such as at night, it is easy for false segmentation and segmentation accuracy to appear. To solve the above problems, a portrait semantic segmentation method based on dual-modal information complementarity is proposed. The encoder adopts a double branch structure, and uses a SK-ASSP module that can adaptively adjust the convolution weights of different receptor fields to extract features in RGB and gray image modes respectively, and carries out cross-modal information complementarity and feature fusion. A hybrid attention mechanism is used in the jump connection phase to capture both the channel and coordinate context information of the image. Experiments on human matting dataset show that the PA and MIoU coefficients of this algorithm model reach 96.58% and 94.48% respectively, which is better than U-net benchmark model and other mainstream semantic segmentation models.

Details

Language :
English
ISSN :
20763417
Volume :
14
Issue :
4
Database :
Directory of Open Access Journals
Journal :
Applied Sciences
Publication Type :
Academic Journal
Accession number :
edsdoj.3381c280fe9a471883c641fc3175f823
Document Type :
article
Full Text :
https://doi.org/10.3390/app14041439