Check patentability & draft patents in minutes with Patsnap Eureka AI!

Multi-modal portrait segmentation method based on separation guide convolution

A multi-modal, portrait technology, applied in image analysis, neural learning methods, image enhancement and other directions, can solve the problems of difficult to obtain high-resolution deep feature images, low segmentation accuracy, low computational efficiency, etc., to improve detection accuracy. The effect of improving the expression ability and speeding up the calculation speed

Pending Publication Date: 2022-05-13
YANGZHOU UNIV
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The network model of the U-shaped network architecture is relatively complex, resulting in low computational efficiency
Although there are other solutions to solve the problem of low computational efficiency, the system cannot perform well due to the low resolution of the deep feature image
[0004] The traditional portrait segmentation algorithm needs to manually select the character area to assist it to achieve portrait segmentation. This algorithm is not only inefficient, but also has low segmentation accuracy in complex scenes.
The semantic segmentation method based on deep learning can solve the accuracy of portrait segmentation in complex scenes, but the existing network models are generally complex, the calculation time is long, and it is not easy to obtain high-resolution deep feature images

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-modal portrait segmentation method based on separation guide convolution
  • Multi-modal portrait segmentation method based on separation guide convolution
  • Multi-modal portrait segmentation method based on separation guide convolution

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0074] refer to figure 1 and figure 2 , which is the first embodiment of the present invention, this embodiment provides a multimodal portrait segmentation method based on separate guided convolution, the multimodal portrait segmentation method based on separate guided convolution includes the following steps:

[0075] S1: Through the decoder, the features output by the encoder are input into the separation-guided convolution for multi-scale learning, and the predicted probability map of the portrait is output to build a portrait segmentation model;

[0076] S2: Input the image of the person to be detected and its depth image to the constructed network model for model training, add depth supervision to the predicted image output from each side, and use the cross-entropy loss to calculate the manually labeled image and the predicted image segmentation image The difference, these errors are fed back to the network to update the model parameters of the entire network;

[0077]...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a multi-modal portrait segmentation method based on separation guide convolution, which comprises the following steps of: inputting features output by an encoder into separation guide convolution through a decoder to carry out multi-scale learning, outputting a prediction probability graph of a portrait, and constructing a portrait segmentation model; inputting a to-be-detected figure image and a depth image thereof into the constructed network model for model training, adding depth supervision to a prediction image output from each side, calculating the difference between a manual annotation image and a predicted portrait segmentation image by using cross entropy loss, and feeding the errors back to the network to update the model parameters of the whole network; and inputting the original image needing to be tested and the depth image into the model for model testing. According to the method, the scale of the side output feature image is reduced, and the multi-scale feature expression capability is improved, so that the reasoning calculation speed is increased, and the detection accuracy in a complex scene is improved.

Description

technical field [0001] The invention relates to the technical field of computer vision and digital image processing, in particular to a multimodal portrait segmentation method based on separation-guided convolution. Background technique [0002] Portrait segmentation is actually to locate and segment the person information in the image from the original image, while multi-modal portrait segmentation is to add a depth map on the basis of the ordinary portrait segmentation algorithm. The depth map is similar to the grayscale image, but Each of its pixel values ​​is the actual distance from the sensor to the object. Multimodal portrait segmentation is more sensitive to the capture of high-level semantic information, which greatly improves the accuracy of portrait segmentation. [0003] Multi-modal portrait segmentation is a binary segmentation task in computer vision. The purpose is to separate the characters from the original image, which is beneficial to the background blur ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06T7/194G06T7/90G06T3/40G06N3/04G06N3/08
CPCG06T7/194G06T7/90G06T3/4007G06N3/08G06T2207/10024G06T2207/20081G06T2207/20084G06T2207/30196G06N3/048G06N3/045
Inventor 陈泽宇陈舒涵徐秀奇俞锦豪陆露汤浩楠
Owner YANGZHOU UNIV
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More