Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

55results about How to "Improving Semantic Segmentation Accuracy" patented technology

Semantic segmentation method and system for RGB-D image

The invention discloses a semantic segmentation method and system for an RGB-D image. The semantic segmentation method comprises the steps: extracting RGB coding features and depth coding features ofan RGB-D image in multiple stages; inputting the RGB coding features and the depth coding features of each stage in the plurality of stages into an attention model to obtain each multi-mode fusion feature corresponding to each stage; extracting context semantic information of the multi-modal fusion features in the fifth stage by using a long short-term memory network; splicing the multi-modal fusion features and the context semantic information in the fifth stage to obtain context semantic features; and performing up-sampling on the context semantic features, and fusing the context semantic features with the multi-modal fusion features of the corresponding stage by using a jump connection mode to obtain a semantic segmentation map and a semantic segmentation model. By extracting RGB codingfeatures and depth coding features of the RGB-D image in multiple stages, the semantic segmentation method effectively utilizes color information and depth information of the RGB-D image, and effectively mines context semantic information of the image by using a long short-term memory network, so that the semantic segmentation accuracy of the RGB-D image is improved.
Owner:HANGZHOU WEIMING XINKE TECH CO LTD +1

Traffic image semantic segmentation method based on multi-feature map

The invention discloses a traffic image semantic segmentation method based on a multi-feature map. The method comprises the following steps: obtaining a multi-feature map training sample: a disparitymap, a height map and an angle map; constructing a network model, training the network model, inputting the trained network model and a six-channel test image into the network model, outputting a probability value that each pixel belongs to each object category in the six-channel image via a multi-class classifier softmax layer, then predicting the object category to which each pixel in the six-channel image belongs, and finally outputting an image semantic segmentation map. By adoption of the traffic image semantic segmentation method based on the multi-feature map provided by the invention,the fusion of a color image with a depth map, the height map and the angle map, more feature information of the image can be obtained, and it is conducive to understanding the road traffic scene and improving the semantic segmentation accuracy. According to the traffic image semantic segmentation method based on the multi-feature map provided by the invention, by means of the learned effective features, the object category to which each pixel in the image belongs can be predicted, and the image semantic segmentation map is output.
Owner:DALIAN UNIV OF TECH

RGB-D image semantic segmentation method based on multi-modal adaptive convolution

The invention relates to an RGB-D image semantic segmentation method based on multi-modal adaptive convolution. The method comprises the steps that an encoding module extracts RGB image features and depth image features; the RGB features and the depth features are sent to a fusion module for fusion; the method comprises the following steps: firstly, inputting multi-modal features into a multi-modal adaptive convolution generation module, and calculating two multi-modal adaptive convolution kernels with different scales; then, enabling the multi-modal feature fusion module to carry out depth separable convolution operation on the RGB features and the depth features and an adaptive convolution kernel to obtain adaptive convolution fusion features; splicing the fusion features with the RGB features and the depth features to obtain final fusion features; enabling the decoding module to perform continuous up-sampling on the final fusion feature, and obtaining a semantic segmentation resultthrough convolution operation. According to the invention, multi-modal features are interacted cooperatively through adaptive convolution, and convolution kernel parameters are dynamically adjusted according to an input multi-modal image, so that the method is more flexible than a traditional convolution kernel with fixed parameters.
Owner:BEIJING UNIV OF TECH

Semantic stereo reconstruction method of remote sensing image

The invention discloses a semantic stereo reconstruction method for a remote sensing image, and mainly solves the problem of low semantic three-dimensional reconstruction precision caused by ignoringrelated information of semantic segmentation and parallax estimation in the prior art. According to the implementation scheme, firstly, experimental data are preprocessed; a semantic segmentation network and a parallax estimation network are trained by using the training data; the trained network is tested on the test image, and test results of different frequency band information are fused to obtain a fused semantic segmentation result and a fused parallax result; the error correction module is used for assisting each other to correct the error part of the opposite side; and the disparity information is calculated to obtain height information, and the semantic segmentation result is combined with the height information to obtain a semantic three-dimensional reconstruction result of the image. According to the method, the proportion of small samples is improved, the influence of data on the network is balanced, the semantic information and the parallax result are fused with each other,the accuracy of semantic three-dimensional reconstruction of the remote sensing image is improved, and the method can be used for urban scene three-dimensional reconstruction.
Owner:XIDIAN UNIV

Large-scene point cloud semantic segmentation method

The invention discloses a large-scene point cloud semantic segmentation method. The method comprises the following steps: carrying out feature splicing on three-dimensional point cloud data containing feature information to obtain point cloud initial features; performing expansion graph convolution and random sampling on the point cloud initial features to obtain multi-layer intermediate features and sampling coding features; performing cross-layer context reasoning on the multi-layer intermediate features to obtain complementary context features, and splicing the complementary context features into the sampling coding features obtained in the last layer to obtain final coding features; decoding the final coding feature to obtain a decoding feature; inputting the decoding feature into a full connection layer classifier to obtain segmentation result prediction; and constructing a loss function, training and optimizing the model, and storing model parameters. According to the method, cross-layer context reasoning is used for aggregating multiple layers of contexts in the coding stage, attention fusion is adopted for feature selection in the decoding stage, information loss can be effectively made up and feature redundancy can be effectively reduced while efficiency is guaranteed, and therefore the accuracy rate is improved.
Owner:SICHUAN UNIV

Road scene semantic segmentation method based on convolutional neural network

The invention discloses a road scene semantic segmentation method based on a convolutional neural network. The method comprises the steps: firstly building the convolutional neural network which comprises an input layer, a hidden layer, and an output layer, and the hidden layer is composed of 13 neural network blocks, 7 upsampling layers, and 8 cascade layers; inputting each original road scene image in the training set into a convolutional neural network for training to obtain 12 semantic segmentation prediction images corresponding to each original road scene image; calculating a loss function value between a set composed of 12 semantic segmentation prediction images corresponding to each original road scene image and a set composed of 12 single hot coding images processed by the corresponding real semantic segmentation image; obtaining an optimal weight vector and an optimal offset item of the convolutional neural network classification training model; inputting the road scene imageto be subjected to semantic segmentation into a convolutional neural network classification training model for prediction to obtain a corresponding predicted semantic segmentation image. The method has the advantage of high semantic segmentation precision.
Owner:ZHEJIANG UNIVERSITY OF SCIENCE AND TECHNOLOGY

Real-time street view image semantic segmentation method based on deep multi-branch aggregation

The invention discloses a real-time street view image semantic segmentation method based on deep multi-branch aggregation, and relates to a computer vision technology. A popular encoder-decoder structure is adopted; the method comprises the following steps: firstly, transforming a lightweight image classification network as a basis to serve as an encoder; dividing the encoder into different sub-networks, and sending features in each sub-network into a designed multi-branch feature aggregation network and a global context module; performing enhancement on spatial details and semantic information on features needing to be aggregated by using a lattice-type enhancement residual module and a feature transformation module in the multi-branch feature aggregation network; and finally, according to the sizes of the feature maps, aggregating the output feature maps of the global context module and the output feature maps of the multi-branch feature aggregation network step by step from small to large so as to obtain a final semantic segmentation result map. While the streetscape image with a large resolution is processed, high streetscape image semantic segmentation precision and real-time prediction speed are maintained.
Owner:XIAMEN UNIV

Semantic segmentation of road scene based on multi-scale perforated convolutional neural network

The invention discloses a road scene semantic segmentation method based on a multi-scale perforated convolutional neural network. In a training stage, a multi-scale perforated convolutional neural network is constructed. The hidden layer thereof comprises nine neural network blocks, five cascade layers and six up-sampling blocks. The original road scene images are input into the multi-scale perforated convolutional neural network for training, and 12 corresponding semantic segmentation prediction maps are obtained. By calculating the loss function value between the set of 12 semantic segmentation prediction maps corresponding to the original road scene image and the set of 12 monothermally coded images processed from the corresponding real semantic segmentation image, The optimal weight vector and bias term of multi-scale perforated convolution neural network classification training model are obtained. In the testing phase, the road scene images to be segmented are input into the multi-scale perforated convolutional neural network classification training model, and the predictive semantic segmentation images are obtained. The invention has the advantages of improving the efficiencyand accuracy of the semantic segmentation of the road scene images.
Owner:ZHEJIANG UNIVERSITY OF SCIENCE AND TECHNOLOGY

Streetscape image semantic segmentation system and segmentation method, electronic equipment and computer readable medium

The invention discloses a streetscape image semantic segmentation system and segmentation method, electronic equipment and a computer readable medium. The segmentation method comprises the following steps: step 1, acquiring a streetscape image and carrying out preprocessing and data enhancement on the streetscape image; step 2, encoding the streetscape image into an output feature map by using an encoder; step 3, collecting features of the last three output feature maps by using a multi-level feature combined up-sampling module, and fusing the features to obtain a second output feature map; 4, converting the second output feature map into a third output feature map; 5, inputting the third output feature map into a convolution classifier to obtain a semantic segmentation feature value; step 6, performing end-to-end training by using a back propagation algorithm to obtain a streetscape image semantic segmentation model; and 7, performing semantic segmentation on the streetscape image by using the streetscape image semantic segmentation model. According to the method, under the condition that semantic segmentation precision is not reduced, the speed of network segmentation is increased, and the real-time response capability of the method in application is enhanced.
Owner:CHANGCHUN UNIV OF TECH

Large-scale point cloud semantic segmentation method and system

The invention relates to a large-scale point cloud semantic segmentation method and system, and the method comprises the steps: extracting point-by-point features of a to-be-recognized point cloud,wherein the to-be-recognized point cloud is composed of a plurality of to-be-recognized points; on the basis of the point cloud space information of each to-be-recognized point, gradually encoding each point-by-point feature to obtain a corresponding point cloud feature; decoding the point cloud features step by step to acquire corresponding decoding feature; and according to each decoding feature, based on a semantic segmentation network model, determining a semantic segmentation prediction result of the to-be-recognized 3D point cloud. The invention extracts point-by-point features of the to-be-recognized point cloud, extracting more effective spatial features from large-scale point cloud information, gradually encodes each point-by-point feature based on point cloud spatial information of each to-be-recognized point to obtain point cloud features, further decodes to obtain decoding features, and determining a semantic segmentation prediction result of the to-be-recognized 3D point cloud according to the decoding features. Therefore, the semantic information of the surrounding space environment is obtained, and semantic segmentation precision is improved.
Owner:INST OF AUTOMATION CHINESE ACAD OF SCI

Remote sensing image semantic segmentation method based on parallel cavity convolution

The invention discloses a remote sensing image semantic segmentation method based on parallel cavity convolution, and relates to the technical field of remote sensing images, and the method comprises the following steps: obtaining a high-resolution remote sensing image in advance, and carrying out the slicing, normalization and standardization of the high-resolution remote sensing image, and obtaining a source high-resolution remote sensing image; initializing a low-layer network of a feature extraction network resnet101 based on a resnet101 parameter pre-trained on an ImageNet, constructing a parallel cavity convolutional network, and extracting a shallow-layer feature of a source high-resolution remote sensing image; inputting the shallow layer features into a parallel dilated convolutional network to obtain multi-scale information, and fusing the multi-scale information; and fusing the fused features with the shallow features again, and repairing image-level information by using a full-connection conditional random field to obtain a semantic segmentation result. Under the condition that extra parameters are not increased, the receptive field of convolution is expanded, and compared with standard convolution achieving the same receptive field, the parallel cavity convolution method can save more video memories.
Owner:HUAZHONG UNIV OF SCI & TECH

Semantic segmentation method and device, electronic equipment and computer readable storage medium

The embodiment of the invention provides a semantic segmentation method and device, electronic equipment and a computer readable storage medium. The method comprises: obtaining a to-be-processed image; and performing semantic segmentation processing on the to-be-processed image by adopting the semantic segmentation model to obtain a semantic segmentation result of the to-be-processed image, wherein the semantic segmentation model is obtained by taking a first transformation feature, which is obtained by performing contour decomposition or enhancement processing on a first intermediate feature output by a reference semantic model, as a reference, and combining with a second transformation feature, which is obtained by performing contour decomposition or enhancement processing on a second intermediate feature output by the to-be-trained semantic segmentation model. The first intermediate feature and the second intermediate feature comprise at least one of the following groups: a first texture feature and a second texture feature; and a first semantic feature and a second semantic feature. The first transformation feature and the second transformation feature comprise at least one of the following groups: a first contour feature and a second contour feature; and a first enhancement feature and a second enhancement feature.
Owner:SHANGHAI SENSETIME INTELLIGENT TECH CO LTD

Road scene image processing method based on double-side dynamic cross fusion

The invention discloses a road scene image processing method based on double-side dynamic cross fusion. The method comprises two processes of a training stage and a testing stage, and includes selecting road scene images and corresponding thermodynamic diagrams and real semantic understanding images to form a training set; constructing a convolutional neural network; performing data enhancement on the training set to obtain an initial input image pair, and inputting the initial input image pair into a convolutional neural network for processing to obtain a corresponding road scene prediction map; calculating a loss function value between the road scene prediction map and the corresponding real semantic segmentation image; repeating the above steps to obtain a convolutional neural network classification training model; and inputting the road scene image to be subjected to semantic segmentation and the corresponding thermodynamic image into the convolutional neural network classification training model to obtain a corresponding predicted semantic segmentation image. According to the invention, the semantic segmentation accuracy of the road scene image is effectively improved, the loss of detail features is reduced, and the edge of an object can be better restored.
Owner:ZHEJIANG UNIVERSITY OF SCIENCE AND TECHNOLOGY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products