Unlock instant, AI-driven research and patent intelligence for your innovation.

Occluded pedestrian re-identification method combining spatial transformation network and multi-scale feature extraction

A multi-scale feature and pedestrian re-identification technology, applied in the field of deep learning and computer vision, can solve the problems of complex models and high training costs, achieve the effect of simple models, easy training, and improved recognition accuracy

Active Publication Date: 2021-02-23
SUN YAT SEN UNIV
View PDF5 Cites 3 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

This method can greatly improve the accuracy of recognition, but due to the simultaneous graph network and human key point estimation, the model is more complex, and the training cost is higher

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Occluded pedestrian re-identification method combining spatial transformation network and multi-scale feature extraction
  • Occluded pedestrian re-identification method combining spatial transformation network and multi-scale feature extraction
  • Occluded pedestrian re-identification method combining spatial transformation network and multi-scale feature extraction

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0058] Such as figure 1 As shown, this embodiment provides a method for re-identifying occluded pedestrians combined with spatial transformation network and multi-scale feature extraction, including the following steps:

[0059] S1. Construct a pedestrian image training set. Specifically, in this embodiment, Market1501 is used as the original pedestrian image data set. First, all the images in the training set of Market1501 are resized into a size of 384*128 and input to the simulated occlusion generator to obtain the corresponding occluded pedestrian image set. Then the original image data set and the newly generated set of occluded pedestrian images are merged into a new data set, including:

[0060] S1.1. Use the analog occlusion generator to generate occlusion. The specific method is: set the size of the original pedestrian image img to w*h, and extract the pixel values ​​img[0,0],img[0,h- of the four vertices of img 1], img[w-1,0] and img[w-1,h-1], to obtain the average...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses an occluded pedestrian re-identification method combining a spatial transformation network and multi-scale feature extraction. The method comprises the following steps: constructing an occluded pedestrian picture set by using a simulated occlusion generator; forming a data set by the original picture and the shielded pedestrian picture, and inputting the data set into a spatial transformation network for spatial transformation correction; performing multi-scale feature extraction on the corrected images through a convolutional neural network and a spatial pyramid pooling layer, and combining the images into a fixed-length one-dimensional feature vector; obtaining a one-dimensional feature vector containing K elements from the fixed-length one-dimensional feature vector through a full connection layer, and performing identity classification training to obtain a trained network; and extracting features of a pedestrian image to be queried by using the trained network and carrying out similarity matching. According to the method, multi-scale feature extraction is carried out, and the model is more robust by combining feature maps of different scales; and a spatial transformation network is introduced and can be directly embedded into any depth network model for end-to-end training.

Description

technical field [0001] The invention belongs to the technical field of deep learning and computer vision, and in particular relates to a re-identification method for occluded pedestrians combined with spatial transformation network and multi-scale feature extraction. Background technique [0002] Pedestrian re-identification is considered to be a sub-problem of image retrieval. It hopes to use computer vision technology to track across cameras. Specifically, given the pedestrian image to be queried under a certain camera, the image library obtained from other non-overlapping cameras Pedestrian images with the same identity are retrieved from . This technology is widely used in public safety fields such as video surveillance and intelligent security. In the past few years, the problem of pedestrian re-identification has been widely studied, but they usually assume that the image database and the image to be queried are complete pedestrian images, and our pedestrians will ine...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/46G06K9/62G06N3/04G06N3/08
CPCG06N3/08G06V20/53G06V10/44G06N3/045G06F18/22G06F18/24147Y02T10/40
Inventor 郑伟诗张镓伟
Owner SUN YAT SEN UNIV