Unlock instant, AI-driven research and patent intelligence for your innovation.

Video pedestrian re-identification method based on channel attention mechanism and application

A pedestrian re-identification and attention technology, applied in the field of video pedestrian re-identification, can solve the problems of inconspicuous changes in optical flow, weak feature modification ability, incomplete pedestrian parts, etc., to reduce the process of re-fusion and better recognition effect, the effect of reducing the amount of calculation

Active Publication Date: 2021-05-25
SOUTH CHINA UNIV OF TECH
View PDF5 Cites 4 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] Defects or deficiencies of existing technologies: 1. The use of prior knowledge to assist in solving the problem of pedestrian re-identification is often limited by the data set, for example, in some data sets, the change of optical flow is not obvious or the parts of pedestrians in the image are not complete
In addition, the acquisition of pedestrian parts based on pedestrian images also depends on the extraction tools trained on other datasets, so the extraction tools will also restrict the acquisition of pedestrian parts
2. The spatial attention mechanism often uses the correlation between the corresponding image feature pixels as the fusion basis, and its granularity can reach the spatial size of the image feature map, such as 16*8, 24*12, and the granularity of the temporal attention mechanism It is often larger, only the number of frames N in a sampled video (generally N<=16), and the larger granularity leads to the weak ability of the attention mechanism to modify features; 3. The existing channel attention mechanism is still relatively Monotonous, such as "Co-segmentation Inspired Attention Networks for Video-based Person Re-identification" directly averages the channel attention coefficients corresponding to each image feature, and then assigns each original image feature as a Hadamard product
4. According to the image quality and then fuse the image features, the quality coefficient is directly calculated from each image feature. On the one hand, the granularity of fusion lies in the number of frames, which is larger than the spatial attention. On the other hand, only through the image itself. The attention coefficient obtained by comparing image features in non-video is difficult to use as a basis for fusion

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Video pedestrian re-identification method based on channel attention mechanism and application
  • Video pedestrian re-identification method based on channel attention mechanism and application
  • Video pedestrian re-identification method based on channel attention mechanism and application

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0064] Such as figure 1 As shown, this embodiment provides a video pedestrian re-identification method based on channel attention mechanism, including the following steps:

[0065] S1: Video sampling to construct training data

[0066] Images for training are selected from each video. Each video contains images with different frame numbers, assuming the i-th video contains L i Frame images, need to select N frames of images, first from the interval [1,L i -N] randomly select a number, and then take out the consecutive N frames of images after this number as training images; if L i

[0067] S2: Obtain temporary video feature map

[0068] Such as figure 2 As shown, N frames of images are respectively input into the convolutional neural netw...

Embodiment 2

[0106] This embodiment provides a video pedestrian re-identification system based on the channel attention mechanism, including: a sampling preprocessing module, a temporary video feature map acquisition module, a channel attention building module, an attention matrix building module, a modification feature map building module, The final video-level feature vector acquisition module, classification module, cross-entropy loss calculation module, model training module, test input module, similarity ranking module and output module;

[0107] In this embodiment, the sampling preprocessing module is used to sample N frames of images from the video, and perform preprocessing to obtain training data;

[0108] In this embodiment, the temporary video feature map acquisition module is used to obtain the temporary video feature map: N frames of images are respectively input into the convolutional neural network, each frame of image corresponds to a feature map, and N feature maps are obta...

Embodiment 3

[0120] This embodiment provides a storage medium, the storage medium can be a storage medium such as ROM, RAM, magnetic disk, optical disk, etc., and the storage medium stores one or more programs. Video person re-identification method with channel attention mechanism.

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a video pedestrian re-identification method based on a channel attention mechanism and application. The method comprises the steps: preprocessing a video sampling image, inputting the preprocessed video sampling image into a convolutional neural network, respectively extracting N image feature maps, and obtaining a temporary video-level feature map through time pooling; inputting the temporary video-level feature map into a channel attention module, and outputting feature vectors; expanding and filling the feature vectors in height and width to obtain an attention matrix, and multiplying the attention matrix by the original N feature maps element by element to obtain modified N feature maps; performing time and space pooling in sequence to obtain a video-level feature vector, adding a classifier to obtain a classification result, calculating cross entropy loss of the classification result, and training and optimizing the whole model; and performing pedestrian re-identification application on the trained pedestrian re-identification model. According to the method, image-level features are better fused into video-level features, the expressive force of the video-level features is increased, and the performance of video pedestrian re-identification is improved.

Description

technical field [0001] The invention relates to the technical field of video pedestrian re-identification, in particular to a video pedestrian re-identification method and application based on a channel attention mechanism. Background technique [0002] In the networking application of public security video surveillance construction, a large number of surveillance cameras are installed in public places, such as schools, hospitals, railway stations, subway stations, airports, etc. The full coverage of public security video surveillance can effectively protect the personal and property safety of the people. However, how to effectively use these large amounts of public safety surveillance video is another difficult problem. The traditional method of relying on the naked eye to search for video to match the target person is often time-consuming and has a low accuracy rate. The executives of the traditional method generally feel exhausted. Faced with the explosive growth of surv...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/62G06N3/04G06N3/08
CPCG06N3/08G06V40/103G06N3/048G06N3/045G06F18/22G06F18/214G06F18/24Y02T10/40
Inventor 顾国强丁长兴
Owner SOUTH CHINA UNIV OF TECH