Video sequence encoding and decoding method for video pedestrian re-identification

A person re-identification and video sequence technology, applied in the field of computer vision image retrieval, can solve the problems of high computational overhead, unsuitable for batch processing, huge storage overhead, etc., and achieve the effect of reducing performance loss and high imaging quality.

Active Publication Date: 2021-02-26
ZHEJIANG UNIV
View PDF2 Cites 3 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, video person re-identification needs to store a large number of video sequences, which will lead to huge storage overhead in practical applications and increase the application cost of video person re-identification.
At the same time, due to the different lengths of each video sequence in the application stage, it is not suitable for batch processing, resulting in a large computational overhead

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Video sequence encoding and decoding method for video pedestrian re-identification
  • Video sequence encoding and decoding method for video pedestrian re-identification
  • Video sequence encoding and decoding method for video pedestrian re-identification

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0047]The present invention is a video sequence encoding and decoding method for video pedestrian re-identification. In the training stage, the tag image feature is fused with the video feature and input to the generator, and then the tag image is used as the reconstruction tag, and the image reconstruction loss is constrained The keyframes generated by the generator. Then the generated key frames are sent to the image feature extraction module for video feature recovery, and the recovered video features and original video features are constrained by the feature reconstruction loss. In the application phase, the HSV-Top-K method is used to select K frames of pictures to generate key frames, and then the generated key frames are stored in the device to reduce storage overhead. When retrieval is required, the image feature extraction module is used to restore the video features of the generated key frames, and the recovered features retain the performance of the video features a...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a video sequence coding and decoding method for video pedestrian re-identification, and the method comprises the following steps: in a training stage, fusing label picture features and video features, inputting the fused features into a generator, then employing a label picture as a reconstruction label, and restraining a key frame generated by the generator through image reconstruction loss; and then sending the generated key frame to an image feature extraction module for video feature recovery, and constraining the recovered video features through feature reconstruction loss, so that the video features are consistent with original video features in performance. In an application stage, K frames of pictures are selected by using an HSVTopK method to be used for generating key frames, and then the generated key frames are stored in equipment so as to reduce the storage overhead. When retrieval is needed, the image feature extraction module is used for carryingout video feature recovery on the generated key frame, and the recovered features reserve the performance of the video features and are used for retrieval matching of pedestrians.

Description

technical field [0001] The invention belongs to the field of computer vision image retrieval, in particular to a video sequence encoding and decoding method for video pedestrian re-identification. Background technique [0002] Pedestrian re-identification is aimed at retrieving pedestrians specified by users from a series of cross-camera surveillance videos; it is widely used in smart cities and security monitoring. [0003] According to the number of different input pictures, person re-identification can be divided into video person re-identification and image person re-identification. Compared with image person re-identification that uses a single frame image as input, video person re-identification uses video sequences as input and is more robust to environmental disturbances. However, video person re-identification needs to store a large number of video sequences, which will lead to huge storage overhead in practical applications and increase the application cost of vid...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/46G06K9/62G06N3/04G06N3/08
CPCG06N3/08G06V20/53G06V10/44G06N3/047G06N3/045G06F18/253
Inventor 潘啸罗浩姜伟
Owner ZHEJIANG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products