Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Online video target tracking method based on depth cross similarity matching

A target tracking and video technology, applied in image data processing, instruments, computing, etc., can solve the problems of difficult integration of multi-task technical solutions, background noise interference, etc.

Active Publication Date: 2019-07-09
DALIAN UNIV OF TECH
View PDF9 Cites 30 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] The technical problems to be solved by the present invention are: first, given any scene and specifying any target therein, even if the appearance of the target changes (such as non-rigid target deformation, turning from the front to the back, etc.) After reproduction, the background noise interference is serious and the situation that is difficult to distinguish from similar appearance objects, the present invention still maintains robust tracking; second, the method of the present invention has portability, and the solution tracking method exists independently and is not easy to be compatible with other related video understanding tasks ( Such as pedestrian search, target detection) integrated multi-task technical solutions

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Online video target tracking method based on depth cross similarity matching
  • Online video target tracking method based on depth cross similarity matching
  • Online video target tracking method based on depth cross similarity matching

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0044] The specific implementation manners of the present invention will be further described below in conjunction with the accompanying drawings and technical solutions.

[0045] Step 1: Collect about 4,000 pieces of video data, intercept templates and search areas of different frames of the same video as positive sample pairs, and use negative sample pairs from different videos to form a training set. Under the constraints of the loss function (4), the invented All parameters of the model are trained offline;

[0046] Step 2: Intercept the target template and the image of the search area at the given initial target position in the first frame, and send them into the designed tracking framework (attached figure 1 ) for forward propagation, and get their respective depth features after the convolutional layer ends;

[0047] Step 3: Calculate all cross-similarities M of two features according to formula (2), and convert them into cross-similarity matrix S according to the mapp...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention belongs to the technical field of video target tracking, and provides an online video target tracking method based on depth cross similarity matching. The method comprises the followingsteps of designing a depth feature cross similarity module, capturing all local similarity information of a template and a sample, wherein obtained similar features are no longer sensitive to displacement and deformation; designing a similarity attention layer, distributing weight coefficients for cross similarity results of different spatial positions, and enabling a tracking algorithm not to respond to edge background interference; and designing a loss function containing a parameter regularization item, and rapidly optimizing the parameter to an optimal value. Based on the above three-pointbasic scheme, the deep learning twin network based on matching is used as a basic framework, and any target in the video is tracked in a precise and robust manner. From the tracking effect, the method provided by the invention has the capabilities of distinguishing similar objects, re-identifying a reappearing target after shielding and coping with rotation and deformation, and can be applied tovideo applications such as automatic driving of a front scene, autonomous flight of an unmanned aerial vehicle, traffic or a safety monitoring video and the like.

Description

technical field [0001] The invention belongs to the technical field of video target tracking. Given any target object specified in the first frame, the position and scale of the target can be framed in subsequent video sequences, and involves related theories of digital image processing, deep learning and linear algebra. Background technique [0002] Thanks to the wide application of artificial intelligence technology and the maturity of deep learning methods, the video target tracking algorithm, as the core basic technology, can be widely used in video surveillance, UAV aerial photography, automatic driving and other fields, and has a certain research basis. [0003] There are currently two mainstream directions for video target tracking methods: correlation filtering methods and deep learning methods. The correlation filtering method is based on a cyclic sampling strategy, and continuously optimizes and updates the filter for the prediction of the next frame by modeling t...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06T7/246
CPCG06T2207/10016G06T2207/20081G06T2207/20084G06T7/246
Inventor 卢湖川王璐瑶
Owner DALIAN UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products