Video analysis method based on cross-modal hash learning
An analysis method and cross-modal technology, applied in the field of video semantic analysis, to achieve the effect of improving accuracy and efficient video positioning
- Summary
- Abstract
- Description
- Claims
- Application Information
AI Technical Summary
Problems solved by technology
Method used
Image
Examples
Embodiment 1
[0037] In step a), take 16 frames as the minimum unit for the kth video data V k Carry out cell division.
Embodiment 2
[0039] In order to ensure that each bidirectional convolution process can obtain a video unit set with a length of R. We need to add padding information for each bidirectional sequential convolution operation. The padding number of the i-th layer is to increase the padding information for each two-way sequential convolution operation, so the formula p i =(ε-1)p Calculate the filling number p of the i-th layer i .
Embodiment 3
[0041] In step e), by the formula Calculate the loss function Γ of the fully connected neural network 1 , where is the Fronius norm, T is the transpose, and Y is the uniform dimension set by the multimodal feature.
PUM
Login to View More Abstract
Description
Claims
Application Information
Login to View More 


