Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

1860 results about "Frame sequence" patented technology

Bidirectional long short-term memory unit-based behavior identification method for video

The invention discloses a bidirectional long short-term memory unit-based behavior identification method for a video. The method comprises the steps of (1) inputting a video sequence and extracting an RGB (Red, Green and Blue) frame sequence and an optical flow image from the video sequence; (2) respectively training a deep convolutional network of an RGB image and a deep convolutional network of the optical flow image; (3) extracting multilayer characteristics of the network, wherein characteristics of a third convolutional layer, a fifth convolutional layer and a seventh fully connected layer are at least extracted, and the characteristics of the convolutional layers are pooled; (4) training a recurrent neural network constructed by use of a bidirectional long short-term memory unit to obtain a probability matrix of each frame of the video; and (5) averaging the probability matrixes, finally fusing the probability matrixes of an optical flow frame and an RGB frame, taking a category with a maximum probability as a final classification result, and thus realizing behavior identification. According to the method, the conventional artificial characteristics are replaced with multi-layer depth learning characteristics, the depth characteristics of different layers represent different pieces of information, and the combination of multi-layer characteristics can improve the accuracy rate of classification; and the time information is captured by use of the bidirectional long short-term memory, many pieces of time domain structural information are obtained and a behavior identification effect is improved.
Owner:SUZHOU UNIV

Self-adaptive voice endpoint detection method

The invention relates to voice detection technology in an automatic caption generating system, in particular to a self-adaptive voice endpoint detection method. The method comprises the following steps: dividing an audio sampling sequence into frames with fixed lengths, and forming a frame sequence; extracting three audio characteristic parameters comprising short-time energy, short-time zero-crossing rate and short-time information entropy aiming at data of each frame; calculating short-time energy frequency values of the data of each frame according to the audio characteristic parameters, and forming a short-time energy frequency value sequence; analyzing the short-time energy frequency value sequence from the data of the first frame, and seeking for a pair of voice starting point and ending point; analyzing background noise, and if the background noise is changed, recalculating the audio characteristic parameters of the background noise, and updating the short-time energy frequency value sequence; and repeating the processes till the detection is finished. The method can carry out voice endpoint detection for the continuous voice under the condition that the background noise is changed frequently so as to improve the voice endpoint detection efficiency under a complex noise background.
Owner:CHINA DIGITAL VIDEO BEIJING

System and Method for Real-Time Super-Resolution

A method and system are presented for real time Super-Resolution image reconstruction. According to this technique, data indicative of a video frame sequence compressed by motion compensated compression technique is processed, and representations of one or more video objects (VOs) appearing in one or more frames of said video frame sequence are obtained. At least one of these representations is utilized as a reference representation and motion vectors, associating said representations with said at least one reference representation, are obtained from said data indicative of the video frame sequence. The representations and the motion vectors are processed, and pixel displacement maps are generated, each associating at least some pixels of one of the representations with locations on said at least one reference representation. The reference representation is re-sampled according to the sub-pixel accuracy of the displacement maps, and a re-sampled reference representation is obtained. Pixels of said representations are registered against the re-sampled reference representation according to the displacement maps, thereby providing super-resolved image of the reference representation of said one or more VOs.
Owner:RAMOT AT TEL AVIV UNIV LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products