Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

482results about How to "Improve decoding speed" patented technology

Decoding method based on deep neural network translation model

The invention relates to the field of the language processing, and provides a decoding method based on a deep neural network translation model. The problems of high model training complexity, large training difficulty and slow decoding speed in a machine translation model are solved. The specific implementation way of the method comprises the following steps: performing word segmentation processing on the to-be-translated sentence to obtain source language vocabulary; step two, performing word alignment on the linguistic data in a preset translation model glossary by using an automatic alignment tool so as to obtain a target language word aligned to the source language vocabulary; and step three, determining a target-side dynamic glossary of the to-be-translated sentence based on the target language word obtained in the step two, and using the sentence decoded by using a column searching method as the output of the translation model according to the pre-constructed translation model, wherein the translation model is the deep neural network based on the threshold residual mechanism and the parallel attention mechanism. Through the decoding method disclosed by the invention, the model translation quality is improved, and the mode coding speed is improved.
Owner:INST OF AUTOMATION CHINESE ACAD OF SCI +1

Two-dimensional recursive network-based recognition method of Chinese text in natural scene images

The invention discloses a two-dimensional recursive network-based recognition method of Chinese text in natural scene images. Firstly, a training sample set is acquired, and a neural network formed bysequentially connecting a deep convolutional network, a two-dimensional recursive network used for encoding, a two-dimensional recursive network used for decoding and a CTC model is trained; test samples are input into the trained deep convolutional network, and feature maps of the test samples are acquired; the feature maps of the test samples are input into the trained two-dimensional recursivenetwork, which is used for encoding, to obtain encoding feature maps of the test samples; the encoding feature maps of the test samples are input into the trained two-dimensional recursive network, which is used for decoding, to obtain a probability result of each commonly used Chinese character in each image of the test samples; and clustering searching processing is carried out, and finally, the overall Chinese text in the test samples is recognized. According to the method of the invention, space/time information and context information of the text images are fully utilized, the text imagepre-segmentation problem can be avoided, and recognition accuracy is improved.
Owner:SOUTH CHINA UNIV OF TECH

All-format media player capable of supporting hardware decoding for digital STB (Set Top Box)

The invention discloses an all-format media player capable of supporting hardware decoding for digital STB (Set Top Box), wherein a hardware decoding interface and a hardware decoding chip which is connected with the hardware decoding interface are arranged between a package format analysis module and an output module; when the hardware decoding manner is needed, demultiplexed PES data or ES data are transmitted to the hardware decoding chip through the hardware decoding interface and decoded into audio / video data. The all-format media player disclosed by the invention has the advantages of coexisting software / hardware decoding, supporting hardware acceleration, supporting various types of channels, having excellent expansibility on software, having excellent generality on source codes and capably transplanting on different hardware platforms. The all-format media player can play media files stored in a local storage medium and network media files, supports the playing of the media files in full package format, and further reads and plays the media files in an external storage medium. According to the all-format media player disclosed by the invention, the disadvantages that the media files in several specific formats only can be played by the present digital television terminal can be effectively solved, therefore the resources are saved, and superior high-definition video playing effect can further be provided to users.
Owner:AVIT

One-dimensional barcode identification method based on image sampling line grey scale information analysis

The invention provides a one-dimensional barcode identification method based on image sampling line grey scale information analysis. The method comprises the following steps that 1, a precise region of a one-dimensional barcode is found out according to image segmentation and calculation to form a sampling interval ABCD; 2, a sampling curve b2 passing through all black and white strip spaces of the one-dimensional barcode is arranged according to the placing position of an image or in the other approaching direction; 3, grey scale values at all positions, from a point b21 to a point b22 on the sampling interval ABCD, on the sampling curve b2 are obtained to form a grey scale curve, grey scale curve crest parts correspond to black strips of the one-dimensional barcode, and trough parts correspond to blanks of the one-dimensional barcode; 4, corresponding processing is carried out on the sampled grey scale curve to obtain a two-value pulse curve of black and white corresponding; 5, barcode bar vacancy boundary points are obtained from the two-value pulse curve, then the width of the barcode bar vacancy sequence is calculated and converted into the width of a barcode bar vacancy unit module, and a coded character set of the corresponding one-dimensional barcode is searched to obtain corresponding codons.
Owner:HANGZHOU SYNOCHIP DATA SECURITY TECH CO LTD

N-gram grammar model constructing method for voice identification and voice identification system

InactiveCN105261358AReduce sparsityControlling the Search PathSpeech recognitionPart of speechSpeech identification
The invention provides an n-gram grammar model constructing method for voice identification and a voice identification system. The method comprises: step (101), training is carried out by using a neural network language model to obtain word vectors, and classification and multi-layer screening is carried out on word vectors to obtain parts of speech; step (102), manual marking is expanded by using a direct word frequency statistic method; and when same-kind-word substitution is carried out, direct statistics of 1-to-n-gram grammar combination units changing relative to an original sentence is carried out, thereby obtaining an n-gram grammar model of the expanding part; step (103), manual marking is carried out to generate a preliminary n-gram grammar model, model interpolation is carried out on the preliminary n-gram grammar model and the n-gram grammar model of the expanding part, thereby obtaining a final n-gram grammar model. In addition, the step (101) includes: step (101-1), inputting a mark and a training text; step (101-2), carrying out training by using a neural network language model to obtain corresponding work vectors of words in a dictionary; step (101-3), carrying out word vector classification by using a k mean value method; and step (101-4), carrying out multi-layer screening on the classification result to obtain parts of speech finally.
Owner:INST OF ACOUSTICS CHINESE ACAD OF SCI +1

Method and apparatus for intra-frame prediction

The invention discloses a method and a device for intra-frame prediction. The method for the intra-frame prediction comprises the steps of (1) determining the row caching index of a lower boundary pixel value which is near to an upper prediction block of a current prediction block, and determining line caching index of a right boundary pixel value which is near to a left prediction block of the current prediction block; (2) taking a corresponding pixel value of the row caching index of the lower boundary pixel value which is near to the upper prediction block obtained in the lower boundary cache as the lower boundary pixel value which is near to the upper prediction block of the current prediction block; regarding the corresponding pixel value of the line caching index of the left boundary pixel value obtained in the right boundary cache as the right boundary pixel value which is near to the left prediction block of the current prediction block; (3) calculating the pixel value of the current prediction block according to the obtained lower boundary pixel value which is near to the upper prediction block and the right boundary pixel value which is near to the left prediction block. The scheme greatly saves prediction time and improves decoding speed of images.
Owner:SHENZHEN COSHIP ELECTRONICS CO LTD

Speech recognition method and system

The invention belongs to the speech recognition technical filed and relates to a speech recognition method and system. The method includes the following steps that: speech signals are acquired; analog-digital conversion is performed on the speech signals, so that corresponding speech digital signals can be generated; preprocessing is performed on the speech digital signals, and speech feature parameters are extracted according to corresponding preprocessing results, and a time sequence of extracting the speech feature parameters is utilized to construct a corresponding feature sequence; the speech feature parameters are matched with speech models in a template library, and the feature sequence is decoded according to a search algorithm, and therefore, a corresponding recognition result can be generated. According to the speech recognition method and system of the invention, time-domain GFCC (gammatone frequency cepstrum coefficient) features are extracted to replace frequency-domain MFCC (mel frequency cepstrum coefficient) features, and DCT conversion is adopted, and therefore, computation quantity can be reduced, and computation speed and robustness can be improved; and the mechanism of weighted finite state transformation is adopted to construct a decoding model, and smoothing and compression processing of the model is additionally adopted, and therefore, decoding speed can be increased.
Owner:徐洋

Video decoding data storage method and calculation method of motion vector data

The invention provides a video decoding data storage method and a calculation method of motion vector data. The video decoding data storage method includes setting a reference frame queue table in a static random access memory, storing basic information of a plurality of reference frames in the reference frame queue table, wherein each unit of the reference frame queue table stores index information corresponding to one reference frame in a decoding buffer area; storing a plurality of groups of frame buffer information corresponding to the index information in the decoding buffer area, wherein each group of frame buffer information contains frame display sequence data and motion vector storage address information; storing the motion vector storage of each reference frame in a dynamic random access memory, wherein the motion vector storage address information is address information of the motion vector data of a coding tree unit of corresponding blocks of the reference frames stored in the dynamic random access memory. The video decoding data storage method and the calculation method of the motion vector data can improve the video decoding efficiency and save hardware expense and bandwidth resources occupied during video decoding.
Owner:ALLWINNER TECH CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products