Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

222 results about "Speech spectrum" patented technology

Speech spectrum - the average sound spectrum for the human voice. acoustic spectrum, sound spectrum - the distribution of energy as a function of frequency for a particular sound source.

A Robust Speech Feature Extraction Method Based on Sparse Decomposition and Reconstruction

The invention discloses a robust speech characteristic extraction method based on sparse decomposition and reconfiguration, relating to a robust speech characteristic extraction method with sparse decomposition and reconfiguration. The robust speech characteristic extraction method solves the problems that 1, the selection of an atomic dictionary has higher the time complexity and is difficult tomeet the sparsity after signal projection; 2, the sparse decomposition of signals has less consideration for time relativity of speech signals and noise signals; and 3, the signal reconfiguration ignores the prior probability of atoms and mutual transformation of all the atoms. The robust speech characteristic extraction method comprises the following detailed steps of: step 1, preprocessing; step 2, conducting discrete Fourier transform and solving a power spectrum; step 3, training and storing the atom dictionary; step 4, conducting sparse decomposition; step 5, reconfiguring the speech spectrum; step 6, adding a Mel triangular filter and taking the logarithm; and step 7, obtaining sparse splicing of Mel cepstrum coefficients and a Mel cepstrum to form the robust characteristic. The robust speech characteristic extraction method is used for the fields of multimedia information processing.
Owner:哈尔滨工业大学高新技术开发总公司

Intelligent wearing equipment for safety and health service for old people and voice recognition method

The invention discloses intelligent wearing equipment for safety and health service for old people and a voice recognition method. The intelligent wearing equipment comprises a moving posture module, a gesture recognition module, a physiological sign module, an environment detection module, a video processing module, a wireless communication module, a charging module, a central processing unit, a memorizer, a display module, a mobile communication module, an alarm module, a wireless transceiving module and a voice recognition module, wherein the voice recognition module is used for studying and storing physiological and abnormal sound of a wearer and frequently-used control and response voice to a standard voice frequency spectrum feature library, studying and storing frequently-used voice of the wearer to a wearer's voice frequency spectrum feature library and recognizing whether the voice of the wearer conforms to the standard voice frequency spectrum feature library or the wearer's voice frequency spectrum feature library. The intelligent wearing equipment is capable of recognizing the voice of the wearer effectively and combines gestures of the wearer or a cloud safety and health service platform for auxiliary recognition, so that voice recognition effects are further improved.
Owner:SHENZHEN GLOBAL LOCK SAFETY SYST ENG +1

Speech language classifying method based on CNN and GRU fused deep neural network

The invention discloses a speech language classifying method based on a CNN and GRU fused deep neural network. The method comprises the following steps that S1, source audio data of a server is obtained, audio preprocessing is conducted, and the source audio data is cut; S2, audio data file information is read, and an audio data inventory CSV file is generated; S3, an audio data file is subjectedto short-time Fourier transformation, and two-dimensional speech spectrums associated with time and frequency domains of expansion of a series of frequency spectrum functions obtained after speech signal time domains are analyzed are obtained; S4, a model is built; S5, two-dimensional speech spectrum image data is input into the CNN and GRU fused speech language classifying deep neural network model, and language classification data is classified and output; S6, the language classification data and source audio data file information are stored. By means of the method, the problem about speechlanguage classification is solved, the method has the advantages of being automatic, high in identification rate, high in robustness, low in cost, high in portability and the like, and the business connection with a third-party system can be facilitated.
Owner:深圳市网联安瑞网络科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products