Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Speech recognition method and device

A speech recognition and speech technology, applied in speech recognition, speech analysis, instruments, etc., can solve the problem of high error rate

Active Publication Date: 2016-08-24
LE SHI ZHI ZIN ELECTRONIC TECHNOLOGY (TIANJIN) LTD
View PDF6 Cites 33 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] It can be seen that the current speech recognition results have a problem of high error rate.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Speech recognition method and device
  • Speech recognition method and device
  • Speech recognition method and device

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0028] refer to figure 2 As shown, in the embodiment of the present invention, the process of the server performing speech recognition includes:

[0029] Step 200: Receive a voice data packet sent by the terminal; wherein, the voice data packet includes voice information.

[0030] In the embodiment of the present invention, the terminal calls the SDK (Software Development Kit; Software Development Kit) to obtain the voice information input by the user through the voice collection part; the terminal generates a voice data packet according to the voice information; and the voice data The package is sent to the server.

[0031] Optionally, a wireless communication network is included between the terminal and the server, and the terminal sends the voice data packet containing the voice information to the server through the wireless communication network.

[0032] Further, after the server receives the voice data packet sent by the terminal, it performs noise removal processing ...

Embodiment 2

[0059] refer to image 3 As shown, in the embodiment of the present invention, the generation process of the user information contained in the database of the server includes:

[0060] Step 300: Receive a voice data packet sent by the terminal; wherein, the voice data packet includes voice information.

[0061] Step 310: Obtain acoustic feature information included in the voice information.

[0062] Step 320: According to the acoustic feature information, determine the age and gender of the user who provided the voice information, and the final text information; according to the determined age and gender of the user who provided the voice information.

[0063] Optionally, the server may also acquire environmental data, such as time and user action range, according to the acoustic feature information.

[0064] Step 330: Analyze the determined age and gender of the user and the final text information, and generate user information according to the analysis result.

[0065] Op...

Embodiment 3

[0068] Based on the above technical solutions, see Figure 4 As shown, in the embodiment of the present invention, a memory space cleaning device is provided, including a receiving unit 40, an acoustic feature information acquiring unit 41, an initial text information acquiring unit 42, a final text information generating unit 43, and a sending unit 44, wherein:

[0069] The receiving unit 40 is configured to receive a voice data packet sent by the terminal; wherein, the voice data packet contains voice information;

[0070] Acoustic feature information acquisition unit 41, configured to acquire acoustic feature information of the voice information; wherein, the acoustic feature information is information representing sound characteristics of the voice information;

[0071] An initial text information acquisition unit 42, configured to sequentially input the acoustic feature information into a preset acoustic model and a language model, and acquire initial text information obt...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a speech recognition method and device. Speech information transmitted by a terminal is received. The acoustic characteristic information of the speech information is acquired. The acoustic characteristic information is sequentially input into an acoustic model and a language model. The acoustic model and the language model recognize the speech information to acquire initial text information. Based on pre-stored user information, the initial text information is corrected to generate final text information. According to the technical scheme provided by the invention, the recognized initial text information is corrected; errors in the initial text information are corrected; the corrected final text information is sent to the terminal; and the terminal can provide accurate services for a user according to the accurate final text information.

Description

technical field [0001] Embodiments of the present invention relate to the technical field of voice information processing, and in particular, to a voice recognition method and device. Background technique [0002] Speech recognition technology is a technology that allows machines to convert voice signals into corresponding commands or texts through the process of recognition and understanding. At present, voice recognition technology is widely used in voice interactive products such as voice control and voice translation. [0003] At present, various terminals have a voice input function, and various application software installed on the terminals need to perform corresponding operations based on voice recognition results, so as to generate information required by the user and present it to the user. Only when the voice recognition of the terminal is good and can accurately recognize the voice information input by the user can the service provided to the user be more accura...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G10L15/30G10L15/22G10L15/10
CPCG10L15/10G10L15/22G10L15/30
Inventor 田伟森赵恒艺
Owner LE SHI ZHI ZIN ELECTRONIC TECHNOLOGY (TIANJIN) LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products