Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

844 results about "Automatic speech" patented technology

Formulaic language (previously known as automatic speech or embolalia) is a linguistic term for verbal expressions that are fixed in form, often non-literal in meaning with attitudinal nuances, and closely related to communicative-pragmatic context. Along with idioms, expletives and proverbs, formulaic language includes pause fillers (e.g., "Like", "Er" or "Uhm") and conversational speech formulas (e.g., "You've got to be kidding," "Excuse me?" or "Hang on a minute").

User interaction with voice information services

An iterative process is provided for interacting with a voice information service. Such a service may permit, for example, a user to search one or more databases and may provide one or more search results to the user. Such a service may be suitable, for example, for searching for a desired entity or object within the database(s) using speech as an input and navigational tool. Applications of such a service may include, for instance, speech-enabled searching services such as a directory assistance service or any other service or application involving a search of information. In one example implementation, an automatic speech recognition (ASR) system is provided that performs a speech recognition and database search in an iterative fashion. With each iteration, feedback may be provided to the user presenting potentially relevant results. In one specific ASR system, a user desiring to locate information relating to a particular entity or object provides an utterance to the ASR. Upon receiving the utterance, the ASR determines a recognition set of potentially relevant search results related to the utterance and presents to the user recognition set information in an interface of the ASR. The recognition set information includes, for instance, reference information stored internally at the ASR for a plurality of potentially relevant recognition results. The recognition set information may be used as input to the ASR providing a feedback mechanism. In one example implementation, the recognition set information may be used to determine a restricted grammar for performing a further recognition.
Owner:MICROSOFT TECH LICENSING LLC

Visual interactive response system and method translated from interactive voice response for telephone utility

A system, method, and computer readable medium storing a software program for translating a script for an interactive voice response system to a script for a visual interactive response system. The visual interactive response system executes the translated visual-based script when a user using a display telephone calls the visual interactive response system. The visual interactive response system then transmits a visual menu to the display telephone to allow the user to select a desired response, which is subsequently sent back to the visual interactive response system for processing. The voice-based script may be defined in voice extensible markup language and the visual-based script may be defined in wireless markup language, hypertext markup language, or handheld device markup language. The translation system and program includes a parser for extracting command structures from the voice-based script, a visual-based structure generator for generating corresponding command structure for the visual-based script, a text prompt combiner for incorporating text translated from voice prompts into command structure generated by the structure generator, an automatic speech recognition routine for automatically converting voice prompts into translated text, and an editor for editing said visual-based script.
Owner:RPX CLEARINGHOUSE

Transparent monitoring and intervention to improve automatic adaptation of speech models

A system and method to improve the automatic adaptation of one or more speech models in automatic speech recognition systems. After a dialog begins, for example, the dialog asks the customer to provide spoken input and it is recorded. If the speech recognizer determines it may not have correctly transcribed the verbal response, i.e., voice input, the invention uses monitoring and if necessary, intervention to guarantee that the next transcription of the verbal response is correct. The dialog asks the customer to repeat his verbal response, which is recorded and a transcription of the input is sent to a human monitor, i.e., agent or operator. If the transcription of the spoken input is correct, the human does not intervene and the transcription remains unmodified. If the transcription of the verbal response is incorrect, the human intervenes and the transcription of the misrecognized word is corrected. In both cases, the dialog asks the customer to confirm the unmodified and corrected transcription. If the customer confirms the unmodified or newly corrected transcription, the dialog continues and the customer does not hang up in frustration because most times only one misrecognition occurred. Finally, the invention uses the first and second customer recording of the misrecognized word or utterance along with the corrected or unmodified transcription to automatically adapt one or more speech models, which improves the performance of the speech recognition system.
Owner:AVAYA INC

Wearable communication system

A portable data input or computer system includes an input / output device such as a keyboard and a display, another data input device such as an optical bar code scanner, and a data processor module. To scan bar code type indicia, the operator points the scanner at the bar code and triggers the scanner to read the indicia. All the system components are distributed on an operator's body and together form a personal area system (PAS). Components may include a scanner or imager, a wrist unit, a headpiece including an eyepiece display, speaker and a microphone. Components within a particular PAS communicate with each other over a personal area network (PAN). Individual PASs may be combined into a network of PASs called a PAS cluster. PASs in a particular PAS cluster can communicate with each other over another wireless communication channel. Individual PAS can gain access to a Local Area Network (LAN) and / or a Wide Area Network (WAN) via an access point. Individual PASs can use devices, such as servers and PCs situated either on the LAN or the WAN to retrieve and exchange information. Individual PAS components can provide automatic speech and image recognition. PAS components may also act a telephone, a pager, or any other communication device having access to a LAN or a WAN. Transmission of digitized voice and / or video data can be achieved over an Internet link.
Owner:SYMBOL TECH INC

Language model adaptation via network of similar users

A language recognition system, method and program product for recognizing language based input from computer users on a network of connected computers. Each computer includes at least one user based language model trained for a corresponding user for automatic speech recognition, handwriting recognition, machine translation, gesture recognition or other similar actions that require interpretation of user activities. Network computer users are clustered into classes of similar users according to user similarities such as, nationality, profession, sex, age, etc. User characteristics are collected by sensors and from databases and, then, distributed over the network during user activities. Language models with similarities among similar users on the network are identified. The language models include a language model domain, with similar language models being clustered according to their domains. Language models identified as similar are modified in response to user production activities. After modification of one language model, other identified similar language models are compared and adapted. Also, user data, including information about user activities and language model data, is transmitted over the network to other similar users. Language models are adapted only in response to similar user activities, when these activities are recorded and transmitted over the network. Language models are given a global context based on similar users that are connected together over the network.
Owner:NUANCE COMM INC
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products