Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

269 results about "Speech generation" patented technology

Speech generation. Speech generation and recognition are used to communicate between humans and machines. Rather than using your hands and eyes, you use your mouth and ears. This is very convenient when your hands and eyes should be doing something else, such as: driving a car, performing surgery, or (unfortunately) firing your weapons at the enemy.

Method and apparatus for improved duration modeling of phonemes

A method and an apparatus for improved duration modeling of phonemes in a speech synthesis system are provided. According to one aspect, text is received into a processor of a speech synthesis system. The received text is processed using a sum-of-products phoneme duration model that is used in either the formant method or the concatenative method of speech generation. The phoneme duration model, which is used along with a phoneme pitch model, is produced by developing a non-exponential functional transformation form for use with a generalized additive model. The non-exponential functional transformation form comprises a root sinusoidal transformation that is controlled in response to a minimum phoneme duration and a maximum phoneme duration. The minimum and maximum phoneme durations are observed in training data. The received text is processed by specifying at least one of a number of contextual factors for the generalized additive model. An inverse of the non-exponential functional transformation is applied to duration observations, or training data. Coefficients are generated for use with the generalized additive model. The generalized additive model comprising the coefficients is applied to at least one phoneme of the received text resulting in the generation of at least one phoneme having a duration. An acoustic sequence is generated comprising speech signals that are representative of the received text.
Owner:APPLE INC

Television voice voting method, television voice voting system and television voice voting terminal

InactiveCN103067754AThe convenient way to participate in TV voting activitiesRealize the function of voice TV votingSelective content distributionComputer hardwareUser participation
The invention discloses a television voice voting method, a television voice voting system and a television voice voting terminal. The television voice voting method comprise the following steps: a vote event is formed by the voting system and sent to a digital television terminal, wherein the vote event comprises a vote event identification (ID), vote content information and validity period information; the vote event is received by the digital television terminal and the vote event within the period of validity is showed; voice input is received, voice is identified, and voice voting information is generated and sent to the voting system; a voting result is updated by the voting system according to the voice voting information, the voting result is sent to the digital television terminal and showed. Due to the fact that the digital television terminal receives the vote event which is sent by the voting system, identifies the voice voting information of a user, sends a vote which is chosen by the user to the voting system, receives and shows the updated voting result which is sent back by the system, the function of voice television voting is achieved. The method through which users participate in a television voting activity is more convenient, faster and more interesting.
Owner:SHENZHEN COSHIP ELECTRONICS CO LTD

Voice generation method and device based on generative adversarial network

The invention discloses a voice generation method based on a generative adversarial network. According to the method, randomly-generated noise data meeting Gaussian distribution is converted into a simulation sample through a generative model; as the simulation sample does not have the language content, when the generative model and a discrimination model are circularly updated, generative capacities required to be learned by the generative model and discrimination capacities required to be learned by the discrimination model are correspondingly increased, and accordingly the generative capacities of the generative model and the discrimination capacities of the discrimination model are improved; when a contrast value between a training sample and the simulation sample is smaller than or equal to a preset threshold value, it is thought that the generative model has the capacity of generating real data; a voice database generated through the generative model has enough reality, and the recognition rate can be increased when the generative model is applied to identity recognition. Correspondingly, the voice generation method, a voice generation device and voice generation equipment based on the generative adversarial network and a computer readable storage medium have the same advantages.
Owner:SPEAKIN TECH CO LTD

Voice authentication system

A voice authentication system includes: a standard template storage part 17 in which a standard template that is generated from a registered voice of an authorized user and featured with a voice characteristic of the registered voice is stored preliminarily in a state of being associated with a personal ID of the authorized user; an identifier input part 15 that allows a user who intends to be authenticated to input a personal ID; a voice input part 11 that allows the user to input a voice; a standard template / registered voice selection part 16 that selects a standard template and a registered voice corresponding to the inputted identifier; a determination part 14 that refers to the selected standard template and determines whether or not the inputted voice is a voice of the authorized user him / herself and whether or not presentation-use information is to be outputted by referring to a predetermined determination reference; a presentation-use information extraction part 19 that extracts information regarding the registered voice of the authorized user corresponding to the inputted identifier; and a presentation-use information output part 18 that presents the presentation-use information to the user in the case where it is determined by the determination part that the presentation-use information is to be outputted to the user.
Owner:FUJITSU LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products