Supercharge Your Innovation With Domain-Expert AI Agents!

A model training method, device, electronic equipment and storage medium

A model training and model technology, applied in the field of computer and model training, can solve problems such as high cost and poor timeliness

Active Publication Date: 2020-10-20
BAIDU ONLINE NETWORK TECH (BEIJIBG) CO LTD
View PDF8 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Using the existing training method of the speech synthesis front-end model based on the pre-trained language model, not only the timeliness is poor, but also the cost is high

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A model training method, device, electronic equipment and storage medium
  • A model training method, device, electronic equipment and storage medium
  • A model training method, device, electronic equipment and storage medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0046] figure 1 It is a schematic flow chart of the model training method provided in the first embodiment of the present application. The method can be executed by a model training device or an electronic device. The device or electronic device can be implemented by software and / or hardware. The device or electronic device can Integrate in any smart device with network communication function. Such as figure 1 As shown, the model training method may include the following steps:

[0047] S101. In the first stage of fine-tuning training, input training samples of each first sample type into the shared layer module of the model to be trained.

[0048] In the specific embodiment of the present application, the training process of the speech synthesis front-end model based on the pre-trained language model may only include the first stage and fine-tuning training, or may include the first-stage fine-tuning training and the second-stage fine-tuning training; Can include more stages of f...

Embodiment 2

[0059] image 3 It is a schematic flowchart of a training method for a speech synthesis front-end model based on a pre-trained language model provided in the second embodiment of the present application. Such as image 3 As shown, the training method of the speech synthesis front-end model based on the pre-trained language model may include the following steps:

[0060] S301: In the first stage of fine-tuning training, input training samples of each first sample type into the shared layer module of the model to be trained.

[0061] In the specific embodiment of the present application, the training process of the speech synthesis front-end model based on the pre-trained language model may only include the first stage and fine-tuning training, or may include the first-stage fine-tuning training and the second-stage fine-tuning training; Can include more stages of fine-tuning training. In the first stage of fine-tuning training, the electronic device can input each training sample o...

Embodiment 3

[0078] Image 6 It is a schematic structural diagram of a training device for a speech synthesis front-end model based on a pre-trained language model provided in the third embodiment of the present application. Such as Image 6 As shown, the device 600 includes: a first input module 601, a first training module 602, a second input module 603, and a second training module 604; wherein,

[0079] The first input module 601 is used to input training samples of each first sample type into the shared layer module of the model to be trained in the first stage of fine-tuning training;

[0080] The first training module 602 is configured to adjust the model parameters in the shared layer module based on the training samples of each first sample type;

[0081] The second input module 603 is configured to input the training samples of each first sample type into the task layer module corresponding to each task type of the model to be trained through the sharing layer module;

[0082] The second...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses model training methods and device, electronic equipment and a memory medium and relates to the field of model training. A specific realization scheme comprises the steps of infine adjustment training of a first stage, inputting each training sample of a first sample type into shared layer modules of a to-be-trained model; adjusting model parameters in the shared layer modules on the basis of each training sample of the first sample type; inputting each training sample of the first sample type into a task layer module corresponding to each task type of the to-be-trainedmodel through utilization of the shared layer modules; extracting own matched training data through utilization of each task layer module; and on the basis of the training data matched with each tasklayer module, adjusting the model parameters in each task layer module. According to the embodiments of the invention, the shared layer modules can be trained uniformly, each task layer module also can be independently trained, single-task performance is improved, and moreover, multi-task training effect is also reserved.

Description

Technical field [0001] This application relates to the field of computer technology, and further relates to the field of model training, especially a model training method, device, electronic equipment, and storage medium. Background technique [0002] In the training process of the existing speech synthesis front-end model based on the pre-trained language model, different language training models are used for different types of input data, which are independent of each other. For example, input the data of the polyphonic character type into the polyphonic character model, use the data of the polyphonic character type to train the polyphonic character model; input the data of the prosody type into the prosody model, and use the data of the prosody type to train the prosody model. Adopting the existing training method of a speech synthesis front-end model based on a pre-trained language model not only has poor timeliness, but also has a high cost. Summary of the invention [0003...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G10L15/06
CPCG10L15/063G10L2015/0635
Inventor 潘政林聂志朋白洁
Owner BAIDU ONLINE NETWORK TECH (BEIJIBG) CO LTD
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More