Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Linguistic model training method and device as well as linguistic model construction method and device

A language model and training method technology, applied in the field of input methods, can solve the problems of private information exposure and difficulty in protecting user privacy, achieve the effect of less data, reduce the risk of user private information exposure, and improve upload efficiency

Pending Publication Date: 2020-01-03
BEIJING SOGOU TECHNOLOGY DEVELOPMENT CO LTD
View PDF0 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Either way will lead to the risk of exposure of the user's private information, such as the specific content entered by the user, making it difficult to protect the user's privacy in the process of training the language model

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Linguistic model training method and device as well as linguistic model construction method and device
  • Linguistic model training method and device as well as linguistic model construction method and device
  • Linguistic model training method and device as well as linguistic model construction method and device

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0058] Embodiments of the present application are described below in conjunction with the accompanying drawings.

[0059] In the traditional language model training method, the input method can train the language model on the network side through a large number of historical input data of users, and the acquired historical input data is generally uploaded or input normal in the process of using the input method collected under. Either way will lead to the risk of exposure of the user's private information, such as the specific content input by the user, making it difficult to protect the user's privacy during the process of training the language model.

[0060] To this end, an embodiment of the present application provides a language model training method. In this method, after the terminal determines the first language model for training, it can train the first language model according to locally generated historical input data, so as to Improve the model parameters of the f...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The embodiment of the invention discloses a linguistic model training method. The method comprises the steps that a terminal determines a first linguistic model; the terminal trains the first linguistic model through historical input data locally generated so that a trained first linguistic model can be obtained; and the terminal sends first model parameters of the trained first linguistic model to a server, wherein the first model parameters are used for determining second model parameters of a second linguistic model, and the second linguistic model is used for displaying candidate items obtained through association for the terminal. Therefore, in the process of obtaining the second linguistic model, the historical input data containing the privacy of a user are trained on the local terminal through the first linguistic model, the original historical input data do not need to be uploaded to the server, and thus the risk of exposure of the private information of the user is reduced; and meanwhile, the terminal only uploads the first model parameters of the trained first linguistic model to the server, thus the uploaded data volume is relatively small, and therefore, the uploadingefficiency of the terminal during the process of obtaining the second linguistic model is improved.

Description

technical field [0001] This application relates to the field of input methods, in particular to a language model training method, construction method and device. Background technique [0002] The user can input a character string through an input method, and select the input method to display the candidates on the screen according to the character string. In addition to displaying candidates according to the input string, the input method can also display corresponding candidates according to the content that the user has selected on the screen in some cases. [0003] The candidates displayed by the input method can be obtained through language model association, and the language model needs to be trained with a large amount of corpus to improve the association accuracy. [0004] In some possible implementations, the input method trains the language model on the network side through a large number of historical input data of users, and the acquired historical input data is ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G10L15/06G10L15/18G10L15/30
CPCG10L15/063G10L15/30G10L15/18
Inventor 黄海兵庞帅张扬
Owner BEIJING SOGOU TECHNOLOGY DEVELOPMENT CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products