Check patentability & draft patents in minutes with Patsnap Eureka AI!

Method and device for obtaining text semantic similarity value, storage medium and electronic equipment

A technology of semantic similarity and similarity value, applied in the field of natural language learning, can solve the problems of increasing the training time and training cost of the BERT model, and achieve the effect of reducing loss and improving acquisition efficiency.

Pending Publication Date: 2021-05-28
JILIN UNIV
View PDF1 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

In the existing technology, the BERT model is usually fine-tuned by adding other unsupervised pre-training models to the BERT model, such as semantic role prediction models, entity recognition models, etc., but these pre-training models will increase the training time and time of the BERT model. training cost

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Method and device for obtaining text semantic similarity value, storage medium and electronic equipment
  • Method and device for obtaining text semantic similarity value, storage medium and electronic equipment
  • Method and device for obtaining text semantic similarity value, storage medium and electronic equipment

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0042] The following will clearly and completely describe the technical solutions in the embodiments of the application with reference to the drawings in the embodiments of the application. Apparently, the described embodiments are only some of the embodiments of the application, not all of them. Based on the embodiments in this application, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the scope of protection of this application.

[0043] In the description of the present application, it should be understood that the terms "first", "second" and so on are used for descriptive purposes only, and should not be understood as indicating or implying relative importance. In the description of the present application, it should be noted that, unless otherwise specified and limited, "including" and "having" and any variations thereof are intended to cover non-exclusive inclusion. For example, a process, method, system,...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The embodiment of the invention discloses a method for obtaining a text semantic similarity value. The method comprises the steps of inputting at least two to-be-detected texts and similarity parameters between the at least two to-be-detected texts into a bidirectional encoder representation BERT model from a transformer; generating a query vector and a key value vector pair based on the at least two to-be-detected texts; generating dot product data based on the query vector, the key value vector pair and the similar parameters; and performing similarity processing on the dot product data to obtain a similarity value between the at least two to-be-detected texts. According to the method, when the similarity value between the at least two to-be-detected texts is obtained, the model effect can be improved on the premise of not increasing the training duration and memory loss.

Description

technical field [0001] The present application relates to the technical field of natural language learning, and in particular to a method, device, storage medium and electronic equipment for acquiring text semantic similarity values. Background technique [0002] Transformer's Bidirectional Encoder Representation from Transformers (Bidirectional Encoder Representation from Transformers, BERT) model, as the best deep language model at present, is one of the most mainstream language models for future Natural Language Process (NLP) research and industrial applications. In order to enable the BERT model to complete its own masked word (Masked Language Mode, MLM) task and predict the next sentence (NextSentence Prediction, NSP) task, it also has other functions such as text semantic similarity tasks. fine-tuning. In the existing technology, the BERT model is usually fine-tuned by adding other unsupervised pre-training models to the BERT model, such as semantic role prediction mo...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F40/30G06F16/33
CPCG06F40/30G06F16/3344
Inventor 夏婷玉王悦田原常毅
Owner JILIN UNIV
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More