Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Information processing method and device and storage medium

An information processing method and technology of an information processing device, which are applied in the fields of electrical digital data processing, special data processing applications, natural language data processing, etc., can solve the problems of strong randomness of word embedding initialization scheme and long word embedding training period, etc.

Pending Publication Date: 2020-02-14
卓尔智联(武汉)研究院有限公司
View PDF7 Cites 4 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

In related technologies, there is a problem that the word embedding initialization scheme is highly arbitrary, which makes the word embedding training cycle long

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Information processing method and device and storage medium
  • Information processing method and device and storage medium
  • Information processing method and device and storage medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0041] In order to make the purpose, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below in conjunction with the drawings in the embodiments of the present invention. Obviously, the described embodiments It is a part of embodiments of the present invention, but not all embodiments.

[0042] In related technologies, word embedding models and training focus on how to design network structures and loss functions after obtaining randomly initialized word embedding results to obtain satisfactory embedding layer parameters. For example, the corresponding network structure in Word2Vec is to use the context vector to predict the central word vector or the central word vector to predict the context vector, and the corresponding network structure in Bert or Xlnet is to use the network visible words to predict hidden words or Predict the n...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The embodiment of the invention discloses an information processing method and device and a storage medium. The method comprises the steps of performing word segmentation processing on a first corpusto obtain a first word set of the first corpus; wherein the first word set comprises at least two words; determining the relevancy between any two words in the first word set; utilizing the determinedrelevancy between any two words in the first word set to construct a first word co-occurrence network; wherein the first word co-occurrence network represents an association relationship between words in the first word set; determining a first feature matrix by utilizing the data of the first word co-occurrence network and combining a graph convolutional neural network (GCN) model; wherein the first feature matrix is a feature matrix corresponding to each node in a first word co-occurrence network; and performing dimension reduction processing on the first feature matrix to obtain a word embedding initialization result. In this way, a reasonable word embedding initialization result can be provided, and therefore the effect of shortening the word embedding training period is achieved through the word embedding initialization result.

Description

technical field [0001] The present invention relates to the field of artificial intelligence, in particular to an information processing method, device and storage medium. Background technique [0002] Word embedding is a general term for language models and representation learning techniques in Natural Language Processing (NLP). Word embedding refers to embedding a high-dimensional space whose dimension is the number of all words into a continuous vector space with a much lower dimension, that is, mapping each word or phrase to a vector on a real number field. The way phrases are represented digitally is considered to be one of the most challenging problems of deep learning in NLP tasks. In related technologies, there is a problem that the word embedding initialization scheme is highly arbitrary, which makes the word embedding training period long. Contents of the invention [0003] In view of this, the embodiments of the present invention expect to provide an informati...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F40/289G06F40/126G06F16/35G06F16/36
CPCG06F16/35G06F16/36
Inventor 林田谦谨
Owner 卓尔智联(武汉)研究院有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products