Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Joint embedded model for zero sample learning

A sample learning and model technology, applied in computer parts, character and pattern recognition, instruments, etc., can solve problems such as high complexity, long training time, and difficulty in association, and achieve the effect of low complexity

Inactive Publication Date: 2016-07-06
TIANJIN UNIV
View PDF4 Cites 12 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The disadvantages of this type of method are: long training time, high complexity, and easy to fall into local optimum
Most of the shallow multimodal models are models composed of one or two layers of structure. The advantage of this method is that it is relatively simple and low in complexity. However, shallow models are used to correlate modal features on some complex data sets. there are still some difficulties

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Joint embedded model for zero sample learning
  • Joint embedded model for zero sample learning
  • Joint embedded model for zero sample learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0020] A joint embedding model for zero-shot learning of the present invention will be described in detail below with reference to embodiments and drawings.

[0021] figure 1 The main flow of a joint embedding model for zero-shot learning of the present invention is described. In the training phase, first extract features from the image and text, extract visual features from the image and use the language model to extract the text vector corresponding to the image from the corpus, and then use the algorithm provided by the present invention to learn to be able to associate different modal features The transformation matrix; in the test phase, first extract the image visual features of unseen categories, and then use the learned feature transformation matrix to transform the visual features into feature descriptions in the text feature space, and correspond to the text features closest to the transformed features The category of is used as the category of the test image.

[0...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

A joint embedded model for zero sample learning comprises the steps of: (1) inputting an image feature X of a training sample, a text feature Y corresponding to an image, and weight parameters alpha, beta and lambda; calculating the sum of all text feature vectors (FORMULA as shown in the description), then calculating a feature conversion matrix M (FORMULA as shown in the description), wherein I is a unit matrix; outputting the feature conversion matrix M. According to the joint embedded model for zero sample learning, a target function is constructed by using the correlativity between different modals so that the correlation of two modal features belonging to an identical category in public space is maximized and meanwhile the correlation of two modals belonging to different categories in public space is minimized. According to the joint embedded model for zero sample learning, a transition matrix is learned by using a training sample so that the similarities between features belonging to different modals can be compared with each other.

Description

technical field [0001] The present invention relates to a joint embedding model. In particular, it concerns a joint embedding model for zero-shot learning. Background technique [0002] With the development of the Internet, multi-modal data on the network continues to grow, and multi-modal learning has gradually become a research hotspot in machine learning and data mining. Multimodal learning is to establish relationships between data features of different modalities. Ideally, multimodal learning can integrate feature information of different modalities into a common representation space to achieve comparison and retrieval at the same semantic level. In the era of big data where multimodal data is constantly growing, with the increasing demand for mining technology for multimodal data, the traditional data mining technology for unimodal models can no longer meet people's requirements. How to mine effective information between different modalities is an important task in t...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06K9/62
CPCG06F18/241G06F18/214
Inventor 冀中于云龙
Owner TIANJIN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products