Systems and methods for evaluating a loss function or a gradient of a loss function via dual decomposition

A loss function and gradient technique, applied in the field of computer systems for evaluating loss functions and/or their gradients, capable of solving a large number of operations, requiring linear running time, etc.

Pending Publication Date: 2021-02-12
GOOGLE LLC
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

In all such problems, the main bottleneck in training the model is the evaluation of the loss function and its gradient
Loss functions for this kind of problem typically need to enumerate all possible outputs and may require linear running time in the number of outputs used for evaluation
This can be a significant bottleneck in iterative methods such as gradient descent for training said model, since each step can be operation-intensive

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Systems and methods for evaluating a loss function or a gradient of a loss function via dual decomposition
  • Systems and methods for evaluating a loss function or a gradient of a loss function via dual decomposition
  • Systems and methods for evaluating a loss function or a gradient of a loss function via dual decomposition

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0028] In general, the present disclosure relates to systems and methods for evaluating loss functions or gradients of loss functions. For problems with large output spaces, evaluating the loss function and its gradients can be computationally expensive, typically taking linear time in the size of the output space. Recently, methods to accelerate learning via efficient data structures for nearest neighbor search (NNS) or maximum inner product search (MIPS) have been developed. However, the performance of such data structures usually degrades in high dimensions. The present disclosure provides systems and methods for reducing an intractable high-dimensional search problem to several much easier to solve lower-dimensional search problems via a dual decomposition of a loss function. The present disclosure further provides a greedy message passing technique that guarantees the convergence of the original loss. In this manner, the disclosed systems and methods can substantially i...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

Systems and methods for evaluating a loss function or a gradient of the loss function. In one example embodiment, a computer-implemented method includes partitioning a weight matrix into a plurality of blocks. The method includes identifying a first set of labels for each of the plurality of blocks with a score greater than a first threshold value. The method includes constructing a sparse approximation of a scoring vector for each of the plurality of blocks based on the first set of labels. The method includes determining a correction value for each sparse approximation of the scoring vector.The method includes determining an approximation of a loss or a gradient of a loss associated with the scoring function based on each sparse approximation of the scoring vector and the correction value associated with the sparse approximation of the scoring vector.

Description

technical field [0001] The present disclosure generally relates to evaluating loss functions or gradients of loss functions. More specifically, the present disclosure relates to computer systems and methods for efficiently evaluating loss functions and / or their gradients for problems with large output spaces via dual decomposition of loss functions. Background technique [0002] Large output spaces are ubiquitous in several machine learning problems today. Such machine learning problems can include, for example, extreme multiclass or multilabel classification problems with many classes, language modeling with large vocabularies, or metric learning with large pairwise distance constraints. In all such problems, the main bottleneck in training the model is the evaluation of the loss function and its gradient. Loss functions for such problems typically need to enumerate all possible outputs and may require linear running time in the number of outputs used for evaluation. Thi...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06F17/11G06N20/00
CPCG06F17/11G06N3/084G06N3/082G06N5/01G06N3/044G06N3/045G06N20/00G06F17/17
Inventor 萨泰恩·钱德拉坎特·卡勒丹尼尔·霍尔特曼-赖斯桑吉夫·库马尔闫恩旭于信男
Owner GOOGLE LLC
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products