Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Model accelerated training method and device based on training data similarity aggregation

A technology of training data and training methods, applied in neural learning methods, biological neural network models, character and pattern recognition, etc., can solve problems such as slowing down model training efficiency, improve training efficiency, and reduce the risk of falling into local optimal solutions Probability, the effect of reducing the number of iterations

Inactive Publication Date: 2021-04-30
北京匠数科技有限公司
View PDF2 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, such an augmentation operation will add a lot of redundant and repeated information to the training data, which will slow down the efficiency of model training.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Model accelerated training method and device based on training data similarity aggregation
  • Model accelerated training method and device based on training data similarity aggregation
  • Model accelerated training method and device based on training data similarity aggregation

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0046] see figure 1 with figure 2, providing a model acceleration training method based on training data similarity aggregation, including the following steps:

[0047] S1: Randomly extract images of the first preset proportion in all training data as the current round of training data;

[0048] S2: Using a model training algorithm for the current round of training data to complete the current round of training and verification, and update the parameters of the image classification model;

[0049] S3: Use the image classification model after the parameter update to perform forward inference on the images in the remaining training data, and extract the training data inconsistent with the label in the inference result. Stop forward reasoning when the number is set;

[0050] S4: Perform similarity aggregation on the training data that is inconsistent with the label in the inference results extracted in step S3; extract a second preset proportion of training data images for ea...

Embodiment 2

[0064] see image 3 , the present invention also provides a model acceleration training device based on training data similarity aggregation, using the above-mentioned model acceleration training method based on training data similarity aggregation, including:

[0065] Training data extraction module 1, for randomly extracting the image of the first preset ratio in all training data as the current round of training data;

[0066] The model training module 2 is used to complete the training and verification of the current round using the model training algorithm for the training data of the current round, and update the parameters of the image classification model;

[0067] The forward inference module 3 is used to perform forward inference on the images in the remaining training data by using the image classification model after the parameter update, and extract the training data inconsistent with the label in the inference result. Stop forward reasoning when the inconsistent...

Embodiment 3

[0083] The present invention provides a computer-readable storage medium. The computer-readable storage medium stores program codes for accelerated training of models based on training data similarity aggregation. Instructions for a model acceleration training method based on training data similarity aggregation among possible implementations.

[0084] The computer-readable storage medium may be any available medium that can be accessed by a computer, or a data storage device such as a server, a data center, etc. integrated with one or more available media. The available medium may be a magnetic medium (for example, a floppy disk, a hard disk, or a magnetic tape), an optical medium (for example, DVD), or a semiconductor medium (for example, a solid state disk (SolidStateDisk, SSD)) and the like.

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a model accelerated training method and device based on training data similarity aggregation, and the method comprises the steps: taking a part of minimized training data as a starting point, extracting data with poor prediction from a prediction result of a current model in a mode of random sampling and random increment in each round, and sampling additional training data in a clustering extraction mode, therefore, the most representative training information is obtained, and the training efficiency of each round is improved. The data set scale of each round of model training is reduced, the training time is greatly shortened, clustering does not need an accurate result, the number of iterations can be reduced or a faster and simpler clustering method is used, and the total training time of each round is still much shorter than that of original full training set training on the whole; the training data selected in each round is targeted, the images with inference errors are selected for training, the back propagation gradient can be obtained to the maximum extent, the probability of falling into the local optimal solution during training is reduced, dynamic adjustment in the training process is facilitated, and the optimal training result is achieved.

Description

technical field [0001] The invention relates to the technical field of image detection, in particular to a model acceleration training method and device based on similarity aggregation of training data. Background technique [0002] Image classification and image detection technologies based on deep learning are widely used technologies in the field of artificial intelligence. Different from traditional methods, image classification and image detection technology need to be extracted based on rules, and then trained for feature vectors. The image detection technology based on deep learning can automatically extract image features through a multi-layer convolutional neural network. The model receives the original pixel matrix input of the image and obtains end-to-end image recognition results. The deep neural network model is trained from a large number of pictures, which has better generalization ability, and the accuracy and anti-interference ability of image classification...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06K9/62G06N3/08
CPCG06N3/08G06F18/23213G06F18/24
Inventor 张乐平侯磊张博李海峰王光曦
Owner 北京匠数科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products