Federal mutual learning model training method for non-independent identically distributed data

A distributed data and learning model technology, which is applied in the field of federated mutual learning model training for non-independent and identically distributed data, can solve the problems of high training performance, lack of communication costs, and reduce direct communication between clients and cloud servers to achieve shortened models the effect of time

Pending Publication Date: 2022-02-25
BEIJING INSTITUTE OF TECHNOLOGYGY
View PDF0 Cites 5 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0008] 1. The neural network will generate a large number of model parameters, and bandwidth limitations will affect the transmission of parameters and model training. Compared with cloud-based federated learning, the client-edge-cloud hierarchical federated learning system can reduce the number of clients by introducing an intermediate edge server. Direct communication with the cloud server reduces model training time, but still needs to transmit a large number of parameters, and there is still a lack of solutions that can reduce communication costs
[0009] 2. In recent years, knowledge distillation technology has been combined with federated learning, but when dealing with non-independent and identically distributed data, it is necessary to pre-generate a teacher model on a suitable proxy data set, but it is difficult to find a suitable data set to generate a teacher model. At present, there is still a lack of federated learning methods to solve the high communication cost and training performance loss encountered in non-independent and identically distributed data

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Federal mutual learning model training method for non-independent identically distributed data
  • Federal mutual learning model training method for non-independent identically distributed data
  • Federal mutual learning model training method for non-independent identically distributed data

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0066] According to the method steps provided in this embodiment, a model training method for solving non-IID federated learning is introduced, taking the MNIST image classification task as an example. Each client is a data owner, at least one terminal participates in training, and intermediate clients participate. The sample data held by each data owner can be the same data set or different data sets. The server and client databases store the MNIST data set. After manual segmentation, the edge clients in each group only contain one or more disjoint data labels, satisfying the situation of non-independent and identical distribution. The server, intermediate client, and edge client use the LeNet-5 convolutional neural network (the input layer dimension is 28×28, the convolution layer includes six 5×5 convolution kernels, and the maximum pooling layer includes a 2 ×2 kernel, the convolution layer includes 16 5×5 convolution kernels, the maximum pooling layer includes 1 2×2 kern...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a federal mutual learning model training method for non-independent identically distributed data, which comprises the following steps: S1, sending initial global model parameters to an intermediate client, generating intermediate client model parameters by the intermediate client, and S2, generating edge client model parameters by an edge client by using a local data set; S3, enabling the intermediate client and the edge client to update parameters by using a mutual learning method; s4, uploading a probability prediction value output by the intermediate client model to a server, and updating the global model and the intermediate client model by the server by utilizing a distillation technology; and S5, executing the steps S3-S4 until the models meet convergence conditions to obtain a final intermediate client model, a final edge client model and a final global model, and then broadcasting the final global model to all edge clients by the server. The problems of federated learning communication bandwidth limitation and model generation of the non-independent identically distributed data are solved through a grouping mutual learning and knowledge distillation method.

Description

technical field [0001] The invention relates to the technical field of federated learning, in particular to a federated mutual learning model training method for non-independent and identically distributed data. Background technique [0002] With the development of artificial intelligence and big data, information resources can achieve high-speed transmission in distributed networks, realizing the global management of physical information. Cloud computing and edge computing have promoted the development of deep learning. Deep learning models usually contain millions of parameters. Usually, expanding the scale of the neural network can effectively improve the accuracy of the model. However, in a distributed network, resource-constrained edge devices cannot When deploying a large-scale neural network model, there will be communication delays and the rejection of low-priority user terminal access. At this time, the powerful computing power and storage space of the central serve...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/08G06N20/00G06N3/04
CPCG06N3/08G06N20/00G06N3/045
Inventor 李侃李洋
Owner BEIJING INSTITUTE OF TECHNOLOGYGY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products