Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

165results about How to "Efficient training process" patented technology

Reinforcement learning training optimization method and device for multi-agent confrontation

The embodiment of the invention provides a reinforcement learning training optimization method and device for multi-agent confrontation. The method comprises the following steps of a rule coupling algorithm training process including the steps of acquiring an initial first state result set of a red party multi-agent for each training step, if the initial first state result set of the red party multi-agent meets a preset action rule, obtaining a decision-making behavior result set according to the preset action rule, and otherwise, acquiring the decision-making behavior result set according toa preset reinforcement training learning algorithm; and performing reinforcement learning training on the red-party multi-agent by utilizing a training sample formed by the decision-making behavior result set and other preset parameters. The embodiment of the invention provides the reinforcement learning training optimization method and device for multi-agent confrontation. In the whole training process, the preset action rule can guide the multiple agents to act, invalid actions are avoided, the problems that in the training process in the prior art, invalid exploration is much, and the training speed is low are solved, and the training efficiency is remarkably improved.
Owner:NAT INNOVATION INST OF DEFENSE TECH PLA ACAD OF MILITARY SCI

Training method and device for financial risk identification model, computer equipment and medium

The invention discloses a training method and device for a financial risk identification model, computer equipment and a medium, and the method comprises the steps: obtaining first user data, with a first credit risk label, of a target domain financial project, inputting the first user data into a meta-learning device for training, and obtaining a classifier corresponding to the target domain financial project for risk identification. According to the method, the classifier corresponding to the target domain financial project is trained in a meta-learning mode, priori knowledge in a source domain task can be effectively migrated, so that the data volume of labeled samples required by model training is small, the generalization performance of the recognition model is improved, and the training process of the model is faster and more efficient. Besides, in the training process of the meta-learner, the source domain correlation among the categories of the task sets is learned, so that prior knowledge can be effectively migrated from tasks closer to the current target domain task during migration learning, and the accuracy of model recognition can be improved. The method can be widelyapplied to the technical field of machine learning.
Owner:TENCENT TECH (SHENZHEN) CO LTD

Target detection method for SAR (Synthetic Aperture Radar) image based on deep reinforcement learning

The invention relates to a target detection method for an SAR (Synthetic Aperture Radar) image based on deep reinforcement learning, which comprises the steps of S1, setting the number of iterations,and sequentially processing images in a training set in each iteration process; S2, inputting an image from the training set, and generating training samples by using the Markov decision process; S3,randomly selecting a certain number of samples, training the Q-network by adopting a gradient descent method, obtaining the state of a reduced observation area, generating a next sample until a presettermination condition is met, and terminating the image processing process; S4, returning back to the step S2, continuing to input the next image from the training set until all images are completelyprocessed, and terminating the current iteration process; S5, continuing the next iteration process until the set number of iterations is met, and determining network parameters of the Q-network; andS6, performing target detection on an image in a test set through the trained Q-network, and outputting a detection result. The target detection method obtains good detection accuracy in target detection for the SAR image.
Owner:BEIHANG UNIV

Rotary machine residual life prediction method of multi-layer bidirectional gating cycle unit network

ActiveCN110866314AEfficiently obtain confidence intervalsHigh precisionGeometric CADNeural learning methodsHealth indexFeature extraction
The invention discloses a rotary machine residual life prediction method of a multi-layer bidirectional gating cycle unit network. The rotary machine residual life prediction method comprises the following steps: acquiring a vibration signal; constructing health indexes; constructing a network training set; constructing a multi-layer bidirectional gating circulation unit network; training a multi-layer bidirectional gating cycle unit network; and carrying out network testing, residual life estimation and confidence interval acquisition; and performing residual life prediction evaluation. According to the rotary machine residual life prediction method, the advantages of strong feature extraction capability of deep learning are combined, and regression prediction is carried out by utilizinga bidirectional gated cycle unit neural network, and the confidence interval of the residual life is obtained through a Bootstrap method. Aiming at the problems that the model precision is sensitive to the value of the learning rate in the training process of a recurrent neural network model, the prediction performance of the model is influenced by too high and too low values, and the neural network is efficiently trained by utilizing the natural exponential decay learning efficiency. The rotary machine residual life prediction method can accurately predict the residual life and confidence interval of the rotary machine, can greatly reduce the expensive unplanned maintenance, and avoids the occurrence of a big disaster.
Owner:SOUTHEAST UNIV

A convolutional echo state network time sequence classification method based on a multi-head self-attention mechanism

InactiveCN109919205AKeep training freeImplement Recode IntegrationCharacter and pattern recognitionNeural architecturesSpacetimeSelf attention
The invention discloses a convolutional echo state network time sequence classification method based on a multi-head self-attention mechanism. The method is realized based on two components of an encoder and a decoder. The encoder is composed of an echo state network of a multi-head self-attention mechanism. An input time sequence is firstly subjected to high-dimensional mapping coding through a plurality of echo state networks to generate original echo state characteristics, and then re-coding of echo state characteristic global space-time information is achieved based on a self-attention mechanism, so that the re-integrated high-dimensional characteristic representation has higher discrimination capability; And finally, taking a shallow convolutional neural network as a decoder to realize high-precision classification. The classification model established by the method inherits the characteristic of high training efficiency of the echo state network, realizes the re-coding of the space-time characteristic information by introducing a self-attention mechanism model without additional parameters, achieves a high-precision time sequence classification effect, and is a simple and efficient model.
Owner:SOUTH CHINA UNIV OF TECH

Classification model application and classification model training method and device

The invention provides a classification model application and a classification model training method and device, and the method comprises the steps: obtaining a to-be-processed corpus, and convertingthe to-be-processed corpus into a word unit sequence; determining a first occurrence frequency of each keyword in a pre-constructed keyword set in the word unit sequence, and generating a first feature vector based on the first occurrence frequency corresponding to each keyword; determining a second occurrence frequency of each word in the word unit sequence in the word unit sequence, and generating a second feature vector based on the second occurrence frequency of each word and a preset inverse document frequency of each word; inputting the first feature vector into a first classification model, and outputting a first classification result of the to-be-processed corpus; and inputting the second feature vector into a second classification model, and outputting a second classification result of the to-be-processed corpus, and determining the category of the to-be-processed corpus based on the first classification result and the second classification result. By means of the method, thecorpus classification accuracy can be improved.
Owner:NEW H3C BIG DATA TECH CO LTD

Distributed depth learning system based on momentum and pruning

The invention provides a distributed depth learning system based on momentum and pruning, which relates to the field of cloud computing and depth learning. A spark distributed cluster is adopted, comprising a main control node and a plurality of working nodes. Any working node is connected with the main control node through a communication link, and the working node stores part of training data. According to the batch of training data stored locally, the work node propagates the training of the depth learning model forward and backward, obtains the update quantity of the weight parameter of the work node, and sends it to the main control node for interactive communication with the global weight parameter of the main control node. The main control node records the node information of the work node, balances the weight parameters of the obtained work node and transmits the obtained weight parameters to each work node. Repeating the interactive communication between the master control node and the working node until the iterative number or convergence condition is reached; The invention solves the problem that the model of the asynchronous algorithm converges slowly under the distributed environment, and at the same time, the operation speed of the asynchronous algorithm is improved.
Owner:ANHUI UNIVERSITY OF TECHNOLOGY +1

Image super-resolution method based on active sampling and Gaussian process regression

The present invention discloses an image super-resolution method based on active sampling and Gaussian process regression, which mainly aims to solve the problem of poor super-resolution effects of a texture region in the prior art. According to the method, a training set is constructed by fully using a large number of existing external natural images, and a sample information amount is measured by using characterization and diversity as indexes, so as to extract a simplified training subset, which enables training of a Gaussian process regression model to be more efficient. The implementation steps are as follows: 1. generating an external training sample set, and performing active sampling on the external training sample set to obtain a training subset; 2. learning a Gaussian process regression model based on the training subset; 3. preprocessing a test image and generating a test sample set; and 5. applying the learned Gaussian process regression model on the test sample set, and predicting and outputting a super-resolution image. The method has a relatively strong super-resolution ability and is capable of restoring more detail information in regions such as texture, so that the method can be used for video monitoring, criminal investigation, aerospace, high definition entertainment and video or image compression.
Owner:XIDIAN UNIV

Artificial neural network predicting method of amorphous alloy thermoplasticity forming performance

ActiveCN108256689AImproved thermoplastic formabilityShorten the timeForecastingNeural learning methodsIndex testPredictive methods
The invention belongs to the field of prediction of amorphous alloy thermoplasticity forming performance, and discloses an artificial neural network predicting method of the amorphous alloy thermoplasticity forming performance. The predicting method comprises the following steps of a, selecting multiple performance parameters and collecting data of the performance parameters, dividing the data into a training sample, a verification sample and a to-be-predicted sample, and testing to obtain feature index test values corresponding to the training sample and the verification sample; b, selectingan artificial neural network model as an initial predicting model for the amorphous alloy thermoplasticity forming performance, adopting the training sample to train the artificial neural network model, and determining an improved predicting model; c, adopting the verification sample to verify the improved predicting model, finally obtaining a final predicting model, and adopting the final predicting model for prediction. By means of the artificial neural network predicting method of the amorphous alloy thermoplasticity forming performance, the amorphous alloy thermoplasticity forming performance is effectively predicted without experiments, guidance is provided for development of an amorphous alloy system suitable for thermoplasticity forming, the time for developing the new amorphous alloy system is greatly shortened, and the money cost for developing of the new amorphous alloy system is greatly reduced.
Owner:HUAZHONG UNIV OF SCI & TECH

Resource management method and system for large-scale distributed deep learning

The invention discloses a resource management method and system for large-scale distributed deep learning, and the method and system achieve the memory resource optimization management of parameters and gradient intermediate data during the training operation of a neural network, and guarantee the reasonable configuration of distributed communication bandwidth resources. Cross-layer memory reuse is realized again, intermediate data required by iterative computation and sparse communication are migrated into a CPU main memory and then migrated back as required, and interlayer memory consumptionis reduced; on the basis of reasonable migration of CPU-GPU data, independence of intra-layer memory reuse, intra-layer calculation mining and memory access operation is achieved, and intra-layer memory consumption is reduced as much as possible. Distributed parameter communication optimization is realized while efficient utilization of memory resources is ensured. Data access in the distributedparameter updating stage is reasonably redirected, a CPU main memory serves as a mirror image access area, the data access to parameters and gradients is completed, and the problems of gradient data missing and parameter writing border crossing are solved.
Owner:HUAZHONG UNIV OF SCI & TECH

LIBS multi-component quantitative inversion method based on deep learning convolutional neural network

ActiveCN110705372AStrong ability to deal with complex nonlinear problemsHigh Quantitative Analysis AccuracyCharacter and pattern recognitionAnalysis by thermal excitationFeature extractionAlgorithm
The invention discloses an LIBS multi-component quantitative inversion method based on a deep learning convolutional neural network, and is suitable for the field of laser spectrum analysis. Accordingto the method, the unique advantages of the convolutional neural network algorithm in the aspect of image feature recognition are utilized, and the method is applied to LIBS spectrum quantitative inversion. According to the convolutional neural network construction scheme designed by the invention, feature extraction and deep learning can be performed on the LIBS spectral line form of the sample,and after the convolutional neural network is trained by using the LIBS spectrum of the known sample, the network can analyze and predict the contents of various chemical components of the unknown sample at the same time. The method has the advantages of being simple and convenient to operate, efficient in training, high in accuracy and good in robustness, is suitable for quantitatively analyzingthe LIBS spectrum, and is particularly suitable for analyzing the LIBS spectrum with relatively high spectral line form complexity and relatively large interference noise.
Owner:SHANGHAI INST OF TECHNICAL PHYSICS - CHINESE ACAD OF SCI

Word vector generation method based on Gaussian distribution

ActiveCN108733647AAvoid Point Estimation FeaturesFix the problem of assuming a fixed number of sensesSemantic analysisCharacter and pattern recognitionInclusion relationAlgorithm
The present invention discloses a word vector generation method based on the Gaussian distribution. The method comprises: firstly, preprocessing the corpus; secondly, using the punctuation to performtext division on the corpus; then combining the local and global information to infer the word meaning, and determining the mapping relationship between the word and the word meaning; and finally, obtaining a word vector by optimizing the objective function. The innovations and beneficial effects of the technical scheme of the present invention are as follows that: 1, words are represented based on the Gaussian distribution, point estimation characteristics of traditional word vectors are avoided, and more abundant information such as probabilistic quality, meaning connotation, an inclusion relationship, and the like can be brought to the word vectors; 2, multiple Gaussian distributions are used to represent the words, so that the linguistic characteristics of a word in the natural language can be coopered with; and 3, the similarity between the Gaussian distributions is defined based on the Hellinger distance, and by combining parameter updating and word meaning discrimination, the number of word meanings can be inferred adaptively, and the problem that the number of hypothetical word meanings of the model in the prior art is fixed is solved.
Owner:SUN YAT SEN UNIV

Watermark recognition online training, sample preparation and removal methods, watermark recognition online training device, equipment and medium

The invention discloses watermark recognition online training, sample preparation and removal methods, a watermark recognition online training device, equipment and a medium. The training method comprises the steps: extracting a background image from a background data set, and randomly cutting the background image into a preset specification; extracting a watermark original image from the watermark data set, and performing image deformation processing on the watermark original image to form a watermark enhanced image; synthesizing the watermark enhancement image and the background image to obtain a corresponding synthesized image; and feeding the composite image as a training sample into a preset image segmentation model which has a coding and decoding structure and is suitable for capturing multi-scale features to implement sample training, so that the image segmentation model is suitable for removing the watermark original image from the to-be-identified image through the training. According to the invention, the watermark-containing training sample can be automatically synthesized on line and immediately used for training the image segmentation model, so that the model has the capability of identifying and removing the watermark in the picture.
Owner:GUANGZHOU HUADUO NETWORK TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products