Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

349 results about "Model compression" patented technology

Progressive block knowledge distillation method for neural network acceleration

The invention discloses a progressive block knowledge distillation method for neural network acceleration. The method specifically comprises the following steps of: inputting an original complex network and related parameters; dividing the original complex network into a plurality of sub-network blocks, designing student sub-network blocks and randomly initializing the parameters; taking the inputoriginal complex network as a teacher network of the first block distillation process and obtaining a student network after the block distillation process is completed, wherein the first student sub-network block has the optimum parameters; taking the student network obtained in the last block distillation process as a teacher network of the next block distillation process so as to obtain a nextstudent network, wherein the student sub-network blocks, the block distillation of which is finished, have the optimum parameters; and obtaining a final simple student network and optimum parameters after all the sub-network block distillation processes are completed. The method is capable of achieving the effect of accelerating model compression on common hardware architecture, is simple to realize, and is an effective, practical and simple deep network model compression acceleration algorithm.
Owner:ZHEJIANG UNIV

Structured network model compression acceleration method based on multistage pruning

The invention discloses a structured network model compression acceleration method based on multistage pruning, and belongs to the technical field of model compression acceleration. The method comprises the following steps: obtaining a pre-training model, and training to obtain an initial complete network model; measuring the sensitivity of the convolution layers, and obtaining a sensitivity-pruning rate curve of each convolution layer through controlling variables; carrying out single-layer pruning from low to high according to a sensitivity sequence, and finely tuning and re-training a network model; selecting a sample as a verification set, and measuring the information entropy of the filter output feature map; performing iterative flexible pruning according to the size sequence of theoutput entropy, and finely tuning and re-training the network model; and hard pruning: carrying out retraining on the network model to recover the network performance, and obtaining and storing a lightweight model. According to the method, the large-scale convolutional neural network can be compressed on the premise of maintaining the original network performance, the local memory occupation of the network can be reduced, the floating point operation and the video memory occupation during operation are reduced, and the lightweight of the network is realized.
Owner:UNIV OF ELECTRONICS SCI & TECH OF CHINA

Model compression method based on pruning sequence active learning

The invention provides a model compression method based on pruning sequence active learning. an end-to-end pruning framework based on sequential active learning is provided; The method has the advantages that the importance of all layers of the network can be actively learned, the pruning priority is generated, a reasonable pruning decision is made, the problem that an existing simple sequential pruning method is unreasonable is solved, pruning is preferentially carried out on the network layer with the minimum influence, pruning is carried out step by step from simplification to difficulty, and the model precision loss in the pruning process is minimized; And meanwhile, the final loss of the model is taken as a guide, and the importance of the convolution kernel is evaluated in a multi-angle, efficient, flexible and rapid manner, so that the compression correctness and effectiveness of the whole-process model are ensured, and technical support is provided for subsequent transplantation of a large model to portable equipment. Experimental results show that the model compression method based on pruning sequence active learning provided by the invention is leading under the conditions of multiple data sets and multiple model structures, can greatly compress the model volume under the condition of ensuring the model precision, and has a very strong practical application prospect.
Owner:TSINGHUA UNIV

Large-scale three-dimensional scene webpage display method based on model compression and asynchronous loading

The invention provides a large three-dimensional scene webpage display method based on model compression and asynchronous loading, and relates to the technical field of computer graphics. The method comprises the following steps of 1, obtaining a two-dimensional picture of a scene to be reconstructed, and generating a three-dimensional model of the scene to be reconstructed by utilizing a three-dimensional reconstruction technology; 2, dividing a point cloud space in the model into small cubes according to a three-dimensional model, and forming structured data by using an octree; 3, calculating the node size and depth of the octree model; 4, constructing an octree model; 5, compressing the octree model requested by the user at the server side, and transmitting the compressed octree model to a browser for rendering display through a network; and 6, dynamically loading the nodes of the octree model on the webpage according to the hierarchical detail technology, and rendering the nodes atthe same time to finally obtain the three-dimensional scene graph of the scene to be reconstructed. According to the method, the real-time performance of model rendering in webpage model display is ensured, and the waiting time of a user is greatly shortened.
Owner:NORTHEASTERN UNIV

Long-distance pipeline inspection method based on YOLOv3 pruning network and deep learning defogging model

ActiveCN111461291AImproving the accuracy of subsequent target detection tasksImprove efficiencyImage enhancementInternal combustion piston enginesImaging processingAlgorithm
The invention belongs to an image processing technology based on deep learning, and particularly relates to a long-distance pipeline inspection method based on a YOLOv3 pruning network and a deep learning defogging model. The method comprises the following steps: 1, constructing and training an AOD-Net defogging network model; 2, designing a YOLOv3 backbone network and a loss function; 3, performing image data acquisition and training on the target area in an unmanned aerial vehicle inspection mode; 4, performing compression and accelerated calculation on the YOLOv3 model through a scaling factor gamma pruning method based on a BN layer; 5, deploying the AOD-Net and YOLOv3 joint model to an embedded module of the unmanned aerial vehicle for target task detection; and 6, returning the inspection task detection result of the long-distance pipeline of the unmanned aerial vehicle to the background system in real time. The system is used for being deployed on an unmanned aerial vehicle embedded module to perform long-distance pipeline inspection work, and the labor cost is greatly reduced while high detection precision, good real-time performance and high efficiency are guaranteed.
Owner:XIAN UNIV OF SCI & TECH

Classification and identification method for low, slow small targets

A disclosed classification and identification method for low, slow and small targets is accurate in identification and short in identification time. The method is realized through the following technical scheme: different target track data is taken as a training sample of a data preprocessing module based on a PPI image formed by an original trace point; the data preprocessing module performs dataprediction and preprocessing on an obtained track-related plot data rule, generates a training set, constructs a deep learning network model and a network optimization module which sequentially adopta deep convolutional network DCNN and a long short-term memory network LSTM, extracts two groups of features from framing data, performs splicing to obtain joint features, and performs deep learning;image recognition target track data is input into a weighting calculation module in real time, model compression and acceleration are conducted on a trained network model structure, model compressionacceleration is achieved through weighting calculation and model pruning, depth features needed by accurate classification are achieved, and accurate recognition of classification and recognition ofsmall birds and unmanned aerial vehicles is completed.
Owner:LINGBAYI ELECTRONICS GRP

Flexible deep learning network model compression method based on channel gradient pruning

The invention discloses a flexible deep learning network model compression method based on channel gradient pruning, and the method comprises the steps: 1, adding a masking layer constraint to an original network, and obtaining a to-be-pruned deep convolutional neural network model; wherein the absolute value of the product of the channel gradient and the weight serves as an importance standard toupdate the masking layer constraint of the channel to obtain a mask and a sparse model, 3, carrying out pruning operation on the sparse model based on the mask, and 4, retraining a compact deep convolutional neural network model. The invention further provides an application effect of the flexible deep learning network model compression method based on channel gradient pruning on an actual objectrecognition APP, the recognition speed of the model to the object after pruning is greatly improved, and the problem that the deep neural network model cannot be applied to the actual object recognition APP due to high storage space occupation and high memory occupation, high computing resources are occupied and cannot be deployed to embedded devices, smart phones and other devices are solved, and the application range of the deep neural network is expanded.
Owner:ZHEJIANG UNIV OF TECH

Method and device for generating neural network model, electronic equipment and storage medium

The embodiment of the invention discloses a method and device for generating a neural network model, electronic equipment and a storage medium, and relates to the technical field of artificial intelligence, deep learning and image processing. According to the specific implementation scheme, the method comprises the steps: executing a plurality of iterative search operations, wherein the iterativesearch operation comprises the following steps of: determining a target compression strategy of a preset neural network model in a search space of a preset compression strategy by adopting a preset compression strategy controller, wherein the compression strategy comprises a pruning and quantifying combined strategy; according to the target compression strategy, pruning and quantifying a preset neural network model to obtain a current compressed model and obtain the performance of the current compressed model; and generating feedback information based on the performance of the compressed model, and determining the current compressed model as a generated target neural network model in response to determining that the feedback information reaches a preset convergence condition. According tothe method, the optimal model compression strategy can be searched.
Owner:BEIJING BAIDU NETCOM SCI & TECH CO LTD

Model compression method and device

ActiveCN110163367AAvoid the influence of human experienceImprove compression efficiencyNeural learning methodsAlgorithmCoincidence
The embodiment of the invention discloses a model compression method and device. The method comprises the steps of obtaining a to-be-compressed model and compression preference configuration for the to-be-compressed model; determining a compression algorithm component and a corresponding algorithm hyper-parameter value according to the model type and the compression preference configuration of a model to be compressed, and using the compression algorithm component and the algorithm hyper-parameter value to carry out first compression on the to-be-compressed model to obtain a candidate compression result corresponding to the first compression; if the coincidence degree of the performance parameter of the candidate compression result corresponding to the first compression and the compressionpreference configuration does not meet the preset condition, executing the second compression; and continuing to generate a parameter adjustment strategy to adjust the compression algorithm componentand the algorithm hyper-parameter value used by the next compression until the coincidence degree of the performance parameter of the candidate compression result corresponding to the certain compression and the compression preference configuration meets a preset condition. A compression algorithm does not need to be adjusted manually, so that the influence caused by human experience is avoided,and the compression efficiency is improved.
Owner:TENCENT TECH (SHENZHEN) CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products