Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

32results about How to "Fast inference" patented technology

Real-time optical flow estimation method based on lightweight convolutional neural network

The invention discloses a real-time optical flow estimation method based on a lightweight convolutional neural network, and the method comprises the steps: giving two adjacent frames of images, and constructing a multi-scale feature pyramid with shared parameters; on the basis of the constructed feature pyramid, constructing a U-shaped network structure of a first frame of image by adopting a deconvolution operation to perform multi-scale information fusion; initializing the lowest resolution optical flow field to be zero, and performing deformation operation based on bilinear sampling on a second frame matching feature after the optical flow estimated by the second low resolution is up-sampled; carrying out local similarity calculation based on an inner product on the features of the first frame and the deformed features of the second frame, constructing a matching cost, and carrying out cost aggregation; taking the multi-scale features, the up-sampled optical flow field and the matching cost features after cost aggregation as the input of an optical flow regression network, and estimating the optical flow field under the resolution; and repeating until the optical flow field under the highest resolution is estimated. According to the invention, optical flow estimation is more accurate, and the model is lightweight, efficient, real-time and rapid.
Owner:SHANGHAI JIAO TONG UNIV

Magic cube color identification method based on artificial neural network

InactiveCN108830908AAvoid the disadvantages of being affected by lightIntelligent recognition processingImage enhancementImage analysisData setArray data structure
The invention discloses a magic cube color identification method based on an artificial neural network. The method includes steps: establishing an artificial neural network training model; acquiring pictures of surfaces of a magic cube, marking color blocks along contours of the color clocks of the magic cube on the pictures by employing rectangular marking frames, marking color categories of thecolor blocks according to category numbers, storing coordinates of upper left corners and lower right corners of the rectangular marking frames and the corresponding category numbers in marking filescorresponding to the pictures; training the artificial neural network training model by employing the marking files and a data set formed by the corresponding pictures, and forming a network model file stored in a mobile platform; and during identification, acquiring pictures of surfaces of a to-be-processed magic cube according to the sequence by employing a camera by the mobile platform, retrieving the network model file for identification to identify a determination picture of the surface at each time of acquisition of the picture of one surface, outputting a color number group of the surface, and finally forming two-dimensional color rectangular output to complete identification. According to the method, the identification rate in a complex environment is increased by introduction of amagic cube robot color identification algorithm into the neural network.
Owner:TIANJIN UNIV

INT8 offline quantization and integer inference method based on Transform model

The invention provides an INT8 offline quantization and integer inference method based on a Transform model. The INT8 offline quantization and integer inference method based on the Transform model comprises the following steps: converting an L2 norm of a normalization layer in an original Transform floating point model into an L1 norm; carrying out model training; performing forward inference through a small amount of data to obtain a quantization coefficient of input data of each layer of matrix operation, and extracting the quantization coefficient as general floating point data; obtaining a weight quantization coefficient of each linear layer in the floating point model, extracting the weight quantization coefficient as general floating point data, and determining an optimal weight quantization coefficient in each layer according to a mean square error calculation method; converting quantization coefficients related to quantization operation in the inference process into 2-n floating-point number forms, and adjusting the quantization coefficients through a joint coefficient adjustment method; and obtaining an integer inference model of INT8 based on the adjusted quantization coefficient in combination with a normalization layer of an L1 norm. According to the invention, errors caused by hardware resources required by model calculation and model quantification can be reduced, hardware resource consumption is reduced, and the deduction speed of the model is increased.
Owner:SOUTH CHINA UNIV OF TECH

Terminal equipment heterogeneous processor inference acceleration method under temperature constraint

The invention provides a terminal equipment heterogeneous processor inference acceleration method under temperature constraint, aims at intelligent terminal equipment equipped with a plurality of heterogeneous processors in an industrial production environment, and solves the problem of low terminal equipment inference efficiency caused by deep neural network interlayer heterogeneity, processor heterogeneity and environment temperature. According to the method, firstly, the environment temperature of industrial production and the power of a terminal device processor are considered, a terminal device dynamic frequency model under temperature constraint is established, and a temperature sensing dynamic frequency algorithm is used for setting the device frequency; secondly, designing a deep neural network single-layer parallel method according to calculation modes and structural characteristics of different layers in the deep neural network; and finally, a heterogeneous processor in the terminal equipment is utilized to design a deep neural network single-layer calculation task allocation method oriented to the heterogeneous processor, and low delay and robustness of collaborative inference of the heterogeneous processor of the terminal equipment are guaranteed.
Owner:SOUTHEAST UNIV +1

Aircraft detection and tracking method based on multi-scale self-adaption and side domain attention

The invention discloses an aircraft detection and tracking method based on multi-scale self-adaption and side domain attention, and the method comprises the steps: extracting an original feature map of a preprocessed original aircraft image through constructing a basic feature extraction network, and extracting a small-target feature map in the original feature map through combining with a small-size target branch network model; and obtaining a detection target feature map set and feature vectors corresponding to the detection target feature maps according to the small target feature maps by using a target prediction model, and performing aircraft detection and tracking by using a multi-aircraft tracking algorithm. Fusion transmission of shallow texture features and deeper semantic features of the feature map is optimized by using a coding and decoding structure and residual connection, the inference speed is improved, information fusion is more sufficient, and the feature extraction capability of a network model is effectively improved by combining a side domain attention mechanism network; and the small-size target branch network model is utilized to reduce the information loss degree, the small-size target detection accuracy is effectively optimized, and the management efficiency of airport scene aircrafts is improved.
Owner:UNIV OF ELECTRONIC SCI & TECH OF CHINA

Attribute recognition method based on knowledge distillation, terminal equipment and storage medium

The invention relates to an attribute recognition method based on knowledge distillation, terminal equipment and a storage medium. The method comprises the steps of collecting question and answer data of different fields to form a training set; constructing a teacher network model, and training the teacher network model through the training set; according to knowledge maps of different fields, extracting all attributes corresponding to each entity; performing entity identification on the question and answer data in the training set, obtaining all attributes possibly corresponding to the question and answer data according to the identified entities and all attributes corresponding to the entities, and obtaining priori knowledge; constructing a student network model, training the student network model through the training set based on knowledge distillation and prior knowledge, and obtaining a final student network model until performance indexes meet requirements; and using the final student network model to carry out attribute identification on the question and answer data. The method overcomes the defect that the precision and speed of a traditional attribute recognition method cannot be both considered.
Owner:厦门渊亭信息科技有限公司

A dynamic vocabulary enhancement combined model distillation method

ActiveCN112699678AWill not increase the burdenMake up for the problem of inaccurate semantic understandingNatural language data processingMachine learningAlgorithmEngineering
The invention relates to the technical field of natural language processing in the field of artificial intelligence, and discloses a dynamic vocabulary enhancement combined model distillation method, which comprises the following steps: on the basis of an ALBert language model, adjusting the language model by combining a fine adjustment technology with a dynamic vocabulary enhancement technology to obtain a finely adjusted language model, and taking the finely adjusted language model as a teacher model; different from the conventional fine adjustment logic, when the language model is finely adjusted, in the fine adjustment process, combining the characteristics of the dictionary information with the output characteristics of the language model, and then performing fine adjustment; and after fine adjustment is finished, distilling the teacher model, and taking an obtained model prediction result as a training basis of the student model. According to the model distillation method provided by the invention, the dictionary information is introduced as the key information, so that the model can still capture the dictionary information as a feature under the condition of greatly reducing the size, thereby achieving the purposes of greatly reducing the size of the model and accelerating the inference speed under the condition of not sacrificing the extraction accuracy.
Owner:达而观数据(成都)有限公司

A Model Distillation Method Combined with Dynamic Vocabulary Augmentation

ActiveCN112699678BWill not increase the burdenMake up for the problem of inaccurate semantic understandingNatural language data processingMachine learningEngineeringData mining
The invention relates to the technical field of natural language processing in the field of artificial intelligence, and discloses a model distillation method combined with dynamic vocabulary enhancement, including: on the basis of the ALbert language model, the language model is adjusted by fine-tuning technology combined with dynamic vocabulary enhancement technology , get the fine-tuned language model, and use it as the teacher model; when fine-tuning the language model, it is different from the conventional fine-tuning logic. In the fine-tuning process, the features of the dictionary information are first combined with the output features of the language model, and then Then fine-tune; after fine-tuning, the teacher model is distilled, and the obtained model prediction results are used as the training basis for the student model. The model distillation method provided by the present invention introduces dictionary information as key information, so that the model can still capture dictionary information as features in the case of greatly reducing the size, so as to greatly reduce the size of the model without sacrificing the accuracy of extraction. The purpose of inferring speed.
Owner:达而观数据(成都)有限公司

Sequence recommendation method and system based on adaptive network depth

The invention discloses a sequence recommendation method and system based on adaptive network depth. The method comprises the steps that a sequence recommendation model is constructed, and the sequence recommendation model is provided with a plurality of cavity convolution residual blocks as a main body network and is provided with a strategy network used for managing the depth of the main body network; taking a set loss function as a target, training the sequence recommendation model by using a sample set to obtain a trained main body network, and for each of the plurality of hole convolutionresidual blocks, outputting a decision indication used for representing reservation or skipping of the hole convolution residual block by the strategy network; and inputting the historical browsing sequence of the to-be-recommended user into the trained sequence recommendation model, and determining a hole convolution residual block needing to be skipped according to the decision indication of the strategy network so as to output a prediction result of the user recommendation item at the subsequent moment. According to the invention, the depth of the main network can be adaptively adjusted byusing the strategy network, and quick and accurate recommendation services can be provided for users.
Owner:SHENZHEN INST OF ADVANCED TECH CHINESE ACAD OF SCI
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products