Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

398results about How to "Reduce training difficulty" patented technology

Human body gesture identification method based on depth convolution neural network

The invention discloses a human body gesture identification method based on a depth convolution neural network, belongs to the technical filed of mode identification and information processing, relates to behavior identification tasks in the aspect of computer vision, and in particular relates to a human body gesture estimation system research and implementation scheme based on the depth convolution neural network. The neural network comprises independent output layers and independent loss functions, wherein the independent output layers and the independent loss functions are designed for positioning human body joints. ILPN consists of an input layer, seven hidden layers and two independent output layers. The hidden layers from the first to the sixth are convolution layers, and are used for feature extraction. The seventh hidden layer (fc7) is a full connection layer. The output layers consist of two independent parts of fc8-x and fc8-y. The fc8-x is used for predicting the x coordinate of a joint. The fc8-y is used for predicting the y coordinate of the joint. When model training is carried out, each output is provided with an independent softmax loss function to guide the learning of a model. The human body gesture identification method has the advantages of simple and fast training, small computation amount and high accuracy.
Owner:UNIV OF ELECTRONICS SCI & TECH OF CHINA

Method of translation, method for determining target information and related devices

The invention discloses a method for determining target information. The method includes the following steps that encoding is conducted on to-be-processed text information to obtain a source-end vector expression sequence; according to the source-end vector expression sequence, a source-end context vector corresponding to the first moment is obtained, wherein the source-end context vector is used for expressing to-be-processed source-end content; according to the source-end vector expression sequence and the source-end context vector, a first translation vector and / or a second translation vector are / is determined, wherein the first translation vector indicates the source-end content which is not translated in the source-end vector expression sequence within the first moment, and the second translation vector indicates the source-end content which is translated in the source-end vector expression sequence within the second moment; decoding is conducted on the first translation vector and / or a second translation vector and the source-end context vector so as to obtain the target information of the first moment. The invention further provides a method of translation and a device for determining the target information. According to the method for determining the target information, the model training difficulty of a decoder can be reduced, and the translation effect of a translation system is improved.
Owner:SHENZHEN TENCENT COMP SYST CO LTD

Image splicing tampering positioning method based on full convolutional neural network

The invention discloses an image splicing tampering positioning method based on a full convolutional neural network. The method includes: establishing a splicing tampered image library; initializing an image splicing tampering positioning network based on a full convolutional neural network, and setting a training process of the network; initializing network parameters; reading a training image, performing training operation on the training image, and outputting a splicing positioning prediction result of the training image; calculating an error value between a training image splicing positioning prediction result and a real label, and adjusting network parameters until the error value meets a precision requirement; performing post-processing on the prediction result of which the precisionmeets the requirement by using a conditional random field, adjusting network parameters, and outputting a final prediction result of the training image; and reading the test image, predicting the test image by adopting the trained network, carrying out post-processing on a prediction result through a conditional random field, and outputting a final prediction result of the test image. The methodhas high splicing tampering positioning precision, the network training difficulty is low, and the network model is easy to converge.
Owner:NANJING UNIV OF INFORMATION SCI & TECH

Bridge crack image barrier detection and removal method based on generative adversarial network

The invention relates to a bridge crack image barrier detection and removal method based on a generative adversarial network. The method comprises the steps that first, multiple barrier pictures are collected, then tags are added, and the pictures with the tags are input into a Faster-RCNN for training; multiple barrier-containing crack pictures are collected, and barrier position calibration is performed through the Faster-RCNN; second, multiple barrier-free crack pictures are collected, and the pictures are turned over to amplify a dataset; third, the amplified dataset is input into the generative adversarial network to train a crack generation model; fourth, information erasure is performed on the positions of barriers in the barrier-containing crack pictures to obtain damaged images; and fifth, the damaged images are input into a cyclic discrimination restoration model for iteration, and then restored crack images are obtained. Through the method, barrier information in the crack pictures can be accurately detected and removed, the peak signal-to-noise ratio of the restored crack images is increased by 0.6-0.9dB compared with before, and therefore a large quantity of crack images with a high restoration degree are generated under a finite crack dataset condition.
Owner:SHAANXI NORMAL UNIV

Video content description method based on text auto-encoder

The invention discloses a video content description method based on a text auto-encoder. The method comprises the following steps: firstly, constructing a convolutional neural network to extract two-dimensional and three-dimensional features of a video; secondly, constructing a text auto-encoder, namely extracting text hidden space features and a decoder-multi-head attention residual error networkreconstruction text by using an encoder-text convolution network; thirdly, obtaining estimated text hidden space features through a self-attention mechanism and full-connection mapping; and finally,alternately optimizing the model through an adaptive moment estimation algorithm, and obtaining corresponding video content description for the new video by using the constructed text auto-encoder andthe convolutional neural network. According to the method, a potential relationship between video content semantics and video text description can be fully mined through the training of the text auto-encoder, and the action time sequence information of the video in a long time span is captured through the self-attention mechanism, so that the calculation efficiency of the model is improved, and the text description more conforming to the real content of the video is generated.
Owner:HANGZHOU DIANZI UNIV

Multi-mode ultrasonic image classification method and breast cancer diagnosis device

The invention provides a multi-mode ultrasonic image classification method and a breast cancer diagnosis device, and the method comprises the steps: S1, segmenting a region-of-interest image from an original gray-scale ultrasonic-elastic imaging image pair, and obtaining a pure elastic imaging image according to the segmented region-of-interest image; s2, extracting single-mode image features of the gray-scale ultrasonic image and the elastic imaging image by using a DenseNet network; s3, constructing a resistance loss function and an orthogonality constraint function, and extracting shared features between the gray-scale ultrasonic image and the elastic imaging image; and S4, constructing a multi-task learning framework, splicing the inter-modal shared features obtained in the S3 and thesingle-modal features obtained in the S2, inputting the inter-modal shared features and the single-modal features into a plurality of classifiers together, and performing benign and malignant classification respectively. According to the method, benign and malignant classification can be carried out on the gray-scale ultrasonic image, the elastic imaging image and the two modal images at the sametime, and the method has excellent performance of high accuracy and wide application range.
Owner:SHANGHAI JIAO TONG UNIV

A multi-sensor attitude data fusion method and system based on a neural network

The invention discloses a multi-sensor attitude data fusion method and system based on a neural network. The method comprises the following steps: generating original attitude data through a pluralityof sensors; Taking the original attitude data as the input of the convolutional neural network, and taking the attitude data output after the convolution layer, the pooling layer, the full connectionlayer and the first activation function as the output of the convolutional neural network for output; Taking the output of the convolutional neural network as the input of the artificial neural network; and according to a preset general kernel structure, not outputting input of a preset node corresponding to any hidden layer of the artificial neural network through a second activation function, and outputting the input of the remaining nodes corresponding to any hidden layer through a second activation function, and outputting the attitude angle data output by the neuron node of the hidden layer at the tail end as the output of the artificial neural network. According to the fusion method, the convolutional neural network and the optimized artificial neural network are effectively combined, so that the measurement precision of the attitude angle data is improved.
Owner:JILIN UNIV

Nano microorganism organic and inorganic compound fertilizer and production method thereof

ActiveCN103030471APromote growthTo achieve the purpose of disease resistanceFertilizer mixturesBacillus megateriumPotassium
The invention relates to the agricultural field and the environment-friendly field, in particular to nano microorganism organic and inorganic compound fertilizer and a production method thereof. The fertilizer comprises the following components in parts by weight: 5 to 40 parts of layered silicate, 0.1 to 0.25 part of bacillus megaterium, 0.1 to 0.25 part of bacillus mucilaginosus, 0.1 to 0.25 part of bacillus subtilis, and 0.1 to 0.25 part of base fertilizer containing nitrogen element, potassium element and phosphor element, wherein the nitrogen element, the potassium element and the phosphor element are 6-18% of total weight of the base fertilizer. The production method comprises the following steps: strains of the bacillus megaterium, the bacillus mucilaginosus and the bacillus subtilis are cultivated respectively to obtain strains in a spore state; the base fertilizer is mixed and smashed; and the smashed base fertilizer, the strains in the spore state and the layered silicate are mixed, pelletized and dried. The fertilizer can provide the nutrition needed by crops, and can decompose nitrogen fertilizer, potassium fertilizer and phosphor fertilizer adsorbed and fixed in the soil, so that soil damage is avoided.
Owner:ZIGONG HUAQI ZHENGGUANG ECOLOGY AGICULTURE FORESTRY +1
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products