Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

39results about How to "Enhanced feature information" patented technology

Attention mechanism-based in-depth learning diabetic retinopathy classification method

The invention discloses an attention mechanism-based in-depth learning diabetic retinopathy classification method comprising the following steps: a series of eye ground images are chosen as original data samples which are then subjected to normalization preprocessing operation, the preprocessed original data samples are divided into a training set and a testing set after being cut, a main neutralnetwork is subjected to parameter initializing and fine tuning operation, images of the training set are input into the main neutral network and then are trained, and a characteristic graph is generated; parameters of the main neutral network are fixed, the images of the training set are adopted for training an attention network, pathology candidate zone degree graphs are output and normalized, anattention graph is obtained, an attention mechanism is obtained after the attention graph is multiplied by the characteristic graph, an obtained result of the attention mechanism is input into the main neutral network, the images of the training set are adopted for training operation, and finally a diabetic retinopathy grade classification model is obtained. Via the method disclosed in the invention, the attention mechanism is introduced, a diabetic retinopathy zone data set is used for training the same, and information characteristics of a retinopathy zone is enhanced while original networkcharacteristics are reserved.
Owner:NANJING UNIV OF AERONAUTICS & ASTRONAUTICS

Vein feature extraction method and method for carrying out identity authentication by utilizing double finger veins and finger-shape features

The invention relates to a method for carrying out identity authentication by utilizing double finger veins and finger-shape features, comprising the following steps of: firstly registering and storing vein features and information of finger-shape features of double fingers of a user; collecting, extracting and authenticating the vine features and the information of the finger-shape features of double fingers of a customer; respectively comparing the extracted double finger vein features and finger-shape features with the stored double finger vein features and information of the finger-shape features, and carrying out decision-level fusion; collecting images of hand shapes and finger veins, establishing a rectangular coordinate system by taking an intersection point between the double fingers as an original point and determining a region of interest (ROI); and finally respectively extracting the vein features and the finger-shape features, comparing with the stored double finger veins and information of the finger-shape features, and carrying out decision-level fusion. In the invention, two kinds of information of veins and finger shapes can be respectively extracted from one image and the two kinds of information are fused; and by introducing the coordinate system with the intersection point of the two fingers as the original point, the method provided by the invention not only improves the positioning precision, but also is convenient for users to use and effectively improves the system performance through double feature fusion.
Owner:BEIJING JIAOTONG UNIV

Method for identifying bamboo hemp fiber by using terahertz time-domain spectroscopy technique

The invention relates to the field of textile materials, and aims at providing a method for identifying a bamboo hemp fiber by using a terahertz time-domain spectroscopy technique. The method comprises the following steps of: respectively mixing and pressing a fiber with polyethylene powder to prepare a disc sheet after preparing the powder from the fiber as a guide sample; testing by adopting a transmission-type terahertz time-domain spectroscopy test device, thus respectively obtaining terahertz pulse time-domain waveforms through each guide sample; calculating the absorption coefficient and the refractive index of each guide sample, so as to respectively draw out an absorption spectrum and a refraction spectrum of the fiber; obtaining an absorption coefficient and a refractive index of the fiber to be identified; and confirming the type of the fiber to be identified according to the corresponding relation with the absorption spectrum and the refraction spectrum of each guide sample. According to the method, required sample quantity is simple, samples are simple and convenient to prepare, test time is short, and chemical pollutants are not caused in a test process; the absorption spectrum and the refraction spectrum of the sample at a terahertz waveband can be obtained at the same time when the sample is tested; the feature information of the samples is increased; the reliability of a test result is also improved.
Owner:ZHEJIANG SCI-TECH UNIV

Sentiment classification method and system based on multi-modal context semantic features

ActiveCN112818861ARich Modal InformationFully acquire emotional semantic featuresCharacter and pattern recognitionNeural architecturesEmotion classificationSemantic feature
The invention discloses a sentiment classification method and system based on multi-modal context semantic features. The method comprises the following steps: segmenting a short video into semantic units with the same number by taking an utterance as a unit, generating corresponding video, voice and text samples, and extracting three characterization features, namely an expression feature, a spectrogram and a sentence vector; respectively inputting the three extracted characterization features into expression, voice and text emotion feature encoders, and extracting corresponding emotion semantic features; constructing corresponding adjacent matrixes by using context relationships of emotion semantic features of expressions, voices and texts; and inputting the expression emotion semantic features, the voice emotion semantic features, the text emotion semantic features and the corresponding adjacent matrixes into corresponding graph convolutional neural networks, extracting corresponding context emotion semantic features, and fusing the context emotion semantic features to obtain multi-modal emotion features for emotion classification and recognition. According to the method, the context relationship of the emotion semantic features is better utilized through the graph convolutional neural network, and the accuracy of emotion classification can be effectively improved.
Owner:NANJING UNIV OF POSTS & TELECOMM

Low-overhead household garbage classification method based on dense convolutional network

The invention discloses a low-overhead household garbage classification method based on a dense convolutional network, and the method comprises the steps: (1), preprocessing data, (2), building the dense convolutional network, (3), dividing a data set into a training set, a verification set and a test set, and (4), selecting a proper optimizer and a loss function in a training process, setting hyper-parameters, and setting an evaluation index as the accuracy rate. According to the method, optimization is carried out in preprocessing of input data, matrix fusion is carried out on a three-channel color image and an edge detection image to serve as input of a model, and feature information is enhanced; a dense convolutional network structure is constructed; a Dropout layer is additionally arranged; a learning rate self-adjusting and hyper-parameter adjusting method is used.According to the method, the model has enough feature extraction capability; the feature mapping of the model is usedas the input of a subsequent layer; the problem of gradient disappearance caused by a deep network is relieved; good balance is realized on the aspects of low overhead and high precision; and 90.8% of precision and 5.08 M of file size are realized.
Owner:HOHAI UNIV

Double-flow face counterfeiting detection method based on Swin Transform

The invention relates to a double-flow face forgery detection method based on Swin Transformer, which is used for detecting a face forgery image by using deep learning. A deep learning network model is integrally built, and the network model is divided into three parts: a double-flow network, a feature extraction network and a classifier. Because all face counterfeit data sets disclosed at present are videos, the videos need to be clipped into frame pictures by using OpenCV. In addition, the frame picture contains a large amount of background information, so that a human face area needs to be cut out by using a human face positioning algorithm. And inputting the obtained face region image into a double-flow network and a feature extraction network to extract and learn features. And finally, inputting the learned features into a classifier, and identifying whether the face image is true or false. The method is used for solving the problem of partial limitation, namely weak generalization ability, of the existing face counterfeiting detection scheme, and meanwhile, the compression resistance of the model is improved through the double-flow framework, so that the method is more in line with the common face video quality in daily life.
Owner:SHANGHAI UNIV

Radar radiation source identification method based on multistage jumper residual network

The invention discloses a radar radiation source identification method based on a multistage jumper residual network, and belongs to the field of image processing. The radar radiation source identification method comprises the following steps: performing time-frequency transformation on a radar radiation source signal to generate a time-frequency image of the radar radiation source signal; carrying out the preprocessing of the time frequency image through an ImageDataGenerator, so as to obtain a one-dimensional time frequency image; and inputting the one-dimensional time-frequency image into a deep residual network for feature extraction and signal classification so as to identify and obtain the radar radiation source. According to the radar radiation source identification method based on the multistage jumper residual network provided by the invention, eight sequentially connected residual blocks are arranged in a deep residual network and are connected by using jumpers to form four residual units, so that the constructed deep residual network with the total convolution layer number of 18 can extract deep information of a signal time-frequency image, and meanwhile, the problems of gradient disappearance, gradient explosion and the like of the network are also avoided.
Owner:中国人民解放军海军航空大学航空作战勤务学院

A sentiment classification method and system based on multimodal contextual semantic features

ActiveCN112818861BRich Modal InformationFully acquire emotional semantic featuresCharacter and pattern recognitionNeural architecturesEmotion classificationSemantic feature
The invention discloses an emotion classification method and system based on multimodal context semantic features. The method includes: dividing a short video into the same number of semantic units by taking utterances as a unit, generating corresponding video, speech and text samples, and extracting three representational features of expression features, spectrograms and sentence vectors; The representation features are respectively input into the expression, speech, and text emotion feature encoders, and the corresponding emotional semantic features are extracted; the contextual relationship between the expression, speech, and text emotion semantic features is used to construct the corresponding adjacency matrix; The semantic features and the corresponding adjacency matrix are input to the corresponding graph convolutional neural network, the corresponding contextual emotional semantic features are extracted, and the multimodal emotional features are obtained by fusion, which is used for the classification and recognition of emotions. The present invention makes better use of the contextual relationship of emotional semantic features through the graph convolutional neural network, and can effectively improve the accuracy of emotional classification.
Owner:NANJING UNIV OF POSTS & TELECOMM

Method and device for determining optimal verbal skill sequence and storage medium

The invention discloses a method and device for determining an optimal verbal skill sequence and a storage medium. The method comprises the following steps: determining user feature information and dialing information of a target user; by utilizing a pre-trained prediction model, according to the user feature information and the dialing information, respectively determining a predicted hang-up rate and a predicted hang-up conversion rate of each verbal skill in the plurality of verbal skill sequences and interaction process information of interaction with the target user, wherein the interaction process information comprises a current verbal skill sequence interacting with the target user, a verbal skill subsequence interacting with the target user and a subsequent verbal skill sequence interacting with the target user; forming a plurality of verbal skill sequences by verbal skill corresponding to each verbal skill node; and determining an optimal verbal skill sequence interacting with the target user according to the predicted hanging-up rate and the predicted hanging-up conversion rate of each verbal skill in the plurality of verbal skill sequences.
Owner:北京有限元科技有限公司

Ultrasonic damage judging method and system for embedded part of train axle

The invention discloses an ultrasonic damage judging method and system for an embedded part of a train axle. The ultrasonic damage judging method comprises the steps: acquiring a single-angle ultrasonic B-scan image of the embedded part of a train wheel; decomposing the ultrasonic B-scan image into a plurality of first images with different scales, calculating the significance value of each pixel block of the plurality of first images, and generating a target area image; and repeating the step S1 and step S2 until a plurality of target area images at all angles are generated, and extracting an overlapping area image of the plurality of target area images as a defect area. According to the method, multiple multi-angle ultrasonic B-scan images of the train wheel embedding part are collected, feature information of wheel set noise and defects is increased, Gaussian pyramid image decomposition is carried out on the ultrasonic B-scan images to carry out significance detection, target defect features are further enhanced, background noise is further weakened, and the detection accuracy is improved. Moreover, an accurate defect area is extracted, a detection result is generated, and the problem that the reliability of the detection result is low in an existing ultrasonic damage judgment method for the axle embedding part is solved.
Owner:CHENGDU TIEAN SCI & TECH

A Method for Imaging Panorama Image of Tube Reactor Furnace

A tubular reacting furnace hearth panoramic image imaging method disclosed by the present invention comprises the steps, such as the image pre-processing, the view angle transformation, the image blind area reproduction, the image splicing, the image fusion, the temperature color coding, etc. According to the tubular reacting furnace hearth panoramic image imaging method provided by the present invention, a hearth panorama top view image is obtained by a plurality of set monitoring probes, and then a hearth local top view image is transformed into a front view image rapidly by combining the hearth spatial size and the spatial solid angles of the pixels of the monitoring probes, thereby visually reflecting the hearth internal real information, and being good in visual effect. According to the method, a tubular reacting furnace hearth panoramic image can be extracted effectively and rapidly, thereby retaining the real-time information of a furnace tube furthest; the infrared radiation information can be extracted by the panoramic image, so that the furnace wall surface temperature of the furnace tube can be calculated accurately, the hearth temperature information and the change variation trend can be displayed panoramically, and the foundation is laid for the tubular reacting furnace panoramic temperature field detection.
Owner:安徽淮光智能科技有限公司

Facial expression recognition method and system, storage medium, computer program and terminal

The invention belongs to the technical field of computer vision. The invention discloses a facial expression recognition method and system, a storage medium, a computer program and a terminal, and themethod comprises the steps: pre-training an image generation model according to the combination of a given depth image and an RGB image, and enabling the trained image generation model to convert aninput depth image into an RGB image according to an RGB image style used for training; generating eyebrows, eyes and mouths of expressions in the RGB images, training a convolutional neural network considering the eyebrows, the eyes and the mouths, and achieving expression recognition through the convolutional neural network. According to the invention, the feature information of eyes, eyebrows and mouths is enhanced, and the recognition accuracy is higher; the effect of the image generation model is good, important information about expressions is reserved through the image generation model,and the forms of RGB images used for expression recognition are unified; the accuracy of expression recognition is higher; when only one channel of the depth map is used for identification, the effectobtained by the method is better.
Owner:SOUTHWEST UNIVERSITY

An underwater target recognition method based on the fusion of GRU and one-dimensional CNN neural network

ActiveCN110807365BSolve the problem that the time feature cannot be extractedEnhanced feature informationCharacter and pattern recognitionNeural architecturesHydrophoneFeature vector
The invention discloses an underwater target recognition method based on the fusion of GRU and one-dimensional CNN neural network, belonging to the field of underwater acoustic target recognition. Aiming at the problem of underwater acoustic target recognition, an underwater target recognition method based on the fusion of GRU and one-dimensional CNN neural network is proposed, and the problem that the traditional neural network cannot extract the temporal characteristics of underwater acoustic signals is solved by using the GRU-based cyclic neural network structure. , while using a one-dimensional CNN convolutional neural network structure to extract the time-domain waveform features of the underwater acoustic signal. The eigenvectors extracted by the fusion of GRU and one-dimensional CNN neural network structure enrich the eigenvalue information of the input classifier. The auxiliary information input that affects the recognition accuracy is expanded, including distance, hydrophone depth, channel depth and other information, and the dropout layer and batch normalization layer are added to avoid the overfitting problem and improve the recognition accuracy of underwater acoustic targets.
Owner:ZHEJIANG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products