Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

80results about How to "Bridging the "Semantic Gap"" patented technology

Visual keyword based remote sensing image semantic searching method

The invention relates to a visual keyword based remote sensing image semanteme searching method. The method comprises the following steps: setting visual keywords which describe image contents in an image base; selecting a training image from the image base; extracting remarkable visual characteristics of each training image, wherein the remarkable visual characteristics include remarkable points, main dominant tone and texture; acquiring a key mode through a cluster center of a cluster algorithm; establishing a visual keyword hierarchical model by adopting a Gaussian mixture model; extracting the remarkable visual characteristics of all images in the image base, setting weight parameters, and constructing a visual keyword characteristic vector describing the image semanteme; and calculating the similarity between an image to be searched and all images according to the similarity criterion, and outputting a search result according to the high-low sequence of the similarity. The method can effectively improve the recall ratio and the precision ratio of image searching by establishing the correlation between low-layer remarkable visual characteristics and high-layer semantic information through the visual keywords, and the technical scheme provided by the invention has excellent expansibility.
Owner:WUHAN UNIV

Image retrieval method based on object detection

The invention discloses an image retrieval method based on object detection. The method is used for solving the problem that multiple objects in an image are not retrieved respectively during image retrieval. According to the implementation process of the method, object detection is performed on an image in an image database, and one or more objects in the image are detected; SIFT features and MSER features of the detected objects are extracted and combined to generate feature bundles; a K mean value and a k-d tree are adopted to make the feature bundles into visual words; visual word indexes of the objects in the image database are established through reverse indexing, and an image feature library is generated; and an object detection method is used to make objects in a query image into visual words, similarity compassion is performed on the visual words of the query image and the visual words of the image feature library, and the image with the highest score is output to serve as an image retrieval result. Through the method, the objects in the image can be retrieved respectively, background interference and image semantic gaps are reduced, and accuracy, retrieval speed and efficiency are improved; and the method is used for image retrieval on a specific object in the image, including a person.
Owner:XIDIAN UNIV

Structured image description method

The invention belongs to the technical field of image retrieval, in particular to a structured image description method. The structured image description method comprises the steps that an image for training is obtained, and three-layer tree-shaped structure label is established for each object in the image, so that a training set is formed; the bottom-layer characteristics of each object of the image in the training set are extracted, all candidate classes, subclasses and classifiers with corresponding attributes are obtained through training, and therefore intermediate data required for modeling of the next step are formed; a conditional random field model is established and model parameters are obtained through training; image segmentation is firstly conducted, objects contained in an image to be described are segmented, and the bottom-layer characteristics of each object of the image to be described are further extracted; tree-shaped structure label of each object of the image to be described is predicated through the established CRF model and the model parameters obtained through training and according to the maximum product belief propagation algorithm. According to the structured image description method, the distinction degree between images can be improved and a good retrieval result is generated.
Owner:TIANJIN UNIV

Image sentiment analysis method based on multi-task learning mode

The invention discloses an image sentiment analysis method based on a multi-task learning mode. The method comprises the following steps: constructing an image sentiment attribute detector and an image sentiment label classifier; using a gradient descent method to train initialization parameters of the image emotion attribute detector; testing the prediction precision of the emotion attributes of the image and judging whether the emotion attributes reach the standard or not, if yes, reasonably designing the training parameters of the detector, otherwise, retraining; taking the output of the image emotion detector and the convolution characteristics of the original image as the input of an emotion label classifier, and training classifier initialization parameters by adopting a gradient descent method; testing the prediction precision of the label classifier and judging whether the prediction precision reaches the standard or not, namely, reasonably designing training parameters of the label classifier when the prediction precision reaches the standard, otherwise, retraining; classifying the image emotion tags, and analyzing the image emotion. According to the method, the influence caused by a semantic gap can be reduced, image emotion prediction is more accurate, and the method is better suitable for large-scale image emotion classification tasks.
Owner:GUANGDONG UNIV OF TECH

Entity linking method for Chinese knowledge graph question-answering system

The invention provides an entity linking method for a Chinese knowledge graph question-answering system. The method comprises the following steps: firstly, performing joint embedding on words and entities in a training corpus to obtain joint embedding vectors of the words and the entities; for an input text of the Chinese knowledge graph question-answering system, firstly, recognizing entity reference items in the input text, and determining a candidate entity list according to the entity reference items; and constructing an entity link model based on an LSTM network, performing vector splicing on the entity representation vector and the entity reference item representation vector to obtain a similarity score of the entity reference item and the candidate entity, and finally obtaining a score rank of the candidate entity, thereby selecting the candidate entity with the highest score as a target entity corresponding to the entity reference item. According to the method, the defect of link model training data redundancy caused by diversity of user questioning modes is effectively solved, and words with similar semantics can be replaced and used in the context, so that the link effectiveness is improved, and the accuracy of a question and answer system is improved.
Owner:NORTHWESTERN POLYTECHNICAL UNIV

Image retrieval method based on characteristic enrichment area set

The invention discloses an image retrieval method based on a characteristic enrichment area. The method comprises the following steps of firstly, acquiring a candidate characteristic point set by calculating a Hessian matrix and non-maximum value restraint, and acquiring a sub-pixel-level characteristic point set by utilizing a three-dimensional linear interpolation method; secondly, calculating a distribution matrix and an adaption matrix of characteristic points according to coordinate positions of the obtained characteristic points of an image, and by utilizing a maximum sub-matrix and an algorithm, solving a sub-matrix of the adaption matrix, namely the most dense distribution area of the characteristic points as the characteristic enrichment area of the image; thirdly, selecting a shape bottom layer characteristic, a texture bottom layer characteristic and a color bottom layer characteristic for the characteristic enrichment area; finally, measuring similarities according to a Gaussian non-linear distance function, and quickly retrieving the image according to the ascending order of the similarities. According to the method, the calculation complexity of image retrieval can be effectively reduced, and the operation efficiency and the accuracy of the image retrieval are improved.
Owner:HEFEI HUIZHONG INTPROP MANAGEMENT

A remote sensing image multi-label retrieval method and system based on a full convolutional neural network

The invention provides a remote sensing image multi-label retrieval method and system based on a full convolutional neural network. multi-label image retrieval is realized by considering multi-category information of remote sensing images, and the method comprises the following steps: inputting a retrieval image library, and dividing the retrieval image library into a training set and a verification set; Constructing a full convolutional neural network model FCN, and performing network training by using the training set; Performing multi-class label prediction on each image in the verificationset by using FCN to obtain a segmentation result; Carrying out up-sampling on each convolutional layer feature map; Extracting local features of each image in the verification set to obtain feature vectors for retrieval; And finally, carrying out coarse-to-fine two-step retrieval based on the extracted multi-scale features and the multi-label information. According to the method, the full convolutional neural network is used for learning multi-scale local features of the image, multi-label information hidden in the image is fully mined, and compared with an existing remote sensing image retrieval method based on a single label, the accuracy of image retrieval is effectively improved.
Owner:WUHAN UNIV

Unsupervised cross-modal retrieval method based on attention mechanism enhancement

The invention belongs to the technical field of artificial intelligence smart community application, and relates to an unsupervised cross-modal retrieval method based on attention mechanism enhancement, which comprises the following steps of: enhancing visual semantic features of an image, then aggregating feature information of different modals, mapping fused multi-modal features to the same semantic feature space, then, on the basis of a generative adversarial network, adversarial learning is carried out on the image modal and text modal features and the same semantic feature after multi-modal fusion, aligning semantic features of different modals, and finally, generating hash codes by the different modal features after alignment of the generative adversarial network; and performing intra-modal feature and Hash code similarity measurement learning and inter-modal feature and Hash code similarity measurement, so a heterogeneous semantic gap problem between different modalities is reduced, a dependency relationship between different modal features is enhanced, a semantic gap between different modal data is reduced, and semantic common characteristics among different modes can be represented more robustly.
Owner:QINGDAO SONLI SOFTWARE INFORMATION TECH

Remote sensing image semantic segmentation method and device, computer equipment and storage medium

The invention discloses a remote sensing image semantic segmentation method and device, computer equipment and a storage medium, and the method comprises the steps of obtaining a preprocessed remote sensing image, carrying out the high-frequency texture feature and low-frequency semantic feature extraction of the preprocessed remote sensing image based on a feature extraction network layer, and taking the extracted features as an input feature set; introducing the low-frequency semantic features into a spatial pyramid pooling module for multi-scale pooling, and obtaining aggregated text features; introducing the input feature set and the aggregated text features into a feature guide alignment module, and obtaining an aligned input feature set according to the difference between the input feature set and the aggregated text features; introducing the aligned input feature set and the aggregated text features into a gating feature selection module for selective fusion to obtain an aligned and fused supplementary feature set; and according to the supplementary feature set and the aggregated text features, performing splicing fusion to generate features, processing the features based on a preset performance function, and performing prediction classification on the processed features to obtain a final feature layer. The segmentation precision is effectively improved.
Owner:HUNAN UNIV

Universal cross-modal retrieval model based on deep hash

The invention discloses a universal cross-modal retrieval model based on deep hash. The universal cross-modal retrieval model comprises an image model, a text model, a binary code conversion model and a Hamming space. The image model is used for the feature and semantic extraction of the image data; the text model is used for the feature and semantic extraction of the text data; the binary code conversion model is used for converting the original features into the binary codes; the Hamming space is a common subspace of images and the text data, and the similarity of the cross-modal data can be directly calculated in the Hamming space. According to the universal model for solving cross-modal retrieval by combining deep learning and Hash learning, the data points in an original feature space are mapped into the binary codes in the public Hamming space, similarity ranking is carried out by calculating the Hamming distance between the codes of the data to be queried and the codes of the original data, and therefore a retrieval result is obtained, and the retrieval efficiency is greatly improved. The binary codes are used for replacing the original data storage, so that the requirement of the retrieval tasks for the storage capacity is greatly reduced.
Owner:CHINA UNIV OF PETROLEUM (EAST CHINA)
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products