Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

75 results about "Visual vocabularies" patented technology

Visual vocabulary consists of images or pictures that stand for words and their meanings. In the same way that individual words make written language possible, individual images make a visual language possible. The term also applies to a theory of visual communication that says pictures and images can be “read” in the same way that words can.

Image object recognition method based on SURF

The invention provides an image object recognition method based on SURF (Speed Up Robust Feature), comprising the following steps: first, preprocessing images; second, extracting SURF corners and SURF descriptors of the images to describe the features of the images; third, processing the features through PCA data whitening and dimension reduction; establishing a bag-of-visual-words model through Kmeans clustering based on the features after processing, and using the bag-of-visual-words model to construct a visual vocabulary histogram of the images; and finally, carrying out training by a nonlinear support vector machine (SVM) classification method, and classifying the images to different categories. After classification model building of different images is completed in the training phase, the images tested in a concentrated way are detected in the testing phase, and therefore, different image objects can be recognized. The method has excellent performance in the aspects of recognition rate and speed, and can reflect the content of images more objectively and accurately. In addition, the classification result of an SVM classifier is optimized, and the error rate of judgment of the classifier and the limitation of the categories of training samples are reduced.
Owner:SHANGHAI JIAO TONG UNIV +1

High-spatial resolution remote-sensing image bag-of-word classification method based on linear words

The invention discloses a high-spatial resolution remote-sensing image bag-of-word classification method based on linear words, which includes first dividing images to be classified into a practice sample and a classification sample. Steps for the practice sample include collecting linear characteristics of the practice image and calculating linear characteristic vector; utilizing K-Means++ arithmetic to generate linear vision word list in cluster mode; segmenting practice images and obtaining linear vision word list column diagram of each segmentation spot block on the base; and conducting class label on the spot block and putting the classification and linear vision word column diagram in storage. After sample practice, steps for the classification sample include collecting linear characteristics of the images to be classified, segmenting the images to be classified, calculating linear characteristics vector on the base, obtaining linear vision word list column diagram of each segmentation spot block and selecting an SVM classifier to classify the images to be classified to obtain classification results. The high-spatial resolution remote-sensing image bag-of-word classificationmethod utilizes linear characteristics to establish bag-of-word models and is capable of obtaining better high spatial resolution remote sensing image classification effect.
Owner:NANJING NORMAL UNIVERSITY

Semantic propagation and mixed multi-instance learning-based Web image retrieval method

The invention belongs to the technical field of image processing and particularly provides a semantic propagation and mixed multi-instance learning-based Web image retrieval method. Web image retrieval is performed by combining visual characteristics of images with text information. The method comprises the steps of representing the images as BoW models first, then clustering the images according to visual similarity and text similarity, and propagating semantic characteristics of the images into visual eigenvectors of the images through universal visual vocabularies in a text class; and in a related feedback stage, introducing a mixed multi-instance learning algorithm, thereby solving the small sample problem in an actual retrieval process. Compared with a conventional CBIR (Content Based Image Retrieval) frame, the retrieval method has the advantages that the semantic characteristics of the images are propagated to the visual characteristics by utilizing the text information of the internet images in a cross-modal mode, and semi-supervised learning is introduced in related feedback based on multi-instance learning to cope with the small sample problem, so that a semantic gap can be effectively reduced and the Web image retrieval performance can be improved.
Owner:XIDIAN UNIV

Remote sensing image land utilization scene classification method based on two-dimension wavelet decomposition and visual sense bag-of-word model

The invention relates to a remote sensing image land utilization scene classification method based on two-dimension wavelet decomposition and a visual sense bag-of-word model. The method comprises the steps that a remote sensing image land utilization scene classification training set is built; scene images in the training set are converted to grayscale images, and two-dimension decomposition is conducted on the grayscale images; regular-grid sampling and SIFT extracting are conducted on the converted grayscale images and sub-images formed after two-dimension decomposition, and universal visual word lists of the converted grayscale images and the sub-images are independently generated through clustering; visual word mapping is conducted on each image in the training set to obtain bag-of-word characteristics; the bag-of-word characteristics of each image in the training set and corresponding scene category serial numbers serve as training data for generating a classification model through an SVM algorithm; images of each scene are classified according to the classification model. The remote sensing image land utilization scene classification method well solves the problems that remote sensing image texture information is not sufficiently considered through an existing scene classification method based on a visual sense bag-of-word model, and can effectively improve scene classification precision.
Owner:INST OF REMOTE SENSING & DIGITAL EARTH CHINESE ACADEMY OF SCI

Picture searching method based on maximum similarity matching

ActiveCN104615676AEliminate multiple matchesEnhanced Visual CompatibilityStill image data indexingCharacter and pattern recognitionFeature setReverse index
The invention relates to a picture searching method based on maximum similarity matching. The method includes the following steps that (1) a training picture set is acquired; (2) feature point detection and description are conducted on acquired pictures in a multi-scale space; (3) feature sets extracted in the second step are clustered and generated into a visual dictionary including k visual vocabularies; (4) each feature extracted in the second step is mapped to the visual vocabulary with the distance being smallest to the current feature l2, the current feature and the normalization residual vector of the corresponding visual vocabulary are stored in a reverse index structure, and accordingly a query database is formed; (5) the pictures to be searched for are acquired, the second step and the fourth step are executed again, the reverse index structure of the pictures to be searched for is acquired, the query database is searched for according to the reverse index structure, and the searching results of the pictures to be searched for are acquired based on the maximum similarity matching. Compared with the prior art, the picture searching method has the advantages of being good in robustness, high in computational efficiency and the like.
Owner:TONGJI UNIV

A pulmonary nodule image classification method for constructing feature representation based on an automatic encoder

The invention provides a pulmonary nodule image classification method for constructing feature representation based on an automatic encoder, and relates to the technical field of computer vision. Themethod comprises the following steps: firstly, segmenting a pulmonary nodule image into local patches through superpixels; Transforming the patches into local feature vectors with a fixed length by using an unsupervised depth auto-encoder; Constructing visual vocabularies on the basis of the local features, and describing global features of the pulmonary nodule image through a visual word bag; Classifying pulmonary nodule types by using a softmax algorithm to complete the design of a model framework for representing pulmonary nodule image characteristics; Performing training by using the designed model framework and the ELCAP data set to obtain an automatic classification model for the pulmonary nodule images; And finally, carrying out pulmonary nodule image classification by using the obtained pulmonary nodule image classification model. According to the pulmonary nodule image classification method for constructing feature representation based on the automatic encoder provided by theinvention, the feature extraction capability of the pulmonary nodule classification model is improved, and the accuracy of pulmonary nodule automatic classification is improved.
Owner:NORTHEASTERN UNIV

Method for automatically classifying forestry service images

The invention relates to a method for automatically classifying forestry service images, comprising the main steps of training and classifying. The step of training is as follows: converting images, calculating the set of key points on a gray-scale image, describing the key points by determining the main direction of the key points and generating eigenvectors, clustering, and producing histograms to express the images. The step of classifying is as follows: expressing the classified images with the histograms, and classifying by a classifier. Therefore, the classification of the forestry service images is finished. The numerous forestry service images collected by the forest rangers are used for constructing a reasonable visual vocabulary book according to the characteristics and the color information of the data of the forestry service images, the forestry service images are divided accurately into seven categories including forest fires, illegal use of forest land, illegal logging, illegal hunting and the like, and the forestry service images of different categories are respectively transferred to the functional management departments to realize the fast, effective and timely management of forest and the information modernization of the management of the forest.
Owner:ZHEJIANG FORESTRY UNIVERSITY

Image retrieval method based on hierarchical convolutional neural network

The invention discloses an image retrieval method based on a hierarchical convolutional neural network, and mainly aims at solving the problem that in existing all-sky aurora image retrieval, the accurate rate is low. The method comprises the implementation steps that 1, local key points of all-sky aurora images are determined by adopting an adaptive polar barrier method; 2, local SIFT features ofthe all-sky aurora images are extracted, and a visual vocabulary is constructed; 3, the convolutional neural network is pre-trained and subjected to fine tuning, and a polar region pooling layer is constructed; 4, region CNN features and global CNN features of the all-sky aurora images are extracted; 5, all the features are subjected to binarization processing, and hierarchical features are constructed; 6, a reverse index table is constructed, and the global CNN features are saved separately; and 7, hierarchical features of a queried image are extracted, the similarity between the queried image and the database images is calculated, and a retrieval result is output. According to the method, matching of the local key points is achieved through the hierarchical features, the problem that inan existing image retrieval method, the false alarm rate is high is solved, the advantage of being high in retrieval accuracy rate is achieved, and the method is suitable for real-time image retrieval.
Owner:XIDIAN UNIV

Rapid and high-efficiency near-duplicate image matching method

The invention discloses a rapid and high-efficiency near-duplicate image matching method. The method comprises the steps that 1) the ORB characteristic of each image in a training image library is extracted and nonlinear mapping is performed on the ORB characteristic of each image so that a visual word table of the training image library is constructed; 2) sparse coding is performed on the nonlinear mapping ORB characteristic of each image in the training image library by utilizing locally-constrained linear coding according to the constructed visual word table; 3) the ORB characteristic of the image to be matched is extracted and nonlinear mapping is performed on the ORB characteristic of the image to be matched, and then sparse coding is performed on the nonlinear mapping ORB characteristic of the image to be matched according to the constructed visual word table; and 4) similarity of sparse coding of the image to be matched and sparse coding of the images in the training image library is calculated, and if similarity exceeds the preset threshold value, matching succeeds, or matching fails. Reconstruction error of a hard quantification method is reduced so that matching speed is greatly enhanced and the method can be used for real-time matching.
Owner:长安通信科技有限责任公司

Method for generating context descriptors of visual vocabulary

The invention relates to a method for generating context descriptors of visual vocabulary. The method comprises following steps: off-line learning, context descriptor generating and context descriptor similarity computing. The off-line learning is used for construction of a visual vocabulary dictionary and evaluation of visual vocabulary. The step of context descriptor generating comprises following sub-steps: 1. extracting local characteristic points and quantifying characteristic descriptors; 2. selecting a context; 3. extracting characteristics of the local characteristic points of the context and generating context descriptors. The context descriptor similarity computing is used for verifying whether local characteristic points of two context descriptors match with each other according to the azimuth and principal direction of the local characteristic points of the context descriptors and consistency of the visual vocabulary, and evaluating the similarity of the two context descriptors through the summation of the inverse document frequency of matched visual vocabulary. The context descriptors established by the invention are adapted to influence brought by conversions such as image clipping, rotation and scale-zooming; the method can be applied in image retrieval and classification, etc.
Owner:杭州远传新业科技股份有限公司

Object recognition method based on semantic feature extraction and matching

The invention provides an object recognition method based on semantic feature extraction and matching and belongs to the field of information retrieval. The object recognition method based on semantic feature extraction and matching includes semantic feature extraction and semantic feature matching. The semantic feature extraction includes firstly extracting SIFT (Scale Invariant Feature Transform) feature points of training images of a class of objects, then performing spatial clustering on the SIFT feature points through k- means clustering, deciding a plurality of efficient points in every space class through a decision-making mechanism based on kernel function, and finally training the efficient points in every space class through a support vector machine classifier; a visual word with semantic features is trained from every space class, and finally a visual vocabulary describing the semantic features of a class of objects is extracted. The semantic feature matching includes firstly extracting SIFT feature points of an image of an object to be detected as the semantic description of the object to be detected, then using the support vector machine classifier for matching and classifying the semantic description of the object to be detected and visual vocabularies of classes of objects, and finally counting a histogram of the visual vocabulary of the object to be detected for determining the class of the object to be detected.
Owner:武汉三际物联网络科技有限公司

Human body detecting method based on SURF (Speed Up Robust Feature) efficient matching kernel

The invention provides a human body detecting method based on an SURF efficient matching kernel, and mainly solves the problem that image background hybridity can not be better processed in the existing method. The method comprises the steps that a negative sample is obtained through bootstrap in an INRIA (Institute National de Recherce en Informatique et Automatique) database, and a training sample set of the whole human body is formed by the negative sample and a positive sample in the database; SURF descriptor feature points are extracted under different image scales for the training sample; feature points are extracted by random sampling to constitute the initial vector basis of a visual vocabulary; constrained singular value decomposition is utilized for the initial vector basis to obtain the maximum kernel function feature; the maximum kernel function feature in different image scales is weighted to obtain the features under all the image scales; the obtained features are trained in different classes by an SVM (Support Vector Machine) classifier, and a detection classifier is obtained; and the image to be detected is input to the classifier to obtain the final detection result. The method disclosed by the invention can be used for accurately detecting the human body, and can be used for intelligent monitoring, driver auxiliary systems and virtual video.
Owner:XIDIAN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products