Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

193 results about "Visual dictionary" patented technology

A visual dictionary is a dictionary that primarily uses pictures to illustrate the meaning of words. Visual dictionaries are often organized by themes, instead of being an alphabetical list of words. For each theme, an image is labeled with the correct word to identify each component of the item in question. Visual dictionaries can be monolingual or multilingual, providing the names of items in several languages. An index of all defined words is usually included to assist finding the correct illustration that defines the word.

Visual sense simultaneous localization and mapping method based on dot and line integrated features

The invention discloses a visual sense simultaneous localization and mapping method based on dot and line integrated features. The method comprehensively utilizes the line features and the dot features extracted and obtained from a binocular camera image and is able to be used for the positioning and the attitude estimation of a robot in both an external environment and an internal environment. As the dot features and the line features are integrated for use, the system becomes more robust and more accurate. For the parameterization of linear features, the Pluck coordinates are used for straight line calculations, including geometric transformations, 3D reconstructions, and etc. In the back-end optimization, the orthogonal representation of the straight line is used to minimize the number of the parameters of the straight line. The off-line established visual dictionary for the dot and line integrated features is used for closed loop detections; and through the method of adding zone bits, the dot characteristics and the line characteristics are treated differently in the visual dictionary and when an image database is created and image similarity is calculated. The invention can be applied to the construction of a scene image both indoors and outdoors. The constructed map integrates the feature dots and the feature lines, therefore, able to provide even richer information.
Owner:ZHEJIANG UNIV

Image appearance based loop closure detecting method in monocular vision SLAM (simultaneous localization and mapping)

The invention discloses an image appearance based loop closure detecting method in monocular vision SLAM (simultaneous localization and mapping). The image appearance based loop closure detecting method includes acquiring images of the current scene by a monocular camera carried by a mobile robot during advancing, and extracting characteristics of bag of visual words of the images of the current scene; preprocessing the images by details of measuring similarities of the images according to inner products of image weight vectors and rejecting the current image highly similar to a previous history image; updating posterior probability in a loop closure hypothetical state by a Bayesian filter process to carry out loop closure detection so as to judge whether the current image is subjected to loop closure or not; and verifying loop closure detection results obtained in the previous step by an image reverse retrieval process. Further, in a process of establishing a visual dictionary, the quantity of clustering categories is regulated dynamically according to TSC (tightness and separation criterion) values which serve as an evaluation criterion for clustering results. Compared with the prior art, the loop closure detecting method has the advantages of high instantaneity and detection precision.
Owner:NANJING UNIV OF POSTS & TELECOMM

Method for re-identifying persons on basis of deep learning encoding models

The invention relates to a method for re-identifying persons on the basis of deep learning encoding models. The method includes steps of firstly, encoding initial SIFT features in bottom-up modes by the aid of unsupervised RBM (restricted Boltzmann machine) networks to obtain visual dictionaries; secondly, carrying out supervised fine adjustment on integral network parameters in top-down modes; thirdly, carrying out supervised fine adjustment on the initial visual dictionaries by the aid of error back propagation and acquiring new image expression modes, namely, image deep learning representation vectors, of video images; fourthly, training linear SVM (support vector machine) classifiers by the aid of the image deep learning representation vectors so as to classify and identify pedestrians. The method has the advantages that the problems of poor effects and low robustness due to poor surveillance video quality and viewing angle and illumination difference of the traditional technologies for extracting features and the problem of high computational complexity of the traditional classifiers can be effectively solved by the aid of the method; the person target detection accuracy and the feature expression performance can be effectively improved, and the pedestrians in surveillance video can be efficiently identified.
Owner:张烜

Three-dimensional image quality objective evaluation method based on sparse representation

The invention discloses a three-dimensional image quality objective evaluation method based on sparse representation. According to the method, in a training stage, left viewpoint images of a plurality of original undistorted three-dimensional images are selected for forming a training image set, Gaussian difference filtering is adopted for carrying out filtering on each image in the training image set to obtain filtered images in different scales, and in addition, a K-SVD method is adopted for carrying out dictionary training operation on a set formed by all sub blocks in all of the filtered images in different scales for constructing a visual dictionary table; and in a test stage, the Gaussian difference filtering is performed on any one tested three-dimensional image and the original undistorted three-dimensional image to obtain filtered images in different scales, then, the filtered images in different scales is subjected to non-overlapped partition processing, and an image quality objective evaluation prediction value of the tested images is obtained. The three-dimensional image quality objective evaluation method has the advantages that a complicated machine learning training process is not needed in the training stage; and the in the test stage, the image quality objective evaluation prediction value only needs to be calculated through a sparse coefficient matrix, and in addition, the consistency with the subjective evaluation value is better.
Owner:创客帮(山东)科技服务有限公司

Three-dimensional face recognition device and method based on three-dimensional point cloud

InactiveCN104504410AStrong detail texture description abilityGood quality adaptabilityImage analysisThree-dimensional object recognitionPoint cloudVisual perception
The invention discloses a three-dimensional face recognition device and method based on three-dimensional point cloud; the device comprises a characteristic area detection unit for positioning a three-dimensional point cloud characteristic area; a mapping unit for normalized mapping the three-dimensional point cloud to a depth image space; a data calculating unit using Gabor filters in different dimensions and directions to calculate responses in different dimensions and directions of three-dimensional face data; a storage unit for training a visual dictionary of the obtained three-dimensional face data; a mapping calculating unit for implementing histogram projection to the visual dictionary with respect to the Gabor response vector obtained by each pixel; a classification calculating unit for coarse classification of the three-dimensional face data; and an identification calculating unit for identifying the three-dimensional face data. By employing the disclosed technical scheme, the detail texture description ability of the three-dimensional data is stronger, and the adaptability to the quality of the input three-dimensional point cloud face data is better, thus the application foreground is better.
Owner:SHENZHEN WEITESHI TECH

Personnel reidentification method based on deep learning and distance metric learning

The invention relates to the field of the identification method, and particularly relates to a personnel reidentification method based on deep learning and distance metric learning. The identificationmethod comprises the steps that (1) a pedestrian target detection method based on the convolutional neural network is adopted to process the video data so as to detect the pedestrian target in the video; (2) the initial characteristics of the pedestrian target are coded by using an unsupervised RBM network through the bottom-up mode so as to obtain a visual dictionary having sparsity and selectivity; (3) supervised fine adjustment is performed on the initial visual dictionary by using error back propagation so as to obtain the new image expression mode of the video image, i.e. the image deeplearning representation vector; and (4) the metric space closer to the real semantics is acquired by using the distance metric learning method of characteristic grouping and characteristic value optimization, and the pedestrian target is identified by using a linear SVM classifier. The essential attributes of the image can be more accurately expressed so as to greatly enhance the accuracy of pedestrian reidentification.
Owner:江苏测联空间大数据应用研究中心有限公司

Visual feature representing method based on autoencoder word bag

ActiveCN104239897ARequirements for reducing the number of training samplesProof of validityCharacter and pattern recognitionNeural learning methodsSlide windowVisual perception
The invention relates to a visual feature representing method based on an autoencoder word bag. The method includes the steps that training samples are input to form a training set; the training samples in the training set are preprocessed, and influences of illumination, noise and the like on the image representing accuracy are reduced; a visual dictionary is generated, an autoencoder is used for extracting random image block features, then the clustering method is used for clustering the random image block features into a plurality of visual words, and the visual dictionary is composed of the visual words; a sliding window mode is used for sequentially collecting image blocks of images in the training set, the collected image blocks serve as input of the autoencoder, and output of the autoencoder is local features of the images; the local features of the images are quantized into the visual words according to the visual dictionary; the frequency of the visual words is counted, a visual word column diagram is generated, and the visual word column diagram is overall visual feature representing of the images. By means of the visual feature representing method, feature representing characteristics are independently studied through the adoption of the autoencoder, and requirements for the quantity of the training samples are reduced through a BoVW framework.
Owner:TIANJIN UNIV

Object classification method and system based on bag of visual word model

ActiveCN104915673AReduce consumptionFlexible and more accurate classification and identification methodsCharacter and pattern recognitionClassification methodsVisual perception
The invention provides an object classification method and system based on a bag of visual word model. The method comprises the following steps: obtaining characteristic points of a sample picture and obtaining position information and description information of each characteristic point, wherein the sample picture comprises a first classification picture and a second classification picture; clustering the description information of each characteristic point so as to generate a visual dictionary taking the description information as visual terms; based on the description information of a target characteristic point in each characteristic point, finding one or more visual terms matching the description information of the target characteristic points; based on the position information of each characteristic point, calculating the weight of the description information of each characteristic point for the target characteristic points on the visual terms matching the target characteristic points; and through combination with all the target characteristic points, based on the position information of all the target characteristic points, generating a characteristic model, which is provided with space information and based on the weights of the visual terms, of the sample picture.
Owner:RICOH KK

Image characteristic extracting and describing method

The invention relates to the field of image processing and computer vision and particularly provides an image characteristic extracting and describing method which is suitable for a BoW (Bag of Words) model and is applied to the field of computer vision. The image characteristic extracting and describing method comprises the following steps of: carrying out format judgment on an input image, not processing if the input image is a gray level image and converting the input image into an HSV (Hue, Saturation, Value) model if the input image is not the gray level image; selecting scale parameters; by adopting a uniform sampling method, according to the selected scale parameters, extracting characteristic points of the image at equal pixel intervals, calculating DF-SIFT (Dense Fast-Scale Invariant Feature Transform) descriptors of an H (Hue) channel, an S (Saturation) channel and a V (Value) channel of the image, applying color information into a classification task and controlling the sampling density by a parameter step to obtain the dense characteristic of the image; and carrying out description on the dense characteristic. According to the invention, by densely sampling, a visual dictionary is more accurate and reliable; and the bilinear interpolation replaces the image and Gaussian kernel function convolution process, so that the implementing process is simpler and more efficient.
Owner:HARBIN ENG UNIV

Three-dimensional face identification device and method based on three-dimensional point cloud

The invention discloses a three-dimensional face identification device and method based on the three-dimensional point cloud. The device comprises a characteristic region detection unit for carrying out positioning on a characteristic region of the three-dimensional point cloud; a depth image mapping unit for carrying out normalization mapping on the three-dimensional point cloud to a depth image space; a Gabor response computation unit for carrying out different-dimension and different-direction response computation on three-dimensional face data by utilizing different-dimension and different-direction Gabor filters; a storage unit for storing visual dictionary of the three-dimensional face data obtained by training; and a histogram mapping computation unit for carrying out histogram mapping with the visual dictionary, for the Gabor response vector obtained by each pixel. To begin with, positioning and registeration are carried out on the characteristic region of a three-dimensional face region; then, point cloud data is mapped to a depth image according to the depth information; next, visual dictionary histogram vector of the three-dimensional data is carried out according to the trained three-dimensional face visual dictionary; and finally, identification can be realized through a classifier, and the identification precision is high.
Owner:SHENZHEN WEITESHI TECH

No-reference fuzzy distorted stereo image quality evaluation method

The invention discloses a no-reference fuzzy distorted stereo image quality evaluation method, which comprises the following steps: in a training stage, selecting a plurality of undistorted stereo images and corresponding fuzzy distorted stereo images to form a training image set; then, carrying out a dictionary training operation by a Fast-ICA (independent component analysis) method, and constructing a visual dictionary table of each image in the training image set; constructing the visual quality table of the visual dictionary table of each distorted stereo image by calculating a distance between the visual dictionary table of each undistorted stereo image and the visual dictionary table of each corresponding fuzzy distorted stereo image; in a testing stage, for any one tested stereo image, carrying out non-overlapping partitioning processing to a left sight point image and a right sight point image of the tested stereo image; and according to the constructed visual dictionary table and the constructed visual quality table, obtaining an objective evaluation prediction value of the image quality of the tested stereo image. The no-reference fuzzy distorted stereo image quality evaluation method has the advantages of low computation complexity and good relevance between an objective evaluation result and subjective perception.
Owner:创客帮(山东)科技服务有限公司

Picture searching method based on maximum similarity matching

ActiveCN104615676AEliminate multiple matchesEnhanced Visual CompatibilityStill image data indexingCharacter and pattern recognitionFeature setReverse index
The invention relates to a picture searching method based on maximum similarity matching. The method includes the following steps that (1) a training picture set is acquired; (2) feature point detection and description are conducted on acquired pictures in a multi-scale space; (3) feature sets extracted in the second step are clustered and generated into a visual dictionary including k visual vocabularies; (4) each feature extracted in the second step is mapped to the visual vocabulary with the distance being smallest to the current feature l2, the current feature and the normalization residual vector of the corresponding visual vocabulary are stored in a reverse index structure, and accordingly a query database is formed; (5) the pictures to be searched for are acquired, the second step and the fourth step are executed again, the reverse index structure of the pictures to be searched for is acquired, the query database is searched for according to the reverse index structure, and the searching results of the pictures to be searched for are acquired based on the maximum similarity matching. Compared with the prior art, the picture searching method has the advantages of being good in robustness, high in computational efficiency and the like.
Owner:TONGJI UNIV

Method for detecting and counting major vegetable pests in South China based on machine vision

The present invention discloses a method for detecting and counting major vegetable pests in South China based on machine vision, comprising the steps of: the pest image preprocessing step; the pest target segmenting and extracting step; and the pest recognizing and counting step. In detail, the method comprises; through the training of SVM classifiers for different pests, optimizing the pest images after standardization; performing HOG characteristic extraction and description; using the K-Means clustering algorithm for clustering the extracted characteristics; constructing a visual dictionary and then using the SPM generated image's histogram representation; and finally, conducting preliminary classification through the morphological data of the target image; selecting different SVM classifiers, outputting the identification results, and identifying and counting the pests. The method of the invention is based on the image's pest automatic identifying and counting technology, which can realize the rapid identifying and counting of the main pest image data of the South China vegetables so as to assist peasants or the grassroots plant protection personnel to carry out the monitoring work of vegetable pests. To a large extent, their labor intensity is reduced.
Owner:SOUTH CHINA AGRI UNIV +1
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products