Target retrieval method based on group of randomized visual vocabularies and context semantic information
A visual dictionary and semantic information technology, applied in computer components, secure communication devices, character and pattern recognition, etc., can solve problems such as high computational complexity, achieve enhanced differentiation, solve computational complexity, and reduce semantic gaps Effect
- Summary
- Abstract
- Description
- Claims
- Application Information
AI Technical Summary
Problems solved by technology
Method used
Image
Examples
Embodiment 1
[0071] Embodiment 1: This embodiment is based on a target retrieval method based on randomizing visual dictionary groups and contextual semantic information. First, in view of the low efficiency of traditional clustering algorithms and the problems of visual word synonymy and ambiguity, E 2 LSH clusters the local feature points of the training image database to generate a set of randomized visual dictionary groups that support dynamic expansion; secondly, select the query image and define the target area with a rectangular frame, and then extract the query image and image database according to Lowe's method. SIFT features and E on them 2 LSH mapping to achieve the matching of feature points and visual words; then, on the basis of the language model, the retrieval score of each visual word in the query image is calculated by using the rectangular box area and image saliency detection, and the target model containing the semantic information of the target context is obtained. ;F...
Embodiment 2
[0073] Example 2: see figure 2 , image 3 , Figure 4 , the target retrieval method based on the randomized visual dictionary group and contextual semantic information of this embodiment adopts the following steps to generate a 2 Randomized visual dictionary set for LSH:
[0074] for each hash function g i (i=1,...,L), use it to hash map the SIFT points of the training image library respectively, and the points that are very close in the space will be stored in the same bucket of the hash table, with the center of each bucket represents a sight word, then each function g i can generate a hash table, that is, a visual dictionary. Then, L functions g 1 ,…,g L can generate a visual dictionary group, the process is as follows figure 2 shown.
[0075] Among them, the detailed process of generating a single visual dictionary can be described as follows:
[0076] (1) SIFT feature extraction of training image library. In this paper, Oxford5K, a commonly used database for ...
Embodiment 3
[0103] Embodiment 3: The difference between this embodiment and Embodiment 2 is that the following steps are used to measure the similarity:
[0104] The similarity between the query image q and any image d in the image library can be measured by the query likelihood p(q|d), then:
[0105] p ( q | d ) = Π i = 1 M q p ( q i | d ) - - - ( 14 )
[0106] Turning it into a risk minimization problem, that is, given a query image q, the risk function that returns an image d is defined as follows:
[0107]
[0108]
[0109] p(θ D |d)p(r|θ Q ,θ D )dθ Q dθ D
[0110] ...
PUM
Login to View More Abstract
Description
Claims
Application Information
Login to View More 