Indexing local features Wed March 30 Prof. Kristen Grauman UT-Austin
43 Slides7.22 MB
Indexing local features Wed March 30 Prof. Kristen Grauman UT-Austin
Matching local features Kristen Grauman
Matching local features ? Image 1 Image 2 To generate candidate matches, find patches that have the most similar appearance (e.g., lowest SSD) Simplest approach: compare them all, take the closest (or closest k, or within a thresholded distance) Kristen Grauman
Matching local features Image 1 Image 2 In stereo case, may constrain by proximity if we make assumptions on max disparities. Kristen Grauman
Indexing local features Kristen Grauman
Indexing local features Each patch / region has a descriptor, which is a point in some high-dimensional feature space (e.g., SIFT) Descriptor’s feature space Kristen Grauman
Indexing local features When we see close points in feature space, we have similar descriptors, which indicates similar local content. Descriptor’s feature space Database images Query image Kristen Grauman
Indexing local features With potentially thousands of features per image, and hundreds to millions of images to search, how to efficiently find those that are relevant to a new image? Kristen Grauman
Indexing local features: inverted file index For text documents, an efficient way to find all pages on which a word occurs is to use an index We want to find all images in which a feature occurs. To use this idea, we’ll need to map our features to “visual words”. Kristen Grauman
Text retrieval vs. image search What makes the problems similar, different? Kristen Grauman
Visual words: main idea Extract some local features from a number of images e.g., SIFT descriptor space: each point is 128-dimensional Slide credit: D. Nister, CVPR 2006
Visual words: main idea
Visual words: main idea
Visual words: main idea
Each point is a local descriptor, e.g. SIFT vector.
Visual words Map high-dimensional descriptors to tokens/words by quantizing the feature space Quantize via clustering, let cluster centers be the prototype “words” Word #2 Descriptor’s feature space Determine which word to assign to each new image region by finding the closest cluster center. Kristen Grauman
Visual words Example: each group of patches belongs to the same visual word Figure from Sivic & Zisserman, ICCV Kristen Grauman
Visual words and textons First explored for texture and material representations Texton cluster center of filter responses over collection of images Describe textures and materials based on distribution of prototypical texture elements. Leung & Malik 1999; Varma & Zisserman, 2002 Kristen Grauman
Recall: Texture representation example Both mean d/dx value mean d/dy value Win. #1 4 10 Win.#2 18 7 20 20 Win.#9 Dimension 1 (mean d/dx value) Windows with small gradient in both directions Windows with primarily vertical edges Dimension 2 (mean d/dy value) Windows with primarily horizontal edges statistics to summarize patterns in small windows Kristen Grauman
Visual vocabulary formation Issues: Sampling strategy: where to extract features? Clustering / quantization algorithm Unsupervised vs. supervised What corpus provides features (universal vocabulary?) Vocabulary size, number of words Kristen Grauman
Inverted file index Database images are loaded into the index mapping words to image numbers Kristen Grauman
Inverted file index When will this give us a significant gain in efficiency? New query image is mapped to indices of database images that share a word. Kristen Grauman
If a local image region is a visual word, how can we summarize an image (the document)?
Analogy to documents Of all the sensory impressions proceeding to the brain, the visual experiences are the dominant ones. Our perception of the world around us is based essentially on the messages that reach the brain from our eyes. For a long time it was thought that the retinal sensory, image was transmitted pointbrain, by point to visual centers in the brain; the cerebral cortex was visual, perception, a movie screen, so to speak, upon which the cerebral cortex, image inretinal, the eye was projected. Through the discoveries ofeye, Hubelcell, and Wiesel we now optical know that behind the origin of the visual image perception in thenerve, brain there is a considerably more complicated course of events. By Hubel, Wiesel following the visual impulses along their path to the various cell layers of the optical cortex, Hubel and Wiesel have been able to demonstrate that the message about the image falling on the retina undergoes a stepwise analysis in a system of nerve cells stored in columns. In this system each cell has its specific function and is responsible for a specific detail in the pattern of the retinal image. China is forecasting a trade surplus of 90bn ( 51bn) to 100bn this year, a threefold increase on 2004's 32bn. The Commerce Ministry said the surplus would be created by a predicted 30% jump in exports to 750bn, compared with a 18% rise in imports to 660bn. The figures are likely to further China, trade, annoy the US, which has long argued that surplus, commerce, China's exports are unfairly helped by a deliberately undervalued yuan. Beijing exports, imports, US, agrees the surplus is too high, but says the yuan, bank, domestic, yuan is only one factor. Bank of China governor Zhou Xiaochuan said the country foreign, increase, also needed to do more to boost domestic value demand so moretrade, goods stayed within the country. China increased the value of the yuan against the dollar by 2.1% in July and permitted it to trade within a narrow band, but the US wants the yuan to be allowed to trade freely. However, Beijing has made it clear that it will take its time and tread carefully before allowing the yuan to rise further in value. ICCV 2005 short course, L. Fei-Fei
Bags of visual words Summarize entire image based on its distribution (histogram) of word occurrences. Analogous to bag of words representation commonly used for documents.
Comparing bags of words Rank frames by normalized scalar product between their (possibly weighted) occurrence counts---nearest neighbor search for similar images. [1 8 1 4] [5 1 1 ⟨ 𝑑 𝑗 , 𝑞⟩ 𝑠𝑖𝑚 ( 𝑑 𝑗 ,𝑞 ) ‖𝑑 𝑗‖‖𝑞‖ 0] 𝑉 𝑑 𝑗 (𝑖 ) 𝑞(𝑖) ¿ dj q 𝑖 1 𝑉 2 𝑑 𝑗 (𝑖) 𝑖 1 𝑉 𝑞( 𝑖) 2 𝑖 1 for vocabulary of V words Kristen Grauman
tf-idf weighting Term frequency – inverse document frequency Describe frame by frequency of each word within it, downweight words that appear often in the database (Standard weighting for text retrieval) Number of occurrences of word i in document d Total number of documents in database Number of words in document d Number of documents word i occurs in, in whole database Kristen Grauman
Bags of words for content-based image retrieval Slide from Andrew Zisserman Sivic & Zisserman, ICCV 2003
Slide from Andrew Zisserman Sivic & Zisserman, ICCV 2003
1. Collect all words within query region 2. Inverted file index to find relevant frames 3. Compare word counts 4. Spatial verification Sivic & Zisserman, ICCV 2003 Demo online at : http://www.robots.ox.ac.uk/ vgg/research/vgoogle/index.h tml K. Grauman, B. Leibe Query region Retrieved frames Visual Perceptual Object and Recognition Sensory Augmented Tutorial Computing Video Google System 32
Scoring retrieval quality Results (ordered): Database size: 10 images Relevant (total): 5 images Query precision #relevant / #returned recall #relevant / #total relevant 1 precision 0.8 0.6 0.4 0.2 0 0 0.2 0.4 0.6 0.8 1 recall Slide credit: Ondrej Chum
Visual Perceptual Object and Recognition Sensory Augmented Tutorial Computing Vocabulary Trees: hierarchical clustering for large vocabularies Tree construction: [Nister & Stewenius, CVPR’06 Slide credit: David Nister
Visual Perceptual Object and Recognition Sensory Augmented Tutorial Computing Vocabulary Tree Training: Filling the tree [Nister & Stewenius, CVPR’06 K. Grauman, B. Leibe Slide credit: David Nister
Visual Perceptual Object and Recognition Sensory Augmented Tutorial Computing Vocabulary Tree Training: Filling the tree [Nister & Stewenius, CVPR’06 K. Grauman, B. Leibe Slide credit: David Nister
Visual Perceptual Object and Recognition Sensory Augmented Tutorial Computing Vocabulary Tree Training: Filling the tree [Nister & Stewenius, CVPR’06 K. Grauman, B. Leibe Slide credit: David Nister
Visual Perceptual Object and Recognition Sensory Augmented Tutorial Computing Vocabulary Tree Training: Filling the tree [Nister & Stewenius, CVPR’06 K. Grauman, B. Leibe Slide credit: David Nister
Visual Perceptual Object and Recognition Sensory Augmented Tutorial Computing Vocabulary Tree Training: Filling the tree [Nister & Stewenius, CVPR’06 K. Grauman, B. Leibe Slide credit: David Nister 39
What is the computational advantage of the hierarchical representation bag of words, vs. a flat vocabulary?
Visual Perceptual Object and Recognition Sensory Augmented Tutorial Computing Vocabulary Tree Recognition RANSAC verification [Nister & Stewenius, CVPR’06 Slide credit: David Nister
Bags of words: pros and cons flexible to geometry / deformations / viewpoint compact summary of image content provides vector representation for sets very good results in practice - basic model ignores geometry – must verify afterwards, or encode via features - background and foreground mixed when bag covers whole image - optimal vocabulary formation remains unclear
Summary Matching local invariant features: useful not only to provide matches for multi-view geometry, but also to find objects and scenes. Bag of words representation: quantize feature space to make discrete set of visual words – Summarize image by distribution of words – Index individual words Inverted index: pre-compute index to enable faster search at query time