1
Hi hi lM hi P if I Hierarchical Matching Pursuit for Image Hierarchical Matching Pursuit for Image Hierarchical Matching Pursuit for Image Cl ifi ti A hit t dF t Al ith Classification: Architecture and Fast Algorithms Classification: Architecture and Fast Algorithms Classification: Architecture and Fast Algorithms 1 2 1 Liefeng Bo 1 Xiaofeng Ren 2 and Dieter Fox 1 Liefeng Bo 1 , Xiaofeng Ren 2 and Dieter Fox 1 Liefeng Bo , Xiaofeng Ren and Dieter Fox 1 Ui it fW hi t 2 ItlLb 1 University of Washington 2 Intel Labs University of Washington, Intel Labs Thi k Thi k Hi hi lM t hi P it E d Hi hi lM t hi P it E d This work This work Hierarchical Matching Pursuit Encoder Hierarchical Matching Pursuit Encoder This work This work Hierarchical Matching Pursuit Encoder Hierarchical Matching Pursuit Encoder Hi hi l hi i b ild f hi h l b l Image Features Hierarchical matching pursuit builds a feature hierarchy layer -by-layer Image Features i ffi i t t hi it d Second layer: sparse codes over using an efficient matching pursuit encoder. Second layer: sparse codes over Hi hi lM t hi P it (HMP) patch-level features are aggregated Hierarchical Matching Pursuit (HMP) patch level features are aggregated h hl i d M t hi it d it f th dl btht th l Matching Pursuit across the whole image to produce Matching pursuit encoder consists of three modules: batch tree orthogonal Encoder -II image level features matching pursuit spatial pyramid max pooling and contrast normalization; Encoder II image-level features. matching pursuit, spatial pyramid max pooling, and contrast normalization; Recursively run matching pursuit encoder; Recursively run matching pursuit encoder; Matching Pursuit Extract features from a typical 300×300 image in less than 1 second; Matching Pursuit Extract features from a typical 300 300 image in less than 1 second; f l i l d k d b di l l Encoder -I Outperform convolutional deep networks and SIFT based single layer sparse First layer: sparse codes over pixel coding in terms of accuracy First layer: sparse codes over pixel l l di h l l Patch coding in terms of accuracy. level are aggregated into patch-level Patch (16x16 pixels) features (16x16 pixels) features. Image K SVD (Dictionary Learning) SVD (Dictionary Learning) K-SVD (Dictionary Learning) SVD (Dictionary Learning) K SVD [1] learns a dictionary D and an associated sparse code matrix X K-SVD [1] learns a dictionary D and an associated sparse code matrix X from observations Y by minimizing the following reconstruction error Object Recognition (Caltech Object Recognition (Caltech-101) 101) from observations Y by minimizing the following reconstruction error Object Recognition (Caltech Object Recognition (Caltech-101) 101) 2 Dictionary is learned by K-SVD on 1,000,000 sampled patches in each layer; K x i t s DX Y F 0 2 || || || || min Dictionary is learned by K SVD on 1,000,000 sampled patches in each layer; i l li h fi d dl i b d i l K x i t s DX Y i F X D 0 , || || , . . || || min Sparsity level in the first and second layers is set to be 5 and 10, respectively; X D , The problem can be solved in an alternating manner In the first stage D is fixed Di ti i i 3ti th filt i i th fi tl d 1000 i th dl The problem can be solved in an alternating manner. In the first stage, D is fixed Dictionary size is 3 times the filter size in the first layer and 1000 in the second layer; and only the sparse codes are computed by orthogonal matching pursuit. Matching pursuit encoder is run on 16×16 image patches over dense grids with a step and only the sparse codes are computed by orthogonal matching pursuit. Matching pursuit encoder is run on 16×16 image patches over dense grids with a step i f i li h fi l dh hl i i h dl size of 4 pixels in the first layer and the whole image in the second layer; K t D 2 || || || || i T i li SVM 30 i dt t th 50 i t K x t s Dx y i i i 0 2 || || . . || || min Train linear SVM on 30 images and test on no more than 50 images per category. i i i x i 0 I th dt h filt i D d it itd d dtd DCT K SVD In the second stage, each filter in D and its associated sparse codes x are updated DCT K-SVD simultaneously by Singular Value Decomposition (||d || =1) simultaneously by Singular Value Decomposition (||d k || 2 1) Filters used in Discrete Cosine 2 2 2 T T T Filters used in Discrete Cosine 2 2 2 || || || || || || F T k k k F T k k T j j F x d E x d x d Y DX Y = = Transformation (DCT) and || || || || || || F k k k F k j k k j j F x d E x d x d Y DX Y Transformation (DCT) and k j When the sparsity level K is set to be 1 and sparse codes are forced to be a filters learned by K-SVD When the sparsity level K is set to be 1 and sparse codes are forced to be a filters learned by K-SVD binary(0/1), K-Means is exactly reproduced (no constraints on d k ). binary(0/1), K Means is exactly reproduced (no constraints on d k ). [1] Aharon, Elad, and Bruckstein, IEEE Transactions on Signal Processing K SVD d DCT ith diff t filt i f th fi tl K-SVD and DCT with different filter sizes for the first layer Batch Tree Orthogonal Matching Pursuit Batch Tree Orthogonal Matching Pursuit Filter size 3×3 4×4 5×5 6×6 7×7 8×8 Batch Tree Orthogonal Matching Pursuit Batch Tree Orthogonal Matching Pursuit Filter size 3×3 4×4 5×5 6×6 7×7 8×8 DCT ( h l) 69 9 70 8 71 5 72 1 73 2 73 1 DCT (orthogonal) 69.9 70.8 71.5 72.1 73.2 73.1 DCT (overcomplete) 69 6 71 8 73 0 74 1 73 7 73 4 DCT (overcomplete) 69.6 71.8 73.0 74.1 73.7 73.4 K-SVD 71.8 74.4 75.9 76.8 76.3 76.1 K SVD 71.8 74.4 75.9 76.8 76.3 76.1 S ti lP id P li d C t tN li ti i ii Spatial Pyramid Pooling and Contrast Normalization improve recognition accuracy by about 2% and 3% respectively Large Dictionary with 10 000 filters in accuracy by about 2% and 3%, respectively. Large Dictionary with 10,000 filters in the second layer is slightly better than standard setting with 1000 filters. Hierarchical Matching Pursuit with K=1 (zero norm): about 74.0%; Hierarchical Matching Pursuit with K 1 (zero norm): about 74.0%; Running Time over a typical 300×300 image Running Time over a typical 300×300 image Al ith HMP (DCT) HMP (K SVD) SIFT+SC DN Algorithms HMP (DCT) HMP (K-SVD) SIFT+SC DN Running time (Seconds) 04 08 22 4 67 5 Running time (Seconds) 0.4 0.8 22.4 67.5 Comparisons with State of the art (Single Feature based Algorithms) Comparisons with State-of-the-art (Single Feature based Algorithms) Layer Single based SIFT Layers Multiple 4 4 4 4 8 4 4 4 4 7 6 4 4 4 4 4 4 8 4 4 4 4 4 4 7 6 K-Means is used to group the whole dictionary into the sub-dictionaries and [1] [2] [3] [4] [5] [6] [7] [8] associate the sub dictionaries with the learned center matrix C; HMP ISPD [1] CDBN [2] DN [3] HSC [4] KDES-G [5] SPM [6] SIFT+SP [7] Macrofeatures [8] associate the sub-dictionaries with the learned center matrix C; Line 5 selects the center filter j that best matches the current residual; 76.8 65.5 65.5 66.9 74.0 75.2 64.4 73.2 75.7 Line 5 selects the center filter j that best matches the current residual; Line 6 selects the filter within the sub-dictionary associated with the center j; [1] Kavukcuoglu, Ranzato, Fergus, and LeCun, CVPR 2009 [2] Lee, Grosse, Ranganath, and Ng, ICML 2009 Line 6 selects the filter within the sub-dictionary associated with the center j; [1] Kavukcuoglu, Ranzato, Fergus, and LeCun, CVPR 2009 [2] Lee, Grosse, Ranganath, and Ng, ICML 2009 [3] Zeiler Krishnan Taylor and Fergus CVPR 2010 [4] Yu Lin and Lafferty CVPR 2011 (Parallel work) Line 8 updates sparse codes with the incremental Cholesky decomposition; [3] Zeiler, Krishnan, Taylor, and Fergus, CVPR 2010 [4] Yu, Lin, and Lafferty, CVPR 2011 (Parallel work) Line 8 updates sparse codes with the incremental Cholesky decomposition; [5]Bo, Ren, and Fox, NIPS 2010 [6] Lazebnik, Schmid, and Ponce, CVPR 2006 Line 9 computes the correlation between each center and the current residual; [7]Yang Yu Gong and Huang CVPR 2009 [8] Boureau Bach LeCun and Ponce CVPR 2010 Line 9 computes the correlation between each center and the current residual; [7] Yang, Yu, Gong, and Huang, CVPR 2009 [8] Boureau, Bach, LeCun, and Ponce, CVPR 2010 If the centers C are set to be the whole dictionary, BTOMP exactly recovers the batch (exact) orthogonal matching pursuit [1] batch (exact) orthogonal matching pursuit [1] . S R iti (MIT S R iti (MIT S ) S ) [1] Rubinstein Zibulevsky and Elad Technical report 2008 Scene Recognition (MIT Scene Recognition (MIT -Scene) Scene) [1] Rubinstein, Zibulevsky , and Elad,Technical report, 2008 Scene Recognition (MIT Scene Recognition (MIT Scene) Scene) Thi d t t ti 15620 i f 67 i d t i This dataset contain 15620 images from 67 indoor scene categories; M t hi P it E d M t hi P it E d Train linear SVM on 80 images and test on 20 images per categor ; Matching Pursuit Encoder Matching Pursuit Encoder Train linear SVM on 80 images and test on 20 images per category; Matching Pursuit Encoder Matching Pursuit Encoder The experimental setting is same as with the Caltech 101 dataset except that the The experimental setting is same as with the Caltech-101 dataset except that the Matching pursuit encoder filter size is 4×4 (cross validation). Matching pursuit encoder filter size is 4 4 (cross validation). consists of three modules: BTOMP Algorithms HMP OB [1] GIST [2] ROI+GIST [2] SIFT+SC consists of three modules: BTOMP, Algorithms HMP OB [1] GIST [2] ROI+GIST [2] SIFT+SC Spatial Pyramid Max Pooling Accuracy 41 8 37 6 22 0 26 0 36 9 Spatial Pyramid Max Pooling Accuracy 41.8 37.6 22.0 26.0 36.9 and Contrast Normalization [1] Li S Xi dFi Fi NIPS 2010 [2] Q tt i dT lb CVPR 2009 and Contrast Normalization. [1] Li, Su, Xing, and Fei-Fei., NIPS 2010 [2] Quattoni and Torralba., CVPR 2009 Spatial Pyramid Max Pooling aggregates the sparse codes which are spatially close Spatial Pyramid Max Pooling aggregates the sparse codes which are spatially close, E tR iti (UIUC E tR iti (UIUC S t) S t) using max pooling in a multi-level patch decomposition. Event Recognition (UIUC Event Recognition (UIUC-Sports) Sports) using max pooling in a multi level patch decomposition. Event Recognition (UIUC Event Recognition (UIUC Sports) Sports) 2 2 4 4 Thi d t t it f8 t t t i ith 137 t 250 i i h } 1 2 2 4 4 4 8 4 7 6 4 4 4 4 4 8 4 4 4 4 4 7 6 × × This dataset consists of 8 sport event categories with 137 to 250 images in each. | | max | | max ) ( L = x x P F T i li SVM 70 i dt t 60 i t 1 | | max , |, | max ) ( L = jm P j j P j x x P F Train linear SVM on 70 images and test on 60 images per category. P j P j Contrast Normalization is helpful since the magnitude of sparse codes varies The experimental setting is same as with the MIT Scene dataset Contrast Normalization is helpful since the magnitude of sparse codes varies i ifi tl d t ill i ti df d b k d t t The experimental setting is same as with the MIT -Scene dataset. significantly due to illumination and foreground-background contrast. Methods HMP OB SIFT+GMM [1] SIFT+SC A 85 7 76 3 73 4 82 7 ) ( ) ( P F P F Accuracy 85.7 76.3 73.4 82.7 = ) ( ) ( P F P F ε + 2 || ) ( || P F [1] Li and Fei-Fei., ICCV 2007 ε + || ) ( || P F

Hi hi l M hi P i f IHierarchical Matching Pursuit for …research.cs.washington.edu/istc/lfb/paper/hmp_poster.pdf · 2012. 1. 31. · Mthi P itE dMatching Pursuit EncoderMatching

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Hi hi l M hi P i f IHierarchical Matching Pursuit for …research.cs.washington.edu/istc/lfb/paper/hmp_poster.pdf · 2012. 1. 31. · Mthi P itE dMatching Pursuit EncoderMatching

Hi hi l M hi P i f IHierarchical Matching Pursuit for ImageHierarchical Matching Pursuit for ImageHierarchical Matching Pursuit for Image g gCl ifi ti A hit t d F t Al ithClassification: Architecture and Fast AlgorithmsClassification: Architecture and Fast AlgorithmsClassification: Architecture and Fast Algorithmsg

1 2 1Liefeng Bo1 Xiaofeng Ren2 and Dieter Fox1Liefeng Bo1, Xiaofeng Ren2 and Dieter Fox1 Liefeng Bo , Xiaofeng Ren and Dieter Fox1U i it f W hi t 2I t l L b1University of Washington 2Intel LabsUniversity of Washington, Intel Labs

Thi kThi k Hi hi l M t hi P it E dHi hi l M t hi P it E dThis workThis work Hierarchical Matching Pursuit EncoderHierarchical Matching Pursuit EncoderThis workThis work Hierarchical Matching Pursuit EncoderHierarchical Matching Pursuit EncoderHi hi l hi i b ild f hi h l b l Image FeaturesHierarchical matching pursuit builds a feature hierarchy layer-by-layer Image Featuresg p y y y y

i ffi i t t hi it d Second layer: sparse codes overusing an efficient matching pursuit encoder. Second layer: sparse codes over Hi hi l M t hi P it (HMP) patch-level features are aggregatedHierarchical Matching Pursuit (HMP) patch level features are aggregated

h h l i dg ( )

M t hi it d i t f th d l b t h t th lMatching Pursuit across the whole image to produce Matching pursuit encoder consists of three modules: batch tree orthogonal

gEncoder-II

g pimage level featuresmatching pursuit spatial pyramid max pooling and contrast normalization;

Encoder II image-level features.matching pursuit, spatial pyramid max pooling, and contrast normalization; Recursively run matching pursuit encoder;Recursively run matching pursuit encoder;

Matching PursuitExtract features from a typical 300×300 image in less than 1 second; Matching Pursuit Extract features from a typical 300 300 image in less than 1 second;f l i l d k d b d i l l

Encoder-IOutperform convolutional deep networks and SIFT based single layer sparse First layer: sparse codes over pixelp p g y pcoding in terms of accuracy

First layer: sparse codes over pixel l l d i h l lPatchcoding in terms of accuracy. level are aggregated into patch-level Patch

(16x16 pixels) gg g pfeatures

(16x16 pixels)features.

Image

KK SVD (Dictionary Learning)SVD (Dictionary Learning)g

KK--SVD (Dictionary Learning)SVD (Dictionary Learning)( y g)( y g)K SVD[1] learns a dictionary D and an associated sparse code matrix XK-SVD[1] learns a dictionary D and an associated sparse code matrix Xfrom observations Y by minimizing the following reconstruction error Object Recognition (CaltechObject Recognition (Caltech--101)101)from observations Y by minimizing the following reconstruction error Object Recognition (CaltechObject Recognition (Caltech--101)101)j g (j g ( ))

2 Dictionary is learned by K-SVD on 1,000,000 sampled patches in each layer;KxitsDXY F ≤∀− 02 ||||||||min Dictionary is learned by K SVD on 1,000,000 sampled patches in each layer;

i l l i h fi d d l i b d i lKxitsDXY iFXD

≤∀ 0,||||,..||||min

Sparsity level in the first and second layers is set to be 5 and 10, respectively;XD,

The problem can be solved in an alternating manner In the first stage D is fixedp y y , p y;

Di ti i i 3 ti th filt i i th fi t l d 1000 i th d lThe problem can be solved in an alternating manner. In the first stage, D is fixed Dictionary size is 3 times the filter size in the first layer and 1000 in the second layer;and only the sparse codes are computed by orthogonal matching pursuit. Matching pursuit encoder is run on 16×16 image patches over dense grids with a stepand only the sparse codes are computed by orthogonal matching pursuit. Matching pursuit encoder is run on 16×16 image patches over dense grids with a step

i f i l i h fi l d h h l i i h d lsize of 4 pixels in the first layer and the whole image in the second layer; KtD ≤2 ||||||||i

p y g y ;T i li SVM 30 i d t t th 50 i t

KxtsDxy iii ≤− 02 ||||..||||min

Train linear SVM on 30 images and test on no more than 50 images per category.y iii

xi0||||||||

I th d t h filt i D d it i t d d d t d DCT K SVDIn the second stage, each filter in D and its associated sparse codes x are updated DCT K-SVDg p psimultaneously by Singular Value Decomposition (||d || =1)simultaneously by Singular Value Decomposition (||dk||2 1)

Filters used in Discrete Cosine222 TTT∑ Filters used in Discrete Cosine 222 |||||||||||| FTkkkF

Tkk

TjjF xdExdxdYDXY −=−−=− ∑

Transformation (DCT) and|||||||||||| FkkkF

kjkkjjF xdExdxdYDXY ∑

≠ Transformation (DCT) andkj≠

When the sparsity level K is set to be 1 and sparse codes are forced to be a filters learned by K-SVDWhen the sparsity level K is set to be 1 and sparse codes are forced to be a filters learned by K-SVDbinary(0/1), K-Means is exactly reproduced (no constraints on dk).binary(0/1), K Means is exactly reproduced (no constraints on dk).[1] Aharon, Elad, and Bruckstein, IEEE Transactions on Signal Processing

K SVD d DCT ith diff t filt i f th fi t lK-SVD and DCT with different filter sizes for the first layery

Batch Tree Orthogonal Matching PursuitBatch Tree Orthogonal Matching Pursuit Filter size 3×3 4×4 5×5 6×6 7×7 8×8Batch Tree Orthogonal Matching PursuitBatch Tree Orthogonal Matching Pursuit Filter size 3×3 4×4 5×5 6×6 7×7 8×8DCT ( h l) 69 9 70 8 71 5 72 1 73 2 73 1

g gg gDCT (orthogonal) 69.9 70.8 71.5 72.1 73.2 73.1

DCT (overcomplete) 69 6 71 8 73 0 74 1 73 7 73 4DCT (overcomplete) 69.6 71.8 73.0 74.1 73.7 73.4K-SVD 71.8 74.4 75.9 76.8 76.3 76.1K SVD 71.8 74.4 75.9 76.8 76.3 76.1

S ti l P id P li d C t t N li ti i i iSpatial Pyramid Pooling and Contrast Normalization improve recognition p y g p gaccuracy by about 2% and 3% respectively Large Dictionary with 10 000 filters inaccuracy by about 2% and 3%, respectively. Large Dictionary with 10,000 filters in the second layer is slightly better than standard setting with 1000 filters.y g y g

Hierarchical Matching Pursuit with K=1 (zero norm): about 74.0%;Hierarchical Matching Pursuit with K 1 (zero norm): about 74.0%;

Running Time over a typical 300×300 imageRunning Time over a typical 300×300 image

Al ith HMP (DCT) HMP (K SVD) SIFT+SC DNAlgorithms HMP (DCT) HMP (K-SVD) SIFT+SC DNRunning time (Seconds) 0 4 0 8 22 4 67 5Running time (Seconds) 0.4 0.8 22.4 67.5

Comparisons with State of the art (Single Feature based Algorithms)Comparisons with State-of-the-art (Single Feature based Algorithms)LayerSingle based SIFTLayersMultiple 4444 84444 76444444 8444444 76 ygyp

K-Means is used to group the whole dictionary into the sub-dictionaries and [1] [2] [3] [4] [5] [6] [7] [8]

8686g p y

associate the sub dictionaries with the learned center matrix C; HMP ISPD[1] CDBN[2] DN[3] HSC[4] KDES-G[5] SPM[6] SIFT+SP[7] Macrofeatures[8] associate the sub-dictionaries with the learned center matrix C;Line 5 selects the center filter j that best matches the current residual; 76.8 65.5 65.5 66.9 74.0 75.2 64.4 73.2 75.7Line 5 selects the center filter j that best matches the current residual;Line 6 selects the filter within the sub-dictionary associated with the center j; [1] Kavukcuoglu, Ranzato, Fergus, and LeCun, CVPR 2009 [2] Lee, Grosse, Ranganath, and Ng, ICML 2009Line 6 selects the filter within the sub-dictionary associated with the center j; [1] Kavukcuoglu, Ranzato, Fergus, and LeCun, CVPR 2009 [2] Lee, Grosse, Ranganath, and Ng, ICML 2009

[3] Zeiler Krishnan Taylor and Fergus CVPR 2010 [4] Yu Lin and Lafferty CVPR 2011 (Parallel work)Line 8 updates sparse codes with the incremental Cholesky decomposition; [3] Zeiler, Krishnan, Taylor, and Fergus, CVPR 2010 [4] Yu, Lin, and Lafferty, CVPR 2011 (Parallel work)Line 8 updates sparse codes with the incremental Cholesky decomposition;[5]Bo, Ren, and Fox, NIPS 2010 [6] Lazebnik, Schmid, and Ponce, CVPR 2006

Line 9 computes the correlation between each center and the current residual; [7] Yang Yu Gong and Huang CVPR 2009 [8] Boureau Bach LeCun and Ponce CVPR 2010Line 9 computes the correlation between each center and the current residual; [7] Yang, Yu, Gong, and Huang, CVPR 2009 [8] Boureau, Bach, LeCun, and Ponce, CVPR 2010

If the centers C are set to be the whole dictionary, BTOMP exactly recovers the y, ybatch (exact) orthogonal matching pursuit[1]batch (exact) orthogonal matching pursuit[1] .

S R iti (MITS R iti (MIT S )S )[1] Rubinstein Zibulevsky and Elad Technical report 2008 Scene Recognition (MITScene Recognition (MIT--Scene)Scene)[1] Rubinstein, Zibulevsky, and Elad,Technical report, 2008 Scene Recognition (MITScene Recognition (MIT Scene)Scene)Thi d t t t i 15620 i f 67 i d t iThis dataset contain 15620 images from 67 indoor scene categories;

M t hi P it E dM t hi P it E d Train linear SVM on 80 images and test on 20 images per categor ;Matching Pursuit EncoderMatching Pursuit Encoder Train linear SVM on 80 images and test on 20 images per category;Matching Pursuit EncoderMatching Pursuit EncoderThe experimental setting is same as with the Caltech 101 dataset except that theThe experimental setting is same as with the Caltech-101 dataset except that the

Matching pursuit encoder filter size is 4×4 (cross validation).Matching pursuit encoder filter size is 4 4 (cross validation).

consists of three modules: BTOMP Algorithms HMP OB[1] GIST[2] ROI+GIST[2] SIFT+SCconsists of three modules: BTOMP, Algorithms HMP OB[1] GIST[2] ROI+GIST[2] SIFT+SCSpatial Pyramid Max Pooling Accuracy 41 8 37 6 22 0 26 0 36 9Spatial Pyramid Max Pooling Accuracy 41.8 37.6 22.0 26.0 36.9

and Contrast Normalization [1] Li S Xi d F i F i NIPS 2010 [2] Q tt i d T lb CVPR 2009and Contrast Normalization. [1] Li, Su, Xing, and Fei-Fei., NIPS 2010 [2] Quattoni and Torralba., CVPR 2009

Spatial Pyramid Max Pooling aggregates the sparse codes which are spatially closeSpatial Pyramid Max Pooling aggregates the sparse codes which are spatially close, E t R iti (UIUCE t R iti (UIUC S t )S t )using max pooling in a multi-level patch decomposition. Event Recognition (UIUCEvent Recognition (UIUC--Sports)Sports)using max pooling in a multi level patch decomposition. Event Recognition (UIUCEvent Recognition (UIUC Sports)Sports)

2244Thi d t t i t f 8 t t t i ith 137 t 250 i i h}12244 4847644444 844444 76 ××

⎤⎡ This dataset consists of 8 sport event categories with 137 to 250 images in each.}||max||max)(

8686L ⎥

⎤⎢⎡= xxPF p g g

T i li SVM 70 i d t t 60 i t1 ||max,|,|max)( L∈∈ ⎥⎦

⎤⎢⎣⎡= jmPjjPj

xxPFTrain linear SVM on 70 images and test on 60 images per category.∈∈ ⎥⎦⎢⎣ PjPj

Contrast Normalization is helpful since the magnitude of sparse codes varies The experimental setting is same as with the MIT Scene datasetContrast Normalization is helpful since the magnitude of sparse codes varies i ifi tl d t ill i ti d f d b k d t t

The experimental setting is same as with the MIT-Scene dataset.significantly due to illumination and foreground-background contrast.

Methods HMP OB SIFT+GMM [1] SIFT+SCet ods O S G S SCA 85 7 76 3 73 4 82 7)()( PFPF Accuracy 85.7 76.3 73.4 82.7=

)()( PFPFε+2||)(||

)(PF

[1] Li and Fei-Fei., ICCV 2007ε+||)(|| PF

[ ] ,