33
Exploiting noisy web data for large- scale visual recognition Lamberto Ballan CVPRW WebVision - Jul 26, 2017 University of Padova, Italy

Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Exploiting noisy web data for large-scale visual recognition

Lamberto Ballan

CVPRW WebVision - Jul 26, 2017

University of Padova, Italy

Page 2: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Datasets drive computer vision progress

2

ImageNet

Slide credit: O. Russakovsky

Page 3: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

ImageNet: ILSVRC results

3

• Result in ILSVRC (classification) over the yearsEr

ror (

Top

5)

0

7.5

15

22.5

30

2010 2011 2012 2013 2014 2014 2015

3.66.77.3

11.7

16.4

25.828.2

NEC Xerox AlexNet Clarifi VGG GoogleNet ResNet

Human error (5.1)

Page 4: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

The long tail• A small number off generic objects/entities/labels

appear very often while most others appear rarely

• There are a few real-world scenarios in which we have access to 1M+ images uniformly belonging to a set of 1000+ classes

4

Labe

l Fre

quen

cy Q: How to scale up to very large vocabularies (infrequent labels) and a scenario where it is hard to collect ground truth data?

Page 5: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Images want to be shared

5

Almost all these services allow users to tag, rate, like, and swipe photos

Page 6: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Daily number of shared photos

6Source: Mary Meeker Internet Trends 2016

Page 7: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

What datasets?• Ideally the entire Web!

• In practice:

7

MIRFLICKR-25K MIRFLICKR-1M

2008

NUS-WIDE (~260K Flickr images)

2009 2015 2017

YFCC100M (100M Flickr images)

WebVision (~2.4M images)

Page 8: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Socializing the Semantic Gap: A Comparative Survey on Image Tag

Assignment, Refinement and Retrieval

Page 9: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

What has been done so far, our survey• An open test-bed for benchmarking image tagging,

tag refinement and image retrieval approaches:

‣ Jingwei: https://github.com/li-xirong/jingwei

‣ Implemented 10 methods; train on 10K-100K-1M Flickr images, test on MIRFLICKR-25K and NUS-WIDE

• Taxonomy of previous works for image tagging:

‣ learning: instance-based, model-based, transduction

‣ media/modality: tag, tag+image, tag+image+user

9

[X.Li*, T.Uricchio*, L.Ballan, M.Bertini, C.Snoek, A.DelBimbo - ACM CSUR 2016 (*equal contrib.)]

Page 10: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Instance-based a.k.a. lazy learning• A popular line of works is based on non-parametric

models where labels are transferred to new samples

‣ “Images similar in appearance are likely to share labels”

‣ e.g. JEC [IJCV’10], TagProp [CVPR’09], 2PKNN [ECCV’12,IJCV’17]

‣ works also in a cross-domain img2video scenario [CVIU’15]

• pros: can adapt to new labels and large vocabularies

• cons: i) it is a memory-based approach so it does not scale well at test time; ii) frequent labels dominate

10

[L.Ballan, M.Bertini, G.Serra, A.DelBimbo - CVIU 2015]

Page 11: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Label transfer in the semantic space• Labels associated to the training images can be

used to re-arrange the original features space

11

[T.Uricchio, L.Ballan, L.Seidenari, A.DelBimbo - PR 2017]

T. Uricchio et al. / Pattern Recognition 71 (2017) 144–157 145 Visual Space

Textual Space

sunset

cloudssea

dog

flower

Semantic Space

Φ(v; t)

Fig. 1. Labels associated to the images can be used to re-arrange the visual fea- tures and induce the semantics not caught by the original features. For instance, the sunset images with the red border should be closer to images of clouds and sea, according to the text space. A projection !( v ; t ) is learned to satisfy correla- tions in visual and textual space. (For interpretation of the references to colour in this figure legend, the reader is referred to the web version of this article.) sual content of the image. Thus the proposed algorithms tend to retrieve just the images whose features are very close in the vi- sual space, but the semantic content is not well preserved. Re- searchers tried to cope with this issue by improving visual fea- tures. To this end, the most significant improvement has been the shift from handcrafting features to end-to-end feature learning, leading to current state-of-the art convolutional neural network representations [26–28] . Nearest neighbors methods may also suf- fer when images are not paired with enough label information, leading to a poor statistical quality of the retrieved neighborhood. This is mostly due to the fact that label frequencies are usually un- balanced. Modern methods address this issue by introducing label penalties and metric learning [7,10,18] .

The image representation can be improved also by shifting to a completely different perspective, namely moving towards a multi- modal representation. A way of bridging the semantic gap might be by designing representations that account not just for the im- age pixels, but also for its textual representation. Here we follow this approach by constructing a framework in which the correla- tion between visual features and labels is maximized. To this end, we present an automatic image annotation approach that relies on Kernel Canonical Correlation Analysis (KCCA) [29] . Our approach strives to create a semantic embedding through the connection of visual and textual modalities. This embedding lives in a latent space that we refer to as semantic space . Images are mapped to this space by jointly considering the visual similarity between images in the original visual space, and label similarities. The projected images are then used to annotate new images by using a nearest- neighbor technique or other standard classifiers. Fig. 1 illustrates our pipeline. The main take-home message is that, as illustrated in the figure, the neighborhood of each image will contain more im- ages associated with the same label (e.g. “sunset”) in the semantic space than in the original visual space (see for example the images with the red border). 1.1. Main contributions

(1) The key contribution of our work is to improve image repre- sentations using a simple multimodal embedding based on KCCA. This approach has several advantages over parametric supervised learning. First, by combining a visual and textual view of the data, we reduce the semantic gap. Thus we can obtain higher similar- ities for images which are also semantically similar, according to their textual representation. Second, we are free from predetermin- ing the vocabulary of labels. This makes the approach well suited for nearest neighbor methods, which for the specific task of image annotation are more robust to label noise. A slight disadvantage

of our method is its inherent batch nature. Although, as shown in our experimental results, learning the semantic projection is also possible on a subset of the training data.

(2) Previous works that learn multimodal representations from language and imagery exist [30] , including prior uses of CCA and KCCA [29,31–33] . However, we are the first to propose a frame- work that combines the two modalities into a joint semantic space which is better exploitable by state-of-the-art nearest neighbor models. Interestingly enough, in our framework the textual infor- mation is only needed at training time, thus allowing to predict labels also for unlabeled images.

(3) We provide extensive experimental validations. Our ap- proach is tested on medium and large scale datasets, i.e. IAPR-TC12 [34] , ESP-GAME [35] , MIRFlickr-25k [36] and NUS-WIDE [37] . We show that our framework is able to leverage recently developed CNN features in order to improve the performance even further. Additionally, we introduce a tag denoising step that allows KCCA to effectively learn the semantic projections also from user-generated tags, which are available at no cost in a social media scenario. The scalability of the method is also validated with subsampling exper- iments.

This paper builds on our previous contribution on cross-modal image representations [38] and improves in many ways. We re- port new experimental evaluations covering the large dataset NUS- WIDE. Validate our pipeline with modern convolutional neural net- work based features. Extend our original approach with a new text filtering method that allows the semantic space to be computed from noisy and sparse tags, such as that from social media. Report new insights on several key aspects such as performance and scal- ability of our approach when subsampling the training set. 2. Related work 2.1. Automatic image annotation: Ideas and main trends

Automatic image annotation is a long standing area of re- search in computer vision, multimedia and information retrieval [14] . Early works often used mixture models to define a joint dis- tribution over image features and labels [1,3,39] . In these mod- els, training images are used as non-parametric density estimators over the co-occurrence of labels and images. Other popular prob- abilistic methods employed topic models, such as pLSA or LDA, to represent the joint distribution of visual and textual features [2,40,41] . They are generative models, thus they maximize the gen- erative data likelihood. They are usually expensive or require sim- plifying assumptions that can be suboptimal for predictive per- formance. Discriminative models such as support vector machines (SVM) and logistic regression have also been used extensively [22–24,42] . In these works, each label is considered separately and a specific model is trained on a per-label basis. In testing, they are used to predict whether a new image should be labeled with the corresponding label. While they are very effective, a major draw- back is that they require to define in advance the vocabulary of labels. Thus, these approaches do not handle well large-scale sce- narios in which you may have thousands of labels and the vocab- ulary may shift over time.

Despite their simplicity, a class of approaches that has gained a lot of attention is that of nearest-neighbor based methods [7,10,17,20] . Their underlying intuition is that similar images are likely to share common labels. Many of these methods start by re- trieving a set of visually similar images and then they implement a label transfer procedure to propagate the most common train- ing labels to the test image. The most recent works usually imple- ment also a refinement procedure, such as metric learning [7,10] or graph learning [43–46] , in order to differently weight rare and common labels or to capture the semantic correlation between la-

i) CCA-based embedding (expert labels or tags)

ii) Label transfer in the “semantic space”

iii) significant improvements in performance at low cost

Page 12: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Model-based methods• Learn a model for each label following a (usually)

fully supervised approach

‣ i.e. train a large CNN network

‣ e.g. WARP: deep convolutional ranking for multi-label image annotation [ICLR’14]

‣ state-of-the-art results on NUS-WIDE using an AlexNet architecture trained on Flickr images and ranking loss

12

[Y.Gong, Y.Jia, T.Leung, A.Toshev, S.Ioffe - ICLR 2014]

Page 13: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Web images are not only pixels• Can we use contextual information such as social-

network metadata to improve image classification?

‣ Image Labeling on a Network: Using Social-Network Metadata for Image Classification [ECCV’12]

13

[J.McAuley, J.Leskovec - ECCV 2012]

Relational network model where each node represents an image, with cliques formed from images sharing common properties

Page 14: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Automatic image annotation by exploiting image metadata and

weak labels

[J.Johnson*, L.Ballan*, L.Fei-Fei - ICCV 2015 (*equal contrib.)]

Page 15: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Motivation• Can you guess what’s in the image?

15

?

petal?

fruit?

tentacle?

Page 16: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition 16

• Let’s try to add more context…

Motivation

Tags:flower petal

closeup water

GPSgroups

?

Page 17: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition 17

• In the context of images which share similar metadata it is easier to give the right answer

Motivation

Page 18: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Approach• For an image x ∈ X and neighborhood z ∈ Zx, we use

a function f parameterized by w to predict labels

‣ We compute hidden state representations for the image and its neighbors

‣ Then we operate on the concatenation of these two representations to compute label scores

• We demonstrate that our model can:

‣ handle different types of image metadata

‣ adapt to changing vocabularies

18

Page 19: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Approach

19

• (1) non-parametric step to build a neighborhood

Page 20: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Approach

20

• (2) deep neural network to blend visual information from the image and its neighbors

Page 21: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Approach

21

Wimage Wneighbors

• In this way the model uses features from both the image and its neighbors

Page 22: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Results• Multi-label image annotation results on the NUS-

WIDE dataset (~240K Flickr images)

22

mAP

labe

l

0

25

50

75

100

Upper bound Tag logistic CNN + kNNvote CNN + logistic Ours Ours (combined)

61.952.8

45.844.043.9

100.0

Page 23: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Results• Multi-label image annotation results on the NUS-

WIDE dataset (~240K Flickr images)

23

Page 24: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Qualitative results

24

Neighborhood

Page 25: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Qualitative results

25

Neighborhood

Page 26: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Qualitative results

26

Neighborhood

Page 27: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Results: ours vs CNN baseline• Experiment 1: evaluates AP for each label of our

model vs the visual-only CNN baseline

27

1 2 3 4Label freq. (log scale)

-0.4

-0.2

0

0.2

0.4

mA

P d

iff.

map

earthquakerainbow

wedding

police

OursV-only

0 0.5 1Visual only

0

0.2

0.4

0.6

0.8

1

Our

model

animal

food

earthquake

map

rainbow

wedding

police

OursV-only

Page 28: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Results: generalization• Experiment 2: vocabulary generalization

28

0 25 50 75 100Overlap percentage (%)

42

44

46

48

50

52

54

mA

P (

pe

r la

be

l)

50.1150.61

51.2351.69

52.78

45.78

49.00

Ours V-only [37]

0 25 50 75 100Overlap percentage (%)

72

74

76

78

80

82

84

mA

P (

pe

r im

ag

e)

79.64 79.83 79.92 80.22 80.34

77.15

Ours V-only

Performance as we vary overlap between tag vocabularies used for training and testing: strong results even in the case of disjoint vocabularies

Page 29: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Results: generalization• Experiment 3: metadata generalization

29

Results using different types of metadata for training and testing

Probability that the k-th neighbor of an image has a label given that the image has the label

Page 30: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Results using label relations• Other recent results on NUS-WIDE by learning label

relations

30

mAP

labe

l

0

25

50

75

100

Upperbound

Taglogistic

CNN &kNNvote

CNN &logistic

Ours Ours(combined)

CNN-RNN[A]

SINN[B]

67.256.1

61.952.8

45.844.043.9

100.0

[A] Wang, Yang, Mao, Huang, Huang, Xu - CVPR 2016 [B] Hu, Zhou, Deng, Liao, Mori - CVPR 2016

Page 31: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Summary• We really need better datasets and evaluation

protocols to evaluate web-vision models

• Visual recognition and learning benefits from:

‣ large collections of noisy web data

‣ good results even when the model is forced to generalize to new types of metadata at test time

31

Page 32: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Next steps• Use a graph(network)-based representation to find

image neighborhoods

• Explore new datasets with a larger label space and noisy annotations

• Visual data mining: share and infer properties based on image similarity on the network

32

Page 33: Exploiting noisy web data for large- scale visual recognition · 2017-08-02 · Lamberto Ballan: Exploiting noisy web data for visual recognition The long tail • A small number

Lamberto Ballan: Exploiting noisy web data for visual recognition

Contact [email protected] www.lambertoballan.net

Xirong Li Tiberio Uricchio Marco Bertini Cees Snoek

Alberto Del Bimbo Lorenzo Seidenari Justin Johnson Li Fei-Fei

Acknowledgements