10
James Z. Wang School of Information Sciences and Technology The Pennsylvania State University University Park, PA 16802 [email protected] Jia Li Department of Statistics The Pennsylvania State University University Park, PA 16802 [email protected]. ABSTRACT 1. INTRODUCTION 1.1 Related work on indexing images

James Z. Wang Jia Li School of Information Sciences and ...infolab.stanford.edu/~wangz/project/imsearch/ALIP/ACM02/wang2.pdf · The ALIP system has three ma jor comp onen ts, the

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: James Z. Wang Jia Li School of Information Sciences and ...infolab.stanford.edu/~wangz/project/imsearch/ALIP/ACM02/wang2.pdf · The ALIP system has three ma jor comp onen ts, the

Learning-based Linguistic Indexing of Pictures

with 2-D MHMMs

James Z. Wang�

School of Information Sciences and TechnologyThe Pennsylvania State University

University Park, PA 16802

[email protected]

Jia LiDepartment of Statistics

The Pennsylvania State UniversityUniversity Park, PA 16802

[email protected].

ABSTRACTAutomatic linguistic indexing of pictures is an importantbut highly challenging problem for researchers in computervision and content-based image retrieval. In this paper, weintroduce a statistical modeling approach to this problem.Categorized images are used to train a dictionary of hun-dreds of concepts automatically based on statistical mod-eling. Images of any given concept category are regardedas instances of a stochastic process that characterizes thecategory. To measure the extent of association between animage and the textual description of a category of images,the likelihood of the occurrence of the image based on thestochastic process derived from the category is computed.A high likelihood indicates a strong association. In our ex-perimental implementation, the ALIP (Automatic Linguis-tic Indexing of Pictures) system, we focus on a particulargroup of stochastic processes for describing images, that is,the two-dimensional multiresolution hidden Markov models(2-D MHMMs). We implemented and tested the system ona photographic image database of 600 di�erent semantic cat-egories, each with about 40 training images. Tested using3,000 images outside the training database, the system hasdemonstrated good accuracy and high potential in linguisticindexing of these test images.Index Terms { Content-based image retrieval, image clas-si�cation, hidden Markov model, computer vision, machinelearning, image segmentation, region matching, wavelets.

1. INTRODUCTIONA picture is worth a thousand words. As human beings, weare able to tell a story from a picture based on what we haveseen and what we have been taught. A 3-year old child iscapable of building models of a substantial number of con-cepts and recognizing them using the learned models storedin her brain. Can a computer program learn a large col-

�James Z. Wang is also with the Department of ComputerScience and Engineering.

lection of semantic concepts from 2-D or 3-D images, buildmodels about these concepts, and recognize them based onthese models? This is the question we attempt to addressin this work.

Automatic linguistic indexing of pictures is essentially im-portant to content-based image retrieval and computer ob-ject recognition. It can potentially be applied to many areasincluding biomedicine, commerce, the military, education,digital libraries, and Web searching. Decades of researchhas shown that designing a generic computer algorithm thatcan learn concepts from images and automatically translatethe content of images to linguistic terms is highly diÆcult.

Much success has been achieved in recognizing a relativelysmall set of objects or concepts within speci�c domains.There is a rich resource of prior work in the �elds of com-puter vision, pattern recognition, and their applications [12].Space limitations do not allow us to present a broad survey.Instead we try to emphasize some of the work that is mostrelated to what we propose. The references below are to betaken as examples of related work, not as the complete listof work in the cited areas.

1.1 Related work on indexing imagesMany content-based image retrieval (CBIR) systems havebeen developed since the early 1990s [11, 7, 20, 22, 18, 4,25, 19, 6]. Most of the above mentioned projects aimed atgeneral-purpose image indexing and retrieval systems focus-ing on searching images visually similar to the query imageor a query sketch. They do not have the capability to assigncomprehensive textual description automatically to pictures,i.e., linguistic indexing, because of the great diÆculties inrecognizing a large number of objects. However, this func-tion is essential for linking images to text and consequentlybroadening the possible usages of an image database.

A growing trend in the �eld of image retrieval is to lin-guistically index images using computer programs relyingon statistical classi�cation methods. The Stanford SIM-PLIcity system [24] uses manually-de�ned statistical clas-si�cation methods to classify the images into rough seman-tic classes, such as textured-nontextured, graph-photograph.Potentially, the categorization enhances retrieval by permit-ting semantically-adaptive searching methods and narrow-ing down the searching range in a database. The approachis limited because these classi�cation methods are problem

Page 2: James Z. Wang Jia Li School of Information Sciences and ...infolab.stanford.edu/~wangz/project/imsearch/ALIP/ACM02/wang2.pdf · The ALIP system has three ma jor comp onen ts, the

speci�c and must be manually developed and coded.

A recent work in associating images explicitly with wordsis that of Barnard and Forsyth of University of Californiaat Berkeley [1]. An object name is associated with a regionin an image based on previously learned region-term asso-ciation probabilities. The work has achieved some successfor certain image types. But as pointed out by the authorsin [1], one major limitation is that the algorithm relies onsemantically meaningful segmentation, which is in generalunavailable to image databases. Automatic segmentation isstill an open problem in computer vision [27, 21, 26]. More-over, some concepts may not be learned from a single regionof an image or a region within an over-segmented image. Forexample, when a tiger object of a test image is segmentedinto the head and the body segments, the computer programassigns the keyword \building" to the head portion of thetiger due to the similarity between the features of the headand those of buildings.

1.2 Our approach

a region of an image the whole image

Figure 1: It is often impossible to accurately de-termine the semantics of an image by looking at asingle region of the image.

Intuitively, human beings recognize many concepts from im-ages based on not just one region of an image. Often weneed to view the image as a whole in order to determinethe semantic meanings of each region and consequently tella complete story about the image. For one example (Fig-ure 1), if we look at only a region of an image, i.e., theface of a person, we would not know that the image depictsthe concept `ski'. But if we see in addition the cloth of theperson, the equipment the person is holding, and the whitesnow in the background, we can recognize easily the concept`ski'. Therefore, treating an image as an entity has the po-tential for modeling relatively high-level concepts as well asimproving the modeling accuracy of low-level concepts.

In our work, we propose to model the entire images statis-tically. In our experimental implementation, we use a 2-Dmultiresolution hidden Markov model (MHMM) [15]. Thisstatistical approach reduces the dependence on correct im-age segmentation because cross pixel and cross resolutiondependencies are captured in the model itself. These mod-els are created automatically by training on sets of imagesrepresenting the same concepts. Machine-generated modelsof the concepts are then stored and used to automaticallyindex images based on linguistic terms. Readers are referredto Li and Gray [16] for details on 2-D MHMM. For clarityof this paper, we present basics regarding the models.

Statistical image modeling is a research topic extensivelystudied in various �elds including image processing and com-

puter vision. Detailed review on some models used in imagesegmentation is provided in [15]. Theories and methodolo-gies related to Markov random �elds (MRFs) [10, 13, 14, 5]have played important roles in the construction of many sta-tistical image models. For a thorough introduction to MRFsand their applications, see Kindermann and Snell [14] andChellappa and Jain [5].

The 2-D multiresolution hidden Markov model has been suc-cessfully applied to image segmentation and compression.This model explores statistical dependence among imagepixels or blocks across multiple resolution as well as withina single resolution. It possesses a exible structure to al-low di�erent proportions of emphasis on inter-resolution andintra-resolution dependency. As a result, users have amplefreedom in choosing a particular form of the model accord-ing to application targeted. Analytic formula for estimat-ing the model by the maximum likelihood criterion and forcomputing likelihood of an instance based on the model areavailable. A 2-D MHMM estimated from training imagessummarizes two types of information: clusters of feature vec-tors at multiple resolutions and the spatial relation betweenthose clusters. The clusters of feature vectors normally re- ect color and texture. Given its modeling eÆciency andcomputational convenience, we consider 2-D MHMM an ap-propriate starting point for exploring the statistical model-ing approach to linguistic indexing.

1.3 Outline of the paperThe remainder of the paper is organized as follows: anoverview of our ALIP (Automatic Linguistic Indexing of Pic-tures) system is introduced in Section 2. The model learn-ing algorithm is described in Section 3. Linguistic indexingmethods are presented in Section 4. In Section 5, experi-ments and results are described. We conclude and suggestfuture research in Section 6.

2. SYSTEM OVERVIEWThe ALIP system has three major components, the featureextraction process, the multiresolution statistical modelingprocess, and the statistical linguistic indexing process. Inthis section, we introduce the basics about these individualcomponents and their relationships.

2.1 Feature extractionThe ALIP system characterizes localized features of train-ing images using wavelets. In this process, an image ispartitioned into small pixel blocks. For our experiments,the block size is chosen to be 4 � 4 to compromise betweenthe texture detail and the computation time. Other similarblock sizes can also be used. The system extracts a featurevector of six dimensions for each block. Three of these fea-tures are the average color components in the block of pixels.The other three are texture features extracted to representenergy in high frequency bands of wavelet transforms [9].Speci�cally, each of the three features is the square root ofthe second order moment of wavelet coeÆcients in one ofthe three high frequency bands. The feature extraction pro-cess is performed in the LUV color space, where L encodesluminance, and U and V encode color information (chromi-nance). The LUV color space is chosen because of its goodperception correlation properties.

Page 3: James Z. Wang Jia Li School of Information Sciences and ...infolab.stanford.edu/~wangz/project/imsearch/ALIP/ACM02/wang2.pdf · The ALIP system has three ma jor comp onen ts, the

LL HL

LH HH

original image wavelet transform

Figure 2: Decomposition of images into frequencybands by wavelet transforms.

To extract the three texture features, we apply either theDaubechies-4 wavelet transform or the Haar transform tothe L component of the image. These two wavelet trans-forms have better localization properties and require lesscomputation compared to Daubechies' wavelets with longer�lters. After a one-level wavelet transform, a 4 � 4 blockis decomposed into four frequency bands as shown in Fig-ure 2. Each band contains 2 � 2 coeÆcients. Without lossof generality, suppose the coeÆcients in the HL band arefck;l; ck;l+1; ck+1;l; ck+1;l+1g. One feature is then computed

as f = 12

qP1i=0

P1j=0 c

2k+i;l+j . The other two texture fea-

tures are computed in a similar manner from the LH andHH bands, respectively.

The motivation for using these features is their re ectionof local texture properties. Unser [23] has shown that mo-ments of wavelet coeÆcients in various frequency bands canbe used to e�ectively discern local texture. Wavelet coeÆ-cients in di�erent frequency bands signal variation in di�er-ent directions. For example, the HL band shows activitiesin the horizontal direction. A local texture of vertical stripsthus has high energy in the HL band of the image and lowenergy in the LH band. The use of this wavelet-based tex-ture feature is a good compromise between computationalcomplexity and e�ectiveness.

2.2 Multiresolution statistical modelingWe �rst manually enlist a series of concepts to be trained forinclusion in the dictionary of concepts. For each concept inthis dictionary, we prepare a training image database witha set of images capturing the concept. These images donot have to be visually similar. We also manually prepare ashort but informative description about any given concept inthis dictionary. Therefore, our approach has the potentialto train a large collection of concepts because we do notneed to manually create descriptions about each image inthe training database.

Block-based features are extracted from each of these train-ing images at several di�erent resolutions. The statisticalmodeling process does not depend on a speci�c feature ex-traction algorithm. The same feature dimensionality is as-sumed for all blocks of pixels.

The statistical modeling process studies the multiresolutionfeatures extracted from each training image in the trainingdatabase. A cross-scale statistical model about a concept isobtained after analyzing all available training images in atraining database. This model is then associated with thetextual description of the concept and stored in the conceptdictionary.

2.3 Statistical linguistic indexingThe ALIP system automatically indexes images with lin-guistic terms based on statistical model comparison. Fora given image to be indexed, we �rst extract multiresolu-tion block-based features in the same manner as the featureextraction process for the training images.

This collection of feature vectors is statistically comparedwith the trained models stored in the concept dictionary toobtain a series of likelihoods representing the statistical sim-ilarity between the image and each of the trained concepts.These likelihoods, along with the stored textual descriptionsabout the concepts, are processed in the signi�cance proces-sor to extract a small set of statistically signi�cant indexterms about the image. These index terms are then storedwith the image in the image database for future keyword-based query processing.

2.4 Major advantagesOur ALIP system has several major advantages:

1. If new images are added to a given concept trainingdatabase, only the particular statistical model of thisgiven concept needs to be retrained. We need notchange the trained models about other concepts in thesame dictionary. This property is very di�erent fromconventional training-based classi�cation approachesbased on neural networks [2], classi�cation and regres-sion trees (CART) [3], and support vector machines(SVM) [8].

2. Because models of di�erent concepts are independentof each others in our system, we can train a relativelylarge number of concepts at once. A statistical model,established for a category of images, serves as a pic-torial description for the entire category and enableseÆcient association of textual annotations with im-age pixel representations. In our experiments, we havetrained the system to automatically create a dictionaryof 600 concepts.

3. In our initial statistical model, spatial relations amongimage pixels and across image resolutions are bothtaken into consideration. This property is especiallyuseful for images with special texture patterns. We canavoid segmenting images or de�ning a similarity dis-tance for any particular set of features. Likelihood canbe used as a universal measure of similarity. With thisstatistical likelihood approach, images used to trainthe same semantic concept do not have to be all visu-ally similar.

3. THE MODEL-BASED LEARNING OFCONCEPTS

In this section, we provide details about our statistical imagemodeling process which learns a dictionary of a large num-ber of concepts automatically. We present here assumptionsof the 2-D MHMM which is modi�ed from the model origi-nally developed for the purpose of image segmentation [15].The model studies the collection of training images withina concept category in their entireties.

Page 4: James Z. Wang Jia Li School of Information Sciences and ...infolab.stanford.edu/~wangz/project/imsearch/ALIP/ACM02/wang2.pdf · The ALIP system has three ma jor comp onen ts, the

3.1 Image ModelingTo describe an image by a multiresolution model, multipleversions of the image at di�erent resolutions are obtained�rst. The original image corresponds to the highest resolu-tion. Lower resolutions are generated by successively �lter-ing out high frequency information. Wavelet transforms [9]naturally provide low resolution images in the low frequencyband (the LL band). Features are extracted at all the res-olutions. The 2-D MHMM aims at describing statisticalproperties of the feature vectors and their spatial depen-dence.

Childblocks

Parentblock

������

������

���������

������������������

���������

������

������

��������

Figure 3: The image hierarchy across resolutions

In the 2-D MHMM, features are regarded as elements ina vector. They can be selected exibly by users and aretreated integratedly as dependent random variables by themodel. Example features include color components andstatistics re ecting texture. To save computation, featurevectors are often extracted from non-overlapping blocks inan image. An element in an image is therefore a blockrather than a pixel. The numbers of blocks in both rowsand columns reduce by half successively at each lower res-olution. Obviously, a block at a lower resolution covers aspatially more global region of the image. As is indicatedby Figure 3, the block at the lower resolution is referred toas a parent block, and the four blocks at the same spatial lo-cation at the higher resolution are referred to as child blocks.We will always assume such a \quad-tree" split in the se-quel since the extension to other hierarchical structures isstraightforward.

We �rst review the basic assumptions of the single resolution2-D HMM as presented in [17]. In the 2-D HMM, featurevectors are generated by a Markov model that may changestate once every block. Suppose there areM states, the stateof block (i; j) being denoted by si;j . The feature vector ofblock (i; j) is ui;j . We use P (�) to represent the probability ofan event. We denote (i0; j0) < (i; j) if i0 < i or i0 = i; j0 < j,in which case we say that block (i0; j0) is before block (i; j).The �rst assumption is that

P (si;j j context) = am;n;l ;

context = fsi0;j0 ; ui0;j0 : (i0; j0) < (i; j)g ;

where m = si�1;j , n = si;j�1, and l = si;j . The secondassumption is that for every state, the feature vectors fol-low a Gaussian distribution. Once the state of a block isknown, the feature vector is conditionally independent ofinformation in other blocks. The covariance matrix �s andthe mean vector �s of the Gaussian distribution vary withstate s.

Given an image, only feature vectors are observable, the rea-son that the model is named as a hidden Markov model. Thestate of a feature vector is conceptually parallel to the clus-ter identity of a vector in unsupervised clustering. As withclustering, the state of vector is not provided directly by thetraining data and hence needs to be estimated. In clustering,feature vectors are considered as independent samples froma given distribution. In the 2-D HMM, feature vectors arestatistically dependent through the underlying states char-acterized by Markovian properties.

For the MHMM, denote the collection of resolutions by R =f1; :::; Rg, with r = R being the �nest resolution. Let thecollection of block indices at resolution r be

N(r) = f(i; j) : 0 � i < w=2R�r ; 0 � j < z=2R�rg :

Images are described by feature vectors at all the resolu-

tions, denoted by u(r)i;j , r 2 R. The underlying state of a

feature vector is s(r)i;j . At each resolution r, the set of states

is f1(r); 2(r); :::; M(r)r g. Note that as states vary across reso-

lutions, di�erent resolutions do not share states.

To structure statistical dependence among resolutions, aMarkov chain with resolution playing a time-like role is as-sumed in the 2-D MHMM. Given the states and the featuresat the parent resolution, the states and the features at thecurrent resolution are conditionally independent of the otherprevious resolutions, so that

P fs(r)i;j ; u

(r)i;j : r 2 R; (i; j) 2 N

(r)g (1)

= Pfs(1)i;j ; u

(1)i;j : (i; j) 2 N

(1)g �

Pfs(2)i;j ; u

(2)i;j : (i; j) 2 N

(2) j s(1)k;l : (k; l) 2 N

(1)g � � � � �

Pfs(R)i;j ; u

(R)i;j : (i; j) 2 N

(R) j s(R�1)k;l : (k; l) 2 N

(R�1)g: (2)

At the coarsest resolution, r = 1, feature vectors are as-sumed to be generated by a single resolution 2-D HMM. Ata higher resolution, the conditional distribution of a featurevector given its state is also assumed to be Gaussian. Theparameters of the Gaussian distribution depend upon thestate at the particular resolution.

Resolution 2

Resolution 3

Resolution 1

������������������������

������������������������

������������������������

������������������������

������������������������

������������������������

������������������������

�������������������������������������������������������������������������������������������������������������

�������������������������������������������������������������������������������������

����������������������������������������

����������������������������������������

������

������

������

������

��������

���

���

��������

������

������

����

��������

������

������

������

������

������

������

���

���

��������

��������

��������

��������

��������

��������

��������

��������

������

������

������

������

�������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������

�������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������

�����������������������������������������������������������������������������������������������������������������������������

�����������������������������������������������������������������������������������������������������������������������������

�������������������������������������

�������������������������������������

���������������

���������������

���������������

���������������

���������������������������������������������������

���������������������������������������������������

����������������������������������

����������������������������������

����������������������������������������������������������������������������������������������������

����������������������������������������������������������������������������������������������������

������������������������������������������������

������������������������������������������������

������������������������������������

������������������������������������

�����������������

�����������������

������������������������������������������������������������������������������������������������������

������������������������������������������������������������������������������������������������������

������

������

���

���

��������

��������

�������������������������������������������������������������������������������������������������������������������������������������������������������������������������������

�������������������������������������������������������������������������������������������������������������������������������������������������������������������������������

��������������������������������������������������������������������������������

��������������������������������������������������������������������������������

����������������������������������

����������������������������������

��������

��������

������

������

Figure 4: The hierarchical statistical dependenceacross resolutions

Given the states at resolution r � 1, statistical dependenceamong blocks at the �ner resolution r is constrained to sib-ling blocks (child blocks descended from the same parentblock). Speci�cally, child blocks descended from di�erent

Page 5: James Z. Wang Jia Li School of Information Sciences and ...infolab.stanford.edu/~wangz/project/imsearch/ALIP/ACM02/wang2.pdf · The ALIP system has three ma jor comp onen ts, the

parent blocks are conditionally independent. In addition,given the state of a parent block, the states of its child blocksare independent of the states of their \uncle" blocks (non-parent blocks at the parent resolution). State transitionsamong sibling blocks are governed by the same Markovianproperty assumed for a single resolution 2-D HMM. Thestate transition probabilities, however, depend on the stateof their parent block. To formulate these assumptions, de-note the child blocks at resolution r of block (k; l) at reso-lution r � 1 by

D (k; l) = f(2k; 2l); (2k + 1; 2l); (2k; 2l + 1); (2k + 1; 2l + 1)g :

According to the assumptions,

Pfs(r)i;j : (i; j) 2 N

(r) j s(r�1)k;l : (k; l) 2 N

(r�1)g

=Y

(k;l)2N(r�1)

Pfs(r)i;j : (i; j) 2 D (k; l) j s

(r�1)k;l g ;

where Pfs(r)i;j : (i; j) 2 D (k; l)j s

(r�1)k;l g can be evaluated by

transition probabilities conditioned on s(r�1)k;l , denoted by

am;n;l(s(r�1)k;l ). We thus have a di�erent set of transition

probabilities am;n;l for every possible state in the parentresolution. The in uence of previous resolutions is exertedhierarchically through the probability of the states, whichcan be visualized in Figure 4. The joint probability of statesand feature vectors at all the resolutions in (2) is then de-rived.

To summarize, a 2-D MHMM re ects both the inter-scaleand intra-scale statistical dependence. The inter-scale de-pendence is modeled by the Markov chain over resolutions.The intra-scale dependence is modeled by the HMM. At thecoarsest resolution, feature vectors are assumed to be gener-ated by a 2-D HMM. Figure 5 illustrates the inter-scale andintra-scale dependencies modeled. At all the higher resolu-tions, feature vectors of sibling blocks are also assumed tobe generated by 2-D HMMs. The HMMs vary according tothe states of parent blocks. Therefore, if the next coarserresolution has M states, then there are, correspondingly, MHMMs at the current resolution.

The 2-D MHMM can be estimated by the maximum likeli-hood criterion using the EM algorithm. Details about theestimation algorithm and the computation of the likelihoodof an image given a 2-D MHMM are referred to [15].

4. THE AUTOMATIC LINGUISTIC INDEX-ING OF PICTURES

In this section, we describe the component of the ALIPsystem that automatically indexes pictures with linguisticterms. For a given image, the system compares the im-age statistically with the trained models in the concept dic-tionary and extract the most statistically signi�cant indexterms to describe the image.

For any given image, a collection of feature vectors at mul-

tiple resolution fu(r)i;j ; r 2 R; (i; j) 2 N

(r)g are computed as

described in Section 3. We regard fu(r)i;j ; r 2 R; (i; j) 2 N

(r)gas an instance of a stochastic process de�ned on a multireso-lution grid. The similarity between the image and a categoryof images in the database is assessed by the log likelihood

of this instance under the model M trained from images inthe category, that is,

log Pfu(r)i;j ; r 2 R; (i; j) 2 N

(r) j Mg :

A recursive algorithm is used to compute the above log like-lihood in an manner described in [15]. After determiningthe log likelihood of the image depicting any given conceptin the dictionary, we sort the log likelihoods to �nd the fewcategories with the highest likelihoods. The short textualdescriptions of these categories are loaded in the program inorder to �nd the proper index terms for this image.

We use the most statistically signi�cant index terms withinthe textual descriptions to index the image. Annotationwords may have vastly di�erent frequencies of appearing inthe categories of an image database. For instance, muchmore categories may be described with the index term \land-scape" than with the term \dessert". Therefore, obtain-ing the index word \dessert" in the top ranked categoriesmatched to an image is in a sense more surprising than ob-taining \landscape" since the word \landscape" may havea good chance of being selected even by random matching.To measure the level of signi�cance when a word appears jtimes in the top k matched categories, we compute the prob-ability of obtaining the word j or more times in k randomlyselected categories. This probability is given by

P (j; k) =

kXi=j

I(i � m)

�m

i

��n�m

k�i

��n

k

�=

kXi=j

I(i � m)m! (n�m)! k! (n� k)!

i! (m� i)! (k � i)! (n�m� k + i)! n!;

where I(�) is the indicator function that equals 1 when theargument is true and 0 otherwise, n is the total numberof image categories in the database, and m is the numberof image categories that are annotated with the given word.The probability P (j; k) can be approximated as follows usingthe binomial distribution if n;m >> k,

P (j; k) =kXi=i

k

i

!pi(1� p)k�i =

kXi=j

k!

i!(k � i)!pi(1� p)k�i ;

where p = m=n is the percentage of image categories in thedatabase that are annotated with this word, or equivalently,the frequency of the word being used in annotation. A lowervalue of P (j; k) indicates a higher level of signi�cance for agiven index term. We rank the index terms within the shortdescriptions of the most likely concept categories accordingto their statistical signi�cance. The terms with high signif-icance are used to index the image.

5. EXPERIMENTSTo validate the methods we have described, we implementedthe components of the ALIP system and tested with a general-purpose image database including about 60; 000 photographs.These images are stored in JPEG format with size 384�256or 256 � 384. The system is written in the C programminglanguage and compiled on two UNIX platforms: LINUX andSolaris. In this section, we describe the training conceptsand show indexing results.

Page 6: James Z. Wang Jia Li School of Information Sciences and ...infolab.stanford.edu/~wangz/project/imsearch/ALIP/ACM02/wang2.pdf · The ALIP system has three ma jor comp onen ts, the

Figure 5: In the statistical modeling process, spatial relations among image pixels and across image resolutionsare both taken into consideration. Arrows, not all drawn, indicate the transition probabilities captured inthe statistical model.

5.1 Training conceptsWe conducted experiments on learning-based linguistic in-dexing with a large number of concepts. The ALIP systemwas trained using a subset of 60; 000 photographs which arebased on 600 CD-ROMs published by COREL Corp. Typi-cally, each COREL CD-ROM of about 100 images representone distinct topic of interest. For our experiment, the dictio-nary of concepts contains all 600 concepts, each associatedwith one CD-ROM of images.

We manually assigned a set of keywords to describe eachCD-ROM collection of 100 photographs. The semantic de-scriptions of these collections of images range from as simpleor low-level as \mushrooms" and \ owers" to as complex orhigh-level as \England, landscape, mountain, lake, Euro-pean, people, historical building" and \battle, rural, people,guard, �ght, grass". On average, 3.6 keywords are used todescribe the content of each of the 600 concept categories. Ittook the authors approximately 10 hours to annotate thesecategories. Table 1 shows a sample of these annotations.

5.2 ResultsAfter the training, a statistical model is generated for each ofthe 600 collections of images. Depending on the complexityof the concept, the training process takes between 15 to40 minutes of CPU time on an 800 MHz Pentium III PCto converge on a model. On average, 30 minutes of CPUtime is spent to train a concept. The training process isconducted only once for each concept in the list.

These models are stored in a fashion similar to a dictionaryor encyclopedia. Essentially, we use computers to create adictionary of concepts that will enable computers to indeximages linguistically. The process is entirely parallelizablebecause the training of one concept is independent from thetraining of other concepts in the same dictionary.

We randomly selected 3,000 test images outside the trainingimage database and processed these images by the linguisticindexing component of the system. For each of the 3,000 test

images, the computer program selected a number of conceptsin the dictionary with the highest likelihood of describing theimage. Next, the most signi�cant index terms for the imageare extracted from the collection of index terms associatedwith the chosen concept categories.

It takes an average of two seconds CPU time on the samePC to compute the likelihood of a test image resemblingone of the concepts in the dictionary. The process is highlyparallelizable because the computation of the likelihood toa concept is independent from the computation of the like-lihood to another concept.

Figure 8 shows the computer indexing results of 21 randomlyselected images outside the training database. The methodappears to be highly promising for automatic learning andlinguistic indexing of images. Some of the computer pre-dictions seem to suggest that one can control what is to belearned and what is not by adjusting the training databaseof individual concepts. As indicated in the second exam-ple, the computer predictions of a wildlife animal pictureinclude \cloth" and \people". It is possible that the com-puter learned the association between animal fur and theclothes of people from the training databases which con-tain images with female super-models wearing fur clothes.Consequently, computer predictions are objective and with-out human subjective biases. Potentially, computer-basedindexing of images eliminates the inconsistency problemscommonly associated with manual image annotations.

5.3 Systematic evaluationTo provide numerical results on the performance, we eval-uated the ALIP system based on a subset of the CORELdatabase, formed by 10 image categories (Africa people andvillages, beach, buildings, buses, dinosaurs, elephants, ow-ers, horses, mountains and glaciers, food), each containing100 pictures. Within this database, it is known whetherany two images are of the same category. We trained eachconcept using 40 images and test the models using 500 im-ages outside the training database. Instead of annotating

Page 7: James Z. Wang Jia Li School of Information Sciences and ...infolab.stanford.edu/~wangz/project/imsearch/ALIP/ACM02/wang2.pdf · The ALIP system has three ma jor comp onen ts, the

ID Category Descriptions0 Africa, people, landscape, animal10 England, landscape, mountain, lake, European, people, historical building20 Monaco, ocean, historical building, food, European, people30 royal guard, England, European, people40 vegetable50 wild life, young animal, animal, grass60 European, historical building, church70 animal, wild life, grass, snow, rock80 plant, landscape, ower, ocean90 European, historical building, grass, people100 painting, European110 ower120 decoration, man-made130 Alaska, landscape, house, snow, mountain, lake140 Berlin, historical building, European, landscape150 Canada, game, sport, people, snow, ice160 castle, historical building, sky170 cuisine, food, indoor180 England, landscape, mountain, lake, tree190 �tness, sport, indoor, people, cloth200 fractal, man-made, texture210 holiday, poster, drawing, man-made, indoor220 Japan, historical building, garden, tree230 man, male, people, cloth, face240 wild, landscape, north, lake, mountain, sky250 old, poster, man-made, indoor260 plant, art, ower, indoor270 recreation, sport, water, ocean, people280 ruin, historical building, landmark290 sculpture, man-made

Table 1: Examples of the 600 categories and their descriptions. Every category has 40 training images.

Figure 6: Training images used to learn a given concept are not necessarily all visually similar. For example,these 40 images were used to train the concept of Paris with the category description: \Paris, European,historical building, beach, landscape, water".

Page 8: James Z. Wang Jia Li School of Information Sciences and ...infolab.stanford.edu/~wangz/project/imsearch/ALIP/ACM02/wang2.pdf · The ALIP system has three ma jor comp onen ts, the

Figure 7: Training images used to learn the concept of male with the category description: \man, male,people, cloth, face".

1 2 3 4 5 6 7 8 9 100

10

20

30

40

50

60

70

80

90

100

Acc

urac

y in

per

cent

age

of c

orre

ctly

cla

ssifi

ed im

ages

Image category

%

Figure 10: Percentages of images classi�ed to thesame category as the manual classi�cation. The ex-periment is conducted on a test image database of10 image categories. (also see Figure 9)

the images, the program was used to select the categorywith the highest likelihood for each test image. That is,we use the classi�cation power of the system as an indica-tion of the accuracy. An image is considered to be classi�edcorrectly if the computer predicts the original category theimage belongs to. This assumption is reasonable since the 10categories were chosen so that each depicts a substantiallydistinct semantic topic.

Figure 9 shows the automatic classi�cation result as com-pared with the original image classi�cation. Figure 10 plotsthe accuracy in each category as the percentage of imagescorrectly classi�ed. As indicated in the plots, some of theconcepts can be related. For example, both the \beach" andthe \mountains and glaciers" categories contain images withrocks, sky, and trees. Moreover, the system is designed forthe purpose of automatically annotating images where the

major challenge is to model hundreds of di�erent conceptsrather than to classify images into a small set of classes withhigh accuracy. As a result, the evaluation method we usehere can be used to assess the lower bounds of the annota-tion accuracy of the system for the given concept categories.

6. CONCLUSIONS AND FUTURE WORKIn this paper, we demonstrated our statistical modeling ap-proach to the problem of automatic linguistic indexing ofpictures for the purpose of image retrieval. In our ALIPsystem, we used categorized images to train a dictionary ofhundreds of concepts automatically. Wavelet-based featuresare used to describe local color and texture in the images.After analyzing all training images for a training concept, atwo-dimensional multiresolution hidden Markov model (2-DMHMM) is created and stored in a concept dictionary. Im-ages in one category is regarded as instances of a stochas-tic process that characterizes the category. To measure theextent of association between an image and the textual de-scription of a category of images, we compute the likelihoodof the occurrence of the image based on the stochastic pro-cess derived from the category. We have demonstrated thatthe proposed methods can be used to train 600 di�erent se-mantic concepts at the same time and these models can beused to index images linguistically.

The major advantages with our approach are (1) modelsfor di�erent concepts can be independently trained and re-trained; (2) a relatively large number of concepts can betrained and stored; (3) spatial relation among image pixelsand across image resolutions is taken into consideration withprobabilistic likelihood as a universal measure.

The current ALIP system has several limitations.

� We are training the concept dictionary with only 2-Dimages without a sense of object size. It is believed

Page 9: James Z. Wang Jia Li School of Information Sciences and ...infolab.stanford.edu/~wangz/project/imsearch/ALIP/ACM02/wang2.pdf · The ALIP system has three ma jor comp onen ts, the

Image Computer pre-dictions

Image Computer pre-dictions

Image Computer pre-dictions

building,sky,lake,landscape, Euro-pean,tree

snow,animal,wildlife,sky,cloth,ice,people

people,European,female

food,indoor, cui-sine,dessert

people,European,man-made,water

lake,Portugal,glacier,mountain,water

skyline, sky, NewYork, landmark

plant, ower, gar-den

modern,parade,people

pattern, ower,red,dining

ocean,paradise,San Diego, Thai-land, beach,�sh

ower, ora,plant,fruit, natu-ral,texture

ancestor,drawing,�tness,history, indoor

hair style,occupation,face,female,cloth

night,cyber, fash-ion,female

Figure 8: Annotations automatically generated by our computer-based linguistic indexing algorithm. Thedictionary with 600 concepts was created automatically using statistical modeling and learning. Test imageswere randomly selected outside the training database.

0 50 100 150 200 250 300 350 400 450 5001

2

3

4

5

6

7

8

9

10

Image indices

Tru

e im

age

clas

ses

0 50 100 150 200 250 300 350 400 450 5001

2

3

4

5

6

7

8

9

10

Image indices

Cla

ssifi

catio

n re

sult

manual image classi�cation machine classi�cation result

Figure 9: Results for the application of ALIP to the automatic image categorization problem. Left: the trueclass identities of the test images indexed from 1 to 500. Every 50 images belong to the same class. For imageswith indices 50(i� 1) + 1 to 50i, the correct class identities are i. Right: the class identities predicted by oursystem for the 500 images. The experiment is conducted on a test image database of 10 image categories.

Page 10: James Z. Wang Jia Li School of Information Sciences and ...infolab.stanford.edu/~wangz/project/imsearch/ALIP/ACM02/wang2.pdf · The ALIP system has three ma jor comp onen ts, the

that the object recognizer of human beings are usuallytrained using 3-D stereo with motion and a sense ofobject sizes. Training with 2-D still images potentiallylimits the ability of accurate learning of concepts. Weare currently attempting to work on training with 3-Dimages.

� Related concepts can sometimes be confusing if themodels are trained with only positive examples. Thecurrent statistical model captures the association be-tween a set of positive training example and their se-mantic concept. However, when a part of an associa-tion is not desired, the model should include a mecha-nism to make necessary corrections. We are exploringthe possibility of adjusting the model based on nega-tive examples provided to the system.

� For very complex concepts, i.e., when images repre-senting the concept have very di�erent appearances,it seems that 40 training images are not enough forthe computer program to build a reliable model. Themore complex the concept, the more training imagesare needed and the more CPU time is required. Thisis expected as it takes human beings more experiencesand longer time to learn more complex concepts.

7. ACKNOWLEDGMENTSThe SIMPLIcity work was supported in part by the USNational Science Foundation under grant IIS-9817511 andStanford University. This work is supported primarily byThe Pennsylvania State University, the US National ScienceFoundation, the PNC Foundation, and SUN Microsystemsunder grant EDUD-7824-010456-US. We would like to ac-knowledge the comments and constructive suggestions fromanonymous reviewers. Conversations with Michael Lesk havebeen very helpful.

8. REFERENCES[1] K. Barnard, D. Forsyth, \Learning the Semantics of Words

and Pictures," Proc. ICCV, vol 2, pp. 408-415, 2001.

[2] C. Bishop, Neural Networks for Pattern Recognition,Oxford University Press, 1995.

[3] L. Breiman, J. Friedman, C.J. Stone, R.A. Olshen,Classi�cation and Regression Trees, Wadsworth Int.Group, Belmont, Calif., 1984.

[4] C. Carson, M. Thomas, S. Belongie, J. M. Hellerstein, J.Malik, \Blobworld: A system for region-based imageindexing and retrieval," Proc. Int. Conf. on VisualInformation Systems, D. P. Huijsmans, A. W.M.Smeulders (eds.), Springer, Amsterdam, The Netherlands,June 2-4, 1999.

[5] R. Chellappa and A. K. Jain, Markov Random Fields:Theory and Applications, Academic Press, 1993.

[6] Y. Chen, J. Z. Wang, \A region-based fuzzy featurematching approach to content-based image retrieval,"IEEE Trans. on Pattern Analysis and MachineIntelligence, vol. 24, 2002.

[7] W. W. Chu, I.T. Leong, R.K. Taira, \A semanticmodeling approach for image retrieval by content," TheVLDB Journal, vol. 3, no. 4, pp. 445-477, 1994.

[8] N. Cristianini, J. Shawe-Taylor, An Introduction toSupport Vector Machines: And Other Kernel-BasedLearning Methods, Cambridge University Press, 2000.

[9] I. Daubechies, Ten Lectures on Wavelets, Capital CityPress, 1992.

[10] R. L. Dobrushin, \The description of a random �eld bymeans of conditional probabilities and conditions of itsregularity," Theory Prob. Appl., vol. 13, pp. 197-224, 1968.

[11] M. Flickner, H. Sawhney, W. Niblack, J. Ashley, Q.Huang, B. Dom, M. Gorkani, J. Hafner, D. Lee, D.Petkovic, D. Steele, P. Yanker, \Query by image and videocontent: The QBIC system," Computer, vol. 28, no. 9, pp.23-32, September 1995.

[12] D. A. Forsyth, J. Ponce, Computer Vision: A ModernApproach, Prentice Hall, 2002.

[13] S. Geman and D. Geman, \Stochastic relaxation, Gibbsdistributions, and the Bayesian restoration of images,"IEEE Trans. Pattern Analysis and Machine Intelligence,vol. PAMI-6, no. 6, pp. 721-741, Nov. 1984.

[14] R. Kindermann and L. Snell, Markov Random Fields andTheir Applications, American Mathematical Society, 1980.

[15] J. Li, R. M. Gray, R. A. Olshen, \Multiresolution imageclassi�cation by hierarchical modeling with twodimensional hidden Markov models," IEEE Trans. onInformation Theory, vol. 46, no. 5, pp. 1826-41, August2000.

[16] J. Li, R. M. Gray, Image Segmentation and CompressionUsing Hidden Markov Models, Kluwer AcademicPublishers, 2000.

[17] J. Li, A. Najmi, R. M. Gray, \Image classi�cation by atwo dimensional hidden Markov model," IEEE Trans. onSignal Processing, vol. 48, no. 2, pp. 517-33, February2000.

[18] W. Y. Ma, B. Manjunath, \NeTra: A toolbox fornavigating large image databases," Proc. Int. Conf. onImage Processing, Santa Barbara, pp. 568-71, 1997.

[19] Y. C. Park, P. K. Kim, F. Golshani, S. Panchanathan,\Technique for eliminating irrelevant terms in termrewriting for annotated media retrieval," Proc. of ACMMultimedia, pp. 582-584, 2001.

[20] A. Pentland, R. W. Picard, S. Sclaro�, \Photobook: Toolsfor content-based manipulation of image databases," Proc.of SPIE, vol. 2185, pp. 34-47, San Jose, February 7-8,1994.

[21] J. Shi, J. Malik, \Normalized cuts and imagesegmentation," IEEE Trans. on Pattern Analysis andMachine Intelligence, vol. 22, no. 8, pp. 888-905, 2000.

[22] J. R. Smith, S.-F. Chang, \An image and video searchengine for the World-Wide Web," Proc. of SPIE, vol.3022, pp. 84-95, 1997.

[23] M. Unser, \Texture classi�cation and segmentation usingwavelet frames," IEEE Trans. Image Processing, vol. 4,no. 11, pp. 1549-1560, Nov. 1995.

[24] J. Z. Wang, Integrated Region-based Image Retrieval,Kluwer Academic Publishers, Dordrecht, 2001.

[25] J. Z. Wang, J. Li, G. Wiederhold, \SIMPLIcity:Semantics-sensitive Integrated Matching for PictureLIbraries," IEEE Trans. on Pattern Analysis andMachine Intelligence, vol. 23, no. 9, pp. 947-963, 2001.

[26] J. Z. Wang, J. Li, R. M. Gray, G. Wiederhold,\Unsupervised multiresolution segmentation for imageswith low depth of �eld," IEEE Trans. on Pattern Analysisand Machine Intelligence, vol. 23, no. 1, pp. 85-91, 2000.

[27] S. Zhu, A. L. Yuille, \Region competition: Unifyingsnakes, region growing, and Bayes/MDL for multi-bandimage segmentation," IEEE Trans. on Pattern Analysisand Machine Intelligence, vol. 18, No. 9, pp. 884-900,1996.