12
VISRET – A Content Based Annotation, Retrieval and Visualization Toolchain Levente Kov´ acs, ´ Akos Utasi, and Tam´ asSzir´anyi Distributed Events Analysis Research Group Computer and Automation Research Institute, Hungarian Academy of Sciences Kende u. 13-17, 1111 Budapest, Hungary {levente.kovacs,utasi,sziranyi}@sztaki.hu http://web.eee.sztaki.hu Abstract. This paper presents a system for content-based video re- trieval, with a complete toolchain for annotation, indexing, retrieval and visualization of imported data. The system contains around 20 feature descriptors, a modular infrastructure for descriptor addition and index- ing, a web-based search interface and an easy-to-use query-annotation- result visualization module. The features that make this system differ from others is the support of all the steps of the retrieval chain, the modular support for standard MPEG-7 and custom descriptors, and the easy-to-use tools for query formulation and retrieval visualization. The intended use cases of the system are content- and annotation-based re- trieval applications, ranging from community video portals to indexing of image, video, judicial, and other multimedia databases. 1 Introduction In the field of content-based multimedia indexing and retrieval research, there are some systems that provide model-based retrieval possibilities in one way or the other. The performance of these systems is highly dependent on the ef- fectiveness of categorization and classification algorithms they contain, yet the most important part from a user’s point of view is the ease of use of the query interface and the relevancy of the first N results presented. Easily understand- able visualization is also a feature that is usually neglected, which is a mistake, since if the results’ relevancy cannot be easily judged the users cannot help in refining the relevancy of these results - which in turn would be important for the researchers. In this work we present a system that provides a toolset for aspects of storage and retrieval of video data: a tool for annotating videos, scenes, shots, frames and regions, a content server with a flexible backend for descriptor and content import, a user frontend for textual and model-based query formulation and re- sult viewing, and a result visualization module that can present the retrieval results in 2D and 3D with images and point clouds, and provides an interface for creating visual queries based on simple clicking through graphs and videos. As a whole, the system can be used to store and index annotated or non-annotated

VISRET – A Content Based Annotation, Retrieval and ...web.eee.sztaki.hu/~ucu/kovacsACIVS09.pdf · modular support for standard MPEG-7 and custom descriptors, and the easy-to-use

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: VISRET – A Content Based Annotation, Retrieval and ...web.eee.sztaki.hu/~ucu/kovacsACIVS09.pdf · modular support for standard MPEG-7 and custom descriptors, and the easy-to-use

VISRET – A Content Based Annotation,

Retrieval and Visualization Toolchain

Levente Kovacs, Akos Utasi, and Tamas Sziranyi

Distributed Events Analysis Research GroupComputer and Automation Research Institute, Hungarian Academy of Sciences

Kende u. 13-17, 1111 Budapest, Hungary{levente.kovacs,utasi,sziranyi}@sztaki.hu

http://web.eee.sztaki.hu

Abstract. This paper presents a system for content-based video re-trieval, with a complete toolchain for annotation, indexing, retrieval andvisualization of imported data. The system contains around 20 featuredescriptors, a modular infrastructure for descriptor addition and index-ing, a web-based search interface and an easy-to-use query-annotation-result visualization module. The features that make this system differfrom others is the support of all the steps of the retrieval chain, themodular support for standard MPEG-7 and custom descriptors, and theeasy-to-use tools for query formulation and retrieval visualization. Theintended use cases of the system are content- and annotation-based re-trieval applications, ranging from community video portals to indexingof image, video, judicial, and other multimedia databases.

1 Introduction

In the field of content-based multimedia indexing and retrieval research, thereare some systems that provide model-based retrieval possibilities in one wayor the other. The performance of these systems is highly dependent on the ef-fectiveness of categorization and classification algorithms they contain, yet themost important part from a user’s point of view is the ease of use of the queryinterface and the relevancy of the first N results presented. Easily understand-able visualization is also a feature that is usually neglected, which is a mistake,since if the results’ relevancy cannot be easily judged the users cannot help inrefining the relevancy of these results - which in turn would be important forthe researchers.

In this work we present a system that provides a toolset for aspects of storageand retrieval of video data: a tool for annotating videos, scenes, shots, framesand regions, a content server with a flexible backend for descriptor and contentimport, a user frontend for textual and model-based query formulation and re-sult viewing, and a result visualization module that can present the retrievalresults in 2D and 3D with images and point clouds, and provides an interface forcreating visual queries based on simple clicking through graphs and videos. Asa whole, the system can be used to store and index annotated or non-annotated

Page 2: VISRET – A Content Based Annotation, Retrieval and ...web.eee.sztaki.hu/~ucu/kovacsACIVS09.pdf · modular support for standard MPEG-7 and custom descriptors, and the easy-to-use

2 VISRET

video content, perform retrievals based on textual and model-based queries, andvisualize the results. We will present all the modules, including the indexingserver, and most of the feature descriptors we use.

2 Related Works

We know of some image/video searching solutions, some based on content in-terpretation, others based on textual searches among annotations (created byhand, by speech-to-text, or from the context of the original source).

The Simplicity engine [1] is one of the most widely known content-basedsearch engines used by many institutions and web sites. It is based on classi-fication of low level features, based mostly on wavelet feature extractors andregion matching. The Amico library [2] provides access to a multimedia collec-tion where every element has been imported with complete catalog and metadatainformation by the partners, and searches are performed over these associateddata. [3] provides content-based search based on color and texture content. TheVideoQ engine [4] besides textual search provides an intuitive interface to drawan arrangement of regions and colors as a query sketch and provides resultswhich match these sketches. [7] is an image search engine where hashes - finger-prints - are extracted from uploaded images and compared to stored hashes ofindexed web images. It is sensitive to rotation and heavy scaling though, and itdoes not use any content-based features. [8] combines extensive manual taggingand machine learning strategies to categorize movies into classes of mood, tone,structure, also including tagging information like awards and user flags. Otherpopular image and video search engines like [5, 6], although backed by large com-panies, still do not provide real content based search possibilities. Their servicescurrently rely heavily on textual annotations, some coming from speech-to-textengines, some from manual work, and from text extraction around the videofrom its original context. Most of these systems exploits only a limited amount -or none - content-based features for retrieval, and this is one of the main issueswe try to address here.

The Visret system that we present in this paper - and the tools it consistsof - is both similar and different to some of the above engines. Similar in that itprovides textual and content-based search possibilities over a video pool. Differ-ent in that it has a modular support for standard and non-standard - i.e. new,custom - feature descriptors, all background processing is automatic (featureextractions, indexing), contains versatile annotation and visualization tools, andprovides visual query, search and browse possibilities. The main advantage ofthe presented framework is its flexibility and versatility, its modular architec-ture provides easy ways of managing, testing, adding new descriptors, and thevisual query interface gives more freedom of browsing.

Page 3: VISRET – A Content Based Annotation, Retrieval and ...web.eee.sztaki.hu/~ucu/kovacsACIVS09.pdf · modular support for standard MPEG-7 and custom descriptors, and the easy-to-use

VISRET 3

3 The Visret System

The Visret toolchain is a set of backend, fronted, descriptor, indexing, retrievaland visualization tools that when put together, can cover the full process ofstorage and content-based retrieval of video data. The elements of the systemare as follows, for their arrangements see Fig. 1:

– Annotation tool: This tool is a pre-processor for the content server. It con-tains automatic shot/scene segmentation, and a series of aids for annotatingscenes, shots, frames, objects of a video. It can also be extended with mod-ules, e.g. automatic face recognition and annotation. It is an applicationsrunning on Microsoft Windows (XP or higher).

– Database and server backend: It provides an administrative interface forimporting videos into the database, and adding descriptor modules to theserver. The available descriptors all run automatically for the videos andtheir results get stored in the database. It is implemented as a set of Enter-prise JavaBean and Message-Driven Bean modules of a Java EE applicationon a Sun Glassfish container server and IBM DB2 database server.

– Descriptors: Descriptors are written as separate executables - based on acommon template -, and then are imported into the server backend. Allimported descriptors are automatically run over the existing videos, andall new imported videos are fed through the descriptors automatically. Newones can easily be added, one just needs to write one conforming to the giventemplate, and provide a metric associated to the new descriptor.

– Indexer: The indexing service builds index trees for all the features and pro-vides a socket-based access for the server to perform content-based queries.

– Web based retrieval interface: An interface for formulating textual queries forsearching among annotated contents, and model images for searching basedon content similarity. Results are presented in relevancy order, can be playedand associated annotations can be viewed. It runs as a JavaServer Faces webmodule of the JAVA EE application on a Sun Glassfish application server.

– Visualization tool: A tool for viewing 2D distribution of images for any 2selected descriptors, and 3D point cloud distributions for any 3 selected de-scriptors, with color-based visualization of associated categories. Any imagecan be selected as a query, annotations can be viewed and edited, new cat-egories can be assigned and existing ones can be edited. It is currently anapplication running on Microsoft Windows (XP or higher) but we are work-ing on integrating it into the server frontend as a JavaScript (AJAX) basedweb application.

3.1 Annotation

The VIND (Video INDexer) application is a tool for automatic segmentation ofvideos into scenes and shots, producing an XML output of the video details. Italso provides a set of tools for assigning textual annotations to parts of videos

Page 4: VISRET – A Content Based Annotation, Retrieval and ...web.eee.sztaki.hu/~ucu/kovacsACIVS09.pdf · modular support for standard MPEG-7 and custom descriptors, and the easy-to-use

4 VISRET

Fig. 1. The Visret toolchain architecture.

(to the whole video, to scenes, to shots, to frames and objects of a frame). It canbe extended with feature extractors, e.g. it automatically extracts representativeframes of shots and human faces on frames of a shot, which also get stored inthe XML output. The video and the produced annotation can be imported intothe database. The most important features of the annotation tool are:

– Automatic shot change detection (cuts, fades, wipes).• For this purpose we use shot, fade and wipe detections developed for

our archive film restoration system [20]. As an example, shot changesare detected by an optical flow analysis method, based on the analysisof motion estimation errors among frames, For the basic cut detector letEi,t,t+1 = ‖b,t − bi,t+1‖ be the error (e.g. mean square error) between ablock bi of frame t and the same block, being motion compensated fromthe next frame t + 1, that is bi,t+1 = Vx,y(bi,t), Vx,y being the motionvector of the ith block, and let Di = 1 if Ei,t,t+1 > γ and 0 otherwise

(where γ is a constant). If∑N

i=0Di > ε (ε usually is half or two-thirds

of the number of blocks) then we signal a shot boundary. This methodruns in multiples of real-time on decompressed video frames.

– Automatic representative frame extraction for shots.• Representative frames are selected by color histogram analysis. Since in

the browsing and retrieval visualization tool these will be the framesstanding for each segmented shot, they need to be some average of theshot frames. We chose a color-based selection, for speed and generallygood performance. The r-frame will be the one closest to the averagehistogram of a shot, i.e. Ri = argmin

j=1...N

‖pj − p‖, where N is the number

of shots, i is the location of the representative frame in shot number j,pj(n) = cn

c, c is the frame pixels, cn is a histogram color.

– Automatic face detection (based on [14]).– Highly visual tools for point-and-click editing (i.e. re-positioning) of scene,

shot, and representative frame locations.– Tools for aiding textual annotation of video structure elements, i.e. whole

video, scenes (group of shots), shots, frames, regions (objects), by free textor by pre-defined word collections.

Fig. 2 left shows a screenshot of the main annotation window, with a videoloaded, shots detected and displayed in browsing mode. Fig. 2 right shows theregion annotation dialog, where rectangular and ellipsoid regions can be definedand textually annotated for further reference and search.

Page 5: VISRET – A Content Based Annotation, Retrieval and ...web.eee.sztaki.hu/~ucu/kovacsACIVS09.pdf · modular support for standard MPEG-7 and custom descriptors, and the easy-to-use

VISRET 5

Fig. 2. Left: The main window of the annotation tool, with a loaded video, detectedshots, annotation tools. Right: The region (rectangular or ellipsoid) definition andannotation dialog of the annotation tool.

3.2 Indexing

For indexing of the imported video content we need feature descriptors. Thedescriptors we use are either standard MPEG-7 ones, or ones that we havedeveloped. These descriptors are modules added to the server backend, and theyautomatically run for each imported video, their output being stored in thedatabase, along the corresponding frames and shots of the videos. Some of themrun on frames (usually on the representative frames of a shot), others run onimage sequences (e.g. motion-based descriptors) which are run for each storedvideo shot. Each descriptor has an associated distance metric, which is used totell how one content element relates to the others. Indexing is done based onthese descriptors and their metrics. Some of the descriptors we use are:

– AverageColor: extracts color samples from frame regions (blocks).– AverageMotion: extracts representative motion directions from frame blocks.– ColorLayout, ColorStructure, DominantColor, EdgeHistogram, Homogeno-

usTexture, ScalableColor: MPEG-7 color, edge and texture descriptors [9].– ColorSegments: color segmentation based on MeanShift [12] classification.– DayNight: a boolean descriptor, decides whether a frame is a day/night shot,

based on color distribution and lighting conditions.– Focus: a relative focus map extractor based on a blind deconvolution ap-

proach [10], for obtaining focused areas as a base for indexing [15].– GrasSky: measures the percentage of grass-like green and sky-like blue color

distributions on frames.– MotionActivity: MPEG-7 motion descriptor, extracts motion-specific infor-

mation on a region-based approach.– SiftPoints: SIFT (Scalable Invariant Feature) [11] descriptor, extracts the

128-dimensional features for the SIFT points found on the frame.

Page 6: VISRET – A Content Based Annotation, Retrieval and ...web.eee.sztaki.hu/~ucu/kovacsACIVS09.pdf · modular support for standard MPEG-7 and custom descriptors, and the easy-to-use

6 VISRET

– Skin: a Boolean descriptor, decides whether a frame contains skin colors ornot; it is based on a statistical learning of possible skin colors.

The indexer builds index-trees for each of the descriptors. The trees we useare customized BK-trees [13], which we will call BK*-trees. The indexer existsas a separate entity, providing a socket-based interface for submitting queries.This is so because we wanted the content server be able to constantly work onnewly imported videos, while new indexes are only built once or twice a day andproviding the last index structure between re-runs. Thus, both the content serverand the indexing service can run in parallel, providing uninterrupted service.

The reason we do not use KD-trees [21] is because KD-trees are best suitedfor partitioning high dimensional spaces, while our solution build index treesfor each descriptor (i.e. dimension), each one having its own metric and distancefunctions. This is how we can easily combine a lot of different feature descriptors,without the need for common normalization (i.e. joint equal contribution or somelogistic regression).

This structure can be used to build quickly searchable index trees for anydescriptor which has a metric. We build these trees for every descriptor we use,separately. The query service will then pick up these index trees and performqueries upon them, when requested by the retrieval interface. Naturally, thesetrees are only able to generate results for a single descriptor at a time, but allof them can be used when performing multi-dimensional queries.

Traditionally BK-trees have been used for string matching algorithms. Es-sentially they are representations of point distributions in discrete metric spaces.That is, if we have feature points with an associated distance metric, then wecan populate a BK*-tree with these points in the following way:

1. Pick one of the points as the root node, R.2. Each node will have a constant number of M child nodes.3. A point Pj will be placed into the child node Ni (i = 0...M − 1), if i ·

dM

< d(Pi, Pj) < (i + 1) · dM

, where d is the maximum distance that twopoints can have (respective the associated metric) and Pj is Pi’s parentnode. Thus, a node will contain a point if its distance from the parent fallsinto the interval specified above; each node representing a difference interval[i · d/M ; (i + 1) · d/M ].

4. Continue recursively until there are no more points left to insert.

The performance of these trees is high; results are generated and orderedin 100-3000ms (depending on the features used in a query, Intel Core2 CPUat 2.4GHz) over the ˜7000 video shots that we currently have in our database.Our video collection consists of video captures from television broadcasts (news,nature films, cartoons, movie clips, ads, etc.) and surveillance cameras.

However, multi-dimensional queries can also be formulated, by selecting mul-tiple features by which the results should be generated, and the results areretrieved from the hypercube cut out from the N -dimensional space of thesefeatures, where the lengths of the cube’s sides are determined by user-editablewindow sizes (which is equivalent with the threshold value t below, as a control

Page 7: VISRET – A Content Based Annotation, Retrieval and ...web.eee.sztaki.hu/~ucu/kovacsACIVS09.pdf · modular support for standard MPEG-7 and custom descriptors, and the easy-to-use

VISRET 7

Fig. 3. Retrieving points from a cube of 3D space defined by 3 feature axes.

mechanism regarding the number of returned results, i.e. a simple more or lesschoice, based on a relative scale between 0 and 1). See Fig. 3.

Given a content-based query (Q), the trees are searched for entries similarto the model image or video shot:

1. If d0 = d(Q,R) < t (t is user-adjustable), the root R element is a result.2. Let Ni be the children of node Pj (P0 = R), and let dk = d(Nk, Pj) for the

child Nk where

k · dM

∈ [dj−1 − t, dj−1 + t] or(k + 1) · d

M∈ [dj−1 − t, dj−1 + t] or

k · dM

≤ dj−1 − t and (k + 1) · dM

≥ dj−1 + t(1)

then if dk < t the element from child Nk is a result.3. Repeat step 2 recursively until the whole tree is visited.4. Sort all the results in the increasing order of their d distances and return the

ordered result list.

3.3 Retrieval

The retrieval over the stored video content can be done in two ways. First,textual queries can be formulated, when we search among the annotations, anddisplay matching results. Logical combinations (and, or) of full or partial wordscan be used as query strings. Secondly, model images can be provided eitherby uploading a query image, or selecting one of the results of a textual query.Results are provided by the indexing service presented above, and results arepresented for the user in the order of relevance, i.e. matches with lower distanceare put at the front.

Fig. 4 shows two samples of model-based queries and first N results, wheretwo query images have been provided, and the results have been generated byusing two features (a color based and an edge based, combined). The resultingimages can be clicked on, then the video shots they are associated with popup and can be played, their annotations can be viewed, and new queries can beformulated by using the specific shot/frame as the model. Fig. 10 shows elementsof the retrieval web interface.

Page 8: VISRET – A Content Based Annotation, Retrieval and ...web.eee.sztaki.hu/~ucu/kovacsACIVS09.pdf · modular support for standard MPEG-7 and custom descriptors, and the easy-to-use

8 VISRET

We do not wish to detail the retrieval performance of the descriptors in thispaper. The performance of the MPEG-7 descriptors we use are known frompractical uses and from many literature sources [18, 19]. The performance ofour relative focus-based descriptor has been detailed in [15]. The evaluationof the other descriptors, and the performance increases of their combined usewill be the subject of another paper. For a quick example though, we includedFig. 5, which shows comparison data for six example content-based retrievals.The six retrievals contain single, and combined queries in the following order:2 (two features), 1 (single feature), 1 (single feature), 2 (two features), 3 (threefeatures), 2 (two features). In the left diagram we show how the precision ofthe retrievals behave when the in-class precision is measured: if the query imageshows a football field with players, the results should also have such a content.The right diagram shows how the prevision of the same retrievals change, whenthe feature content’s precision is measured: if the the query contains a certaincolor and edge/texture content, the results should also have such a content. Thecategory recognition [16, 17] step is what our current on-going research is focusedon.

Fig. 4. Top: Query images. Bottom: Results.

3.4 Visualization

Besides presenting the results in decreasing order of relevance to the user on theweb interface above, we created a tool which provides a flexible way for brows-ing among database shots and images, easily selecting query shots/images, per-forming textual queries, viewing/editing/assigning annotations and categoriesto shots/images, 2D and 3D visualization of results and images belonging to aspecific category. The main window of the tool is in Fig. 6.

Page 9: VISRET – A Content Based Annotation, Retrieval and ...web.eee.sztaki.hu/~ucu/kovacsACIVS09.pdf · modular support for standard MPEG-7 and custom descriptors, and the easy-to-use

VISRET 9

Fig. 5. Precision values for 6 queries. Left: in-class precision, middle: descriptor preci-sion, right: sample query images.

Fig. 6. Main window of the visualization tool, with a subset of the representativeframes.

Fig. 7. Selecting any displayed image (left) to be the query, and re-arrange the imageswith the new image as the base (right).

Page 10: VISRET – A Content Based Annotation, Retrieval and ...web.eee.sztaki.hu/~ucu/kovacsACIVS09.pdf · modular support for standard MPEG-7 and custom descriptors, and the easy-to-use

10 VISRET

The most important functions this tool provides are:

– Display the representative frames in a 2D browsing mode where the two axescan be any combination of the available descriptors.

– Images can be zoomed and dragged around.– Select any of the displayed images, use them as a query image for a content-

based retrieval and re-display the images in a form where this new queryimage is the base of the 2D representation (Fig. 7).

– View and edit the annotation of any selected image.– Assign categories to a selection of images, add new categories, edit categories

(Fig. 8).– Display a 3D point cloud of the images (Fig. 9) where colors represent dif-

ferent categories; the axes of the 3D plot can be any combination of theavailable descriptors; the 3D plots can be zoomed, rotated, points can beselected (larger yellow points in the plot), any selection of images from the2D view can be displayed in the 3D view as well.

– Display all 2D combinations of descriptor pairs for easy selection (simplepoint&click) to change the image display perform text queries on the anno-tations, and the result frames will be displayed in the 2D view.

Fig. 8. Left: 3D plot of frame distributions for 3 specific features, different colors meandifferent categories. Right: List of - color coded - categories that can be assigned,displayed, and new ones can be added; and all the 2D plots variations for the selectedfeatures.

3.5 Conclusions, Future Work

The system we presented contains a series of tools aiding the import, process-ing, indexing and querying a video database, and the visualization of the results.For each imported video the descriptors are run automatically, and every newlyadded descriptor is also automatically run for all previously imported videos,indexing is also automatic. What the system is lacking, is a high level catego-rization algorithm, which would learn features specific for certain categories, andthen automatically classify newly imported videos and shots into the learnt cat-egories. This certainly is a highly desired feature, that we are working on. Still,

Page 11: VISRET – A Content Based Annotation, Retrieval and ...web.eee.sztaki.hu/~ucu/kovacsACIVS09.pdf · modular support for standard MPEG-7 and custom descriptors, and the easy-to-use

VISRET 11

in its current form, the architecture provides an easy platform for testing newdescriptors, already usable for general video retrieval tasks - e.g. judicial mediasearch in one of our projects -, and the performance of the retrieval engine is inthe range of milliseconds. We are also working on extending it to support stan-dalone images besides videos, and importing Corel, Microsoft and other imagedatabases into our data pool.

The system can easily be extended with additional descriptors, and the addi-tion of other modalities - e.g. audio - can also easily be done without modifyingthe elements of the system. We are also working on integrating the visualizationtool into the web interface as an AJAX application, and making an instance ofthe system publically accessible as a community video server.

Acknowledgments Our work has been partially supported by the JUMAS EUFP7 project, and by the Hungarian Scientific Research Fund under grant number76159.

Fig. 9. 3 samples for displaying 3D point cloud distributions of images, any of thedescriptors can be selected to be one of the 3 axes. 3D plots can be rotated, zoomed,tilted; points can be selected (with click&move) and selections shown in the 2D imagebrowsing pane.

References

1. Wang, J. Z., Li, J., Wiederhold, G.:SIMPLIcity: Semantics-sensitive IntegratedMatching for Picture LIbraries. IEEE Trans. on Pattern Analysis and Machine In-telligence, vol 23, no.9, 947–963 (2001)

2. The Art Museum Image Consortium, http://www.amico.org/.3. IBM’s Query by Image Content, IBM QBIC, http://wwwqbic.almaden.ibm.com/.4. Chang, S. F., Chen, W., Meng, H. J., Sundaram, H., Zhong, D.: VideoQ: An Auto-

matic Content-Based Video Search System Using Visual Cues. Proceedings of ACMMultimedia (1997)

5. Google video search, http://video.google.com/.6. Yahoo video search, http://video.search.yahoo.com/.

Page 12: VISRET – A Content Based Annotation, Retrieval and ...web.eee.sztaki.hu/~ucu/kovacsACIVS09.pdf · modular support for standard MPEG-7 and custom descriptors, and the easy-to-use

12 VISRET

Fig. 10. Search web interface. Left-to-right: user search window; view video data andannotation; administrative interface with imported videos and descriptors.

7. Tineye, http://tineye.com/.8. Jinni, http://www.jinni.com/.9. Manjunath, B. S., Ohm, J. R., Vasudevan, V. V., Yamada, A.: Color and Texture

Descriptors. IEEE Tr. on Circuits and Systems for Video Technology, vol. 2, no. 6,703–715 (2001)

10. Kovacs, L., Sziranyi, T.: Focus Area Extraction by Blind Deconvolution for Defin-ing Regions of Interest. IEEE Tr. on Pattern Analysis and Machine Intelligence, vol.29, no. 6, 1080–1085 (2007)

11. Lowe, D. G.: Object recognition from local Scale-Invariant Features. Proceedingsof ICCV, 1150–1157 (1999)

12. Comaniciu, D., Meer, P.: Mean Shift: A Robust Approach Toward Feature SpaceAnalysis. IEEE Tr. on Pattern Analysis and Machine Intelligence, vol. 24, no. 5,603–619 (2002)

13. Burkhard, W., Keller, R.: Some Approaches to Best-Match File Searching, Pro-ceedings of CACM (1973)

14. Viola, P., Jones, M.: Robust Real-Time Face Detection. International Journal ofComputer Vision (IJCV) vol. 57, no. 2, 137–154 (2004)

15. Kovacs, L., Sziranyi, T.: Evaluation of Relative Focus Map Based Image Indexing.Proceedings of CBMI, 181–191, (2007)

16. Ion, A., Stanescu, L., Burdescu, D., Udristoiu, S.: Mapping Image Low-Level De-scriptors to Semantic Concepts. Proceedings of ICCGI, 154–159 (2008)

17. Fergus, R., Perona, P., Zisserman, A.: Weakly Supervised Scale-Invariant Learningof Models for Visual Recognition. Int. J. Comput. Vision, vol. 71, no. 3, 273–303(2007)

18. Annesley, J., Orwell, J., Renno, J. P.: Evaluation of MPEG7 Color Descriptors forVisual Surveillance Retrieval. Proceedings of ICCCN, 105–112 (2005).

19. Ojala, T., Maenpaa, T., Viertola, J., Kyllonen, J., Pietikainen, M.: Empirical Eval-uation of MPEG-7 Texture Descriptors with A Large-Scale Experiment. Proceeed-ings of the Workshop on Texture Analysis in Machine Vision, 99–102 (2002)

20. Czuni, L., Hanis, A., Kovacs, L., Kranicz, B., Licsar, A., Sziranyi, T., Kas,I., Kovacs, Gy., Manno, S.: Digital Motion Picture Restoration System for FilmArchives (DIMORF). SMPTE Motion Imaging Journal, May/June 2004, 170–178(2004).

21. Friedman, J. H., Bentley, J. L., Finkel, R. A..: An Algorithm for Finding BestMatches in Logarithmic Expected Time. ACM Trans. on Mathematical Software,vol. 3, no. 3, 209–226 (1977).