• Aucun résultat trouvé

The LIG Multi-Criteria System for Video Retrieval

N/A
N/A
Protected

Academic year: 2021

Partager "The LIG Multi-Criteria System for Video Retrieval"

Copied!
2
0
0

Texte intégral

(1)

HAL Id: hal-00953858

https://hal.inria.fr/hal-00953858

Submitted on 28 Feb 2014

HAL is a multi-disciplinary open access archive for the deposit and dissemination of sci- entific research documents, whether they are pub- lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

The LIG Multi-Criteria System for Video Retrieval

Stéphane Ayache, Georges Quénot, Laurent Besacier

To cite this version:

Stéphane Ayache, Georges Quénot, Laurent Besacier. The LIG Multi-Criteria System for Video

Retrieval. International Conference on Image and Video Retrieval (demo session, VideOlympics),

2009, Santorini, Greece. �hal-00953858�

(2)

The LIG Multi-Criteria System for Video Retrieval

Stéphane Ayache

Laboratoire d’Informatique Fondamentale de Marseille

Georges Quénot, Laurent Besacier

Laboratoire d’Informatique de Grenoble

Georges.Quenot@imag.fr

ABSTRACT

The LIG search system uses a user-controlled combination of six criteria: keywords, phonetic string, similarity to example images, semantic categories, similarity to already identified positive images, and temporal closeness to already identified positive images.

Categories and Subject Descriptors: H.3.3 [Informa- tion Storage and Retrieval]: Information Search and Re- trieval

General Terms: Algorithms, Experimentation.

Keywords: Video retrieval, multi-criteria search.

1. THE LIG VIDEO RETRIEVAL SYSTEM

The LIG search system [1] uses a user-controlled combina- tion of six criteria: keywords, phonetic string (new in 2009), similarity to example images, semantic categories, similarity to already identified positive images, and temporal closeness to already identified positive images (Figure 1). The system outputs an ordered list of relevant shots for each topic after interaction with the user (initial query and multiple rele- vance feedback). This system was used for TRECVID [2], VideOlympics [3] and the Star Challenge [4].

Figure 1: The LIG video retrieval system Criterion 1: Keyword based search. The keyword based search is done using a vector space model. The words present in the ASR-MT transcription are used as vector

Copyright is held by the author/owner(s).

CIVR’09,July 8–10, 2009 Santorini, GR ACM 978-1-60558-480-5/09/07.

space dimensions. Stemming ans stopword list are used.

Relevance is first assigned to speech segments and then pro- jected onto overlapping shots.

Criterion 2: Phonetic string based search. Video doc- uments are transcribed at the phonetic level using a subset of the International Phonetic Alphabet (IPA). Queries are made in the same way and the search is done using a pho- netic string spotting algorithm.

Criterion 3: Similarity to image examples. Visual similarity between key frames and image examples is looked for using color and texture characteristics (4×4×4 color his- tograms and 8×5 Gabor transforms). normalized and the turned into a relevance value for each feature.

Criterion 4: Feature based search. Video shots are ranked according to the 36 pre-classified categories used in the TRECVID 2007 high level feature detection task.

Criteria 5 and 6: Visual similarity and temporal closeness. Visual similarity and temporal closeness (within the video stream) to already retrieved images can be used for the search. These images have to be marked by the user as positive examples for similarity based search and/or as positive examples for temporal closeness based search (rele- vance feedback).

Combination of search criteria. The user can dynami- cally define his search strategy according to the topic and/or the looking of the retrieved images. Each search criterion can be independently configured and given a global weight for the search. The per-criterion relevances are linearly com- bined according to the criteria weights to produce the final key frame relevance.

2. REFERENCES

[1] St´ephane Ayache and Georges Qu´enot. CLIPS-LSR Experiments at TRECVID 2006. InProceedings of the TRECVID 2006, 13-14 November 2006.

[2] Alan F. Smeaton, Paul Over, and Wessel Kraaij.

Evaluation campaigns and TRECVid. InMIR ’06:

Proceedings of the 8th ACM International Workshop on Multimedia Information Retrieval, pages 321–330, New York, NY, USA, 2006. ACM Press.

[3] Cees G. M. Snoek, Marcel Worring, Ork de Rooij, Koen E. A. van de Sande, Rong Yan, and Alexander G.

Hauptmann. VideOlympics: Real-time evaluation of multimedia retrieval systems.IEEE MultiMedia, 15(1):86–91, January-March 2008.

[4] http://hlt.i2r.a-star.edu.sg/starchallenge.

Références

Documents relatifs

The values of the new BRCA1/2 mutation scores for the differ- ent members of a family, according to cancer localisation and type (TN status for breast cancer), age class at

In the last few years many efforts have been spent to extract information contained in text doc- uments, and a large number of resources have been developed that allow exploring

Semantic similarity is used to include similarity measures that use an ontology’s structure or external knowledge sources to determine similarity between

This method allows an efficient combination of machine transla- tion systems, by rescoring the log-linear model at the N-best list level according to auxiliary systems: the

In this section, we briefly list out all the features used in WMT 2013 (Luong et al., 2013) that were kept for this year, followed by some proposed features taking advantage of

More precisely, we compared the performance of a SLT system where multiple ASR 1-best are combined before translation (source combination), with a SLT system

Un outil motivationnel pour les fumeurs atteints de BPCO peut être les résultats positifs des études dans lesquelles il y a eu un fort taux de sevrage tabagique des patients

However, this technique is much more used on text-based resources (where content analysis is not expensive), while in the field of CoPE, resources are of various