• Aucun résultat trouvé

Multi-level fusion for semantic video content indexing and retrieval

N/A
N/A
Protected

Academic year: 2022

Partager "Multi-level fusion for semantic video content indexing and retrieval"

Copied!
9
0
0

Texte intégral

(1)

Multi-level Fusion for Semantic Video Content Indexing and Retrieval

Rachid Benmokhtar and Benoit Huet D´epartement Communications Multim´edias

Institut Eur´ecom 2229, route des crˆetes 06904 Sophia-Antipolis - France

(Rachid.Benmokhtar, Benoit.Huet)@eurecom.fr

Abstract. In this paper, we present the results of our work on the analysis of an automatic semantic video content indexing and retrieval system based on fus- ing various low level visual and edges descriptors. Global MPEG-7 features, ex- tracted from video shots, are described via IVSM signature (Image Vector Space Model) in order to have a compact description of the content. Both static and dy- namic feature fusion are introduced to obtain effective signatures. Support Vec- tor Machines (SVMs) are employed to perform classification (One classifier per feature). The task of the classifiers is to detect the video semantic content. Then, classifier outputs are fused using a neural network based on evidence theory (NN- ET) in order to provide a decision on the content of each shot. The experimental results are conducted in the framework of the TrecVid feature extraction task.

1 Introduction

To respond to the increase in audiovisual information, various methods for indexing, classification and fusion have emerged. The need to analyse the content has appeared to facilitate understanding and contribute to a better automatic video content indexing and retrieval.

The retrieval of complex semantic concepts requires the analysis of many features per modalities. The task consisting of combining of all these different parameters is far from trivial. The fusion mechanism can take place at different levels of the classification process. Generally, it is either applied on signatures (feature fusion) or on classifier outputs (classifier fusion).

This paper presents our research conducted toward a semantic video content in- dexing and retrieval system aimed at the TrecVid high level feature detection task.

It starts with a description of our automatic system architecture. We distinguish four steps: feature extraction, feature fusion, classification and classifier fusion. The overall processing chain of our system is presented in figure 1. The feature extraction step con- sists in creating a set of global MPEG-7 low level descriptors (based on color, texture and edges). Two feature fusion approaches are used: Static and dynamic. The static ap- proach is based on simple operators while the dynamic approach consist in reducing the data dimensionality using Principal Component Analysis (PCA). Both are implemented and evaluated with the aim to obtain effective signature for each shot. The classification

(2)

Clustering

(x) feature number Color Layout

Classification (SVM, GMM,

NN,…)

Video shot Semantic detection MPEG 7 feature

extraction

Visual dictionary

IVSM

Feature Fusion

Classifier Fusion (NNET,DT,NN…)

Color Structure Scalable Color Homogenous Texture Edge histogram Dominant Color

Merged feature

TrecVid videos

Annotation

Fig. 1.General framework of the application.

step is used to estimate the video semantic content. Support Vector Machine (SVMs) are employed. In the final stage of our system, fusion of classifier outputs is performed thanks to a neural network based on evidence theory (NN-ET).

The experimental results presented in this paper are conducted on the TrecVid col- lection, varying the automatic generation techniques and combination strategies. Fi- nally, we examine the outcomes of our experiments, detail our continuing work on how dynamic feature fusion could be used to complement, rather than replace existing ap- proaches.

2 System Architecture

The MPEG-7 standard defines a comprehensive, standardized set of audiovisual de- scription tools for still images as well as movies. The aim of the standard is to facil- itate quality access to content, which implies efficient storage, identification, filtering, searching and retrieval of media [1]. We have used the following still image features:

– Dominant color (DC)represents the most dominant colors,

– Color layout (CL)specifies a spatial distribution of colors. The image is divided into (8x8) blocks and the dominant colors are solved for each block in the YCbCr color system. Discrete Cosine Transform is applied to the dominant colors in each channel and the DCT coefficients are used as a descriptor.

– Color structure (CS)slides a structuring element over the image, the numbers of positions where the element contains each particular color is recorded and used as a descriptor.

– Scalable color (SC)is a 256-bin color histogram in HSV color space, which is encoded by a Haar transform.

– Edge histogram (EH)calculates the amount of vertical, horizontal, 45 degree, 135 degree and non-directional edges in 16 sub-images of the picture.

(3)

– Homogeneous texture (HT)descriptor filters the image with a bank of orientation and scale tuned filters that are modeled using Gabor functions. The first and second moments of the energy in the frequency domain in the corresponding sub-bands are then used as the components of the texture descriptor.

Fig. 2.Example of key-frames illustrating three semantic concepts (Person, airplane and car).

The obtained vectors over the complete database are clustered to find the N most representative elements. The clustering algorithm used in our experiments is the well- known k-means with the Euclidean distance. Representative elements are then used as visual keywords to describe video shot content. Then, the occurrence vector of the visual keywords in the shots are built and this vector is called the IVSM signature (Image Vector Space Model). The number of visual terms used in our experiments is 70.

2.1 Static feature fusion

In this work, experiments describe an automatic detection of semantic concepts. Four color descriptors, edges histogram and homogenous texture descriptor are extracted from the visual content of a video shot. The main objective of feature fusion step is to reduce redundancy, uncertainty and ambiguity of signatures, in order to obtain a complete information of better quality, for take better decision and act.

Concatenation of features In the first fusion strategy, all global MPEG-7 descriptors are merged into a unique vector, that is calledmerged fusion(Dmerged)as follow :

Dmerged= [DC|CL|CS|SC|EH|HT] (1)

All descriptors must have more or less the same numerical values to avoid scale effects [2].

Average of features This approach builds an average of the different descriptors. It requires no compilation of data, a simple normalization step is required before data can be added. It is interesting to give a weight or confidence level to each of the descriptors.

This method is commonly used, in particular in the automatic video concepts detec- tion of the TrecVid project [3], where we observe the good contribution of the fusion operators asMinand average.

(4)

2.2 Dynamic feature fusion using PCA

Many techniques for dimensionality reduction have been proposed in the literature.

However, Principal Component Analysis (PCA), latent semantic analysis (LSA) [4] and recently Independent Component Analysis (ICA) are the most frequently used. PCA extracts the features as the projections on the principal subspace whose basis vectors correspond to the maximum variance directions in the original space, while discarding the complementary subspace as a noise subspace. In some cases, PCA can obtain sat- isfactory performance. However, no theory can prove the complementary subspace is useless for recognition, and, on the contrary, experiments show that using the comple- mentary subspace properly may improve recognition performance [5,6]. In our work, the dimensionm∈[10,450]evolve per step of20.

2.3 Support Vector Machines Classification

SVMs were widely used in the past ten years and they have been proved efficient in many classification applications. They have the property to allow a non linear separa- tion of classes with very good generalization capacities. They were first introduced by Vapnik [7] for the text recognition task. The main idea is similar to the concept of a neuron: separate classes with a hyperplane. However, samples are indirectly mapped into a high dimensional space thanks to a kernel function. To this end, the selected ker- nel denotedK(.)is a radial basis function which normalization parameterσis chosen depending on the performance obtained on a validation set. The radial basis kernel is chosen for his good classification results comparing to polynomial and sigmoidal ker- nels [4].

K1(x, y) = exp

−||x−y||2 σ

(2)

2.4 Classifier fusion : Neural network based on evidence theory (NN-ET) Classifier fusion is a necessary step to efficiently classify the video semantic content from multiple cues. For this aim, an improved version of RBF neural network based on evidence theory [8] witch we call NN-ET is used [9], with one input layerLinput, two hidden layersL2andL3and one output layerLoutput(figure 3). Each layer corresponds to one step of the procedure described in following:

1. LayerL1:ContainsNunits (prototypes). It is identical to the RBF network input layer with an exponential activation functionφandda distance computed using training data.α∈[0,1]is a weakening parameter associated to prototypei, where = 0at the initialization stage [9]:

siiφ(di)

φ(di) = exp (−γi(di)2) αi=1+exp (−1 i)

(3) whereγiis a positive parameter defining the receptive field size of prototypei= {1, ..., N}.

(5)

L3

O1 OM OM+1

+1

+

= q M

q O O

P Classifier 1

Classifier 2

Classifier i

Classifier L

m1

mi

mN j

Pi, u1,q

q

ui,

q

uN,

Lin L2 Lout

s1

si

sN

μ1

μi

μN

Normalized output

Plausibility

Fig. 3.Neural network based on evidence theory (NN-ET) classifier fusion structure

2. LayerL2:Computes the belief massesmi(Equ. 4) associated to each prototype. It is composed ofNmodules ofM+ 1units each (Equ. 5). The units of moduleiare connected to neuroniof the previous layer. Knowing that each image can belong to only one class (annotation clauses), we write:

mi({wq}) =αiuiqφ(di) mi({Ω}) = 1−PM

q=1mi({wq}) (4)

hence,

mi= (mi({w1}), ..., mi({wM+1}))

= (ui1si, ..., uiMsi,1−si) (5) whereuiqis the membership degree to each classwq,qclasse indexq={1, ..., M}

3. Layer L3: The Dempster-Shafer combination rule combines N different mass functions in one single mass. It’s given by the following conjunctive combination:

m(A) = (m1⊕...⊕mN) = X

B1T...TBN=A N

Y

i=1

mi(Bi) (6) TheNmass functionmiare composed ofNmodules ofM+ 1units. The activa- tions vector of modulesiis defined as

µi. µi=Ti

k=1mki−1Tmi

µ1=m1 (7)

The activation vectors fori = {2, ..., N} can be recursively computed using the following formula:

(6)

µiji−1j miji−1j miM+1i−1M+1mij

µiM+1i−1M+1miM+1

(8)

4. Ouput Layer:We build the normalized outputOdefined as:

Oj=

PN i=1µij PN

i=1

PM+1

j=1 µij (9)

The different parameters (∆u,∆γ,∆α,∆P,∆s) can be determined by gradient descent of output error for an input patternx(more explanations see [9]). Finally, we compute the maximum ofPq(i.e the plausibility of each classwq) as follow:

Pq=Oq+OM+1 (10)

3 Experiments

Experiments are conducted on TrecVid videos [10]. The main goal of TrecVid is to promote progress in content-based retrieval from digital video via open, metrics-based evaluation. TrecVid is a laboratory-style evaluation that attempts to model real world situations or significant component tasks involved in such situations. It will test on news reports, science news, documentaries, educational programming, and archival video, to see how well the technologies apply to new sorts of data. In this work, about 5 hours of video (4000 shots) are used to train the feature extraction system and 1 hour (800 shots) are used for evaluation purpose. The training set is divided into two subsets in order to train both classifiers and subsequently determine through learning the fusion parameters. Detection performance was measured using the standard precision and recall metrics. We are interested by the precision to have a measure of the ability of a system to present only relevant shots. Average precision is given in follow:

AP=

number of relevent shots retrieved total number of shots retrieved

total number of relevent shots (11)

The feature extraction task consists in retrieving video shots expressing one of the following six concepts (Sports, outdoor, building, mountain, waterscape, maps) among 36 proposed semantic concepts. Table 1 provides some insight about the composition in terms of our selected semantic concepts.

We start the experimentations with the description of three following system con- figurations:

– System 1:System without feature fusion step (see figure 1, we build one SVM classifier per feature, then fuse all classifier outputs by NN-ET);

– System 2:System with concatenation static feature fusion (we add in the system 1, the merged feature by concatenation to be classified);

– System 3:System with average static feature fusion (we add in the system 1, the merged feature by average to be classified).

(7)

Id Concepts test train

1 Sports 19 86

2 Outdoor 260 512 3 Building 90 205 4 Mountain 12 45 5 Waterscape 23 108

6 Maps 13 29

Table 1.key-frames distribution of the video key-frames in the various sets by semantic concepts. The relative quantity of every class is clarified to give an idea of the lower border of the performances to be obtained

Fig. 4.Comparison of system configurations results

Figure 4 shows average precision results for the three distinct experiences. It can be seen that thesystem 3improves the average precision for all semantic concepts and obtains better scores (4%improvement on MAP) comparing tosystem 1. Contrary to thesystem 2, that decreases precision.

The average precisionAP ∈[8,60%], for exemple the semantic concept (outdoor) obtain an average precision of60%. This is can be explained by the high number of positive samples in the test set. Here, almost all positive samples are retrieved in the 100 first video shots returned by systems.

For semantic concepts (mountain,maps), thesystem 3 obtain (14%,11%), which can be explained per the low number of positive samples in the training and test sets.

On average, the MAP oscillates around29% using average feature fusion step, which represents a good performance considering the video shots annotation complex- ity.

Figure 5 shows the variation of average precision results using PCA feature fusion vs dimension for each concept. The dimensiondim ∈ [10,450]per step of 20. The system with PCA step improve the precision of all concepts. The bestM AP = 36.32

(8)

is obtained usingdim= 410. In the test set, we have several monochrome video shots.

We notice that the descriptors are highly redundant. This is not very surprising, because four of six investigated MPEG-7 descriptors are colour descriptors.

So, smaller dimensionsdim ∈ [10,170]lead to loss of information and explains the bad performances with the small dimensions, and a high dimension raises the cal- culation time problem and also the relevance of our signatures. Observe that the stable dimension interval isdim∈[190,450].

Fig. 5.NN-ET results using PCA feature fusion step. Mean Average Precision for PCA feature fusion, from 10 to 450 dimension by step of 20.

Systems Without Average PCA

Feature Fusion Feature Fusion Feature Fusion MAP 25.61% 29.16% 36.32%

Table 2.Mean Average Precision (MAP) for different systems.

Finally, the table 2 summarizes the mean average precision (MAP) for different systems. We notice that PCA feature fusion system obtain superior results to those obtained by the static feature fusion for all semantic concepts and to system without feature fusion step. This shows the importance of feature fusion.

4 Conclusion

In this paper, both static and dynamic feature fusion approaches have been evaluated.

Six global MPEG-7 visual descriptors are being employed for this difficult task. The aim of this feature fusion step is to provide a compact and effective representation for an SVM classifier which is trained to solve the challenging task of video content detection. A further classifier fusion step, featuring a neural network based on evidence

(9)

theory, is also employed within the proposed system in order to combine in the most effective way the output of the SVMs.

We have demonstrated through empirical testing the potential of feature fusion, to be exploited in video shots retrieval. Our model, achieves respectable performance, particularly, for certain semantic concepts like outdoor and building, when the variety of the quality of features used is considered.

Of course, these tests (Small data) do not guarantee the usability of global MPEG-7 descriptors in the general case, but they imply that global MPEG-7 descriptors are worth experimenting with. We start to investigate the effect of TrecVid’07 data in our system (50 hours of videos) with new semantic concepts like (Person, face, car, explosion,...).

We believe that the dynamic feature fusion of different MPEG-7 descriptors based on dimensionnality reduction has a positive impact on our system. Other statistical ap- proaches, such as LDA and SOM are under investigation. In parrallel, we are extending the global MPEG-7 descriptors used in this paper with local MPEG-7 descriptor in or- der to enrich the low level representation and to study the effect of their addition to a fusion system like ours.

Acknowledgement

The work presented here is supported by the European Commission under contract FP6- 027026-K-SPACE. This work is the view of the authors but not necessarily the view of the community.

References

1. M. Mottaleb and S. Krishnamachari, “Multimedia descriptions based on MPEG-7: Extrac- tion and applications,”Proceeding of IEEE Multimedia, vol. 6, pp. 459–468, 2004.

2. E. Spyrou, H. Leborgne, T. Mailis, E. Cooke, Y. Avrithis, and N. O’Connor, “Fusing MPEG- 7 visual descriptors for image classification,”Proceedings of ICANN, vol. 3697, pp. 847–852, 2005.

3. M. Rautiainen and T. Seppanen, “Comparison of visual features and fusion techniques in au- tomatic detection of concepts from news video based on gabor filters,”Proceeding of ICME, pp. 932–935, 2005.

4. F. Souvannavong, B. Merialdo, and B. Huet, “Latent semantic analysis for an effective region based video shot retrieval system,”Proceedings of ACM MIR, pp. 243–250, 2004.

5. I. Jolliffe, “Principle component analysis,”Springer-Verlag, 1986.

6. W. Zhang, S. Shan, W. Gao, Y. Chang, B. Cao, and P. Yang, “Information fusion in face identification,”Proceedings of IEEE ICPR, vol. 3, pp. 950–953, 2004.

7. V. Vapnik, “The nature of statistical learning theory,”Springer, 1995.

8. G. Shafer, “A mathematical theory of evidence,”Princeton University Press, 1976.

9. R. Benmokhtar and B. Huet, “Neural network combining classifier based on Dempster- Shafer theory,”Proceedings of Multimedia MMM, vol. 4351, pp. 196–205, 2007.

10. TrecVid, “Digital video retrieval at NIST,”http://www-nlpir.nist.gov/projects/trecvid/.

Références

Documents relatifs

Run 7 – Mixed: A visual and a semantic retrieval are applied as in the runs 5 and 6 and the results are combined using percentages(80-20) for the text retrieval and image

In order to perform efficient content-based retrieval in the database, it is necessary to construct a hier- archical index representation which, given a query image described by

Content-based Video Retrieval systems (CBVRS) are less common and seem at a first glance to be a natural extension of CBIRS.. In this document, we summarise advances made in

Temporal alignment of primary video data (soccer match videos) with semantically organized events and entities from the textual and structured complementary

Performances of the K-Nearest Neighbors and Gaussian Mixture Models classifiers are compared and provide a framework to evaluate the efficiency of Latent Semantic Indexing for

Our previous work on Latent Semantic Analysis revealed the high potential of this simple method for object retrieval and semantic content estimation.. In this paper we have presented

We propose a system to efficiently index visual features in or- der to extract the semantic content of video shots.. The first step is conducted with an adaptation of Latent

The Image Vector Space Model is the basic approach that con- sists in indexing the content with a count vector of prede- fined visual terms.. The EMD-based method directly uses