• Aucun résultat trouvé

Adaptive and online speaker diarization for meeting data

N/A
N/A
Protected

Academic year: 2022

Partager "Adaptive and online speaker diarization for meeting data"

Copied!
5
0
0

Texte intégral

(1)

ADAPTIVE AND ONLINE SPEAKER DIARIZATION FOR MEETING DATA Giovanni Soldi

1

, Christophe Beaugeant

2

and Nicholas Evans

1

1

Multimedia Communications Department, EURECOM, Sophia Antipolis, France

2

Intel Mobile Communications, Sophia Antipolis, France,

1{soldi,evans}@eurecom.fr,2christophe.beaugeant@intel.com

ABSTRACT

Speaker diarization aims to determine ‘who spoke when’ in a given audio stream. Different applications, such as document structuring or information retrieval have led to the exploration of speaker di- arization in many different domains, from broadcast news to lec- tures, phone conversations and meetings. Almost all current diariza- tion systems are offline and ill-suited to the growing need for online or real-time diarization, stemming from the increasing popularity of powerful, mobile smart devices. While a small number of such sys- tems have been reported, truly online diarization systems for chal- lenging and highly spontaneous meeting data are lacking. This paper reports our work to develop an adaptive and online diarization sys- tem using the NIST Rich Transcription meetings corpora. While not dissimilar to those previously reported for less challenging domains, high diarization error rates illustrate the challenge ahead and lead to some ideas to improve performance through future research.

Index Terms— Speaker diarization, clustering and segmenta- tion, online diarization

1. INTRODUCTION

Speaker diarization [1–3] is an unsupervised statistical pattern recog- nition task which aims to determine ‘who spoke when’ in a given audio stream. Speaker diarization has become a key, enabling tech- nology in a wide variety of tasks including document processing, structuring and navigation, information retrieval, meta-data extrac- tion and copyright detection. Data domains include broadcast news, lectures, phone conversations and meetings. The latter is generally considered to be the most challenging on account of the high level of speaker overlap, spontaneity and short speaker turns.

Historically, the state-of-the art in speaker diarization for meet- ings has evolved around the implementation of offline systems, such as bottom-up and top-down hierarchical clustering approaches [3–5].

In both cases, speakers are modelled with Gaussian mixture mod- els (GMMs) which are interconnected to form an ergodic hidden Markov model (HMM) in which the transitions represent speaker turns. These approaches, usually coupled with other segmenta- tion and clustering techniques are applied offline to an entire audio stream to generate a segmentation hypothesis which resembles the real number of speaker and speaker turns.

With the increasing popularity of powerful, mobile smart de- vices, there is now a growing interest to develop online speaker di- arization systems. Due to their computational complexity and high latency, the existing state-of-art diarization techniques are not easily adapted to online processing. There is thus an interest to develop en- tirely new, online approaches. A small number have been reported previously, but the majority, e.g. [6–8], focus on applications involv- ing plenary speeches and broadcast news. Inspired by the approach

in [6], this paper presents our efforts to develop such a system for the online diarization of meetings.

The remainder of this paper is organized as follows. Section 2 outlines previous, related work. Section 3 describes the online di- arization system. Section 4 describes the experimental setup used to obtain results presented in Section 5. Our conclusions are presented in Section 6.

2. PRIOR WORK

Driven by the increasing popularity of powerful, mobile smart de- vices and the need for real-time information extraction in diverse human interaction scenarios, online diarization has attracted increas- ing interest in recent years. Although real-time diarization can be performed efficiently with the aid of multiple microphones and cam- eras [9, 10], scenarios in which only a single microphone is available remain challenging. Online diarization performance is typically far from what can be achieved with offline approaches. This section presents a review of the past work.

Liu et al. [11] present an approach in the context of broadcast news diarization. Speech activity detection (SAD) is applied to iden- tify speech segments which are clustered via one of two different algorithms in order to perform online diarization. The study, in- volving leader-follower, dispersion-based and combined clustering algorithms was conducted with the NIST Hub4 1996 broadcast news database.

Markov et al. [7, 8] investigated more conventional speaker di- arization using GMMs. Non-speech segments are discarded using a suitably trained GMM whereas diarization is performed upon the comparison of speech segments to a set of speaker models. New speaker models are introduced using an incremental expectation- maximization (EM) algorithm. The system was assessed on a database of European Parliament plenary speeches for which a diarization error rate (DER) of 8% was reported.

A similar approach was reported by Geiger et al. [6] for broad- cast news. Here, speaker models were learned through maximum-a- posteriori adaptation (MAP) of a universal backgroud model (UBM).

The same UBM is used to control the attribution of speech segments to existing speakers and the addition of new speaker models. A DER of 39% was reported.

Vaquero et al. [12] present a hybrid system composed of of- fline diarization [5] and online speaker identification. An initial of- fline diarization stage is used to learn speaker models. An online speaker identification system is then used for subsequent diarization.

Speaker model adaptation is performed in parallel. Performance is dependent on the latency and accuracy of the offline process. A DER of 38% is reported for a set of 26 meetings from the NIST Rich Tran- scription (RT) evaluation corpora.

(2)

Oku et al. [13] report a low-latency, online speaker diarization system by exploiting phonetic information in order to estimate more discriminative speaker models. Phone boundaries are considered as potential speaker turns. Features are initially clustered into prede- fined acoustic classes. GMM speaker models have the same number of components as the number of acoustic classes. A traditional delta- BIC-like criterion is then used for speaker clustering and segmenta- tion. Performance is assessed using Japanese TV talk shows where conversations are characterized by short speaker turns and only few silence intervals.

Although there is a growing body of work with a focus on online diarization, a truly online, usable system for meeting data is lacking.

This paper presents our efforts to develop such a system using stan- dard meeting data from the NIST RT (RT) evaluation corpora.

3. ONLINE DIARIZATION SYSTEM IMPLEMENTATION The online diarization system is illustrated in Fig. 1. It is based on the top-down or divisive hierarchical clustering approach to of- fline diarization reported in [4] and the online diarization approach reported in [6]. Aside from background modelling, there are two stages: (i) feature extraction; (ii) speech activity detection and (iii) online classification.

3.1. Feature extraction

Audio files are initially treated with Wiener filtering noise reduc- tion. The signal is then frame blocked, windowed and parametrised with Mel-frequency cepstral coefficients plus energy. Parameters are augmented with delta and acceleration coefficients to produce a se- ries of acoustic observationso1, . . . ,oT. Critically, for any time τ∈1, . . . , Tonly those observations fort < τare used for diariza- tion.

3.2. Speech activity detection and online classification

Non-speech segments are removed according to the output of a conventional model-based speech activity detector (SAD). The re- maining speech segments are then divided into smaller sub-segments whose duration is no longer than an a-priori fixed maximum dura- tionTS. Online classification is then applied in sequence to each segment.

Segments are either attributed to an existing speaker model, or a new speaker model is created. This procedure is controlled with a universal background model (UBM) denoted bys0which is trained on external data. New speaker models are introduced if the current segmentigenerates a higher log-likelihood when compared to the UBM than to a set of speaker modelssj, wherej = 1, . . . , Nand whereNindicates the number of speakers in the current hypothesis.

Segments are attributed according to:

sj= arg max

l∈(0,...,N) K

X

k=1

L(ok|sl) (1) whereok is thek-th acoustic feature in the segmenti, K repre- sents the number of acoustic features in thei-th segment and where L(ok|sl)denotes the log-likelihood of thek-th feature in segmenti given the GMM modelsl. If the segment is attributed tos0then a new speaker modelsN+1is learned by MAP adaptation of the UBM models0 using the features contained in segmenti. The segment is then labelled according to the newly introduced speaker andN is increased by one. When a segment is attributed to an existing

speaker, then the corresponding model is adapted through maximum a-posteriori (MAP) adaptation. The segment is then labelled accord- ing to the recognised speakerjas per Eq. 1.

4. EXPERIMENTAL SETUP

In contrast to previous work this paper is concerned with the online diarization of the most challenging meeting data amassed from the set of NIST RT corpora. The full experimental setup is described here.

4.1. Databases, feature extraction and UBM

Experiments were performed using three independent datasets:

1. RTubm: a set of 16 meeting shows from the NIST RT’04 evaluations;

2. RTdev:a set of 15 meeting shows from the RT’05 and RT’06 evaluations, and

3. RTeval:a set of 15 meeting shows from the RT’07 and RT’09 evaluations.

The average show duration within the RTubm, RTdev and RTe- val sets is 10, 15 and 24 minutes respectively. The average num- ber of speakers within each set is 5, 5 and 4 respectively. All ex- periments concern the most challenging, single distant microphone (SDM) condition. Each acoustic signal is characterized by 12 Mel- frequency cepstral coefficients (MFCCs) augmented by energy, delta and acceleration coefficients thereby obtaining feature vectors with a total of 39 coefficients and computed every 10ms using a 20ms window.

The UBM models0 is trained offline using data in the RTubm set. Experiments were performed with different model sizes: 8, 16, 32, 64 and 128 Gaussian components. In all cases, non-speech seg- ments are removed according to ground-truth transcriptions; UBMs are trained using only the remaining speech segments.

4.2. Online diarization

The system was developed and optimised using only the RTdev set.

The dependence of the segment duration and model size on online di- arization performance is assessed and presented here independently for both the RTdev and RTeval sets; both sets are small and the ac- knowledged differences between them means that it is of interest to analyse performance on both sets. Experiments are reported for segment durations of0.25,0.5,1, . . . ,10seconds and model sizes of between 8 and 128. Speech activity detection is applied in the same way, with diarization being applied only to remaining speech segments.

4.3. Assessment and metrics

Diarization performance is assessed using the standard diarization error rate (DER) metric with the usual 0.25 second collar applied to each speaker turn; diarization errors within the collar are ignored.

Performance is first assessed as a function of segment duration for differing model sizes. Dynamic convergence performance is then as- sessed periodically at each minuteTi. For maximum statistical sig- nificance, assessment is nonetheless performed on the entire show.

While this approach toassessmentmight appear more in keeping with offline diarization, results still reflect online performance; it uses speaker models learned through the online process only up until minuteTi. While diarization performance should improve naturally

(3)

7.85 7.9 7.95 8 8.05 8.1

o1,1 o1,2 o1,3 o1,N

o2,1 o2,2 o2,3 o2,N

ot,1 ot,2 ot,3 ot,N

oT,1 oT,2 oT,3

oT,N

o1 o2 ot oT

S S S S

SPK 1

SPK 1

SPK 1

S

SPK 1

S

SPK 1

SPK 2 update

update SEGMENTS (1,2,… seconds) CALCULATE THE LIKELIHOOD OF THE ENTIRE SEGMENT FOR EACH OF THE MODELS

o3,1 o3,2 o3,3

o3,N o3 1. FEATURE

EXTRACTION

2. SAD

1

NS S

2 3 4 5 6 7

update SPK

2

update

S

SPK

1 update

update SPK

2 S

3. ONLINE CLASSIFICATION

(Speech segments longer than a fixed maximum duration TS are split in smaller sub-segments)

TS

Fig. 1: An illustration of the online speaker diarization system.

as the full set of speakers is gradually introduced into the online pro- cess, as illustrated in Fig. 4 (red line) 90% of speakers appear in the first 2 to 3 minutes of each show. This approach to assessment is therefore still representative of online performance.

Each segment is attributed to one of the speaker models ac- cording to the highest likelihood criteria, without any model adap- tation or re-segmentation. While the application of adaptation and re-segmentation would improve performance, they would also intro- duce latency not in keeping with online diarization. Adaptation and re-segmentation would furthermore add to the computational com- plexity and so we did not investigate their use.

5. EXPERIMENTAL RESULTS

DER results are assessed as a function of segment duration, model size and the amount of training data available as a function of time Ti. Also presented is an analysis of dynamic speaker statistics in addition to some ideas for future work.

5.1. Segment duration and model size

Figure 2 illustrates online diarization performance in terms of DER as a function of segment duration and model size. Plots are illus- trated for the RTdev set (left) and RTeval set (right). The optimal model size is either 32 or 64 Gaussian components, with the larger model size being the most consistent across the two sets. Perfor- mance deteriorates as the model size is increased further and is due to a lack of sufficient data for reliable learning. Initially, the DER tends to decrease as the segment duration increases. As the segment size increases beyond the optimum, then more and more speaker turns

are missed causing the DER to increase. Across the two datasets, the minimum DER is between 40% and 45%. This is a high error rate, but one not dissimilar to that reported in previous work performed using broadcast new data, e.g. [6].

5.2. Adaptive speaker modelling and convergence

Figure 3 illustrates the dynamic convergence of the DER as a func- tion of timeTi. Plots are again illustrated for the RTdev set (left) and RTeval set (right) and for segment durations of between 1 and 6 seconds. All plots are for GMMs with 64 components. Both fig- ures show that the DER decreases as the amount of data available for model training increases. The plots also show that segment durations of less than 2 seconds are largely insufficient for reliable diarization, with optimal performance being achieved with 2 or 3 second seg- ments. Again, as the segment size increases, then more and more mid-segment speaker turns are missed, thus leading to higher DERs.

5.3. Dynamic speaker statistics

Figure 4 illustrates the evolution in speaker numbers for the RTdev set. Profiles are shown for the ground truth references (red line) and the corresponding number appearing in the automatically generated diarization hypothesis (blue line). The hypothesis corresponding to GMMs of 64 components and segments length of 2 seconds. For the first five to six minutes, the hypothesis contains fewer speakers than the ground truth whereas, beyond, the hypothesis contains more speakers than the ground truth.

(4)

0.25 0.50 1.00 2.00 3.00 4.00 5.00 6.00 7.00 8.00 9.00 10.00

Segment duration (sec)

40 50 60 70 80 90 100

DER (%)

G128 G64 G32 G16 G8

0.25 0.50 1.00 2.00 3.00 4.00 5.00 6.00 7.00 8.00 9.00 10.00

Segment duration (sec)

40 50 60 70 80 90

DER (%)

G128 G64 G32 G16 G8

Fig. 2: An illustration of DER as a function of segment duration (0.25,0.5,1-10 sec) and for different model sizes (8-128). Results shown for the RTdev (left) and RTeval (right) datasets.

1 2 3 4 5 6 7 8 9 10 11

Minutes (min)

35 40 45 50 55 60 65

DER (%)

1 sec 2 sec 3 sec 4 sec 5 sec 6 sec

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18

Minutes (min)

35 40 45 50 55 60 65

DER (%)

1 sec 2 sec 3 sec 4 sec 5 sec 6 sec

Fig. 3: An illustration of the evolution in DER against time (minutes) and for different segment durations. Results shown for the RTdev (left) and RTeval (right) datasets

5.4. Future work

These observations suggest that a speaker penalty might be useful to favour the introduction of fewer speakers, or at least reduce the rate at which new speakers are added. Of course, this is likely to reduce the overall convergence rate; the full number of speakers will take longer to appear in the hypothesis.

Two alternative strategies might help to reach a better compro- mise between the overestimation of speaker numbers and slower

convergence. First, an adaptive speaker penalty may be used initially to favour the introduction of more speakers, but penalise their intro- duction later. Second, parallelised agglomerative hierarchical clus- tering may be applied periodically to merge similar models which might correspond to the same speaker. Since the computational im- plications of this approach may be too great for our intended appli- cation, their investigation has been left for future work.

Finally, since the use of longer segments introduces latency, it will also be of interest to combine online diarization with phone

(5)

20 18 16 14 12 10 8 6 4 2

0 Time (min)

0 20 40 60 80 100 120 140

Percentage of speakers (%)

Groundtruth Hypothesis

Fig. 4: An illustration of the evolution in speaker numbers for the RTdev dataset. Profiles shown for the ground-truth reference (red profile) and diarization hypothesis (blue profile).

adaptive training (PAT) [14, 15] which can improve speaker mod- elling when dealing with short utterances or segments. Since PAT in- volves only a linear feature transform, the additional computational requirements are minimal.

6. CONCLUSION

This paper presents a new adaptive, online approach to speaker di- arization. In contrast to the past work, this paper addresses the most challenging of domains, namely meeting data characterised by high speaker overlap, spontaneity and short speaker turns. Experiments show that the best performance implies a latency in the order of 2 seconds and that speaker models convergence as the amount of train- ing data increases. While results are in line with those reported for less challenging data, diarization error rates remain high. The paper also presents some strategies to improve performance through future work. This should focus not only on reducing diarization errors, but also on the rate of convergence, more reliable estimates of speaker numbers and improved diarization in the case of short segments.

7. REFERENCES

[1] S.E. Tranter and D.A. Reynolds, “An overview of automatic speaker diarization systems,”IEEE Trans. Audio, Speech, Lan- guage Processing, vol. 14, no. 5, pp. 1557–1565, Sept 2006.

[2] X. Anguera, S. Bozonnet, N. Evans, C. Fredouille, G. Fried- land, and O. Vinyals, “Speaker diarization: A review of recent research,” IEEE Trans. Audio, Speech, Language Processing, vol. 20, no. 2, pp. 356–370, 2012.

[3] N. Evans, S. Bozonnet, Dong Wang, C. Fredouille, and R. Troncy, “A comparative study of bottom-up and top- down approaches to speaker diarization,” IEEE Trans. Au-

dio, Speech, Language Processing, vol. 20, no. 2, pp. 382–392, 2012.

[4] S. Bozonnet, N.W.D. Evans, and C. Fredouille, “The Lia- Eurecom RT’09 speaker diarization system: enhancements in speaker modelling and cluster purification,” inProc. Int.

Conf. Acoustics, Speech, Signal Processing (ICASSP), 2010, pp. 4958–4961.

[5] G. Friedland, A. Janin, D. Imseng, X. Anguera, L. Gottlieb, M. Huijbregts, M.T. Knox, and O. Vinyals, “The ICSI RT’09 speaker diarization system,”IEEE Trans. Audio, Speech, Lan- guage Processing, vol. 20, no. 2, pp. 371–381, 2012.

[6] J. T. Geiger, F. Wallhoff, and G. Rigoll, “GMM-UBM based open-set online speaker diarization.,” in Proc. Ann.

Conf. International Speech Communication Association (IN- TERSPEECH), 2010, pp. 2330–2333.

[7] K. Markov and S. Nakamura, “Never-ending learning sys- tem for on-line speaker diarization,” inIEEE Workshop Auto- matic Speech Recognition Understanding (ASRU), Dec 2007, pp. 699–704.

[8] K. Markov and S. Nakamura, “Improved novelty detection for online gmm based speaker diarization,” in Proc. Ann.

Conf. International Speech Communication Association (IN- TERSPEECH), 2008, pp. 363–366.

[9] T. Hori, S. Araki, T. Yoshioka, M. Fujimoto, S. Watanabe, T. Oba, A. Ogawa, K. Otsuka, D. Mikami, K. Kinoshita, T. Nakatani, A. Nakamura, and J. Yamato, “Low-latency real- time meeting recognition and understanding using distant mi- crophones and omni-directional camera,” IEEE Trans. Audio, Speech, Language Processing, vol. 20, no. 2, pp. 499–513, Feb 2012.

[10] S. Araki, T. Hori, M. Fujimoto, S. Watanabe, T. Yoshioka, T. Nakatani, and A. Nakamura, “Online meeting recognizer with multichannel speaker diarization,” inConf. Signals, Sys- tems and Computers (ASILOMAR), Nov 2010, pp. 1697–1701.

[11] D. Liu and F. Kubala, “Online speaker clustering,” inProc. Int.

Conf. Acoustics, Speech, Signal Processing (ICASSP), May 2004, vol. 1, pp. I – 333–6.

[12] C. Vaquero, O. Vinyals, and G. Friedland, “A hybrid approach to online speaker diarization,” inProc. Ann. Conf. Interna- tional Speech Communication Association (INTERSPEECH), 2010, pp. 2638–2641.

[13] T. Oku, S. Sato, A. Kobayashi, S. Homma, and T. Imai, “Low- latency speaker diarization based on bayesian information cri- terion with multiple phoneme classes,” in Proc. Int. Conf.

Acoustics, Speech, Signal Processing (ICASSP), March 2012, pp. 4189–4192.

[14] S. Bozonnet, R. Vipperla, and N. Evans, “Phone adaptive train- ing for speaker diarization,” inProc. Ann. Conf. International Speech Communication Association (INTERSPEECH), 2012.

[15] G. Soldi, S. Bozonnet, Alegre F., C. Beaugeant, and N. Evans,

“Short-duration speaker modelling with phone adaptive train- ing,”Odyssey, 2014.

Références

Documents relatifs

Novel contributions include: (i) a new approach to BK extraction based on multi-resolution spectral analysis; (ii) an approach to speaker change detection using BKs; (iii)

Discrepancies between performance for oracle, automatic and segmental diarization systems are, however, dependent to some degree on the fixed speaker latency, especially for the

Select 1st Gaussian

5.7 An illustration of DER for the semi-supervised on-line diarization system as a function of the speaker model training duration T SP K and for different maximum segment durations

While know- ing the number of speakers and the use of labelled data is also at odds with the traditional definition of diarization, many meeting scenarios involve a short

Ex- periments carried out on the REPERE phase 1 test set of TV broadcast programs show the effectiveness of the introduced similarity measures and the NAP compensation method,

On the development data, the overall speaker diarization error was reduced from 24.8% for the best setting of the baseline system to 13.2% using BIC clustering, and to 7.1% with

Error rates, expressed in terms of missed speaker (MiE), false alarm speaker (F aE), speaker (SE) and diarization speaker (DER) error rates (in %), obtained by each speaker