• Aucun résultat trouvé

Convolutive Bottleneck Network with Dropout for Dysarthric Speech Recognition

N/A
N/A
Protected

Academic year: 2021

Partager "Convolutive Bottleneck Network with Dropout for Dysarthric Speech Recognition"

Copied!
16
0
0

Texte intégral

(1)

HAL Id: hal-01301040

https://hal.archives-ouvertes.fr/hal-01301040

Submitted on 27 Mar 2017

HAL is a multi-disciplinary open access

archive for the deposit and dissemination of sci-entific research documents, whether they are pub-lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

Dysarthric Speech Recognition

Toru Nakashika, Toshiya Yoshioka, Tetsuya Takiguchi, Yasuo Ariki, Stefan

Duffner, Christophe Garcia

To cite this version:

(2)

Convolutive Bottleneck Network with Dropout for

Dysarthric Speech Recognition

Toru Nakashika1, Toshiya Yoshioka1, Tetsuya Takiguchi1, Yasuo Ariki1, Stefan Duffner2, Christophe Garcia2

1Department of System Informatics, Kobe University, JAPAN 2Université de Lyon, CNRS, INSA-Lyon, LIRIS, UMR 5205, FRANCE

1nakashika@me.cs.scitec.kobe-u.ac.jp

ABSTRACT

In this paper, we investigate the recognition of speech produced by a person with an articulation disorder resulting from athetoid cerebral palsy. The articulation of the first spoken words tends to become unstable due to strain on speech muscles, and that causes degradation of speech recognition. Therefore, we propose a robust feature extraction method using a convolutive bottleneck network (CBN) instead of the well-known MFCC. The CBN stacks multiple various types of layers, such as a convolution layer, a subsampling layer, and a bottleneck layer, forming a deep network. Applying the CBN to feature extraction for dysarthric speech, we expect that the CBN will reduce the influence of the unstable speaking style caused by the athetoid symptoms. Furthermore, we also adopt dropout in the output layer since automatically-assigned labels to the dysarthric speech are usually unreliable due to ambiguous phonemes uttered by the person with speech disorders. We confirmed its effectiveness through word-recognition experiments, where the CNN-based feature extraction method outperformed the conventional feature extraction method.

Keywords: Articulation disorders, Feature extraction, Convolutional neural network, Bottleneck feature, Dropout, Dysarthric speech.

1 INTRODUCTION

Recently, the importance of information technology in the welfare-related fields has increased. For example, sign language recognition using image recognition technology [1], text reading systems from natural scene images [2], and the design of wearable speech synthesizers for voice disorders [3, 4] have been studied.

(3)

persons with a speech impediment, and so on, and there has been very little research on orally-challenged people, such as those with speech impediments. It is hoped that speech recognition systems will one day be able to recognize their voices.

One of the causes of speech impediments is cerebral palsy. Cerebral palsy results from damage to the central nervous system, and the damage causes movement disorders. Three general times are given for the onset of the disorder: before birth, at the time of delivery, and after birth. Cerebral palsy is classified as follows: 1) spastic type 2) athetoid type 3) ataxic type 4) atonic type 5) rigid type, and a mixture of types [6].

In this paper, we focused on a person with an articulation disorder resulting from the athetoid type of cerebral palsy as in [7]. Athetoid symptoms develop in about 10-15% of cerebral palsy sufferers. In the case of a person with this type of articulation disorder, the first movements are sometimes more unstable than usual. That means, the case of movements related to speaking, the first utterance is often unstable or unclear due to the athetoid symptoms. Therefore, we recorded speech data for a person with a speech impediment who uttered a given word several times, and we investigated the influence of the unstable speaking style caused by the athetoid symptoms.

In speech recognition technology, frame-wise features such as mel-frequency cepstral coefficients (MFCC), linear predictive coding (LPC), and an autoregressive model (AR) have been widely used so far. However, these features do not capture the temporal information unless delta features are used. Especially for dysarthric speech, where the signals fluctuate more obviously than the signals uttered by a physically unimpaired person, spectral transition in the short term is considered to be an important factor in capturing the local temporal-dimensional characteristics. In this paper, we employ a convolutional neural network (CNN) [8, 9])-based approach to extract disorder- dependent features from a segment MFCC map. The CNN is regarded as a successful tool and has been widely used in recent years for various tasks, such as image analysis [10, 11, 12], a spoken language [13], and music recognition [14]. A CNN consists of a pipeline of convolution and pooling operations followed by a multi-layer perceptron. In dysarthric speech, the key points in time-spectral local areas of an input feature map are often shifted slightly due to the fluctuation of the speech uttered by a person with an articulation disorder. However, thanks to the convolution and pooling operations, we can train the CNN robustly to deal with the small local fluctuations. Furthermore, we expect that the CNN extracts specific features associated with the articulation disorder we are targeting when we train the network using only the speech data of the articulation disorder.

(4)

is expected that it can aggregate the propagated information and extract fundamental features included in an input map. Furthermore, we also use a dropout technique [16] to prevent over-fitting appeared in the output layer. Since the label data (teaching signal) is obtained by forced-alignment from dysarthric speech where each phoneme fluctuates largely, the teaching signal for the training of the network is very unreliable. The dropout in the output layer ignores some information in the teaching signal, and thus will deal with the problem of the wrong alignment.

This paper is organized as follows: we briefly review the fundamental method, CNN in Section 2. The proposed feature extraction method and the structure of the CBN used in the experiments are presented in Section 3. In Section 4, we show the experimental results, and we give our conclusions regarding in Section 5.

2 CONVOLUTIONAL NEURAL NETWORK

In this section, we describe a convolution layer and a pooling layer, which are fundamental components of a CNN (convolutional neural network).

Figure 1: Feature maps in a convolutional layer (right) feed input feature maps (left) 2.1.1 Convolution layer

Assuming that we have a two-dimensional input feature map and a

convolutive filter , the output of a convolutive operation also

becomes a two-dimensional feature with the size of ( and vice

versa). A CNN generally has a number of such filters in a convolutive layer, and

feeds an input using each filter to create the corresponding outputs , which is

referred to as a feature map.

Given all of the feature maps in the th layer , the th

(5)

,

(1)

where and indicate a predictable filter from the th feature map in th layer

to the th map in the th layer and a bias map of the th map in the th layer, respectively. In this paper, we used an element-wise sigmoid function for the activation function f as follows

, (2)

where the fraction bar indicates element-wise division.

Each unit in a convolution layer is connected to the units in the corresponding local area of size in the previous layer (local receptive field). In other words, the convolution layer in a CNN captures local patterns in an input map using various filters (Figure 1).

2.1.2 Pooling layer

Followed by the convolution layer, a pooling procedure is generally used in a CNN, creating what is called a pooling layer. Each unit in the pooling layer aggregates responses in the local

subregion in the previous convolution layer. As a result, a feature map in the

pooling layer has the size of .

There exist various pooling methods (e.g. max-pooling), but we use average-pooling in this paper, calculated as follows

,

(3)

where and are weight parameter and a bias parameter of the th feature map in

the pooling layer ( th layer), respectively. represents the unit in the corresponding

subregion identified with in the feature map in the th layer.

(6)

3 PROPOSED METHOD

In our approach, we use a convolutional neural network (CNN) that has a bottleneck layer in the network, referred to as a convolutive bottleneck network (CBN) [15], for capturing speaker-dependent features from a dysarthric speech signal.

3.1 Convolutive Bottleneck Network

A CBN consists of an input layer, convolution layer and pooling layer pairs, fully-connected MLPs (multi-layer perceptrons) with a bottleneck structure, and an output layer in the order shown in Figure 2. In our approach, the CBN feeds a mel map (two-dimensional acoustic features in time-mel-frequency) and outputs 54 phoneme labels. The MLP shown in Figure 2 stacks three layers (m1 , m2 , m3 ), where we give 108 units, 30 bottleneck units, and 108 units in each layer, respectively. The filter sizes in the convolution layer and the pooling layer will be discussed in the experimental section. Since the bottleneck layer has reduced the number of units for the adjacent layers, we can expect that each unit in the bottleneck layer aggregates information and behaves as a compact feature descriptor that represents an input with a small number of bases, similar to other feature descriptors, such as MFCC, LDA (linear discriminant analysis) or PCA (principal component analysis). It is reported that a feature extraction method using bottleneck features improved speech recognition accuracy from a well-known ABN (American Broad News) corpus [17], and in this paper as well, we use such features (the unit values in the bottleneck layer) for speech recognition, instead of using MFCC. The extracted features are obtained from the statistically-trained speaker-dependent CBN; hence, it is expected that it better represents characteristics in the speech of the target articulation disordered speech than MFCC does.

(7)

3.2 Bottleneck feature extraction from dysarthric speech

Figure 3 shows a system flowchart of our method, where speaker-specific features are extracted using a CBN. First, we prepare the input features for training a CBN from a speech signal uttered by person with an articulation disorder. After calculating short-term mel spectra from the signal, we obtain mel maps by dividing the mel spectra into segments with several frames (13 frames in our experiments) allowing overlaps. For the output units of the CBN, we use phoneme labels that correspond to the input mel-map. For example, when we have a mel map with the label /i/, only the unit corresponding to the label /i/ is set to 1, and the others are set to 0 in the output layer. The label data is obtained by forced alignment using a hidden Marcov model (HMM) from the speech data. The parameters of the CBN are trained by back-propagation with stochastic gradient descent, starting from random values. The bottleneck (BN) features in the trained CBN are then used in the training of another HMM for speech recognition.

In the test stage (“Feature extraction” in Figure 3), we extract features using the CBN, which feeds the mel maps obtained from test data and tries to produce the appropriate phoneme labels in the output layer. Again, note that we do not use the output (estimated) labels for the following procedure, but we use the BN features in the middle layer, where it is considered that information in the input data is aggregated. Finally, the system recognizes dysarthric speech by feeding the extracted BN features into HMMs.

(8)

4 EXPERIMENTAL EVALUATION 4.1 Word recognition experiments (speaker A)

In this section, we discuss and evaluate our experiments of word recognition using a CBN presented in the previous section. Since we want to investigate the effectiveness of the CBN, we did not use dropout in the output layer of the CBN in these experiments.

4.1.1 Conditions

Our feature extraction method was evaluated on word recognition tasks for one male person (identified with “speaker A”) with an articulation disorder. We recorded 216 words included in the ATR Japanese speech database [18] repeating each word five times (Figure 4). The utterance signal was sampled at 16 kHz and windowed with a 25-msec Hamming window every 10 msec. Then we clipped each utterance manually. In our experiments, the first utterances of each word were used for evaluation, and the other utterances (the 2nd through 5th utterances) were used for the training of both a CBN and an acoustic model. We used the HMMs (54 context-independent phonemes) with 5 states and 8 mixtures of Gaussians for the acoustic model.

We trained and evaluated three CBNs: 28 units, 30 units, and 32 units in the bottleneck (BN) layer. The extracted BN features from each CBN were compared with 28-, 30-, and 32-dimensional MFCC+∆MFCC, respectively.

Figure 2: Example of recorded speech data 4.1.2 Training of CBN

In this section, we explain the training conditions of the CBN in detail. We trained the network using pairs of a mel map from the 2nd through 5th utterances and the label, as shown in Figure 3.

Each value of the convolutive filter (and the weight ) is initialized by [19] as

follows:

,

(9)

where denotes uniform distribution at the interval from to . and indicate the numbers of input dimensions and output dimensions at the th layer, respectively. The bias parameters and were set to 0 as initial values.

These parameters were trained so as to minimize the errors between the target labels and the output values using a back-propagation algorithm. We iterated batch-based training with 50 frames in a mini-batch 100 times, with a fixed learning rate of 0.1.

4.1.3 CBN architectures

In this section, we discuss our preliminary experiment in which we change the architecture of a CBN (such as the number of feature maps and the size of a convolution filter) as shown in Table 1. In this preliminary experiment, we will see which architecture produced the best recognition accuracy. All the architectures have two pairs of a convolution layer and a pooling layer followed by three-layer MLPs with a bottleneck layer, forming a nine-layer network in total. We used 108, 30 and 108 units in the MLP part, in this order, for all the architectures (the bottleneck feature had 30 dimensions in this preliminary experiment). For the input layer, we used a 39-dimensional mel map with 13 frames without overlapping. The size of the map was 39×13. When we use ‘Arc1’ in Table 1, for example, the feature maps in each convolution and pooling layer have 36×12, 12×4, 9×3, and 3×1 sizes, in this order. In this case, 3×27(= 81) units are fully connected to the first layer of the MLP part.

Table 2 shows recognition accuracies obtained from each architecture. As shown in Table 2, we obtained the best word recognition accuracies from ‘Arc1’, although it did not always outperform the other architectures with respect to the MSE and label classification accuracy. This is considered to be due to the fact that the extracted bottleneck features of ‘Arc1’ were more abstract and suited to the acoustic model in word recognition. For the remaining experiments, we used ‘Arc1’ for the CBN architecture except for the number of units in the bottleneck layer.

Table 1: Filter size and number of feature maps for each architecture. The values for C1 (and C2) indicate filter size of the first (and the second) convolution layer that has #1 maps (and #2 maps), respectively. Each convolution layer is associated with the pooling layer (S1 and S2). The values for S1 and S2 mean the pooling

(10)

Arc5 8×6 2 13 8×2 3 27

Table 2: Word recognition accuracies using the bottleneck features (Word-Acc.) obtained from each architecture, along with the mean squared error of the closed data (MSE), and the open classification accuracies

of the phoneme labels using a CBN only (without using the acoustic models) (Phoneme-Acc.)

Arc1 Arc2 Arc3 Arc4 Arc5

MSE ( ) 2.42 2.01 2.32 2.43 2.09

Phoneme-Acc. (%) 48.7 49.3 47.9 49.4 49.5

Word-Acc. (%) 88.0 87.7 82.4 84.3 83.8

4.1.4 Results and discussion

Figure 5 shows the word recognition accuracies comparing our CBN features with the conventional MFCC features, when changing the number of feature dimensions. As shown in Figure 5, the use of bottleneck features in a convolutive network improved the accuracies from 84.3% to 88.0% (30 dimensions). This is due to the robustness of the CBN features to small local fluctuations in a time-melfrequency map, caused by the dysarthric speech.

(11)

Figure 5: Word recognition accuracies (%) for the utterances of “speaker A” using bottleneck features in a convolutional network

Table 3: Word recognition accuracies as the number of convolution layers changed.

# of conv v.s. full layers Acc. (%)

No conv, 5 full (DBN) 83.3 1 conv, 4 full (CBN) 84.7 2 conv, 3 full (CBN) 88.0 Baseline (MFCCs) 84.3

4.2 Word recognition experiments (speaker B)

We confirmed improvement of recognition accuracies by using a CBN feature extraction method in the previous experiments. However, the symptom of the articulation disorder and the tendency of the fluctuation in dysarthric speech vary on the speaker. In this section, we show experimental results using speech uttered by another person (female; “speaker B”) with an articulation disorder.

4.2.1 Conditions

(12)

4.2.2 Results and discussion

Figure 6: Word recognition accuracies (%) for the utterances of “speaker B” using bottleneck features in a convolutional network

(13)

Figure 7: Example of forced-alignment for “speaker A”

(14)

4.2.3 Dropout effect

To solve the problem discussed in the previous subsection, we employ a dropout technique [16] when training the CBN, in this paper. The dropout gives the network generalization capabilities by ignoring some randomly-selected units (setting the weights to zero) in a layer. We expect that the network with dropout in the output layer (training signal) prevents over-fitting to the label data that is unreliable because of forced-alignment. Specifically, given an

input vector , the output of a network is obtained by applying a binary mask as

follows:

(4) where denotes element-wise product, and each value of is randomly set to 1 with a probability (0 with a probability ). When each unit in the output layer randomly takes the value of 0 even when the value is activated, the network tries to capture abstracts hidden in

the data. In our experiments, we used .

Figure 9 shows experimental results, adding the result of a CBN with dropout from Figure 6. As shown in Figure 9, the accuracy was highly improved when we apply dropout. Compared with the baseline (MFCC+∆MFCC), the dropout method (CBN+DR) achieved an up-to-5-point improvement. Through these experiments, “ambiguous” training using dropout worked well especially for the speech where phonemes fluctuate largely and it is difficult to prepare the exact labels, as the speech uttered by “speaker B”.

Figure 9: Word recognition accuracies (%) for the utterances of “speaker B” using bottleneck features and dropout (DR)

5 CONCLUSIONS

(15)

extraction method using a convolutive bottleneck network (CBN). In word recognition experiments, our method achieved an approximately 4-point improvement compared with the conventional MFCC features for a male speaker (“speaker A”). For the another speaker (“speaker B”), we could not obtain better performance from the same technique as “speaker A”, because the speech fluctuates largely and it was difficult to estimate boundaries of phonemes correctly by forced-alignment. However, when we realize ambiguous training using dropout, our method achieved a 5-point improvement compared with MFCC features for “speaker B”.

Since the tendency of the fluctuation in dysarthric speech depends on the speaker, we would like to apply and investigate our method to more persons with speech disorders in the future.

REFERENCES

[1]. S. Cox and S. Dasmahapatra, “High-Level Approaches to Confidence Estimation in Speech Recognition,” IEEE Trans. on SAP, vol. 10, No. 7, pp. 460-471, 2002.

[2]. M. K. Bashar, T. Matsumoto, Y. Takeuchi, H. Kudo and N. Ohnishi, “Unsupervised Texture Segmentation via Wavelet-based Locally Orderless Images (WLOIs) and SOM,” 6th IASTED International Conference COMPUTER GRAPHICS AND IMAGING, 2003.

[3]. T. Ohsuga and Y. Horiuchi and A. Ichikawa, “Estimating Syntactic Structure from Prosody in Japanese Speech,” IEICE Trans. on Information and Systems, 86(3), pp. 558―564, 2003.

[4]. K. Nakamura and T. Toda and H. Saruwatari and K. Shikano, “Speaking Aid System for Total Laryngectomees Using Voice Conversion of Body Transmitted Artificial Speech,” Interspeech 2006, pp. 1395―1398, 2006. [5]. D. Giuliani and M. Gerosa, “Investigating recognition of children’s speech,” ICASSP2003, pp. 137―140, 2003. [6]. S. T. Canale and W. C. Campbell, “Campbell ’s Operative Orthopaedics,” Mosby-Year Book, 2002.

[7]. H. Matsumasa, T. Takiguchi, Y. Ariki, I. LI, T. Nakabayashi, “Integration of Metamodel and Acoustic Model for Speech Recognition,” Interspeech 2008, pp.2234-2237, 2008.

[8]. Y. Lecun et al., “Gradient-based learning applied to document recognition,” Proc. IEEE, pp. 2278-2324, 1998. [9]. H. Lee et al., “Unsupervised feature learning for audio classification using convolutional deep belief

networks,” in Advances in Neural Information Processing Systems 22, pp. 1096-1104, 2004.

[10]. C. Garcia and M. Delakis, “Convolutional Face Finder: A Neural Architecture for Fast and Robust Face Detection,” Pattern Analysis and Machine Intelligence, 2004.

(16)

[12]. R. Hadsell et el., “Learning long-range vision for autonomous off-road driving,” Journal of Field Robotics, 2009.

[13]. G. Montavon, “Deep learning for spoken language identification,” NIPS Workshop on Deep Learning for Speech Recognition and Related Applications, 2009.

[14]. T. Nakashika, C. Garcia, and T. Takiguchi, “Local-feature-map Integration Using Convolutional Neural Networks for Music Genre Classication,” Interspeech 2012, 2012.

[15]. K. Vesely et al. “Convolutive bottleneck network features for LVCSR,” in ASRU, pp. 42-47, 2011.

[16]. C. Plahl, R. Schly ̈ter, and H. Ney, “Improving neural networks by preventing co-adaptation of feature detectors,” CoPR, 2012.

[17]. C. Plahl et al., “Hierarchical bottle neck features for LVCSR,” Interspeech 2010, pp. 1197-1200, 2010.

[18]. A. Kurematsu et al., “ATR japanese speech database as a tool of speech recognition and synthesis,” Speech Communication, No.4, pp.357–363, 1990.

Références

Documents relatifs

By applying the lattice decomposition algorithm presented above, all words and sub-words in different lattices are decomposed into a unique sub-word set.. Then, we

In order to merge these primitive components into characters and to find optimum character segmentation of a word, dynamic programming (DP) is applied using the total likelihood

To achieve such a goal, two categories of features are considered: high level features including, global, histogram and segmenta- tion features are extracted from word segments

This paper presents a word extraction approach based on the use of a confidence index to limit the total number of segmentation hypotheses in order to further extend our

Four experiments using a lexical decision task showed systematic effects of the anagram relationship between lexical units as well as between prime and target stimuli, even though

This inspires a new model of spoken word recognition that combines time-specific phoneme representations similar to those in TRACE with higher-level representations based on

Our contributions are the following: 1) We apply for the first time the Fully Convolutional Networks (FCN) classifier from TSC on activity recognition in smart homes. 2) We propose

In order to determine the minimum amount of information for written word recognition, correct and incorrect answers of all participants were compared at each