14
Phone-level pronunciation scoring and assessment for interactive language learning S.M. Witt * , S.J. Young Engineering Department, Cambridge University, Trumpington Street, Cambridge CB2 1PZ, UK Received 20 February 1998; received in revised form 26 November 1998; accepted 2 December 1998 Abstract This paper investigates a method of automatic pronunciation scoring for use in computer-assisted language learning (CALL) systems. The method utilises a likelihood-based ‘Goodness of Pronunciation’ (GOP) measure which is ex- tended to include individual thresholds for each phone based on both averaged native confidence scores and on re- jection statistics provided by human judges. Further improvements are obtained by incorporating models of the subjectÕs native language and by augmenting the recognition networks to include expected pronunciation errors. The various GOP measures are assessed using a specially recorded database of non-native speakers which has been an- notated to mark phone-level pronunciation errors. Since pronunciation assessment is highly subjective, a set of four performance measures has been designed, each of them measuring dierent aspects of how well computer-derived phone-level scores agree with human scores. These performance measures are used to cross-validate the reference annotations and to assess the basic GOP algorithm and its refinements. The experimental results suggest that a like- lihood-based pronunciation scoring metric can achieve usable performance, especially after applying the various en- hancements. Ó 2000 Elsevier Science B.V. All rights reserved. Zusammenfassung In diesem Artikel wird eine Methode zur automatischen Bewertung der Aussprache innerhalb eines Systems f ur computergest utztes Fremdsprachenlernen vorgestellt, welche anhand eines Wahrscheinlichkeitsmaßes, Goodness of Pronunciation (GOP), einen Aussprachewert f ur jedes Phoneme in einer Außerung berechnet. Liegt ein solcher Auss- prachewert oberhalb eines Schwellwertes, wurde ein Aussprachefehler detektiert. Die Methode wird im folgendem durch individuelle Schwellwerte f ur jedes Phoneme, durch die Einbindung von Modellen der Muttersprache des Fre- mdsprachensch ulers und durch Erweiterung der Erkennungsnetzwerke mit zu erwartenden Aussprachefehlern ver- bessert. Die Evaluation der GOP Methode erfolgt mit Hilfe einer speziell f ur diese Zwecke aufgenommenen Datenbank englischer Sprache mit ausl andischen Akzent, die Phonetikern in Bezug auf Aussprachefehler kommentierten. Da Bewertung von Aussprache h ochst subjektiv ist, sind vier Meßmethoden zur Evaluation verschiedener Aspekte der Ubereinstimmung verschiedener Bewertungen eines Datensatzes entwickelt worden. Die Anwendung dieser Meßmethoden erm oglicht, die Leistung der GOP Methode mit Phonetikern zu vergleichen. Die experimentiellen Er- gebnisse deuten darauf hin, daß eine auf Wahrscheinlichkeitsmetrik zur Aussprachebewertung in der Lage ist, in der Praxis anwendbare Ergebnisse zu liefern; dies gilt insbesondere nach der Anwendung der Verbesserungen. Ó 2000 Elsevier Science B.V. All rights reserved. www.elsevier.nl/locate/specom Speech Communication 30 (2000) 95–108 * Corresponding author. E-mail addresses: [email protected] (S.M. Witt), [email protected] (S.J. Young). 0167-6393/00/$ - see front matter Ó 2000 Elsevier Science B.V. All rights reserved. PII: S 0 1 6 7 - 6 3 9 3 ( 9 9 ) 0 0 0 4 4 - 8

Phone-level pronunciation scoring and assessment for ...mi.eng.cam.ac.uk/research/dialogue/papers/wiyo00.pdfPhone-level pronunciation scoring and assessment for interactive language

  • Upload
    others

  • View
    7

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Phone-level pronunciation scoring and assessment for ...mi.eng.cam.ac.uk/research/dialogue/papers/wiyo00.pdfPhone-level pronunciation scoring and assessment for interactive language

Phone-level pronunciation scoring and assessment forinteractive language learning

S.M. Witt *, S.J. Young

Engineering Department, Cambridge University, Trumpington Street, Cambridge CB2 1PZ, UK

Received 20 February 1998; received in revised form 26 November 1998; accepted 2 December 1998

Abstract

This paper investigates a method of automatic pronunciation scoring for use in computer-assisted language learning

(CALL) systems. The method utilises a likelihood-based `Goodness of Pronunciation' (GOP) measure which is ex-

tended to include individual thresholds for each phone based on both averaged native con®dence scores and on re-

jection statistics provided by human judges. Further improvements are obtained by incorporating models of the

subjectÕs native language and by augmenting the recognition networks to include expected pronunciation errors. The

various GOP measures are assessed using a specially recorded database of non-native speakers which has been an-

notated to mark phone-level pronunciation errors. Since pronunciation assessment is highly subjective, a set of four

performance measures has been designed, each of them measuring di�erent aspects of how well computer-derived

phone-level scores agree with human scores. These performance measures are used to cross-validate the reference

annotations and to assess the basic GOP algorithm and its re®nements. The experimental results suggest that a like-

lihood-based pronunciation scoring metric can achieve usable performance, especially after applying the various en-

hancements. Ó 2000 Elsevier Science B.V. All rights reserved.

Zusammenfassung

In diesem Artikel wird eine Methode zur automatischen Bewertung der Aussprache innerhalb eines Systems f�ur

computergest�utztes Fremdsprachenlernen vorgestellt, welche anhand eines Wahrscheinlichkeitsmaûes, Goodness of

Pronunciation (GOP), einen Aussprachewert f�ur jedes Phoneme in einer �Auûerung berechnet. Liegt ein solcher Auss-

prachewert oberhalb eines Schwellwertes, wurde ein Aussprachefehler detektiert. Die Methode wird im folgendem

durch individuelle Schwellwerte f�ur jedes Phoneme, durch die Einbindung von Modellen der Muttersprache des Fre-

mdsprachensch�ulers und durch Erweiterung der Erkennungsnetzwerke mit zu erwartenden Aussprachefehlern ver-

bessert. Die Evaluation der GOP Methode erfolgt mit Hilfe einer speziell f�ur diese Zwecke aufgenommenen Datenbank

englischer Sprache mit ausl�andischen Akzent, die Phonetikern in Bezug auf Aussprachefehler kommentierten. Da

Bewertung von Aussprache h�ochst subjektiv ist, sind vier Meûmethoden zur Evaluation verschiedener Aspekte der�Ubereinstimmung verschiedener Bewertungen eines Datensatzes entwickelt worden. Die Anwendung dieser

Meûmethoden erm�oglicht, die Leistung der GOP Methode mit Phonetikern zu vergleichen. Die experimentiellen Er-

gebnisse deuten darauf hin, daû eine auf Wahrscheinlichkeitsmetrik zur Aussprachebewertung in der Lage ist, in der

Praxis anwendbare Ergebnisse zu liefern; dies gilt insbesondere nach der Anwendung der Verbesserungen. Ó 2000

Elsevier Science B.V. All rights reserved.

www.elsevier.nl/locate/specomSpeech Communication 30 (2000) 95±108

* Corresponding author.

E-mail addresses: [email protected] (S.M. Witt), [email protected] (S.J. Young).

0167-6393/00/$ - see front matter Ó 2000 Elsevier Science B.V. All rights reserved.

PII: S 0 1 6 7 - 6 3 9 3 ( 9 9 ) 0 0 0 4 4 - 8

Page 2: Phone-level pronunciation scoring and assessment for ...mi.eng.cam.ac.uk/research/dialogue/papers/wiyo00.pdfPhone-level pronunciation scoring and assessment for interactive language

Keywords: Speech recognition; Computer-assisted language learning; Pronunciation assessment; Pronunciation teaching

1. Introduction

Computer-assisted language learning (CALL)systems can provide many potential bene®ts to boththe language learner and teacher. They allow con-tinuous feedback to the student without requiringthe sole attention of the teacher, they facilitate self-study and encourage interactive use of the languagein preference to rote-learning. Finally, they can beused to streamline assessment procedures.

To be e�ective, a CALL system requires theability to accurately measure pronunciation qual-ity both to enable the immediate correction of er-rors and to provide longer term feedback onoverall language competence. The aim of the workdescribed in this paper is to study acoustic likeli-hood-based methods for automatic pronunciationassessment within the framework of a hiddenMarkov model (HMM) speech recognition system.

Existing work on automatic pronunciationscoring has mainly been focussed on the word andphrase level, possibly augmented by measures ofintonation, stress and rhythm (Goddijn and deKrom, 1997; Hiller et al., 1993; Hamada et al.,1993; Rogers et al., 1994). These systems typicallyrequire several recordings of native utterances totrain the models for each word in the teachingmaterial. They are therefore text-dependent withthe disadvantage that the teaching material cannotbe adjusted without making additional recordings.Systems aimed at teaching selected phonemic er-rors are described in (Kawai and Hirose, 1997;Kim et al., 1997; Ronen et al., 1997), where eitherdurational information or models trained on non-native speech have been employed. Automaticspeech recognition with HMMs has been used toscore complete sentences rather than smaller unitsof speech (Bernstein et al., 1990; Neumeyer et al.,1996). The system described by (Eskenazi, 1996)produces scores for each phone 1 in an utterance,

but there is no attempt to relate this to humanjudgements of pronunciation quality. A dialogsystem aimed at teaching spoken Japanese is pre-sented in (Ehsani et al., 1997) where speech rec-ognition is used to analyse the studentÕs answer ateach stage of the dialog.

The system described here is focussed on mea-suring pronunciation quality of non-native speechat the phone level. The aims are to locate pro-nunciation errors, to assess how close the pro-nunciation is to that of a native speaker and toidentify systematic di�erences when compared to apronunciation dictionary.

The remainder of this paper is organised asfollows. In Section 2, the basic Goodness of Pro-nunciation (GOP) scoring algorithm is describedalong with a number of re®nements. Section 3 thenpresents a set of four performance measures whichcan be used both to validate pronunciation as-sessments made by human judges and to assess theperformance of the GOP algorithms. Section 4describes the non-native database which was spe-cially recorded for this work. Finally, Sections 5and 6 present performance assessments of thehuman judges who annotated the database and theautomatic GOP algorithms. The paper concludeswith a discussion of the results and some com-ments on future directions.

2. GOP scoring

2.1. Basic GOP algorithm

The aim of the GOP measure is to provide ascore for each phone of an utterance. In comput-ing this score it is assumed that the orthographictranscription is known and that a set of HMMs isavailable to determine the likelihood p�O�q�jq� ofthe acoustic segment O�q� corresponding to eachphone q. Under these assumptions, the quality ofpronunciation for any phone p is de®ned to bethe duration normalised log of the posterior

1 Throughout this paper a ``phone'' denotes a sound unit used

to model speech with HMMs, which roughly corresponds to a

phoneme as de®ned by linguists.

96 S.M. Witt, S.J. Young / Speech Communication 30 (2000) 95±108

Page 3: Phone-level pronunciation scoring and assessment for ...mi.eng.cam.ac.uk/research/dialogue/papers/wiyo00.pdfPhone-level pronunciation scoring and assessment for interactive language

probability P �pjO�p�� that the speaker utteredphone p given the corresponding acoustic segmentO�p�. That is,

GOP1�p� � log P�pjO�p��ÿ ��� ��=NF�p�; �1�

� logp�O�p�jp�P �p�P

q2Q p�O�p�jq�P�q�

!����������,

NF�p�;

�2�where Q is the set of all phone models and NF(p)the number of frames in the acoustic segment O�p�.

Assuming all phones are equally likely(P(p)�P(q)) and that the sum in the denominatorcan be approximated by its maximum, the basicGOP measure becomes

GOP1�p� � logp�O�p�jp�

maxq2Q p�O�p�jq�� ��������� �

NF�p�:

�3�The acoustic segment boundaries and the corre-sponding likelihoods are determined from Viterbialignments. Firstly, the numerator of Eq. (3) iscomputed using a forced alignment in which thesequence of phone models is ®xed by the knowntranscription and secondly, the denominator isdetermined using an unconstrained phone loop.This is the same arrangement as is commonly usedin word spotting (Knill and Young, 1994). Onedi�culty in Eq. (3) is that if a mispronunciation

has occurred, it is not reasonable to constrain theacoustic segment used to compute the maximumlikelihood phone q to be identical to the assumedphone O�p�. Hence, the denominator score is de-termined by simply summing the log likelihood perframe over the duration of O�p�. In practice, thiswill often mean that more than one phone in theunconstrained phone sequence has contributed tothe computation of maxq2Q p�O�p�jq�.

A block diagram of the resulting scoringmechanism is shown in Fig. 1. The front-end fea-ture extraction converts the speech waveform to asequence of mel-frequency cepstral coe�cients(MFCC) and these are used in two recognitionpasses: the forced alignment pass and the phonerecognition pass where each phone can follow theprevious one with equal probability. Based onthese results, the individual GOP scores are cal-culated for each phone as de®ned in the previousequations. Finally, a threshold is applied to eachGOP score to reject badly pronounced phones.The choice of threshold depends on the level ofstrictness required. The selection of suitablethresholds is further discussed in Section 6.

The quality of the GOP scoring procedure de-scribed above depends on the quality of theacoustic models used. Since the aim of the GOPmeasure is to assess pronunciation quality withrespect to native speaker performance, it is

Fig. 1. Block-diagram of the pronunciation scoring system: phones whose scores are above the prede®ned threshold are assumed to be

badly pronounced and are therefore rejected.

S.M. Witt, S.J. Young / Speech Communication 30 (2000) 95±108 97

Page 4: Phone-level pronunciation scoring and assessment for ...mi.eng.cam.ac.uk/research/dialogue/papers/wiyo00.pdfPhone-level pronunciation scoring and assessment for interactive language

reasonable to use native speakers to train theacoustic models. However, non-native speech ischaracterised by di�erent formant structures com-pared to those from a native speaker for the samephones (Arslan and Hansen, 1997) and this canlead to phone recognition errors. Hence, some de-gree of speaker adaptation may be justi®ed. To testthis hypothesis, the GOP measure can be computedusing models whose Gaussian means have beenadapted using Maximum Likelihood Linear Re-gression (MLLR) (Leggetter and Woodland,1994). In order to achieve speaker normalisationwithout adapting to speci®c phone error patterns,this adaptation is limited to a single global trans-form of the HMM mixture component means.

2.2. Phone dependent thresholds

So far a single threshold for all phones has beenassumed. However, in practice, the acoustic ®t ofphone-based HMMs di�ers from phone to phone.For example, fricatives tend to have lower loglikelihoods than vowels suggesting that a higherthreshold should be used for these.

A simple phone-speci®c threshold can be com-puted from the global GOP statistics. For exam-ple, the threshold for a phone p can be de®ned interms of the mean lp and variance rp of all theGOP scores for phone p in the training data,

Tp1 � lp � arp � b; �4�where a and b are empirically determined scalingconstants. The assumption here is that averagingthe native GOP scores will reduce the a�ect oferrors in the phone recogniser.

A reasonable target for an automatic pronun-ciation system is to perform as well as a humanjudge. One way to approximate human perfor-mance is to learn from human labelling behaviour.Let cn(p) be the total number of times that phone puttered by speaker n was marked as mispro-nounced by one of the human judges in thetraining database. Then a second phone dependentthreshold can be de®ned by averaging the nor-malised rejection counts over all speakers,

Tp2� log

1

N

XN

n�1

cn�p�XM

m�1

cn�m�, !

; �5�

where M is the total number of distinct phones andN the total number of speakers in the training set.

2.3. Explicit error modelling

Pronunciation errors can be grouped into twomain error classes. The ®rst class contains indi-vidual mispronunciations when a student is notfamiliar with the pronunciation of a speci®c word.The second class consists of substitutions of nativesounds for sounds of the target language, which donot exist in the native language. This error typewill be called systematic mispronunciations. Be-cause the GOP method described so far does notemploy models of a studentÕs native phones, in-correct acoustic modelling of the non-nativespeech will occur especially in the case of system-atic mispronunciations. The detection of these er-rors might be improved if knowledge of the nativetongue of the learner can be included in the GOPscoring.

For this purpose a recognition network hasbeen implemented incorporating both correctpronunciation and common pronunciation errorsin the form of error sublattices for each phone,using phone model sets of both the target and thenative language. Concatenating these sublatticesaccording to target transcriptions yields the de-sired error network for any utterance. For exam-ple, Fig. 2 shows the resulting network for theword ``but''. The list of possible errors of aSpanish speaker learning English has been takenfrom (Kenworthy, 1987), some examples of which

Fig. 2. Example of error network for the word `but', created

through concatenating the sublattice of possible errors for each

phone, the topmost phones correspond to the target tran-

scription. (Phoneme names with subscript `s' denote Spanish

models.)

98 S.M. Witt, S.J. Young / Speech Communication 30 (2000) 95±108

Page 5: Phone-level pronunciation scoring and assessment for ...mi.eng.cam.ac.uk/research/dialogue/papers/wiyo00.pdfPhone-level pronunciation scoring and assessment for interactive language

are listed in Table 1. The recognition output ofsuch an error network will be a sequence of phoneseither corresponding to the target transcriptionp� pt in the case that the target pronunciation wasmore likely or otherwise to an error phone p� pe.

A straightforward detector of systematic mis-pronunciations based on an error network couldconsist of rejecting all phone segments where anerror phone has been recognised. However, suchan approach would ignore the information aboutthe likelihood of the occurrence of such an error.Hence, the posterior likelihood of each errorphone P �pejO�p�� is computed by normalising withthe recognition results of a phone loop networkincluding acoustic models of both the target lan-guage and the speaker's native language applyingEq. (3).

Knowledge of P �pejO�p�� permits calculating theposterior probability of the target phones pt in allphone segments containing systematic mispro-nunciations:

P �ptjO�p�� � 1ÿXq 6�pt

P �qjO�p��

� 1ÿmaxq6�pt

P �qjO�p��

� 1ÿ P �pejO�p��: �6�

Again the assumption has been made that theabove sum can be approximated by its maximum.Thus, scores for systematic mispronunciationsGOPe(p) are de®ned as

GOPe�p� � j log�1ÿ P�pejO�p���j if p � pe;0:0 otherwise:

��7�

Combining the basic GOP1 with GOPe yields asecond GOP metric which includes additionalpenalties for scores of phone segments where sys-tematic error was recognised.

GOP2�p� � GOP1�p� � K GOPe�p�; �8�where K is a scaling constant.

3. Performance measures

In order to assess the e�ectiveness of the GOPscoring for detecting pronunciation errors, a set offour new performance measures has been designed.These are based on similarity measurements be-tween reference transcriptions produced by humanjudges and the output of the GOP metric. Since theproduction of reference transcriptions must bedone by human judges and is highly subjective, thesame performance measures are also used to cross-validate the judges. Note that the performancemeasures are only concerned with the detection ofpronunciation errors. They do not take account ofthe type of error which has occurred.

To cover all aspects of performance, four dif-ferent dimensions are considered.· Strictness ± how strict was the judge in marking

pronunciation errors?· Agreement ± what is the overall agreement be-

tween the reference transcription and the auto-matically derived transcription? This measuretakes account of all phones whether mispro-nounced or not.

· Cross-correlation ± what is the overall agreementbetween the errors marked in the reference andthe automatically detected errors? This measureonly takes account of phones for which an errorhas been marked in one or both transcriptions.

· Overall phone correlation ± how well do the over-all rejection statistics for each phone correlatebetween the reference and the automatic system?

The next section describes the form of the errortranscriptions in more detail and then the mea-sures themselves are de®ned.

Table 1

Expected errors of a Spanish speaker for some British-English

phones (Phone names with subscript `s' denote Spanish models)

British Expected errors

b b, v, b_s

d dh, d_s

th f, s, f_s, s_s

s hh, del

ch ch_s

jh ch,ch_s

k del, k_s

l l_s

ah a_s, aw, ae

uh uw, u_s

ae eh, e_s

oh o_s

S.M. Witt, S.J. Young / Speech Communication 30 (2000) 95±108 99

Page 6: Phone-level pronunciation scoring and assessment for ...mi.eng.cam.ac.uk/research/dialogue/papers/wiyo00.pdfPhone-level pronunciation scoring and assessment for interactive language

3.1. The transcription of pronunciation errors

The non-native database used for assessmentconsists of target transcriptions based on a pro-nunciation dictionary and transcriptions whichhave been annotated by human judges to containthe phone sequence actually spoken. The utterancetranscriptions marked with corrections will be re-ferred to as corrected transcriptions and the tran-scriptions derived directly from the pronunciationdictionary will be referred to as the dictionary-based transcriptions. Finally, transcriptions inwhich each phone correction has been replaced bya single rejection symbol are referred to as rejec-tion-marked transcriptions.

Two corrected transcriptions of the same utter-ance are di�cult to align with each other due toinsertions and deletions of phones. Therefore, allperformance measures compare transcriptions on aframe by frame basis. With this approach, mea-suring the similarity of two di�erently correctedtranscriptions of the same utterance becomesequivalent to comparing the rejection/acceptancemarking of corresponding speech frames.

Based on the rejection-marked transcriptions,the frame level markings are calculated as follows:1. The phone level segmentation for each sentence

is calculated by forced alignment of the acousticwaveform with the corrected transcriptions.

2. All frames corresponding to substituted, insert-ed or deleted phones are marked with ``1'', allother ones with ``0''. This yields a vector x oflength N with x�i� 2 f0; 1g. These vectors willbe called transcription vectors.

3. The transitions between ``0'' and ``1'' in thetranscription vectors are abrupt whereas inpractice the precise location of the boundariesbetween correctly and incorrectly pronouncedspeech segments are uncertain. Moreover,segmentation based on forced alignmentscan be erroneous due to the poor acousticmodelling of non-native speech. For thesetwo reasons, the vectors representing correct-ed transcriptions are smoothed by a Ham-ming window

x0�n� �XN=2

k�ÿN=2

x�k�w�nÿ k�: �9�

Using a frame period of 10 ms, the length of avowel tends to extend over 6±20 frames, whereasconsonants can be much shorter. Also, if rejectedframes in one transcription are immediately fol-lowed by rejected frames in the other transcription,the rejections can be considered to have beencaused by the same pronunciation error. Based onthese considerations, a window length of N� 15was selected for all experiments. The e�ect of thesmoothing window is illustrated in Fig. 3.

3.2. Performance measures

This section de®nes the four performancemeasures used to compare transcriptions correctedby two judges or one judge and the automaticGOP scoring system.

Firstly, human correction of the pronunciationof non-native speakers depends on personaljudgement. There will always exist a large numberof phones whose pronunciation is on the border-line between correct and incorrect, a stricter judgemight declare more borderline cases as incorrectthan another judge who is more benign. In the caseof computer-based scoring, the choice of a rejec-tion threshold determines how strict the scoringsystem will be. This strictness of labelling, S, can bede®ned as the overall fraction of phones which arerejected,

Fig. 3. Smoothing e�ect of the windowing, overlapping regions

denote areas where both judges decided to reject the pronun-

ciation.

100 S.M. Witt, S.J. Young / Speech Communication 30 (2000) 95±108

Page 7: Phone-level pronunciation scoring and assessment for ...mi.eng.cam.ac.uk/research/dialogue/papers/wiyo00.pdfPhone-level pronunciation scoring and assessment for interactive language

S � Count of rejected phonemes

Total count of phonemes: �10�

As an example, the database used for assessment(described below) contains a set of calibrationsentences which were labelled by six di�erentjudges. Fig. 4 shows the strictness of the judges forthese calibration sentences where the mean andstandard deviation are lS � 0.18 and rS � 0.05,respectively.

A simple way to compare the strictness of twojudges J1 and J2 is to use the di�erence betweenstrictness levels for the two, i.e.,

dS � jSJ1 ÿ SJ2j: �11�The overall Agreement (A) between two utterancesis de®ned in terms of the cityblock distance be-tween the corresponding transcription vectors, i.e.,

AJ1;J2 � 1ÿ 1

NkxJ1 ÿ xJ2kC; �12�

where kxkC �PNÿ1

i�0 jx�i�j:Agreement measures overall similarity of two

transcriptions by comparing all frames of an ut-terance. In contrast, the Cross-Correlation (CC)measure takes into account only those frameswhere there exists a rejection in either of them,

CCJ1;J2 � xTJ1xJ2

kxJ1kEkxJ2kE

; �13�

where kxkE �����������������������PNÿ1

i�0 x�i�2q

is the standard Eu-clidean distance. Cross-correlation measures thesimilarity between all segments which contain re-

jections in either of the two transcriptions. Becausesimilarity of the rejection patterns with a humanjudge is the main design objective of the GOPscoring system, this measure has the highestimportance.

Finally, Phoneme Correlation (PC) measures theoverall similarity of the phone rejection statistics.Given a vector c whose elements contain the countof rejections for each phone in a complete speakerset, phone correlation is de®ned as

PCJ1;J2 �PM

m�0�cJ1�m� ÿ lcJ1��cJ2�m� ÿ lcJ2�PMm�0

�����������������������������������������������������������������cJ1�m� ÿ lcJ2�2�cJ1�m� ÿ lcJ2�2

q ;

�14�

where lc denotes the mean rejection counts.

4. Collection of a non-native database

In order to evaluate the pronunciation scoringmethods described above, a database of non-na-tive speech from second-language learners hasbeen recorded and annotated.

The recording guidelines for this database col-lection were based on the procedures used for theWSJCAM0 corpus (Fransen et al., 1994). Studentsof English as a second language read promptingtexts composed of a limited vocabulary of 1500words in a quiet room environment. The compe-tence level of the speakers was intermediate. Theywere all able to understand the prompting textsand instructions and they were able to read thesentences with relatively few hesitations. On theother hand, their competence level was low enoughto ensure that they produced a signi®cant numberof easily detectable mispronunciations.

Each prompting session consisted of a commonset of 40 phonetically balanced sentences read byall subjects and an additional 80 sentences whichvaried from session to session. Extracts from``Penguin Readers'' (Fine, 1995; Chandler, 1991)were used as the source of the prompting texts.These texts have been speci®cally written for thepurpose of teaching English as a foreign language.They employ a limited vocabulary and simpli®edsentence and grammatical structures.

Fig. 4. Relative strictness for all human judges measured on the

calibration sentences.

S.M. Witt, S.J. Young / Speech Communication 30 (2000) 95±108 101

Page 8: Phone-level pronunciation scoring and assessment for ...mi.eng.cam.ac.uk/research/dialogue/papers/wiyo00.pdfPhone-level pronunciation scoring and assessment for interactive language

The subjects for the database collection con-sisted of 10 students, six of them female and fourof them male, speaking as their mother-tonguesLatin-American Spanish, Italian, Japanese andKorean. Additionally, 20 sentences from a femaleSpanish speaker were recorded to serve as cali-bration sentences. These were annotated by all sixjudges participating in the labelling task. The re-sulting six sets of calibration transcriptions canthus be used to compare the human labellingconsistency.

The database was annotated by trained pho-neticians (the ``judges''), all of them native speak-ers of British English. This annotation wasperformed at three di�erent levels. Firstly, theoriginal transcriptions were annotated with allsubstitution, deletion and insertion errors made bythe non-native speaker. Since the non-nativespeech contained a range of sounds which do notexists in British English, the judges had the op-portunity to extend the supplied British Englishphone set with phones taken from the speaker'snative language. These labelling instructionsyielded transcriptions resembling the non-nativespeech as closely as possible. Secondly, each wordwas scored on a scale of 1±4, with 1 representingbarely intelligible speech and 4 representing native-like pronunciation. Finally, each sentence wasscored on the same scale. Of these three levels of

annotation, only the ®rst phone error correctionlevel is used for the experiments reported here.

5. The labelling consistency of the human judges

In order to properly interpret the results of as-sessing a computer-based pronunciation systemusing manually derived transcriptions as the ref-erence, it is necessary to measure the inter-judgelabelling consistency and to obtain an under-standing of how the judges label the data. Theirlabelling is characterised by the phones they con-sider important for good pronunciation and thustend to correct, by the consistency of the rejectionpatterns across di�erent judges and ®nally by theirstrictness. In this section, the four performancemeasures described above are used in conjunctionwith the 20 calibration sentences to determinethese characteristics.

Fig. 5 shows averaged results of all the mea-sures for each judge. These results have been cal-culated by averaging A, CC, PC and dS betweenthe respective judge and all other ones. All resultsvary within an acceptable range, that is0.85 < A < 0.95, 0.40 < CC < 0.65, 0.70 < PC < 0.85and 0.03 < dS < 0.14. Therefore, the labelling bydi�erent human judges can be considered as beingreasonably consistent although Judge 5 is a slight

Fig. 5. A, CC, PC and dS for each judge based on averaging the measures between the respective judge and all the other judges.

102 S.M. Witt, S.J. Young / Speech Communication 30 (2000) 95±108

Page 9: Phone-level pronunciation scoring and assessment for ...mi.eng.cam.ac.uk/research/dialogue/papers/wiyo00.pdfPhone-level pronunciation scoring and assessment for interactive language

outlier in that he has a lower average cross-cor-relation with the other judges. The total meanvalues over all pairs of judges of all four measuresare shown in Table 2. These mean values will beused as benchmark values against which to mea-sure the performance of the automatic scoringpresented in later sections.

Table 3 shows the similarity between the humanjudges and the baseline GOP scoring method foreach non-native speaker in that judge's group. Itcan be seen that the intra-judge results are quiteconsistent. However, Judge 4 had a high accep-tance of non-native pronunciation, and thus cor-rected a signi®cantly smaller portion of the data.For such a degree of strictness, the automaticscoring performs considerably worse.

Fig. 6 shows the CC and PC measures for eachspeaker grouped according to their native lan-guages. Also shown on this ®gure are the gendersof each speaker. From this ®gure and the datashown in Fig. 5, it appears that the labelling of the

human judges does not depend signi®cantly on themother tongue or the gender of the subjects, butdepends mostly on the variability of human judges.

Finally, the rejection patterns for three of thejudges are shown in Fig. 7, which depicts the re-jection counts for all phones for the judges. Thestrong correlation between the rejection pattern ofthese three judges is clearly evident.

The above analysis of human judgement char-acteristics shows that although there is signi®cantvariability in the labelling of each judge, there isnevertheless su�cient common ground to form abasis for assessing the performance of the variousautomatically derived pronunciation metrics.

6. Experimental results

This section presents performance resultsfor both the basic GOP scoring method and the

Table 2

Averaged A, CC, PC and dS results based on correlating all

possible pairs of judges (these values are the baseline against

which to measure automatic scoring performance)

A CC PC dS

0.91 0.47 0.78 0.06

Table 3

Similarity results between judges and the baseline GOP scoring

grouped according to the judge who labelled the respective

speaker sets (the speaker name Cal. denotes the calibration

sentences)

Judge Speaker Strictness CC PC

J1 Cal. 0.25 0.51 0.77

ss 0.25 0.56 0.73

ts 0.21 0.49 0.84

J2 Cal. 0.19 0.53 0.81

yp 0.16 0.49 0.62

J3 Cal. 0.21 0.50 0.68

mk 0.13 0.38 0.57

J4 Cal. 0.13 0.37 0.62

sk 0.07 0.12 0.61

as 0.11 0.37 0.50

J5 Cal. 0.16 0.22 0.71

ay 0.19 0.50 0.61

¯ 0.16 0.43 0.56

pc 0.19 0.50 0.62

ky 0.23 0.48 0.34

Fig. 6. CC and PC results grouped according to each student's

mother-tongue.

S.M. Witt, S.J. Young / Speech Communication 30 (2000) 95±108 103

Page 10: Phone-level pronunciation scoring and assessment for ...mi.eng.cam.ac.uk/research/dialogue/papers/wiyo00.pdfPhone-level pronunciation scoring and assessment for interactive language

various re®nements described in Section 2. Allspeech recognition is based on multiple mixturemonophone models trained on the British Englishcorpus WSJCAM0 (Fransen et al., 1994). TheHMMs were built using the HTK Toolkit (Younget al., 1996).

For the case of automatic GOP scoring, agree-ment A, cross-correlation CC and phone correla-tion PC vary according to the level of strictnessapplied, which again depends on the thresholdlevels set. In Fig. 8, the GOP scores for an examplesentence are shown. Varying the threshold deter-mines the number of rejections. For the markedthreshold in the example ®gure, both human andmachine judgements agree on which phones toaccept and to reject with two exceptions. The ®rst

phone of the sentence is not rejected by the humanjudge but it is rejected by the GOP metric, this isprobably due to bad acoustic modelling at thesentence beginning. Further, the `ae' in `carry' hasbeen rejected by the human judge method but notby the GOP scoring.

In the work reported here, the range of rejectionthresholds studied was restricted to lie within onestandard deviation of the judges strictness i.e.,jdS j6 rS where in this case rS � 0.05. Within thisrange, the variation of A, CC and PC for onespeaker as a function of the threshold level isshown in Fig. 9. In this ®gure, the vertical linesdenote the acceptable range of threshold settingsand, as can be seen, the performance values do notvary greatly within this range.

Fig. 7. Rejection counts of all phones for all judges based on the calibration sentences to show the correlation between the rejection

pattern of di�erent judges.

104 S.M. Witt, S.J. Young / Speech Communication 30 (2000) 95±108

Page 11: Phone-level pronunciation scoring and assessment for ...mi.eng.cam.ac.uk/research/dialogue/papers/wiyo00.pdfPhone-level pronunciation scoring and assessment for interactive language

Table 4 shows optimal values of A, CC and PCachievable for each speaker within the allowedthreshold range. As can be seen, the optimalthreshold is speaker dependent. However, apartfrom speakers `sk', `as', a threshold of 4.5 wouldbe close to optimal for all speakers. Since `sk' and`as' were the two speakers whose transcriptionswere annotated by the very strict judge (Judge 4),these two speakers have not been included in theresults presented in the following.

The performance results for the automatic GOPscoring metrics as discussed in Section 2 are sum-marised in Fig. 10. The ®rst bar on the left marked``Baseline'' shows the performance of the basicGOP 1 metric with a ®xed overall threshold asdiscussed above. The ®nal bar on the left shows thehuman±human performance on the calibration

sentences for comparison. As can be seen, thescores for A and CC are similar whereas for PC,the automatic scoring is worse by about 20%. Thesecond bar marked ``MLLR'' shows the e�ect ofapplying speaker adaptation. For Group 1 animprovement of 5% has been obtained for PC atthe cost of a small decrease in CC. The third andfourth bars show the e�ects of using individualthresholds for each phone based on averagingnative scores Tp1

and on averaging the judgesscores Tp2

. As can be seen, thresholds derived fromthe statistics of the judge's scoring appear to pro-vide the best performance. This is probably be-cause these are directly related to the desiredrejection statistics.

Finally, Table 5 summarises the e�ects ofincorporating error modelling into the GOP

Fig. 8. GOP scoring results for an example sentence, `ss' denotes the location of a rejection, the automatically rejected phones

correspond to GOP scores above the threshold.

S.M. Witt, S.J. Young / Speech Communication 30 (2000) 95±108 105

Page 12: Phone-level pronunciation scoring and assessment for ...mi.eng.cam.ac.uk/research/dialogue/papers/wiyo00.pdfPhone-level pronunciation scoring and assessment for interactive language

algorithm. The British English HMMs were aug-mented by a set of similar Spanish models trainedon a database of Latin-American Spanish. The datafrom the three Spanish speakers in the database wasanalysed using the extended GOP2 metric with thescale factor K adjusted to give optimal performance.Instead of averaging over all eight speakers, resultsshown in this table are only averaged over the threeSpanish speakers in the database, the averagedbaseline performance of which is shown in the ®rstline of Table 5. Comparision of these results withthose for the second GOP metric demonstrate that aslight improvement can be obtained by includingthe extra information relating to systematic mis-pronunciations. Finally, the results of combining allproposed re®nements of the baseline algorithm, i.e.,

adaptation, judge-based individual thresholds anderror modelling, are as high as the human bench-mark values.

Table 6 compares human rejections with allautomatically detected systematic mispronuncia-tions, i.e., all rejections of phone segments where anative phone had been more likely than the targetphone. The relatively high values for CC and PCindicate that a large proportion of pronunciationerrors are due to systematic mispronunciationsand that a signi®cant proportion of these can alsobe detected by the use of error networks. Addi-tionally, this metric provides information aboutwhich type of mispronunciations occurred andwhether the pronunciation of a given phonemesounds more Spanish than English. This

Fig. 9. Dependency of A, CC, PC and dS on threshold variation, based on data for `¯', a male Spanish speaker. The range inside the

bold lines is the range of valid dS .

106 S.M. Witt, S.J. Young / Speech Communication 30 (2000) 95±108

Page 13: Phone-level pronunciation scoring and assessment for ...mi.eng.cam.ac.uk/research/dialogue/papers/wiyo00.pdfPhone-level pronunciation scoring and assessment for interactive language

information might be used in future work to pro-vide additional feedback about error types in ad-dition to detecting error locations within anutterance or a word.

7. Conclusions

This paper has presented a likelihood-basedmethod for GOP scoring and proposed a numberof metrics which can be used to assess perfor-mance in comparison to human judges. Using a

Table 4

Thresholds yielding optimal performance for all non-native speakers of the database

ID Thres A CC PC dS

¯ 5 0.91 0.43 0.56 0.02

pc 4.5 0.87 0.50 0.62 0.04

yp 4.0 0.90 0.49 0.62 0.02

ts 4 0.87 0.49 0.84 0.03

ky 5 084 0.48 0.34 0.04

sk 7 0.90 0.12 0.61 0.06

ss 4.5 0.85 0.56 0.73 0.05

as 7 0.90 0.37 0.50 0.07

mk 4.5 0.90 0.38 0.57 0.07

ay 4.5 0.90 0.50 0.61 0.05

j1 4.5 0.86 0.51 0.77 0.01

j2 5 0.87 0.53 0.81 0.04

j3 4.5 0.86 0.50 0.68 0.05

j4 7.5 0.91 0.43 0.62 0.00

j5 5.5 0.88 0.22 0.71 0.05

j6 4 0.85 0.52 0.65 0.03

GOP mean 0.88 0.47 0.60 0.05

Human mean 0.91 0.47 0.78 0.05

Fig. 10. Comparison of the A, CC and PC performance mea-

sures using: (a) the basic GOP scoring (Baseline); (b) basic GOP

with adaptation (MLLR); (c) individual thresholds based on

native scores (Ind-Nat); (d) individual thresholds based on

human judges (Ind-Jud); (e) human±human average perfor-

mance (Human).

Table 5

Performance for experiments with and without error modelling

(all experiments include adaptation)

Experimental setup A CC PC

Baseline 0.89 0.46 0.71

Ind-Judge 0.89 0.48 0.76

Error modelling 0.88 0.48 0.72

Ind-Judge + Error modelling 0.90 0.49 0.78

Human mean 0.91 0.47 0.78

Table 6

Performance results when using an error network to detect

systematic mispronunciations

Speaker ID A CC PC SID dS

¯ 0.90 0.34 0.42 0.18 0.01

pc 0.88 0.39 0.57 0.23 0.04

ts 0.89 0.27 0.40 0.19 0.02

S.M. Witt, S.J. Young / Speech Communication 30 (2000) 95±108 107

Page 14: Phone-level pronunciation scoring and assessment for ...mi.eng.cam.ac.uk/research/dialogue/papers/wiyo00.pdfPhone-level pronunciation scoring and assessment for interactive language

specially recorded database of non-native speech,the basic GOP method has been investigated andthe e�ectiveness of the performance measuresstudied.

The combination of the baseline method withseveral re®nements yielded improvements in theautomatic scoring performance, which then be-came comparable to the human±human bench-mark values. Applying speaker adaption andindividual thresholds trained on human judge-ments has improved the phone correlation fromPC� 0.62 to 0.72, this being only about 7.7%worse than the averaged human performance ofPC� 0.78. For the Spanish speaker sets in thedatabase, application of the error modelling tech-nique yielded performance as high as the bench-mark values.

In conclusion, this work indicates that a com-puter based pronunciation scoring system is likelyto be capable of providing similar feedback to astudent as a human judge with regard to whichphonetic segments in an utterance can be acceptedas correct or not. Future work will concentrate onexpanding the algorithm to inform the studentabout which mistake he or she has made.

References

Arslan, L., Hansen, J., 1997. Frequency characteristics of

foreign accented speech. In: ICASSP'97. Munich, Germa-

ny.

Bernstein, J., Cohen, M., Murveit, H., Ritschev, D., Weintraub,

M., 1990. Automatic evaluation and training in English

pronunication. In: ICSLP'90. Kobe, Japan, pp. 1185±1188.

Chandler, R., 1991. Farewell, My Lovely. Penguin Readers

Series, Penguin, Harmondsworth.

Ehsani, F., Bernstein, J., Najmi, A., Todic, O., 1997. Subarashi:

Japanese interactive spoken language education. In: Pro-

ceedings EUROSPEECH'97. Rhodes, Greece.

Eskenazi, M., 1996. Detection of foreign speakers' pronunci-

ation errors for second language training ± preliminary

results. In: ICSLP'96. Philadelphia, PA, USA.

Fine, A., 1995. Madame Doubt®re. Penguin Readers Series,

Penguin, Harmondsworth.

Fransen, J., Pye, D., Robinson, A., Woodland, P., Young, S.,

1994. WSJCAM0 corpus and recording description. Tech-

nical Report CUED/F-INFENG/TR 192, Cambridge

University Engineering Department, Cambridge, UK.

Goddijn, A., de Krom, G., 1997. Evaluation of second language

learners' pronunciation using hidden Markov models. In:

Proceedings EUROSPEECH'97. Rhodes, Greece.

Hamada, H., Miki, S., Nakatsu, R., 1993. Automatic evaluation

of English pronunciation based on speech recognition

techniques. IEICE Trans. Inform. Syst. E76-D (3), 352±359.

Hiller, S., Rooney, E., Laver, J., Jack, M., 1993. SPELL: An

automated system for computer-aided pronunciation

teaching. Speech Communication 13, 463±473.

Kawai, G., Hirose, K., 1997. A call system using speech

recognition to train the pronunciation of japanese long

vowels, the mora nasal and mora obstruent. In: Proceed-

ings EUROSPEECH'97. Rhodes, Greece.

Kenworthy, J., 1987. Teaching English Pronunciation. Long-

man, New York.

Kim, Y., Franco, H., Neumeyer, L., 1997. Automatic pronun-

ciation scoring of speci®c phone segments for language

instruction. In: Proceedings EUROSPEECH'97. Rhodes,

Greece.

Knill, K., Young, S., 1994. Speaker dependent keyword

spotting for accessing stored speech. Technical Report

CUED/F-INFENG/TR 193. Cambridge University Engi-

neering Department, Cambridge, UK.

Leggetter, C., Woodland, P., 1994. Speaker adaptation of

HMMs using linear regression. Technical Report CUED/

F-INFENG/TR. 181. Cambridge University Engineering

Department, Cambridge, UK.

Neumeyer, L., Franco, H., Weintraub, M., Price, P., 1996.

Pronunciation scoring of foreign language student speech.

In: ICSLP'96, Philadelphia, PA, USA.

Rogers, C., Dalby, J., DeVane, G., 1994. Intelligibility training

for foreign-accented speech: A preliminary study. J. Acoust.

Soc. Amer. 96 (4), pt. 2.

Ronen, O., Neumeyer, L., Franco, H., 1997. Automatic

detection of mispronunciation for language instruction.

In: Proceedings EUROSPEECH'97. Rhodes, Greece.

Young, S., Odell, J., Ollason, D., Woodland, P., 1996. The

HTK Book. Entropic Cambridge Research Laboratory.

108 S.M. Witt, S.J. Young / Speech Communication 30 (2000) 95±108