cu-htk march 2001 hub5 system
TRANSCRIPT
CU-HTK March 2001 Hub5 system
Phil Woodland, Thomas Hain, Gunnar Evermann & Dan Povey
May 3rd 2001
Cambridge University Engineering Department
Hub5 Workshop
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Overview
• CU-HTK 2000 system
• Cellular data
• MMIE training
• Lattice MLLR
• 2001 system & results
• Conclusions
• HTK3
Cambridge UniversityEngineering Department
Hub5 Workshop 1
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
CU-HTK 2000 System: Basic Features
• Front-end
– Reduced bandwidth 125–3800 Hz– 12 MF-PLP cepstral parameters + C0 and 1st/2nd derivatives– Side-based cepstral mean and variance normalisation– Vocal tract length normalisation in training and test
• Decision tree state clustered, context dependent triphone & quinphone models:MMIE and MLE versions
• Generate lattices with MLLR-adapted models
• Rescore using iterative MLLR + Full-Variance transform adaptation
• Posterior probability decoding via confusion networks
• System combination
Cambridge UniversityEngineering Department
Hub5 Workshop 2
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Acoustic Training/Test Data
h5train00 248 hours Switchboard (Swbd1), 17 hours CallHome English (CHE)
h5train00sub 60 hours Swbd1, 8 hours CHE
Development test sets
dev01 40 sides Swbd2 (eval98), 40 sides Swbd1 (eval00), 38 sides Swbd2 cellular(dev01-cell)
dev01sub half of the dev01 selected to give similar WER to full set
eval98 40 sides Swbd2 (eval98-swbd2), 40 sides of CHE (eval98-che)
eval97sub 20 side subset of eval97 evaluation set (Swbd2 + CHE)
Earlier development used eval98/eval97sub and later work on dev01sub.
Cambridge UniversityEngineering Department
Hub5 Workshop 3
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
CU-HTK 2000/1 Systems: MLE acoustic models
• MLE triphone models
– Initial models trained on h5train00sub using VTLN data(6168 states / 12 mix)
– Extended training using 265 hour set h5train00 (16 mix comps)– Soft-tying of closest states for each phone– Create gender dependent (GD) versions
• MLE quinphone models
– ±2 phone context + word boundary clustering on h5train00 VTLN data– Trained up to 16 mix comps (9642 states)– Soft-tied gender dependent versions
Cambridge UniversityEngineering Department
Hub5 Workshop 4
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
CU-HTK 2000 System: MMIE acoustic models
• Starting point: MLE models (triphone/quinphone, GI, VTLN)
• Trained using extended Baum-Welch algorithm with lattice-based MMIE
• Lattices on training data using MLE models and a bigram language model
– Numerator/denominator word level lattices with model alignment and Hub5unigram LM probabilities.
– Scaling of acoustic likelihoods (instead of LM)
• Best performance after 2 iterations
Cambridge UniversityEngineering Department
Hub5 Workshop 5
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
CU-HTK 2000/1 Systems: Vocabulary & LanguageModelling
• Vocabulary
– 54537 words: Hub5 vocabulary plus top 50k words of Broadcast News data(0.30% OOV rate on eval00)
– Multiple pronunciation dictionary (based on LIMSI’93 + TTS)– Pronunciation probabilities estimated from forced alignment
• Language models
– Training data∗ 204MW Broadcast News∗ 3MW 1998 Hub5 data∗ 3MW Hub5 data (CHE + Jan 2000 MSU transcriptions)
– 3-fold interpolated/merged bigram, trigram, and 4-gram word LMs– Class based trigram model (400 classes) to smooth word LM
Cambridge UniversityEngineering Department
Hub5 Workshop 6
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
CU-HTK 2000/1 System: Decoding
• Lattice generation/rescoring with time-synchronous Viterbi decoder
• Post-process lattices to yield confusion networks
• Find 1-best min word error rate hypothesis from confusion network
• Combine networks from different stages using Confusion Network Combination
• Confidence scores estimated using confusion networks
• Piecewise linear mapping of word posteriors to confidence scores via decisiontree.
Cambridge UniversityEngineering Department
Hub5 Workshop 7
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
CU-HTK 2000 System Stages
P2
P1
P3
VTLN warp factors, gender
MLLR
4−gram Lattices
GI, MMIE triphones, 54k, bigram
GI, MMIE triphones, 54k, 4−gram
GI, MLE triphones, 27k, trigramP4b
2 MLLR Trans.P6a
P5a P5b
4−gram Lattices
P4a
1−best
CN
Lattice
MLLR
FV
PPROB
CN
��
��
Quinphones
Triphones
CNC
GI, MMIE GD, MLE, ST
Final result
Cambridge UniversityEngineering Department
Hub5 Workshop 8
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
CU-HTK 2000 performance on eval00
Models Ctxt CN MLLR FV Swbd2 CHE TotalP1 1998 P1 31.7 45.4 38.6P2 MMIE tri n - n 25.5 38.1 31.8P3 MMIE tri n 1 n 22.9 35.7 29.3P4a MMIE tri y 1 y 20.9 33.5 27.2P4b MLE/ST/GD tri y 1 y 21.9 33.7 27.8P5a MMIE quin y 1 y 20.3 32.7 26.6P5b MLE/ST/GD quin y 1 y 21.0 32.8 26.9P6a MMIE quin y 2 y 20.3 32.6 26.5CNC P4a+P6a+P4b+P5b 19.3 31.4 25.4
%WER of CUHTK 2000 system on eval00
Cambridge UniversityEngineering Department
Hub5 Workshop 9
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
CU-HTK 2000 performance on dev01
dev01 Swbd1 Swbd2 Swdb2 cell TotalP1 31.7 46.9 48.1 42.1P2 25.5 40.1 41.7 35.7P3 22.9 37.5 38.3 32.8P4a 20.9 34.5 34.9 30.0P4b 21.9 35.6 35.9 31.0P5a 20.3 33.9 34.5 29.5P5b 21.0 34.5 35.1 30.1P6a 20.3 33.6 34.3 29.4
CNC (cuhtk1) 19.3 32.5 33.2 28.3%WER of CUHTK 2000 system on dev01
Cambridge UniversityEngineering Department
Hub5 Workshop 10
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Dealing with Cellular Data-I
• No cellular training data that is really appropriate (all available data sets haveproblems)
• Investigated simulating the GSM channel with the “toast” simulator
• GSM coding/decoding of the eval98-swbd2 data
eval98-swdb2 dev01-cellularoriginal GSM-simulated
MMI 40.0 43.6 41.7MMI+MLLR 37.5 41.4 38.3
%WER in cellular data using MMI triphones trained on h5train00
Cambridge UniversityEngineering Department
Hub5 Workshop 11
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Dealing with Cellular Data-II
• Try GSM coding/decoding the h5train00sub training data and re-training
• Used MLE models trained on 68 hours of data and single-pass retraining
eval98-gsm dev01-cellularSwbd2 CHE Total
baseline MLE 46.4 52.7 49.6 44.3GSM simul. training 45.8 51.8 48.8 44.6%WER for single-pass decoding with tg LM, h5train00sub models trained on original or simulated GSM data
• Training on simulated GSM coded data helps when test uses same process butnot real cellular data!
• Decided to stick with baseline system for cellular data
Cambridge UniversityEngineering Department
Hub5 Workshop 12
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Review of CU-HTK MMIE Training
• Maximum mutual information estimation (MMIE) maximises thesentence level posterior : in log form
Fλ =R
∑
r=1
logPλ (Or|Mwr)P (wr)
∑
w Pλ (Or|Mw) P (w)
– Numerator is likelihood of data given correct transcription– Denominator is total likelihood, calculated as sum over all word sequences– Need to optimise rational function: use extended Baum-Welch algorithm– Compute using word lattices for numerator and denominator
Cambridge UniversityEngineering Department
Hub5 Workshop 13
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Extended Baum-Welch Algorithm
EBW re-estimation formulae are of the form:
µ̂j,m =
{
θj,m(O)− θdenj,m(O)
}
+ Dµj,m{
γj,m − γdenj,m
}
+ D
• Due to high computational load need to ensure fast convergence
– Gaussian-specific D setting with flooring
• Improve MMIE generalisation
– Acoustic scaling by inverse of normal language model scale factor– Weakened language model (unigram) helps focus on acoustic distinctions
Cambridge UniversityEngineering Department
Hub5 Workshop 14
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Lattice Based MMIE
• MLE triphone models used to generate word lattices
• Model-marked lattices for triphone/quinphone
• Run EBW with model-boundaries with margin for pruning
• Best performance after two iterations
• Applied to gender-independent non-soft-tied triphones and quinphones
• WER reductions of 2-3% absolute with 265 hours of training
• Also works well with MLLR, confusion networks etc.
• Investigate MMIE (& variants) for different styles of models
Cambridge UniversityEngineering Department
Hub5 Workshop 15
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Fixed Boundary Estimation (“Exact Match”)
• Exact match scheme scheme relies on boundaries in model-marked lattices
• Comparisons show equal or better performance to using boundaries with margin
– More efficient—runs about twice as fast– Allows more exact acoustic scaling– Fixed boundary estimation used for all experiments here– Used larger D flooring values to reduce overtraining after two iterations
Iteration h5train00sub h5train00Number eval97sub eval98 eval97sub eval980 (MLE) 46.0 46.6 44.4 45.61 44.4 45.4 42.6 44.02 43.7 44.7 41.9 42.93 43.9 44.4 41.6 42.74 43.9 44.3 41.4 42.2
%WER GI models rescoring 1998 trigram lattices
Cambridge UniversityEngineering Department
Hub5 Workshop 16
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Interpolated Objective Functions
• Maximising MMIE criterion tends to over-train, especially for smaller data sets
• Alternative: use modified objective function that combines MLE and MMIEobjective function: a type of “H-criterion”
• Function of the form αFMMIE + (1− α)FMLE. Typically α in range 0.6-0.9
• Unlike pure MMIE, test data WER minimised as objective function maximised
α = fraction MMIE1.0 0.9 0.8 0.7 0.5
eval97sub 44.2 44.0 44.1 43.6 43.9eval98 44.3 44.0 44.0 44.2 44.9
%WER of h5train00sub models with H-criterion training
Cambridge UniversityEngineering Department
Hub5 Workshop 17
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
MMIE Training for Gender Dependent Models
• Std method of training GD models starts from GI models
– splits training data for male/female– update gender dependent mean and mix weights only
• Used H-criterion GD models with α = 0.75
• On eval97sub with h5train00sub training WER reduced from 43.7% after 3iterations to 43.1% after 2 iterations of GD updating
• With h5train00 training only 0.1% reduction in WER from GD modelling
• MMIE GD models not used in 2001 eval system
Cambridge UniversityEngineering Department
Hub5 Workshop 18
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
MMIE Training for Soft-Tied Models
• MLE models in 2000 evaluation system used soft-tying, but MMIE did not
• Initial investigations using h5train00sub triphones:
– similar gains from soft-tying for MMIE as for MLE models– 1% reduction in WER on eval97sub and 0.5% on eval98
• Attempted to extend this to h5train00 using realigned lattices
– improvements appeared to be very small on eval98– insufficient time to fully investigate– not used for MMIE models in 2001 eval system
Cambridge UniversityEngineering Department
Hub5 Workshop 19
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
MMIE-based Model Alignments
• Regenerate model-marked lattices from MMIE-trained triphones/quinphones
• Continue MMIE training for several iterations
triphones quinphonesTraining type eval97sub eval98 eval97subMLE 44.4 45.6 42.02000 MMIE 41.9 42.7 39.8fixed bound 41.4 42.2 39.2realigned 40.9 41.5 38.6
%WER rescoring 1998 tg lattices, h5train00 non-adapted models, quinphones also use pronunciation probabilities
• More than 1% total reduction in WER due to new model training
• MMIE models now 3.4% to 4.1% better than MLE
Cambridge UniversityEngineering Department
Hub5 Workshop 20
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Summary of MMIE Developments
• Lattice MMIE using fixed boundaries
• Use slower convergence—typically used 4 iterations
• Can use interpolated objective function to aid generalisation: more importantwith smaller training sets
• Worthwhile benefits from gender dependent models and soft-tying on 68 hourtraining appears not to scale to 265 hours: hence still using vanilla GI MMIEmodels
• Realigning lattices helps: overall MMIE triphones are 1-1.2% lower WER andquinphones 1.2% lower WER than 2000 evaluation MMIE models.
Cambridge UniversityEngineering Department
Hub5 Workshop 21
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Lattice MLLR
• Unsupervised MLLR requires a transcription from a recogniser
– transcription assumed correct– used to derive a single model sequence for MLLR forward-backward pass
• Gains from adaptation reduced due to supervision errors
– Can estimate fewer transform parameters– Effect is most important when adaptation needed most!
• Lattice MLLR (Uebel & Woodland, ICASSP 2001) uses a recognition latticefor forward-backward pass rather than single model sequence
– Take into account all model sequences found in lattice weighted by posteriorprobability
– Use acoustic scaling to broaden posterior as for MMIE– Similar principles to Padmanabhan, Saon & Zweig, ASR 2000.
Cambridge UniversityEngineering Department
Hub5 Workshop 22
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Lattice MLLR: Triphone Results
• Triphone MMIE models on eval98 set using 2000 system and lattices
• Generate phone-marked lattices once and iteratively update MLLR transformsin sequence 1,2,4,8 MLLR transforms
• Iteratively update FV transform: true interleaved updates of MLLR and FVtransforms
#MLLR(+FV) %WER (Viterbi) %WER (Conf-Net)std MLLR 1 38.7 37.1lattice MLLR 1 38.5 36.9lattice MLLR 2 38.2 36.7lattice MLLR 4 38.0 36.6lattice MLLR 8 37.7 36.7
%WER on eval98: iterative lattice MLLR vs std MLLR
Cambridge UniversityEngineering Department
Hub5 Workshop 23
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Lattice MLLR: Triphone Dev Results
#MLLR FV iterative Swbd1 Swbd2 Swbd2 cell Total1 Y 18.8 34.3 34.7 29.24 Y 18.7 34.2 34.4 29.08 Y 18.8 33.9 34.3 28.94 N 18.7 34.0 34.3 28.98 N 18.6 34.0 34.5 28.9
%WER (confnet) on dev01sub: single vs iterative FV estimation
• Iterative estimation of FV transform not required
• 2001 system used up to 4 MLLR transforms with FV transform estimated once
Cambridge UniversityEngineering Department
Hub5 Workshop 24
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Resegmentation of Audio Data
• Concerned that MSU-style silence “bracketing” (0.5s) would effectnormalisation factors relative to training
– silence at segment boundaries reduced to 200ms
• Initially investigated VTLN: implementation very stable wrt amount of silence
• Cepstral Mean/Variance normalisation is more of an issue
– re-segment test data using same rules as training– use aligned P1 output for segmentation points– re-compute mean/variance normalisation factors
• Investigated effect of re-segmentation on adapted and non-adapted models
Cambridge UniversityEngineering Department
Hub5 Workshop 25
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Resegmentation of Audio Data: Results
MLLR dev01-cellular eval00-sw1original seg N 42.4 26.2new seg for normalisation N 41.2 25.1original seg Y 39.2 —new seg for normalisation Y 38.6 —
%WER 2000 MMIE/VTLN models models, tg LM
• More than 1% abs reduction in WER without adaptation for mismatchedtraining/test segmentation
– variance normalisation is rather sensitive to segmentation!
• Considerably smaller improvement when using adaptation
• Negligible impact on eval98
Cambridge UniversityEngineering Department
Hub5 Workshop 26
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
2001 system – Lattice Generation
• Stages similar to last year
• New MMIE models
• Normalisation after resegmentation based onP1 output
P1
Resegmentation Gender detection
VTLN,CMN, CVN
P2
P3
4−gram Lattices
GI, MMIE triphones, 54k, fgint00
GI, MMIE triphones, 54k, fgintcat00
GI, MLE triphones, 27k, tgint98
MLLR, 1 speech transform
Cambridge UniversityEngineering Department
Hub5 Workshop 27
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
2001 system – Rescoring & Combination
• MLLR in rescoring replaced by iterative Lattice MLLR (4 transforms)
4−gram Lattices
LATMLLR
P4a
2−4 trans.
LATMLLR2−4 trans.
P5a
P4b
MLLR
P5b
MLLR
1 trans.
1 trans.
1−best
CN
Lattice��
��
��
��Quinphones
Triphones
GI, MMIE GD, MLE, ST
FV
PPROB
CN
Final result cu−htk1
CNC
Cambridge UniversityEngineering Department
Hub5 Workshop 28
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Results on dev01 set
Swbd1 Swbd2 Swbd2 cell TotalP1 VTLN/gender det 31.7 46.9 48.1 42.1P2 initial trans. 23.5 38.6 39.2 33.7P3 lat gen 21.1 36.0 36.7 31.2P4a MMIE tri 20.0 33.5 34.0 29.1P4b MLE tri 21.3 35.0 35.4 30.5P5a MMIE quin 19.8 33.2 33.4 28.7P5b MLE quin 20.2 34.0 34.2 29.4CNC P5a+P4a+P5b 18.3 31.9 32.1 27.3Rover vote 18.9 32.5 32.6 27.9Rover conf 18.6 32.3 32.4 27.6
%WER on dev01 for all stages of 2001 system
• final confidence scores have NCE 0.254
Cambridge UniversityEngineering Department
Hub5 Workshop 29
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Results on eval01 set
Swbd1 Swbd2 Swbd2 cell TotalP1 VTLN/gender det 31.9 39.2 45.5 39.1P2 initial trans. 24.3 29.9 36.6 30.4P3 lat gen 22.5 27.8 33.9 28.2P4a MMIE tri 21.3 26.3 31.6 26.5P4b MLE tri 22.6 27.8 32.9 27.9P5a MMIE quin 21.5 26.1 30.8 26.2P5b MLE quin 21.3 26.7 32.0 26.8CNC P5a+P4a+P5b 19.8 24.5 29.2 24.6Rover vote 20.1 25.2 30.2 25.3Rover conf 19.9 24.6 29.8 24.9
%WER on eval01 for all stages of 2001 system
• final confidence scores have NCE 0.294
Cambridge UniversityEngineering Department
Hub5 Workshop 30
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Lattice MLLR & System combination
• Effect of lattice MLLR for quinphone models
– Compare with std iterative MLLR based on triphone output.– Lattice MLLR is much more independent of previous stages– No cross-system adaptation effects– Possible explanation for relatively little gain from quinphones
• Include the effects of confusion networks and system combination.
• Results on dev01sub
– cross-system adaptation important for quinphones– system combination means that single best quinphone system less important!
Cambridge UniversityEngineering Department
Hub5 Workshop 31
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Quinphone Lattice MLLR & System Combinations
#MLLR(+FV) %WER (Vit) %WER (CN)Q1 lattice MLLR (P5a-1) 1 30.2 29.1Q2 std MLLR (tri adapt) 1 29.9 28.6Q3 lattice MLLR 4 29.9 28.8Q4 std MLLR (quin adapt) 2 29.5 28.5
%WER on dev01sub for MMI quinphone models
Swbd1 Swbd2 Swbd2 cell TotalQ1 + P4a + P5b 16.9 32.5 32.8 27.3Q2 + P4a + P5b 17.0 32.4 32.7 27.3Q3 + P4a + P5b 17.1 32.5 32.6 27.3Q4 + P4a + P5b 17.0 32.5 32.6 27.3
%WER on dev01sub for various adapted quinphone combinations
Cambridge UniversityEngineering Department
Hub5 Workshop 32
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Computation
Pass Speed (×RT) Memory (MB)P1 12 357P2 13 280P3 39 335P4a 30 280P4b 34 299P5a 27 380P5b 36 391
Times based on Pentium III 1GHz
• MLLR/Full-Variance 4xRT
• Time marked lattices for LatMLLR (tri+quin) 48xRT+43xRT
• Lattice MLLR/Full-Variance 10xRT
• Overall 298xRT
Cambridge UniversityEngineering Department
Hub5 Workshop 33
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Faster Contrast
• Contrast cuhtk2 is first part of the full system
• Confusion network output of P3 lattices (only MMIE models)
• No rescoring with multiple models, LatMLLR or sys combination
• Runs in a fraction of the time of cuhtk1 (65xRT vs. 298xRT)
Swbd1 Swbd2 Swbd2 cell Total NCEcuhtk1 19.8 24.5 29.2 24.6 0.294cuhtk2 21.6 27.0 32.6 27.2 0.308
%WER on eval01 for primary and contrast systems
Cambridge UniversityEngineering Department
Hub5 Workshop 34
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
Conclusions
• Improved MMIE training 1% abs lower WER
• Improved adaptation using Lattice MLLR
– Allows use of more transforms - improvements of 1-best decoding 1% abs– Doesn’t exploit cross-adaptation effects and overall probably little win
• Re-segmentation for improved normalisation 0.1-0.7% better with adaptation
• Overall system improvement
– 1% improvement over 2000 system: rather less than sum of parts!– Still lowest overall WER
• Faster contrast system
– no rescoring passes and sys combination still yields competitive system– improved 1.5% abs over corresponding 2000 system
Cambridge UniversityEngineering Department
Hub5 Workshop 35
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
HTK3
• Available free of charge since September 2000
• Includes full C source & 300 page HTK book
• Aims to lower entry barrier for ASR research
• Web site got hits from 25k unique IP addresses
• 5000 registered users
• User/developer mailing lists (100 posts/month)
• Meeting: 10th May 7pm Hilton Salt Lake City
http://htk.eng.cam.ac.uk
Cambridge UniversityEngineering Department
Hub5 Workshop 36
Woodland, Hain, Evermann & Povey: CU-HTK March 2001 Hub5 system
HTK3 features
• Discrete and (semi-)continuous HMMs
• Decision-tree state clustering of cross-word triphones
• Baum-Welch training, Viterbi recognition & alignment
• Bigram language models and finite state grammars
• Lattice generation & rescoring
• MLLR and MAP adaptation
• New version supports PLP, VTLN as used in eval
• Future plans: lattice tools, LM toolkit, MMIE, LVCSR decoder
Cambridge UniversityEngineering Department
Hub5 Workshop 37