1 / 23

USING CLASS WEIGHTING IN INTER-CLASS MLLR

USING CLASS WEIGHTING IN INTER-CLASS MLLR. Sam-Joo Doh and Richard M. Stern Department of Electrical and Computer Engineering and School of Computer Science Carnegie Mellon University October 20, 2000. Outline. Introduction Review Transformation-based adaptation Inter-class MLLR

sharne
Download Presentation

USING CLASS WEIGHTING IN INTER-CLASS MLLR

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. USING CLASS WEIGHTING IN INTER-CLASS MLLR Sam-Joo Doh and Richard M. Stern Department of Electrical and Computer Engineering and School of Computer Science Carnegie Mellon University October 20, 2000

  2. Outline • Introduction • Review • Transformation-based adaptation • Inter-class MLLR • Application of weights • For different neighboring classes • Summary Robust Speech Group

  3. Introduction • We would like to achieve • Better adaptation using small amount of adaptation data  Enhance recognition accuracy • Current method • Reduce the number of parameters • Assume transformation function  Transformation-based adaptation • Example: Maximum likelihood linear regression Robust Speech Group

  4. Introduction (cont’d) • Transformation-based adaptation • Transformation classes are assumed to be independent • It does not achieve reliable estimates for multiple classes using a small amount of adaptation data • Better idea ? • Utilize inter-class relationship to achieve more reliable estimates for multiple classes Robust Speech Group

  5. Transformation-Based Adaptation • Estimate each target parameter(mean vector) Robust Speech Group

  6. Transformation-Based Adaptation (cont’d) • Estimate each transformation function Robust Speech Group

  7. Better estimation of transformation function More details of target parameters Quality Number of transformation classes Transformation-Based Adaptation (cont’d) • Trade-off Robust Speech Group

  8. Previous Works • Consider Correlations among model parameters • Mostly in Bayesian framework • Considering a few neighboring models: • Not effective • Considering all neighboring models: • Too much computation • It is difficult to apply correlation on multi-Gaussian mixtures: No explicit correspondence Robust Speech Group

  9. Previous Works (cont’d) • Using correlationsamong model parameters Robust Speech Group

  10. Inter-Class Relationship • Inter-class relationshipamong transformation functions ? Robust Speech Group

  11. Inter-Class Relationship (cont’d) • Two classes are independent Class 1Class 2 Robust Speech Group

  12. Transform class 2 parameters f2(.) m2k(12) f1(.) g12(.) Inter-Class Relationship (cont’d) • If we know aninter-class transformation g12(.) Class 1Class 2 • Now class 2 data contribute to the estimation off1(.)  More reliable estimationoff1(.) while it keeps the characteristics of Class 1 • f2(.) can be estimated by transforming class 1 parameters Robust Speech Group

  13. Estimate (A1, b1) to minimize Q Where Inter-class MLLR • Use Linear Regression for inter-class transformation Robust Speech Group

  14. Application of Weights • Neighboring classes have different contributions to the target class Robust Speech Group

  15. Weighted least squares estimation: • Use the variance of the error for weight • Large error  Small weight • Small error  Large weight Application of Weights (cont’d) • Application of weights to the neighboring classes • We assume in neighboring class n • The error using (A1, b1) in neighboring class n Robust Speech Group

  16. Number of Neighboring Classes • Limit the number of neighboring classes • Sort neighboring classes • Set threshold for the number of samples • Use “closer” neighboring class first • Count the number of samples used • Use next neighboring classes until the number of samples exceed the threshold Robust Speech Group

  17. Experiments • Test data • 1994 DARPA, Wall Street Journal (WSJ) task • 10 Non-native speaker x 20 test sentences (Spoke 3: s3-94) • Baseline System: CMU SPHINX-3 • Continuous HMM, 6000 senones • 39 dimensional features • MFCC cepstra + delta + delta-delta + power • Supervised/Unsupervised adaptation • Focus on small amounts of adaptation data • 13 phonetic-based classes for inter-class MLLR Robust Speech Group

  18. Experiments Experiments (cont’d) • Supervised adaptation • Word Error Rates Robust Speech Group

  19. Experiments Experiments (cont’d) • Unsupervised adaptation • Word Error Rates Robust Speech Group

  20. Experiments (cont’d) • Limit the number of neighboring classes • Supervised adaptation: 10 adaptation sentences Robust Speech Group

  21. Summary • Application of weights • Use weighted least square estimation • Was helpful for supervised case • Was not helpful for unsupervised case (with small amount of adaptation data) • Number of neighboring classes • Use smaller number of neighboring classes as more adaptation data are available Robust Speech Group

  22. Summary • Inter-class transformation • It can have speaker-dependent information • We may prepare several sets of inter-class transformations  Select appropriate set for a new speaker • Combination with Principal Component MLLR • Did not provide additional improvement Robust Speech Group

  23. Thank you ! Robust Speech Group

More Related