Lecture image placeholder

Premium content

Access to this content requires a subscription. You must be a premium user to view this content.

Monthly subscription - $9.99Pay per view - $4.99Access through your institutionLogin with Underline account
Need help?
Contact us
Lecture placeholder background
VIDEO DOI: https://doi.org/10.48448/t6fg-sp69

poster

ACL 2024

August 14, 2024

Bangkok, Thailand

To Distill or Not to Distill? On the Robustness of Robust Knowledge Distillation

keywords:

efficiency

speech recognition

knowledge distillation

Arabic is known to present unique challenges for Automatic Speech Recognition (ASR). On one hand, its rich linguistic diversity and wide range of dialects complicate the de- velopment of robust, inclusive models. On the other, current multilingual ASR models are compute-intensive and lack proper com- prehensive evaluations. In light of these challenges, we distill knowledge from large teacher models into smaller student variants that more efficient. We also introduce a novel human-annotated dataset covering five under- represented Arabic dialects for evaluation. We further evaluate both our models and existing SoTA multilingual models on both standard available benchmarks and our new dialectal data. Our best-distilled model’s overall perfor- mance (45.0% WER) surpasses that of a SoTA model twice its size (SeamlessM4T-large-v2, WER=47.0%) and its teacher model (Whisper- large-v2, WER=55.1%), and its average perfor- mance on our new dialectal data (56.9% WER) outperforms all other models. To gain more in- sight into the poor performance of these models on dialectal data, we conduct an error analysis and report the main types of errors the different models tend to make. The GitHub repository for the project is available at https://github. com/UBC-NLP/distill-whisper-ar.

Downloads

SlidesTranscript English (automatic)

Next from ACL 2024

LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding
poster

LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding

ACL 2024

+10
Mostafa Elhoushi and 12 other authors

14 August 2024

Stay up to date with the latest Underline news!

Select topic of interest (you can select more than one)

PRESENTATIONS

  • All Lectures
  • For Librarians
  • Resource Center
  • Free Trial
Underline Science, Inc.
1216 Broadway, 2nd Floor, New York, NY 10001, USA

© 2023 Underline - All rights reserved