Hwang, Mei-Yuh
DISGO: Automatic End-to-End Evaluation for Scene Text OCR
Hwang, Mei-Yuh, Shi, Yangyang, Ramchandani, Ankit, Pang, Guan, Krishnan, Praveen, Kabela, Lucas, Seide, Frank, Datta, Samyak, Liu, Jun
This paper discusses the challenges of optical character recognition (OCR) on natural scenes, which is harder than OCR on documents due to the wild content and various image backgrounds. We propose to uniformly use word error rates (WER) as a new measurement for evaluating scene-text OCR, both end-to-end (e2e) performance and individual system component performances. Particularly for the e2e metric, we name it DISGO WER as it considers Deletion, Insertion, Substitution, and Grouping/Ordering errors. Finally we propose to utilize the concept of super blocks to automatically compute BLEU scores for e2e OCR machine translation. The small SCUT public test set is used to demonstrate WER performance by a modularized OCR system.
Training Augmentation with Adversarial Examples for Robust Speech Recognition
Sun, Sining, Yeh, Ching-Feng, Ostendorf, Mari, Hwang, Mei-Yuh, Xie, Lei
This paper explores the use of adversarial examples in training speech recognition systems to increase robustness of deep neural network acoustic models. During training, the fast gradient sign method is used to generate adversarial examples augmenting the original training data. Different from conventional data augmentation based on data transformations, the examples are dynamically generated based on current acoustic model parameters. We assess the impact of adversarial data augmentation in experiments on the Aurora-4 and CHiME-4 single-channel tasks, showing improved robustness against noise and channel variation. Further improvement is obtained when combining adversarial examples with teacher/student training, leading to a 23% relative word error rate reduction on Aurora-4.