Minimally Supervised Number Normalization
Abstract
We propose two models for verbalizing numbers, a key component in speech recognition and synthesis systems. The first model uses an end-to-end recurrent neural network. The second model, drawing inspiration from the linguistics literature, uses finite-state transducers constructed with a minimal amount of training data. While both models achieve near-perfect performance, the latter model can be trained using several orders of magnitude less data than the former, making it particularly useful for low-resource languages.
Full Text:
PDF (presented at EMNLP 2016)Refbacks
- There are currently no refbacks.
Copyright (c) 2016 Association for Computational Linguistics

This work is licensed under a Creative Commons Attribution 4.0 International License.