Abstract
Training neural sequence-to-sequence models with simple token-level log-likelihood is now a standard approach to historical text normalization, albeit often outperformed by phrase-based models. Policy gradient training enables direct optimization for exact matches, and while the small datasets in historical text normalization are prohibitive of from-scratch reinforcement learning, we show that policy gradient fine-tuning leads to significant improvements across the board. Policy gradient training, in particular, leads to more accurate normalizations for long or unseen words.
Originalsprog | Engelsk |
---|---|
Titel | Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics |
Forlag | Association for Computational Linguistics |
Publikationsdato | 2019 |
Sider | 1614-1619 |
DOI | |
Status | Udgivet - 2019 |
Begivenhed | 57th Annual Meeting of the Association for Computational Linguistics - Florence, Italien Varighed: 1 jul. 2019 → 1 jul. 2019 |
Konference
Konference | 57th Annual Meeting of the Association for Computational Linguistics |
---|---|
Land/Område | Italien |
By | Florence, |
Periode | 01/07/2019 → 01/07/2019 |