Historical Text Normalization with Delayed Rewards

Simon Flachs, Marcel Bollmann, Anders Søgaard

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningpeer review

85 Downloads (Pure)

Abstract

Training neural sequence-to-sequence models with simple token-level log-likelihood is now a standard approach to historical text normalization, albeit often outperformed by phrase-based models. Policy gradient training enables direct optimization for exact matches, and while the small datasets in historical text normalization are prohibitive of from-scratch reinforcement learning, we show that policy gradient fine-tuning leads to significant improvements across the board. Policy gradient training, in particular, leads to more accurate normalizations for long or unseen words.
OriginalsprogEngelsk
TitelProceedings of the 57th Annual Meeting of the Association for Computational Linguistics
ForlagAssociation for Computational Linguistics
Publikationsdato2019
Sider1614-1619
DOI
StatusUdgivet - 2019
Begivenhed57th Annual Meeting of the Association for Computational Linguistics - Florence, Italien
Varighed: 1 jul. 20191 jul. 2019

Konference

Konference57th Annual Meeting of the Association for Computational Linguistics
Land/OmrådeItalien
ByFlorence,
Periode01/07/201901/07/2019

Citationsformater