Historical Text Normalization with Delayed Rewards

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Standard

Historical Text Normalization with Delayed Rewards. / Flachs, Simon; Bollmann, Marcel; Søgaard, Anders.

Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, 2019. p. 1614-1619.

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Harvard

Flachs, S, Bollmann, M & Søgaard, A 2019, Historical Text Normalization with Delayed Rewards. in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, pp. 1614-1619, 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 01/07/2019. https://doi.org/10.18653/v1/P19-1157

APA

Flachs, S., Bollmann, M., & Søgaard, A. (2019). Historical Text Normalization with Delayed Rewards. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (pp. 1614-1619). Association for Computational Linguistics. https://doi.org/10.18653/v1/P19-1157

Vancouver

Flachs S, Bollmann M, Søgaard A. Historical Text Normalization with Delayed Rewards. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics. 2019. p. 1614-1619 https://doi.org/10.18653/v1/P19-1157

Author

Flachs, Simon ; Bollmann, Marcel ; Søgaard, Anders. / Historical Text Normalization with Delayed Rewards. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, 2019. pp. 1614-1619

Bibtex

@inproceedings{f56b041d68144f9a9d7a246e533680a5,
title = "Historical Text Normalization with Delayed Rewards",
abstract = "Training neural sequence-to-sequence models with simple token-level log-likelihood is now a standard approach to historical text normalization, albeit often outperformed by phrase-based models. Policy gradient training enables direct optimization for exact matches, and while the small datasets in historical text normalization are prohibitive of from-scratch reinforcement learning, we show that policy gradient fine-tuning leads to significant improvements across the board. Policy gradient training, in particular, leads to more accurate normalizations for long or unseen words",
author = "Simon Flachs and Marcel Bollmann and Anders S{\o}gaard",
year = "2019",
doi = "10.18653/v1/P19-1157",
language = "English",
pages = "1614--1619",
booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics",
publisher = "Association for Computational Linguistics",
note = "57th Annual Meeting of the Association for Computational Linguistics ; Conference date: 01-07-2019 Through 01-07-2019",

}

RIS

TY - GEN

T1 - Historical Text Normalization with Delayed Rewards

AU - Flachs, Simon

AU - Bollmann, Marcel

AU - Søgaard, Anders

PY - 2019

Y1 - 2019

N2 - Training neural sequence-to-sequence models with simple token-level log-likelihood is now a standard approach to historical text normalization, albeit often outperformed by phrase-based models. Policy gradient training enables direct optimization for exact matches, and while the small datasets in historical text normalization are prohibitive of from-scratch reinforcement learning, we show that policy gradient fine-tuning leads to significant improvements across the board. Policy gradient training, in particular, leads to more accurate normalizations for long or unseen words

AB - Training neural sequence-to-sequence models with simple token-level log-likelihood is now a standard approach to historical text normalization, albeit often outperformed by phrase-based models. Policy gradient training enables direct optimization for exact matches, and while the small datasets in historical text normalization are prohibitive of from-scratch reinforcement learning, we show that policy gradient fine-tuning leads to significant improvements across the board. Policy gradient training, in particular, leads to more accurate normalizations for long or unseen words

U2 - 10.18653/v1/P19-1157

DO - 10.18653/v1/P19-1157

M3 - Article in proceedings

SP - 1614

EP - 1619

BT - Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics

PB - Association for Computational Linguistics

T2 - 57th Annual Meeting of the Association for Computational Linguistics

Y2 - 1 July 2019 through 1 July 2019

ER -

ID: 239617712