Locke’s Holiday: Belief Bias in Machine Reading

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Standard

Locke’s Holiday : Belief Bias in Machine Reading. / Søgaard, Anders.

Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 2021. p. 8240–8245.

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Harvard

Søgaard, A 2021, Locke’s Holiday: Belief Bias in Machine Reading. in Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, pp. 8240–8245, 2021 Conference on Empirical Methods in Natural Language Processing, 07/11/2021. https://doi.org/10.18653/v1/2021.emnlp-main.649

APA

Søgaard, A. (2021). Locke’s Holiday: Belief Bias in Machine Reading. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (pp. 8240–8245). Association for Computational Linguistics. https://doi.org/10.18653/v1/2021.emnlp-main.649

Vancouver

Søgaard A. Locke’s Holiday: Belief Bias in Machine Reading. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics. 2021. p. 8240–8245 https://doi.org/10.18653/v1/2021.emnlp-main.649

Author

Søgaard, Anders. / Locke’s Holiday : Belief Bias in Machine Reading. Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 2021. pp. 8240–8245

Bibtex

@inproceedings{18dbf9a48bfa4f6487af069ebd3453a6,
title = "Locke{\textquoteright}s Holiday: Belief Bias in Machine Reading",
abstract = "I highlight a simple failure mode of state-of-the-art machine reading systems: when contexts do not align with commonly shared beliefs. For example, machine reading systems fail to answer What did Elizabeth want? correctly in the context of {\textquoteleft}My kingdom for a cough drop, cried Queen Elizabeth.{\textquoteright} Biased by co-occurrence statistics in the training data of pretrained language models, systems predict my kingdom, rather than a cough drop. I argue such biases are analogous to human belief biases and present a carefully designed challenge dataset for English machine reading, called Auto-Locke, to quantify such effects. Evaluations of machine reading systems on Auto-Locke show the pervasiveness of belief bias in machine reading.",
author = "Anders S{\o}gaard",
year = "2021",
doi = "10.18653/v1/2021.emnlp-main.649",
language = "English",
pages = "8240–8245",
booktitle = "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing",
publisher = "Association for Computational Linguistics",
note = "2021 Conference on Empirical Methods in Natural Language Processing ; Conference date: 07-11-2021 Through 11-11-2021",

}

RIS

TY - GEN

T1 - Locke’s Holiday

T2 - 2021 Conference on Empirical Methods in Natural Language Processing

AU - Søgaard, Anders

PY - 2021

Y1 - 2021

N2 - I highlight a simple failure mode of state-of-the-art machine reading systems: when contexts do not align with commonly shared beliefs. For example, machine reading systems fail to answer What did Elizabeth want? correctly in the context of ‘My kingdom for a cough drop, cried Queen Elizabeth.’ Biased by co-occurrence statistics in the training data of pretrained language models, systems predict my kingdom, rather than a cough drop. I argue such biases are analogous to human belief biases and present a carefully designed challenge dataset for English machine reading, called Auto-Locke, to quantify such effects. Evaluations of machine reading systems on Auto-Locke show the pervasiveness of belief bias in machine reading.

AB - I highlight a simple failure mode of state-of-the-art machine reading systems: when contexts do not align with commonly shared beliefs. For example, machine reading systems fail to answer What did Elizabeth want? correctly in the context of ‘My kingdom for a cough drop, cried Queen Elizabeth.’ Biased by co-occurrence statistics in the training data of pretrained language models, systems predict my kingdom, rather than a cough drop. I argue such biases are analogous to human belief biases and present a carefully designed challenge dataset for English machine reading, called Auto-Locke, to quantify such effects. Evaluations of machine reading systems on Auto-Locke show the pervasiveness of belief bias in machine reading.

U2 - 10.18653/v1/2021.emnlp-main.649

DO - 10.18653/v1/2021.emnlp-main.649

M3 - Article in proceedings

SP - 8240

EP - 8245

BT - Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing

PB - Association for Computational Linguistics

Y2 - 7 November 2021 through 11 November 2021

ER -

ID: 299822827