Common Sense Bias in Semantic Role Labeling

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Standard

Common Sense Bias in Semantic Role Labeling. / Lent, Heather Christine; Søgaard, Anders.

Proceedings of the Seventh Workshop on Noisy User-generated Text (W-NUT 2021). Association for Computational Linguistics, 2021. p. 114–119.

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Harvard

Lent, HC & Søgaard, A 2021, Common Sense Bias in Semantic Role Labeling. in Proceedings of the Seventh Workshop on Noisy User-generated Text (W-NUT 2021). Association for Computational Linguistics, pp. 114–119, 7th Workshop on Noisy User-generated Text (W-NUT 2021), Online, 11/11/2021. https://doi.org/10.18653/v1/2021.wnut-1.14

APA

Lent, H. C., & Søgaard, A. (2021). Common Sense Bias in Semantic Role Labeling. In Proceedings of the Seventh Workshop on Noisy User-generated Text (W-NUT 2021) (pp. 114–119). Association for Computational Linguistics. https://doi.org/10.18653/v1/2021.wnut-1.14

Vancouver

Lent HC, Søgaard A. Common Sense Bias in Semantic Role Labeling. In Proceedings of the Seventh Workshop on Noisy User-generated Text (W-NUT 2021). Association for Computational Linguistics. 2021. p. 114–119 https://doi.org/10.18653/v1/2021.wnut-1.14

Author

Lent, Heather Christine ; Søgaard, Anders. / Common Sense Bias in Semantic Role Labeling. Proceedings of the Seventh Workshop on Noisy User-generated Text (W-NUT 2021). Association for Computational Linguistics, 2021. pp. 114–119

Bibtex

@inproceedings{4a9f987a87754c4cb50e3b6695e05d08,
title = "Common Sense Bias in Semantic Role Labeling",
abstract = "Large-scale language models such as ELMo and BERT have pushed the horizon of what is possible in semantic role labeling (SRL), solving the out-of-vocabulary problem and enabling end-to-end systems, but they have also introduced significant biases. We evaluate three SRL parsers on very simple transitive sentences with verbs usually associated with animate subjects and objects, such as, “Mary babysat Tom”: a state-of-the-art parser based on BERT, an older parser based on GloVe, and an even older parser from before the days of word embeddings. When arguments are word forms predominantly used as person names, aligning with common sense expectations of animacy, the BERT-based parser is unsurprisingly superior; yet, with abstract or random nouns, the opposite picture emerges. We refer to this as “common sense bias” and present a challenge dataset for evaluating the extent to which parsers are sensitive to such a bias. Our code and challenge dataset are available here: github.com/coastalcph/comte",
author = "Lent, {Heather Christine} and Anders S{\o}gaard",
year = "2021",
doi = "10.18653/v1/2021.wnut-1.14",
language = "English",
pages = "114–119",
booktitle = "Proceedings of the Seventh Workshop on Noisy User-generated Text (W-NUT 2021)",
publisher = "Association for Computational Linguistics",
note = "7th Workshop on Noisy User-generated Text (W-NUT 2021) ; Conference date: 11-11-2021 Through 11-11-2021",

}

RIS

TY - GEN

T1 - Common Sense Bias in Semantic Role Labeling

AU - Lent, Heather Christine

AU - Søgaard, Anders

PY - 2021

Y1 - 2021

N2 - Large-scale language models such as ELMo and BERT have pushed the horizon of what is possible in semantic role labeling (SRL), solving the out-of-vocabulary problem and enabling end-to-end systems, but they have also introduced significant biases. We evaluate three SRL parsers on very simple transitive sentences with verbs usually associated with animate subjects and objects, such as, “Mary babysat Tom”: a state-of-the-art parser based on BERT, an older parser based on GloVe, and an even older parser from before the days of word embeddings. When arguments are word forms predominantly used as person names, aligning with common sense expectations of animacy, the BERT-based parser is unsurprisingly superior; yet, with abstract or random nouns, the opposite picture emerges. We refer to this as “common sense bias” and present a challenge dataset for evaluating the extent to which parsers are sensitive to such a bias. Our code and challenge dataset are available here: github.com/coastalcph/comte

AB - Large-scale language models such as ELMo and BERT have pushed the horizon of what is possible in semantic role labeling (SRL), solving the out-of-vocabulary problem and enabling end-to-end systems, but they have also introduced significant biases. We evaluate three SRL parsers on very simple transitive sentences with verbs usually associated with animate subjects and objects, such as, “Mary babysat Tom”: a state-of-the-art parser based on BERT, an older parser based on GloVe, and an even older parser from before the days of word embeddings. When arguments are word forms predominantly used as person names, aligning with common sense expectations of animacy, the BERT-based parser is unsurprisingly superior; yet, with abstract or random nouns, the opposite picture emerges. We refer to this as “common sense bias” and present a challenge dataset for evaluating the extent to which parsers are sensitive to such a bias. Our code and challenge dataset are available here: github.com/coastalcph/comte

U2 - 10.18653/v1/2021.wnut-1.14

DO - 10.18653/v1/2021.wnut-1.14

M3 - Article in proceedings

SP - 114

EP - 119

BT - Proceedings of the Seventh Workshop on Noisy User-generated Text (W-NUT 2021)

PB - Association for Computational Linguistics

T2 - 7th Workshop on Noisy User-generated Text (W-NUT 2021)

Y2 - 11 November 2021 through 11 November 2021

ER -

ID: 300076700