Do Language Models Know the Way to Rome?

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Documents

  • Fulltext

    Final published version, 428 KB, application/octet-stream

The global geometry of language models is important for a range of applications, but language model probes tend to evaluate rather local relations, for which ground truths are easily obtained. In this paper we exploit the fact that in geography, ground truths are available beyond local relations. In a series of experiments, we evaluate the extent to which language model representations of city and country names are isomorphic to real-world geography, e.g., if you tell a language model where Paris and Berlin are, does it know the way to Rome? We find that language models generally encode limited geographic information, but with larger models performing the best, suggesting that geographic knowledge can be induced from higher-order co-occurrence statistics.
Original languageEnglish
Title of host publicationProceedings of the Fourth BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP
PublisherAssociation for Computational Linguistics
Publication date2021
Pages510–517
DOIs
Publication statusPublished - 2021
EventFourth BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP - Online
Duration: 11 Nov 202111 Nov 2021

Conference

ConferenceFourth BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP
ByOnline
Periode11/11/202111/11/2021

ID: 300078921