Lowering the Language Barrier
Investigating Deep Transfer Learning and Machine Translation for Multilingual Analyses of Political Texts
Keywords:text-as-data, machine learning, multilingualism, computational social sciences
The social science toolkit for computational text analysis is still very much in the making. We know surprisingly little about how to produce valid insights from large amounts of multilingual texts for comparative social science research. In this paper, we test several recent innovations from deep transfer learning to help advance the computational toolkit for social science research in multilingual settings. We investigate the extent to which prior language and task knowledge stored in the parameters of modern language models is useful for enabling multilingual research; we investigate the extent to which these algorithms can be fruitfully combined with machine translation; and we investigate whether these methods are accurate, practical and valid in multilingual settings – three essential conditions for lowering the language barrier in practice. We use two datasets with texts in 12 languages from 27 countries for our investigation. Our analysis shows, that, based on these innovations, supervised machine learning can produce substantively meaningful outputs. Our BERT-NLI model trained on only 674 or 1,674 texts in only one or two languages can validly predict political party families’ stances towards immigration in eight other languages and ten other countries.
How to Cite
Copyright (c) 2023 Moritz Laurer, Wouter van Atteveldt, Andreu Casas, Kasper Welbers
This work is licensed under a Creative Commons Attribution 4.0 International License.