
of Finlayson et al. (2021) on stimuli in languages
typologically related to English, such that we can
observe whether there exist syntax neurons that
are shared across a set of languages that are all
relatively high-resource and grammatically similar.
Our contributions include the following:
1.
We causally probe for syntactic agreement
neurons in an autoregressive language model,
XGLM (Lin et al.,2021); a masked language
model, multilingual BERT (Devlin et al.,
2019); and a series of monolingual BERT-
based models. We find two distinct layer-wise
effect patterns, depending on whether the sub-
ject and verb are separated by other tokens.
2.
We quantify the degree of neuron overlap
across languages and syntactic structures, find-
ing that many neurons are shared across struc-
tures and fewer are shared across languages.
3.
We analyze the sparsity of syntactic agree-
ment representations for individual structures
and languages, and find that syntax neurons
are more sparse in MLMs than ALMs, but also
that the degree of sparsity is similar across
models and structures.
Our data and code are publicly available.1
2 Related Work
Multilingual language modeling.
Multilingual
language models enable increased parameter effi-
ciency per language, as well as cross-lingual trans-
fer to lower-resource language varieties (Wu and
Dredze,2019). This makes both training and de-
ployment more efficient when support for many
languages is required. A common approach for
training multilingual LMs is to concatenate train-
ing corpora for many languages into one corpus,
often without language IDs (Conneau et al.,2020;
Devlin et al.,2019).
These models present interesting opportunities
for syntactic analysis: Do multilingual models
maintain similar syntactic abilities despite a de-
creased number of parameters that can be dedi-
cated to each language? Current evidence suggests
slight interference effects, but also that identical
models maintain much of their monolingual per-
formance when trained on multilingual corpora
(Mueller et al.,2020). Is syntactic agreement, in
particular, encoded independently per language or
1https://github.com/aaronmueller/
multilingual-lm-intervention
shared across languages? Some studies suggest
that syntax is encoded in similar ways across lan-
guages (Chi et al.,2020;Stanczak et al.,2022),
though these rely on correlational methods based
on dependency parsing, which introduce confounds
and may not rely on syntactic information per se.
Syntactic probing.
Various behavioral probing
studies have analyzed the syntactic behavior of
monolingual and multilingual LMs (Linzen et al.,
2016;Marvin and Linzen,2018;Ravfogel et al.,
2019;Mueller et al.,2020;Hu et al.,2020). Re-
sults from behavioral analyses are generally eas-
ier to interpret and present clearer evidence for
what models’ preferences are given various con-
texts. However, these methods do not tell us where
or how syntax is encoded.
A parallel line of work employs parametric
probes. Here, a linear classifier or multi-layer per-
ceptron probe is trained to map from a model’s
hidden representations to dependency attachments
and/or labels (Hewitt and Manning,2019) to locate
syntax-sensitive regions of a model. This approach
has been applied in multilingual models (Chi et al.,
2020), and produced evidence for parallel depen-
dency encodings across languages. However, if
such probes are powerful, they may learn the target
task themselves rather than tap into an ability of
the underlying model (Hewitt and Liang,2019),
leading to uninterpretable results. When control-
ling for this, even highly selective probes may not
need access to syntactic information to achieve
high structural probing performance (Sinha et al.,
2021). There are further confounds when analyzing
individual neurons using correlational methods; for
example, probes may locate encoded information
that is not actually used by the model (Antverg and
Belinkov,2022).
Causal probing has recently become more com-
mon for interpreting various phenomena in neu-
ral models of language. Lakretz et al. (2019) and
Lakretz et al. (2021) search for syntax-sensitive
units in English and Italian monolingual LSTMs
by intervening directly on activations and evalu-
ating syntactic agreement performance. Vig et al.
(2020) propose causal mediation analysis for lo-
cating neurons and attention heads implicated in
gender bias in pre-trained language models; this
method involves intervening directly on the inputs
or on individual neurons. Finlayson et al. (2021)
extend this approach to implicate neurons in syn-
tactic agreement. This study extends their data and