
To test ordering preferences, we generated
meaning-equivalent grammatical variants (Exam-
ples 1b and 1c above) of reference sentences (Ex-
ample 1a) from the Hindi-Urdu Treebank corpus
of written text (HUTB; Bhatt et al.,2009) by per-
muting their preverbal constituent ordering. Sub-
sequently, we used a logistic regression model to
distinguish the original reference sentences from
the plausible variants based on a set of cognitive
predictors. We test whether fine-tuning a neural
language model on preceding sentences improves
predictions of preverbal Hindi constituent order in
later sentences over other cognitive control mea-
sures. The motivation for our fine-tuning method is
that, during reading, encountering a syntactic struc-
ture eases the comprehension of subsequent sen-
tences with similar syntactic structures as attested
in a wide variety of languages (Arai et al.,2007;
Tooley and Traxler,2010) including Hindi (Husain
and Yadav,2020). Our cognitive control factors are
motivated by recent works which show that Hindi
optimizes processing efficiency by minimizing lex-
ical and syntactic surprisal (Ranjan et al.,2019)
and dependency length (Ranjan et al.,2022a) at
the sentence level.
Our results indicate that discourse predictabil-
ity is maximized by reference sentences compared
with alternative orderings, indicating that discourse
predictability influences Hindi word-order prefer-
ences. This finding corroborates previous find-
ings of adaptation/priming in comprehension (Fine
et al.,2013;Fine and Jaeger,2016) and produc-
tion (Gries,2005;Bock,1986). Generally, this
effect is influenced by lexical priming, but we also
find that certain object-fronted constructions prime
subsequent object-fronting, providing evidence
for self-priming of larger syntactic configurations.
With the introduction of neural model surprisal
scores, dependency length minimization effects re-
ported to influence Hindi word order choices in
previous work (Ranjan et al.,2022a) disappear ex-
cept in the case of direct object fronting, which we
interpret as evidence for the Information Locality
Hypothesis (Futrell et al.,2020). Finally, we dis-
cuss the implications of our findings for syntactic
priming in both comprehension and production.
Our main contribution is that we show the im-
pact of discourse predictability on word order
choices using modern computational methods and
naturally occurring data (as opposed to carefully
controlled stimuli in behavioural experiments).
Cross-linguistic evidence is imperative to validate
theories of language processing (Jaeger and Nor-
cliffe,2009), and in this work we extend existing
theories of how humans prioritize word order deci-
sions to Hindi.
2 Background
2.1 Surprisal Theory
Surprisal Theory (Hale,2001;Levy,2008) posits
that comprehenders construct probabilistic inter-
pretations of sentences based on previously encoun-
tered structures. Mathematically, the surprisal of
the
kth
word,
wk
, is defined as the negative log
probability of wkgiven the preceding context:
Sk=−log P(wk|w1...k−1) = log P(w1...wk−1)
P(w1...wk)(1)
These probabilities can be computed either over
word sequences or syntactic configurations and
reflect the information load (or predictability) of
wk
. High surprisal is correlated with longer read-
ing times (Levy,2008;Demberg and Keller,2008;
Staub,2015) as well as longer spontaneous spoken
word durations (Demberg et al.,2012;Dammalap-
ati et al.,2021). Lexical predictability estimated us-
ing n-gram language models is one of the strongest
determinants of word-order preferences in both En-
glish (Rajkumar et al.,2016) and Hindi (Ranjan
et al.,2022a,2019;Jain et al.,2018).
2.2 Dependency Locality Theory
Dependency locality theory (Gibson,2000) has
been shown to be effective at predicting the com-
prehension difficulty of a sequence, with shorter de-
pendencies generally being easier to process than
longer ones (Temperley,2007;Futrell et al.,2015;
Liu et al.,2017, cf. Demberg and Keller,2008).
3 Data and Models
Our dataset comprises 1996 reference sentences
containing well-defined subject and object con-
stituents from the HUTB
1
corpus of dependency
trees (Bhatt et al.,2009). The HUTB corpus,
1https://verbs.colorado.edu/hindiurdu/