arxiv_dump / txt /2204.03592.txt
billxbf's picture
Upload 101 files
8f1929a verified
raw
history blame
118 kB
Testing the limits of natural language models for
predicting human language judgments
Tal Golan1,2∗†, Matthew Siegelman3∗,
Nikolaus Kriegeskorte1,3,4,5, Christopher Baldassano3
1Zuckerman Mind Brain Behavior Institute, Columbia University, New York, NY , USA
2Department of Cognitive and Brain Sciences, Ben-Gurion University of the Negev, Be’er-Sheva, Israel
3Department of Psychology, Columbia University, New York, NY , USA
4Department of Neuroscience, Columbia University, New York, NY , USA
5Department of Electrical Engineering, Columbia University, New York, NY , USA
∗The first two authors contributed equally to this work.
†To whom correspondence should be addressed; E-mail: [email protected]
Neural network language models appear to be increasingly aligned with how hu-
mans process and generate language, but identifying their weaknesses through ad-
versarial examples is challenging due to the discrete nature of language and the
complexity of human language perception. We bypass these limitations by turning
the models against each other. We generate controversial sentence pairs for which
two language models disagree about which sentence is more likely to occur. Con-
sidering nine language models (including n-gram, recurrent neural networks, and
transformers), we created hundreds of controversial sentence pairs through syn-
thetic optimization or by selecting sentences from a corpus. Controversial sentence
pairs proved highly effective at revealing model failures and identifying models that
aligned most closely with human judgments of which sentence is more likely. The
most human-consistent model tested was GPT-2, although experiments also revealed
significant shortcomings of its alignment with human perception.
Keywords— Language Models, Human Acceptability Judgments, Controversial Stimuli, Adversarial Attacks
in NLP
1 Introduction
Neural network language models are not only key tools in natural language processing (NLP) but are also draw-
ing an increasing scientific interest as potential models of human language-processing. Ranging from recurrent
neural networks [1, 2] to transformers [3–7], each of these language models (explicitly or implicitly) defines
a probability distribution over strings of words, predicting which sequences are likely to occur in natural lan-
guage. There is substantial evidence from measures such as reading times [8], functional MRI [9], scalp EEG
1arXiv:2204.03592v3 [cs.CL] 12 Sep 2023[10], and intracranial ECoG [11] that humans are sensitive to the relative probabilities of words and sentences as
captured by language models, even among sentences that are grammatically correct and semantically meaning-
ful. Furthermore, model-derived sentence probabilities can also predict human graded acceptability judgments
[12, 13]. These successes, however, have not yet addressed two central questions of interest: (1) Which of the
models is best-aligned with human language processing? (2) How close is the best-aligned model to the goal
of fully capturing human judgments?
A predominant approach for evaluating language models is to use a set of standardized benchmarks such
as those in the General Language Understanding Evaluation (GLUE) [14], or its successor, SuperGLUE [15].
Though instrumental in evaluating the utility of language models for downstream NLP tasks, these benchmarks
prove insufficient for comparing such models as candidate explanations of human language-processing. Many
components of these benchmarks do not aim to measure human alignment but rather the usefulness of the mod-
els’ language representation when tuned to a specific downstream task. Some benchmarks challenge language
models more directly by comparing the probabilities they assign to grammatical and ungrammatical sentences
(e.g., BLiMP [16]). However, since such benchmarks are driven by theoretical linguistic considerations, they
might fail to detect novel, unexpected ways in which language models may diverge from human language un-
derstanding. Lastly, an additional practical concern is that the rapid pace of NLP research has led to quick
saturation of these kinds of static benchmarks, making it difficult to distinguish between models [17].
One proposed solution to these issues is the use of dynamic human-in-the-loop benchmarks in which peo-
ple actively stress-test models with an evolving set of tests. However, this approach faces the major obstacle
that “finding interesting examples is rapidly becoming a less trivial task” [17]. We propose to complement
human-curated benchmarks with model-driven evaluation. Guided by model predictions rather than experi-
menter intuitions, we would like to identify particularly informative test sentences, where different models
make divergent predictions. This approach of running experiments mathematically optimized to “put in jeop-
ardy” particular models belongs to a long-standing scientific philosophy of design optimization [18]. We can
find these critical sentences in large corpora of natural language or synthesize novel test sentences that reveal
how different models generalize beyond their training distributions.
We propose here a systematic, model-driven approach for comparing language models in terms of their
consistency with human judgments. We generate controversial sentence pairs : pairs of sentences designed
such that two language models strongly disagree about which sentence is more likely to occur. In each of these
sentence pairs, one model assigns a higher probability to the first sentence than the second sentence, while the
other model prefers the second sentence to the first. We then collect human judgments of which sentence in
each pair is more probable to settle this dispute between the two models.
This approach builds on previous work on controversial images for models of visual classification [19].
That work relied on absolute judgments of a single stimulus, which are appropriate for classification responses.
However, asking the participants to rate each sentence’s probability on an absolute scale is complicated by
between-trial context effects common in magnitude estimation tasks [20–22], which have been shown to impact
judgments like acceptability [23]. A binary forced-choice behavioral task presenting the participants with a
choice between two sentences in each trial, the approach we used here, minimizes the role of between-trial
context effects by setting an explicit local context within each trial. Such an approach has been previously
used for measuring sentence acceptability [24] and provides substantially more statistical power compared to
designs in which acceptability ratings are provided for single sentences [25].
Our experiments demonstrate that (1) it is possible to procedurally generate controversial sentence pairs for
all common classes of language models, either by selecting pairs of sentences from a corpus or by iteratively
modifying natural sentences to yield controversial predictions; (2) the resulting controversial sentence pairs
enable efficient model comparison between models that otherwise are seemingly equivalent in their human
2consistency; and (3) all current NLP model classes incorrectly assign high probability to some non-natural
sentences (one can modify a natural sentence such that its model probability does not decrease but human
observers reject the sentence as unnatural). This framework for model comparison and model testing can give
us new insight into the classes of models that best align with human language perception and suggest directions
for future model development.
2 Results
We acquired judgments from 100 native English speakers tested online. In each experimental trial, the partic-
ipants were asked to judge which of two sentences they would be “more likely to encounter in the world, as
either speech or written text”, and provided a rating of their confidence in their answer on a 3-point scale (see
Extended Data Fig. 1 for a trial example). The experiment was designed to compare nine different language
models (Supplementary Section 6.1): probability models based on corpus frequencies of 2-word and 3-word
sequences (2-grams and 3-grams) and a range of neural network models comprising a recurrent neural network
(RNN), a long short-term memory network (LSTM), and five transformer models (BERT, RoBERTa, XLM,
ELECTRA, and GPT-2).
2.1 Efficient model comparison using natural controversial pairs
As a baseline, we randomly sampled and paired 8-word sentences from a corpus of Reddit comments. How-
ever, as shown in Fig. 1a, these sentences fail to uncover meaningful differences between the models. For
each sentence pair, all models tend to prefer the same sentence (Extended Data Fig. 2), and therefore perform
similarly in predicting human preference ratings (see Supplementary Section 7.1).
Instead, we can use an optimization procedure (Supplementary Section 6.2) to search for controversial
sentence pairs, in which one language model assigns a high probability (above the median probability for
natural sentences) only to sentence 1 and a second language model assigns a high probability only to sentence
2; see examples in Table 1. Measuring each model’s accuracy in predicting human choices for sentence pairs
in which it was one of the two targeted models indicated many significant differences in terms of model-
human alignment (Fig. 1b), with GPT-2 and RoBERTa showing the best human consistency and 2-gram the
worst. We can also compare each model pair separately (using only the stimuli targeting that model pair),
yielding a similar pattern of pairwise dominance (Extended Data Fig. 3a). All models except GPT-2, RoBERTa,
and ELECTRA performed significantly below our lower bound on the noise ceiling (the accuracy obtained
by predicting each participant’s responses from the other participants’ responses), indicating a misalignment
between these models’ predictions and human judgments which was only revealed when using controversial
sentence pairs.
30 25% 50% 75% 100%
human-choice prediction accuracyGPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-gram
0 50 100
GPT-2
p(sentence) percentile020406080100RoBERTa
p(sentence) percentile
0 25% 50% 75% 100%
human-choice prediction accuracyGPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-gram
0 50 100
GPT-2
p(sentence) percentile020406080100RoBERTa
p(sentence) percentile
aRandomly sampled natural-sentence pairs
bControversial natural-sentence pairsFigure 1: Model comparison using natural sentences. (a) (Left) Percentile-transformed sentence probabili-
ties for GPT-2 and RoBERTa (defined relative to all sentences used in the experiment) for randomly-sampled
pairs of natural sentences. Each pair of connected dots depicts one sentence pair. The two models are highly
congruent in their rankings of sentences within a pair (lines have upward slope). (Right) Accuracy of model
predictions of human choices, measured as the proportion of trials in which the same sentence was preferred
by both the model and the human participant. Each dot depicts the prediction accuracy of one candidate model
averaged across a group of 10 participants presented with a unique set of trials. The colored bars depict grand-
averages across all 100 participants. The gray bar is the noise ceiling whose left and right edges are lower
and upper bounds on the grand-average performance an ideal model would achieve (based on the consistency
across human subjects). There were no significant differences in model performance on the randomly sam-
pled natural sentences. (b)(Left) Controversial natural-sentence pairs were selected such that the models’
sentence probability ranks were incongruent (lines have downward slope). (Right) Controversial sentence pairs
enable efficient model comparison, revealing that BERT, XLM, LSTM, RNN and the n-gram models perform
significantly below the noise ceiling (asterisks indicate significance—two-sided Wilcoxon signed-rank test,
controlling the false discovery rate for nine comparisons at q<.05). On the right of the plot, each closed circle
indicates a model significantly dominating alternative models indicated by open circles (two-sided Wilcoxon
signed-rank test, controlling the false discovery rate for all 36 model pairs at q<.05). GPT-2 outperforms all
models except RoBERTA at predicting human judgments.
4sentence log probability (model 1) log probability (model 2) # human choices
n1: Rust is generally caused by salt and sand. logp(n1|GPT-2 ) =−50.72 log p(n1|ELECTRA ) =−38.54 10
n2: Where is Vernon Roche when you need him. logp(n2|GPT-2 ) =−32.26 logp(n2|ELECTRA ) =−58.26 0
n1: Excellent draw and an overall great smoking experience. logp(n1|RoBERTa ) =−67.78 log p(n1|GPT-2 ) =−36.76 10
n2: I should be higher and tied to inflation. logp(n2|RoBERTa ) =−54.61 logp(n2|GPT-2 ) =−50.31 0
n1: You may try and ask on their forum. logp(n1|ELECTRA ) =−51.44 log p(n1|LSTM ) =−44.24 10
n2: I love how they look like octopus tentacles. logp(n2|ELECTRA ) =−35.51 logp(n2|LSTM ) =−66.66 0
n1: Grow up and quit whining about minor inconveniences. logp(n1|BERT ) =−82.74 log p(n1|GPT-2 ) =−35.66 10
n2: The extra a is the correct Sanskrit pronunciation. logp(n2|BERT ) =−51.06 logp(n2|GPT-2 ) =−51.10 0
n1: I like my password manager for this reason. logp(n1|XLM ) =−68.93 log p(n1|RoBERTa ) =−49.61 10
n2: Kind of like clan of the cave bear. logp(n2|XLM ) =−44.24 logp(n2|RoBERTa ) =−67.00 0
n1: We have raised a Generation of Computer geeks. logp(n1|LSTM ) =−66.41 log p(n1|ELECTRA ) =−36.57 10
n2: I mean when the refs are being sketchy. logp(n2|LSTM ) =−42.04 logp(n2|ELECTRA ) =−52.28 0
n1: This is getting ridiculous and ruining the hobby. logp(n1|RNN) =−100.65 log p(n1|LSTM ) =−43.50 10
n2: I think the boys and invincible are better. logp(n2|RNN) =−45.16 logp(n2|LSTM ) =−59.00 0
n1: Then attach them with the supplied wood screws. logp(n1|3-gram ) =−119.09 log p(n1|GPT-2 ) =−34.84 10
n2: Sounds like you were used both a dog. logp(n2|3-gram ) =−92.07 logp(n2|GPT-2 ) =−52.84 0
n1: Cream cheese with ham and onions on crackers. logp(n1|2-gram ) =−131.99 log p(n1|RoBERTa ) =−54.62 10
n2: I may have to parallel process that drinking. logp(n2|2-gram ) =−109.46 logp(n2|RoBERTa ) =−70.69 0
Table 1: Examples of controversial natural-sentence pairs that maximally contributed to each model’s
prediction error. For each model (double row, “model 1”), the table shows results for two sentences on which
the model failed severely. In each case, the failing model 1 prefers sentence n2(higher log probability bolded),
while the model it was pitted against (“model 2”) and all 10 human subjects presented with that sentence pair
prefer sentence n1. (When more than one sentence pair induced an equal maximal error in a model, the example
included in the table was chosen at random.)
2.2 Greater model disentanglement with synthetic sentence pairs
Selecting controversial natural-sentence pairs may provide greater power than randomly sampling natural-
sentence pairs, but this search procedure considers a very limited part of the space of possible sentence pairs.
Instead, we can iteratively replace words in a natural sentence to drive different models to make opposing
predictions, forming synthetic controversial sentences that may lay outside any natural language corpora, as il-
lustrated in Fig. 2 (see Methods, “Generating synthetic controversial sentence pairs” for full details). Examples
of controversial synthetic-sentence pairs that maximally contributed to the models’ prediction error appear in
Table 2.
We evaluated how well each model predicted the human sentence choices in all of the controversial synthetic-
sentence pairs in which the model was one of the two models targeted (Fig. 3a). This evaluation of model-
human alignment resulted in an even greater separation between the models’ prediction accuracies than was
obtained when using controversial natural-sentence pairs, pushing the weaker models (RNN, 3-gram, and 2-
gram) far below the 50% chance accuracy level. GPT-2, RoBERTa, and ELECTRA were found to be signifi-
cantly more accurate than the alternative models (BERT, XLM, LSTM, RNN, 3-gram, and 2-gram) in predicting
the human responses to these trials (with similar results when comparing model pair separately, see Extended
Data Fig. 3b). All of the models except for GPT-2 were found to be significantly below the lower bound on the
noise ceiling, demonstrating misalignment with human judgments.
50 20 40 60 80 100
GPT-2
p(sentence) percentile020406080100ELECTRA
p(sentence) percentileNothing has
a world of
excitement
and joys.
Diddy has a wealth of experience
with grappling.Luke has a ton of experience
with winning.
a
0 20 40 60 80 100
RoBERTa
p(sentence) percentile0204060801003-gram
p(sentence) percentileYou have to realize
is that noise again.
I wait to see how
it shakes out.I need to see how
this played out.
bFigure 2: Synthesizing controversial sentence pairs. The small open dots denote 500 randomly sampled
natural sentences. The big open dot denotes the natural sentence used for initializing the controversial sentence
optimization, and the closed dots are the resulting synthetic sentences. (a)In this example, we start with the
randomly sampled natural sentence “Luke has a ton of experience with winning”. If we adjust this sentence to
minimize its probability according to GPT-2 (while keeping the sentence at least as likely as the natural sentence
according to ELECTRA), we obtain the synthetic sentence “Nothing has a world of excitement and joys”. By
repeating this procedure while switching the roles of the models, we generate the synthetic sentence “Diddy
has a wealth of experience with grappling”, which decreases ELECTRA’s probability while slightly increasing
GPT-2’s. (b)In this example, we start with the randomly sampled natural sentence “I need to see how this
played out”. If we adjust this sentence to minimize its probability according to RoBERTa (while keeping the
sentence at least as likely as the natural sentence according to 3-gram), we obtain the synthetic sentence “You
have to realize is that noise again”. If we instead decrease only 3-gram’s probability, we generate the synthetic
sentence “I wait to see how it shakes out”.
2.3 Pairs of natural and synthetic sentences uncover blindspots
Last, we considered trials in which the participants were asked to choose between a natural sentence and one of
the synthetic sentences which was generated from that natural sentence. If the language model is fully aligned
with human judgments, we would expect humans to agree with the model, and select the synthetic sentence
at least as much as the natural sentence. In reality, human participants showed a systematic preference for the
natural sentences over their synthetic counterparts (Fig. 3b), even when the synthetic sentences were formed
such that the stronger models (i.e., GPT-2, RoBERTA, or ELECTRA) favored them over the natural sentences;
see Extended Data Table 1 for examples. Evaluating natural sentence preference separately for each model-
pairing (Extended Data Fig. 4), we find that these imperfections can be uncovered even when pairing a strong
model with a relatively weak model (such that the strong model “accepts” the synthetic sentence and the weak
model rejects it).
60 25% 50% 75% 100%
human-choice prediction accuracyGPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-gram
0 50 100
GPT-2
p(sentence) percentile020406080100RoBERTa
p(sentence) percentile
0 25% 50% 75% 100%
human-choice prediction accuracyGPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-gram
0 50 100
GPT-2
p(sentence) percentile020406080100RoBERTa
p(sentence) percentile
aSynthetic controversial sentence pairs
bSynthetic vs. natural sentences
natural sentence synthetic sentenceFigure 3: Model comparison using synthetic sentences. (a) (Left) Percentile-transformed sentence probabil-
ities for GPT-2 and RoBERTa for controversial synthetic-sentence pairs. Each pair of connected dots depict
one sentence pair. (Right) Model prediction accuracy, measured as the proportion of trials in which the same
sentence was preferred by both the model and the human participant. GPT-2, RoBERTa and ELECTRA sig-
nificantly outperformed the other models (two-sided Wilcoxon signed-rank test, controlling the false discovery
rate for all 36 model comparisons at q<.05). All of the models except for GPT-2 were found to perform below
the noise ceiling (gray) of predicting each participant’s choices from the majority votes of the other participants
(asterisks indicate significance—two-sided Wilcoxon signed-rank test, controlling the false discovery rate for
nine comparisons at q<.05). (b)(Left) Each connected triplet of dots depicts a natural sentence and its derived
synthetic sentences, optimized to decrease the probability only under GPT-2 (left dots in a triplet) or only under
RoBERTa (bottom dots in a triplet). (Right) Each model was evaluated across all of the synthetic-natural sen-
tence pairs for which it was targeted to keep the synthetic sentence at least as probable as the natural sentence
(see Extended Data Fig. 6 for the complementary data binning). This evaluation yielded a below-chance pre-
diction accuracy for all of the models, which was also significantly below the lower bound on the noise ceiling.
This indicates that, although the models assessed that these synthetic sentences were at least as probable as the
original natural sentence, humans disagreed and showed a systematic preference for the natural sentence. See
Fig. 1’s caption for details on the visualization conventions used in this figure.
7sentence log probability (model 1) log probability (model 2) # human choices
s1: You can reach his stories on an instant. logp(s1|GPT-2 ) =−64.92 log p(s1|RoBERTa ) =−59.98 10
s2: Anybody can behead a rattles an an antelope. logp(s2|GPT-2 ) =−40.45 logp(s2|RoBERTa ) =−90.87 0
s1: However they will still compare you to others. logp(s1|RoBERTa ) =−53.40 log p(s1|GPT-2 ) =−31.59 10
s2: Why people who only give themselves to others. logp(s2|RoBERTa ) =−48.66 logp(s2|GPT-2 ) =−47.13 0
s1: He healed faster than any professional sports player. logp(s1|ELECTRA ) =−48.77 log p(s1|BERT ) =−50.21 10
s2: One gets less than a single soccer team. logp(s2|ELECTRA ) =−38.25 logp(s2|BERT ) =−59.09 0
s1: That is the narrative we have been sold. logp(s1|BERT ) =−56.14 log p(s1|GPT-2 ) =−26.31 10
s2: This is the week you have been dying. logp(s2|BERT ) =−50.66 logp(s2|GPT-2 ) =−39.50 0
s1: The resilience is made stronger by early adversity. logp(s1|XLM ) =−62.95 log p(s1|RoBERTa ) =−54.34 10
s2: Every thing is made alive by infinite Ness. logp(s2|XLM ) =−42.95 logp(s2|RoBERTa ) =−75.72 0
s1: President Trump threatens to storm the White House. logp(s1|LSTM ) =−58.78 log p(s1|RoBERTa ) =−41.67 10
s2: West Surrey refused to form the White House. logp(s2|LSTM ) =−40.35 logp(s2|RoBERTa ) =−67.32 0
s1: Las beans taste best with a mustard sauce. logp(s1|RNN) =−131.62 log p(s1|RoBERTa ) =−60.58 10
s2: Roughly lanes being alive in a statement ratings. logp(s2|RNN) =−49.31 logp(s2|RoBERTa ) =−99.90 0
s1: You are constantly seeing people play the multi. logp(s1|3-gram ) =−107.16 log p(s1|ELECTRA ) =−44.79 10
s2: This will probably the happiest contradicts the hypocrite. logp(s2|3-gram ) =−91.59 logp(s2|ELECTRA ) =−75.83 0
s1: A buyer can own a genuine product also. logp(s1|2-gram ) =−127.35 log p(s1|ELECTRA ) =−40.21 10
s2: One versed in circumference of highschool I rambled. logp(s2|2-gram ) =−113.73 logp(s2|ELECTRA ) =−92.61 0
Table 2: Examples of controversial synthetic-sentence pairs that maximally contributed to each model’s
prediction error. For each model (double row, “model 1”), the table shows results for two sentences on which
the model failed severely. In each case, the failing model 1 prefers sentence s2(higher log probability bolded),
while the model it was pitted against (“model 2”) and all 10 human subjects presented with that sentence pair
prefer sentence s1. (When more than one sentence pair induced an equal maximal error in a model, the example
included in the table was chosen at random.)
81.0
0.5
0.0 0.5 1.0
ordinal correlation between human ratings and models'
sentence pair probability log-ratio (signed-rank cosine similarity)GPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-gram
Figure 4: Ordinal correlation of the models’ sentence probability log-ratios and human Likert ratings.
For each sentence pair, model prediction was quantified by logp(s1|m)
p(s2|m). This log-ratio was correlated with the
Likert ratings of each particular participant, using signed-rank cosine similarity (see Methods). This analysis,
taking all trials and human confidence level into account, indicates that GPT-2 performed best in predicting
human sentence probability judgments. However, its predictions are still significantly misaligned with the
human choices. See Fig. 1’s caption for details on the visualization convention.
2.4 Evaluating the entire dataset reveals a hierarchy of models
Rather than evaluating each model’s prediction accuracy with respect to the particular sentence pairs that were
formed to compare this model to alternative models, we can maximize our statistical power by computing the
average prediction accuracy for each model with respect to all of the experimental trials we collected. Fur-
thermore, rather than binarizing the human and model judgments, here we measure the ordinal correspondence
between the graded human choices (taking confidence into account) and the log ratio of the sentence proba-
bilities assigned by each candidate model. Using this more sensitive benchmark (Fig. 4), we found GPT-2 to
be the most human-aligned, followed by RoBERTa; then ELECTRA; BERT; XLM and LSTM; and the RNN,
3-gram, and 2-gram models. However, all of the models (including GPT-2) were found to be significantly less
accurate than the lower bound on the noise ceiling.
One possible reason for the poorer performance of the bidirectional transformers (RoBERTa, ELECTRA,
BERT, and XLM) compared to the unidirectional transformer (GPT-2) is that computing sentence probabilities
in these models is complex, and the probability estimator we developed (see Methods, “Evaluating sentence
probabilities in transformer models”) could be non-optimal; Indeed, the popular pseudo-log-likelihood (PLL)
approach yields slightly higher accuracy for randomly sampled natural-sentence pairs (Extended Data Fig. 5a).
And yet, when we directly compared our estimator to PLL by means of generating and administrating new
synthetic controversial sentences, our estimator was found to be markedly better aligned to human judgments
(Extended Data Fig. 5b and Extended Data Table 2).
Finally, a control analysis employing probability measures normalized by token count revealed that such
normalization had minimal influence on the observed differences among models (Supplementary Section 7.2
and Supplementary Fig. S1).
93 Discussion
In this study, we probed the ability of language models to predict human relative sentence probability judgments
using controversial sentence pairs, selected or synthesized so that two models disagreed about which sentence
was more probable. We found that (1) GPT-2 (a unidirectional transformer model trained on predicting up-
coming tokens) and RoBERTa (a bidirectional transformer trained on a held-out token prediction task) were the
most predictive of human judgments on controversial natural-sentence pairs (Fig. 1b); (2) GPT-2, RoBERTa,
and ELECTRA (a bidirectional transformer trained on detecting corrupted tokens) were the most predictive of
human judgments on pairs of sentences synthesized to maximize controversiality (Fig. 3a); and (3) GPT-2 was
the most human-consistent model when considering the entire behavioral dataset we collected (Fig. 4). And yet,
all of the models, including GPT-2, exhibited behavior inconsistent with human judgments; using an alternative
model as a counterforce, we could corrupt natural sentences such that their probability under a model did not
decrease, but humans tended to reject the corrupted sentence as unlikely (Fig. 3b).
3.1 Implications for computational psycholinguistic modeling
Unlike convolutional neural networks, whose architectural design principles are roughly inspired by biological
vision [26], the design of current neural network language models is largely uninformed by psycholinguistics
and neuroscience. And yet, there is an ongoing effort to adopt and adapt neural network language models
to serve as computational hypotheses of how humans process language, making use of a variety of different
architectures, training corpora, and training tasks [11, 27–35]. We found that recurrent neural networks make
markedly human-inconsistent predictions once pitted against transformer-based neural networks. This find-
ing coincides with recent evidence that transformers also outperform recurrent networks for predicting neural
responses as measured by ECoG or fMRI [11, 32], as well as with evidence from model-based prediction
of human reading speed [33, 36] and N400 amplitude [36, 37]. Among the transformers, GPT-2, RoBERTa,
and ELECTRA showed the best performance. These models are trained to optimize only word-level pre-
diction tasks, as opposed to BERT and XLM which are additionally trained on next-sentence prediction and
cross-lingual tasks, respectively (and have the same architecture as RoBERTa). This suggests that local word
prediction provides better alignment with human language comprehension.
Despite the agreement between our results and previous work in terms of model ranking, the significant
failure of GPT-2 in predicting the human responses to natural versus synthetic controversial pairs (Fig. 3b)
demonstrates that GPT-2 does not fully emulate the computations employed in human processing of even short
sentences. This outcome is in some ways unsurprising, given that GPT-2 (like all of the other models we con-
sidered) is an off-the-shelf machine learning model that was not designed with human psycholinguistic and
physiological details in mind. And yet, the considerable human inconsistency we observed seems to stand in
stark contrast with the recent report of GPT-2 explaining about 100 percent of the explainable variance in fMRI
and ECoG responses to natural sentences [32]. Part of this discrepancy could be explained by the fact that
Schrimpf and colleagues [32] mapped GPT-2 hidden-layer activations to brain data by means of regularized
linear regression, which can identify a subspace within GPT-2’s language representation that is well-aligned
with brain responses even if GPT-2’s overall sentence probabilities are not human-like. More importantly,
when language models are evaluated with natural language, strong statistical models might capitalize on fea-
tures in the data that are distinct from, but highly correlated with, features that are meaningful to humans.
Therefore, a model that performs well on typical sentences might employ computational mechanisms that are
very distinct from the brain’s, which will only be revealed by testing the model in a more challenging domain.
Note that even the simplest model we considered—a 2-gram frequency table—actually performed quite well
10on predicting human judgments for randomly-sampled natural sentences, and its deficiencies only became ob-
vious when challenged by controversial sentence pairs. We predict that there will be substantial discrepancies
between neural representations and current language models when using stimuli that intentionally stress-test
this relationship, using our proposed sentence-level controversiality approach or complementary ideas such as
maximizing controversial transition probabilities between consecutive words [38].
Using controversial sentences can be seen as a generalization test of language models: can models predict
what kinds of changes to a natural sentence will lead to humans rejecting the sentence as improbable? Humans
are sometimes capable of comprehending language with atypical constructions (e.g. in cases when pragmatic
judgments can be made about a speaker’s intentions from environmental and linguistic context [39]), but none of
the models we tested were fully able to predict which syntactic or semantic perturbations would be accepted or
rejected by humans. One possibility is that stronger next-word prediction models, using different architectures,
learning rules, or training data, might close the gap between models and humans. Alternatively, it might be
that optimizing for other linguistic tasks, or even non-linguistic task demands (in particular, representing the
external world, the self, and other agents) will turn out to be critical for achieving human-like natural language
processing [40].
3.2 Controversial sentence pairs as adversarial attacks
Machine vision models are highly susceptible to adversarial examples [41, 42]. Such adversarial examples are
typically generated by choosing a correctly classified natural image and then searching for a minuscule (and
therefore human-imperceptible) image perturbation that would change the targeted model’s classification. The
prospect that similar covert model failure modes may exist also for language models has motivated proposed
generalizations of adversarial methods to textual inputs [43]. However, imperceptible perturbations cannot be
applied to written text: any modified word or character is humanly perceptible. Prior work on adversarial
examples for language models have instead relied on heuristic constraints aiming to limit the change in the
meaning of the text, such as flipping a character [44, 45], changing number or gender [46], or replacing words
with synonyms [47–49]. However, since these heuristics are only rough approximations of human language
processing, many of these methods fail to preserve semantic meaning [50]. Interactive (“human-in-the-loop”)
adversarial approaches allow human subjects to repeatedly alter model inputs such that it confuses target models
but not secondary participants [17, 51], but these approaches are inherently slow and costly and are limited by
mental models the human subjects form about the evaluated language models.
By contrast, testing language models on controversial sentence pairs does not require approximating or
querying a human ground truth during optimization—the objective of controversiality is independent of cor-
rectness. Instead, by designing inputs to elicit conflicting predictions among the models and assessing human
responses to these inputs only once the optimization loop has terminated, we capitalize on the simple fact that
if two models disagree with respect to an input, at least one of the models must be making an incorrect predic-
tion. Pitting language models against other language models also can be conducted by other approaches such
as “red-teaming”, where an alternative language model is used as a generator of potential adversarial examples
for a targeted model and a classifier is used to filter the generated examples such that the output they induce
in the targeted model is indeed incorrect [52]. Our approach shares the underlying principle that an alternative
language model can drive a more powerful test than handcrafted heuristics, but here the models have symmetric
roles (there are no “attacking” and “attacked” models) and we can optimize stimuli directly without relying on
filtering.
113.3 Limitations and future directions
While our results demonstrate that using controversial stimuli can identify subtle differences in language mod-
els’ alignment with human judgments, our study was limited in a number of ways. Our stimuli were all 8-word
English sentences, limiting our ability to make cognitively meaningful claims that apply to language use glob-
ally. 8-word sentences are long enough to include common syntactic constructions and convey meaningful
ideas but may not effectively probe long-distance syntactic dependencies [53]. Future work may introduce
additional sentence lengths and languages, as well as (potentially adaptive) controversial sentence optimization
procedures that consider large sets of candidate models, allowing for greater model coverage than our sim-
pler pairwise approach. Future work may also complement the model-comparative experimental design with
procedures designed to identify potential failure modes common to all models.
A more substantial limitation of the current study is that, like any comparison of pre-trained neural networks
as potential models of human cognition, there could be multiple reasons (i.e., training data, architecture, training
tasks, learning rules) why particular models are better aligned with human judgments. For example, as we
did not systematically control the training corpora used for training the models, it is possible that some of
the observed differences are due to differences in the training sets rather than model architecture. Therefore,
while our results expose failed model predictions, they do not readily answer why these failed predictions arise.
Future experiments could compare custom-trained or systematically manipulated models, which reflect specific
hypotheses about human language processing. In Extended Data Fig. 5, we demonstrate the power of using
synthetic controversial stimuli to conduct sensitive comparisons between models with subtle differences in how
sentence probabilities are calculated.
It is important to note that our analyses considered human relative probability judgments as reflecting
a scalar measure of acceptability. We made this assumption in order to bring the language models (which
assign a probability measure to each sentence) and the human participants onto a common footing. However,
it is possible that different types of sentence pairs engage different human cognitive processes. For pairs of
synthetic sentences, both sentences may be unacceptable in different ways (e.g. exhibit different kinds of
grammatical violations), requiring a judgment that weighs the relative importance of multiple dimensions [54]
and could therefore produce inconsistent rankings across participants or across trials [55]. By contrast, asking
participants to compare a natural and a synthetic sentence (Fig. 3b, Extended Data Table 1) may be more
analogous to previous work measuring human acceptability judgments for sentence pairs [24]. Nonetheless, it
is worth noting that for allof the controversial conditions, the noise ceiling was significantly above the models’
prediction accuracy, indicating non-random human preferences unexplained by current models that should be
accounted for by future models, which may have to be more complex and capture multiple processes.
Finally, the use of synthetic controversial sentences can be extended beyond probability judgments. A
sufficiently strong language model may enable constraining the experimental design search-space to particular
sentence distributions (e.g., movie reviews or medical questions). Given such a constrained space, we may
be able to search for well-formed sentences that elicit contradictory predictions in alternative domain-specific
models (e.g., sentiment classifiers or question-answering models). However, as indicated by our results, the
task of capturing distributions of well-formed sentences is less trivial than it seems.
124 Methods
4.1 Language models
We tested nine models from three distinct classes: n-gram models, recurrent neural networks, and transformers.
The n-gram models were trained with open source code from the Natural Language Toolkit [56], the recurrent
neural networks were trained with architectures and optimization procedures available in PyTorch [57], and
the transformers were implemented with the open-source repository HuggingFace [58]. For full details see
Supplementary Section 6.1.
4.2 Evaluating sentence probabilities in transformer models
We then sought to compute the probability of arbitrary sentences under each of the models described above.
The term “sentence” is used in this context in its broadest sense—a sequence of English words, not necessarily
restricted to grammatical English sentences. Unlike some classification tasks in which valid model predictions
may be expected only for grammatical sentences (e.g., sentiment analysis), the sentence probability comparison
task is defined over the entire domain of eight-word sequences.
For the set of unidirectional models, evaluating sentence probabilities was performed simply by summing
the log probabilities of each successive token in the sentence from left to right, given all the previous tokens.
For bidirectional models, this process was not as straightforward. One challenge is that transformer model
probabilities do not necessarily reflect a coherent joint probability; the summed log sentence probability result-
ing from adding words in one order (e.g. left to right) does not necessarily equal the probability resulting from a
different order (e.g. right to left). Here we developed a novel formulation of bidirectional sentence probabilities
in which we considered all permutations of serial word positions as possible construction orders (analogous to
the random word visitation order used to sample serial reproduction chains, [59]). In practice, we observed that
the distribution of log probabilities resulting from different permutations tends to center tightly around a mean
value (for example, for RoBERTa evaluated with natural sentences, the average coefficient of variation was
approximately 0.059). Therefore in order to efficiently calculate bidirectional sentence probability, we evaluate
100 different random permutations and define the overall sentence log probability as the mean log probability
calculated from each permutation. Specifically, we initialized an eight-word sentence with all tokens replaced
with the “mask” token used in place of to-be-predicted words during model training. We selected a random
permutation Pof positions 1 through 8, and started by computing the probability of the word at first of these
positions P1given the other seven “mask” tokens. We then replaced the “mask” at position P1with the actual
word at this position and computed the probability of the word at P2given the other six “mask” tokens and the
word at P1. This process was repeated until all “mask” tokens had been filled by the corresponding word.
A secondary challenge in evaluating sentence probabilities in bidirectional transformer models stems from
the fact that these models use word-piece tokenizers (as opposed to whole words), and that these tokenizers are
different for different models. For example, one tokenizer might include the word “beehive” as a single token,
while others strive for a smaller library of unique tokens by evaluating “beehive” as the two tokens “bee” and
“hive”. The model probability of a multi-token word—similar to the probability of a multi-word sentence—
may depend on the order in which the chain rule is applied. Therefore, all unique permutations of token order
for each multi-token word were also evaluated within their respective “masks”. For example, the probability of
the word “beehive” would be evaluated as follows:
13logp(w=beehive ) =0.5
logp(w1=bee|w2=MASK ) + log p(w2=hive|w1=bee)
+0.5
logp(w2=hive|w1=MASK ) + log p(w1=bee|w2=hive) (1)
This procedure aimed to yield a more fair estimate of the conditional probabilities of word-piece tokens
and therefore the overall probabilities of multi-token words by 1) ensuring that the word-piece tokens were
evaluated within the same context of surrounding words and masks, and 2) eliminating the bias of evaluating
the word-piece tokens in any one particular order in models which were trained to predict bidirectionally.
One more procedure was applied in order to ensure that all models were computing a probability distribution
over sentences with exactly 8 words. When evaluating the conditional probability of a masked word in models
with word-piece tokenizers, we normalized the model probabilities to ensure that only single words were being
considered, rather than splitting the masked tokens into multiple words. At each evaluation step, each token was
restricted to come from one of four normalized distributions: i) single-mask words were restricted to be tokens
with appended white space, ii) masks at the beginning of a word were restricted to be tokens with preceding
white space (in models with preceding white space, e.g. BERT), iii) masks at the end of words were restricted
to be tokens with trailing white space (in models with trailing white space, e.g. XLM), and iv) masks in the
middle of words were restricted to tokens with no appended white space.
4.3 Assessing potential token count effects on sentence probabilities
Note that, because tokenization schemes varied across models, the number of tokens in a sentence could dif-
fer for different models. These alternative tokenizations can be conceived of as different factorizations of the
modeled language distribution, changing how a sentence’s log probability is additively partitioned across the
conditional probability chain (but not affecting its overall probability) [60]. Had we attempted to normalize
across models by dividing the log probability by the number of tokens, as is often done when aligning model
predictions to human acceptability ratings [12, 13], our probabilities would have become strongly tokenization-
dependent [60]. To empirically confirm that tokenization differences were not driving our results, we sta-
tistically compared the token counts of each model’s preferred synthetic sentences with the token counts of
their non-preferred counterparts. While we found significant differences for some of the models, there was
no systematic association between token count and model sentence preferences (Supplementary Table S1). In
particular, lower sentence probabilities were not systematically confounded by higher token counts.
4.4 Defining a shared vocabulary
To facilitate the sampling, selection, and synthesis of sentences that could be evaluated by all of the candidate
models, we defined a shared vocabulary of 29,157unique words. Defining this vocabulary was necessary in
order to unify the space of possible sentences between the transformer models (which can evaluate any input
due to their word-piece tokenizers) and the neural network and n-gram models (which include whole words as
tokens), and to ensure we only included words that were sufficiently prevalent in the training corpora for all
models. The vocabulary consisted of the words in the subtlex database [61], after removing words that occurred
fewer than 300 times in the 300M word corpus (see Supplementary Section 6.1) used to train the n-gram and
recurrent neural network models (i.e., with frequencies lower than one in a million).
144.5 Sampling of natural sentences
Natural sentences were sampled from the same four text sources used to construct the training corpus for the
n-gram and recurrent neural network models, while ensuring that there was no overlap between training and
testing sentences. Sentences were filtered to include only those with eight distinct words and no punctuation
aside from periods, exclamation points, or question marks at the end of a sentence. Then, all eight-word
sentences were further filtered to include only the words included in the shared vocabulary and to exclude those
included in a predetermined list of inappropriate words and phrases. To identify controversial pairs of natural
sentences, we used integer linear programming to search for sentences that had above-median probability in
one model and minimum probability rank in another model (see Supplementary Section 6.2).
4.6 Generating synthetic controversial sentence pairs
For each pair of models, we synthesized 100 sentence triplets. Each triplet was initialized with a natural
sentence n(sampled from Reddit). The words in sentence nwere iteratively modified to generate a synthetic
sentence with reduced probability according to the first model but not according to the second model. This
process was repeated to generate another synthetic sentence from n, in which the roles of the two models
were reversed. Conceptually, this approach resembles Maximum Differentiation (MAD) competition [62],
introduced to compare models of image quality assessment. Each synthetic sentence was generated as a solution
for a constrained minimization problem:
s∗= argmin
slogp(s|mreject )
subject to logp(s|maccept )≥logp(n|maccept )(2)
mreject denotes the model targeted to assign reduced sentence probability to the synthetic sentence compared
to the natural sentence, and maccept denotes the model targeted to maintain a synthetic sentence probability
greater or equal to that of the natural sentence. For one synthetic sentence, one model served as maccept and
the other model served as mreject , and for the other synthetic sentence the model roles were flipped.
At each optimization iteration, we selected one of the eight words pseudorandomly (so that all eight posi-
tions would be sampled Ntimes before any position would be sampled N+ 1times) and searched the shared
vocabulary for the replacement word that would minimize the logp(s|mreject )under the constraint. We ex-
cluded potential replacement words that already appeared in the sentence, except for a list of 42 determiners and
prepositions such as “the”, “a”, or “with”, which were allowed to repeat. The sentence optimization procedure
was concluded once eight replacement attempts (i.e., words for which no loss-reducing replacement has been
found) have failed in a row.
4.7 Word-level search for bidirectional models
For models for which the evaluation of logp(s|m)is computationally cheap (2-gram, 3-gram, LSTM, and
the RNN), we directly evaluated the log-probability of the 29,157sentences resulting from each of the 29,157
possible word replacements. When such probability vectors were available for both models, we simply chose
the replacement minimizing the loss. For GPT-2, whose evaluation is slower, we evaluated sentence probabil-
ities only for word replacements for which the new word had a conditional log-probability (given the previous
words in the sentence) of no less than −10; in rare cases when this threshold yielded fewer than 10 candidate
words, we reduced the threshold in steps of 5until there were at least 10 words above the threshold. For the
15bi-directional models (BERT, RoBERTa, XLM, and ELECTRA), for which the evaluation of logp(s|m)is
costly even for a single sentence, we used a heuristic to prioritize which replacements to evaluate.
Since bi-directional models are trained as masked language models, they readily provide word-level com-
pletion probabilities. These word-level log-probabilities typically have positive but imperfect correlation with
the log-probabilities of the sentences resulting from each potential completion. We hence formed a simple
linear regression-based estimate of logp(s{i} ← w|m), the log-probability of the sentence swith word w
assigned at position i, predicting it from logp(s{i}=w|m, s{i} ←mask ), the completion log-probability
of word wat position i, given the sentence with the i-th word masked:
log ˆp(s{i} ←w|m) =β1logp(s{i}=w|m, s{i} ←mask ) +β0 (3)
This regression model was estimated from scratch for each word-level search. When a word was first
selected for being replaced, the log-probability of two sentences was evaluated: the sentence resulting from
substituting the existing word with the word with the highest completion probability and the sentence resulting
from substituting the existing word with the word with the lowest completion probability. These two word-
sentence log-probability pairs, as well as the word-sentence log-probability pair pertaining to the current word,
were used to fit the regression line. The regression prediction, together with the sentence probability for the
other model (either the exact probability, or approximate probability if the other model was also bi-directional)
was used to predict logp(s|mreject )for each of the 29,157potential replacements. We then evaluated the
true (non-approximate) sentence probabilities of the replacement word with the minimal predicted probability.
If this word indeed reduced the sentence probability, it was chosen to serve as the replacement and the word-
level search was terminated (i.e., proceeding to search a replacement for another word in the sentence). If it
did not reduce the probability, the regression model (Eq. 3) was updated with the new observation, and the
next replacement expected to minimize the sentence probability was evaluated. This word-level search was
terminated after five sentence evaluations that did not reduce the loss.
4.8 Selecting the best triplets from the optimized sentences
Since the discrete hill-climbing procedure described above is highly local, the degree to which this succeeded in
producing highly-controversial pairs varied depending on the starting sentence n. We found that typically, nat-
ural sentences with lower than average log-probability gave rise to synthetic sentences with greater controver-
siality. To better represent the distribution of natural sentences while still choosing the best (most controversial)
triplets for human testing, we used stratified selection.
First, we quantified the controversiality of each triplet as
cm1,m2(n, s1, s2) = logp(n|m1)
p(s1|m1)+ logp(n|m2)
p(s2|m2), (4)
where s1is the sentence generated to reduce the probability in model m1ands2is the sentence generated to
reduce the probability in model m2.
We employed integer programming to choose the 10 most controversial triplets from the 100 triplets opti-
mized for each model pair (maximizing the total controversiality across the selected triplets), while ensuring
that for each model, there was exactly one natural sentence in each decile of the natural sentences probability
distribution. The selected 10 synthetic triplets were then used to form 30 unique experimental trials per model
pair, comparing the natural sentence with one synthetic sentence, comparing the natural sentence with the other
synthetic sentence, and comparing the two synthetic sentences.
164.9 Design of the human experiment
Our experimental procedures were approved by the Columbia University Institutional Review Board (protocol
number IRB-AAAS0252) and were performed in accordance with the approved protocol. All participants
provided informed consent prior. We presented the controversial sentence pairs selected and synthesized by
the language models to 100 native English-speaking, US-based participants (55 male) recruited from Prolific
(www.prolific .co), and paid each participant $5.95. The average participant age was 34.08 ± 12.32. The
subjects were divided into 10 groups, and each ten-subject group was presented with a unique set of stimuli.
Each stimulus set contained exactly one sentence pair from every possible combination of model pairs and
the four main experimental conditions: selected controversial sentence pairs; natural vs. synthetic pair, where
one model served as maccept and the other as mreject ; a natural vs. synthetic pair with the reverse model
role assignments; and directly pairing the two synthetic sentences. These model-pair-condition combinations
accounted for 144 (36 ×4) trials of the task. In addition to these trials, each stimulus set also included nine trials
consisting of sentence pairs randomly sampled from the database of eight-word sentences (and not already
included in any of the other conditions). All subjects also viewed 12 control trials consisting of a randomly
selected natural sentence and the same natural sentence with the words scrambled in a random order. The order
of trials within each stimulus set as well as the left-right screen position of sentences in each sentence pair
were randomized for all participants. While each sentence triplet produced by the optimization procedure (see
subsection “Generating synthetic controversial sentence pairs”) gave rise to three trials, these were allocated
such that no subject viewed the same sentence twice.
On each trial of the task, participants were asked to make a binary decision about which of the two sentences
they considered more probable (for the full set of instructions given to participants, see Supplementary Fig. S2).
In addition, they were asked to indicate one of three levels of confidence in their decision: somewhat confident,
confident, or very confident. The trials were not timed, but a 90-minute time limit was enforced for the whole
experiment. A progress bar at the bottom of the screen indicated to participants how many trials they had
completed and had remaining to complete.
We rejected the data of 21 participants who failed to choose the original, unshuffled sentence in at least
11 of the 12 control trials, and acquired data from 21 alternative participants instead, all of whom passed this
data-quality threshold. In general, we observed high agreement in sentence preferences among our participants,
though the level of agreement varied across conditions. There was complete or near-complete agreement (at
least 9/10 participants with the same binary sentence preference) in 52.2% of trials for randomly-sampled
natural-sentence pairs, 36.6% of trials for controversial natural-sentence pairs, 67.6% of trials for natural-
synthetic pairs, and 60.0% of trials for synthetic-synthetic pairs (versus a chance rate of 1.1%, assuming a
binomial distribution with p= 0.5).
4.10 Evaluation of model-human consistency
To measure the alignment on each trial between model judgments and human judgments, we binarized both
measures; we determined which of the two sentences was assigned with a higher probability by the model,
regardless of the magnitude of the probability difference, and which of the two sentences was favored by the
subject, regardless of the reported confidence level. When both the subject and the model chose the same
sentence, the trial was considered as correctly predicted by that model. This correctness measure was averaged
across sentence pairs and across the 10 participants who viewed the same set of trials. For the lower bound on
the noise ceiling, we predicted each subject’s choices from a majority vote of the nine other subjects who were
presented with the same trials. For the upper bound (i.e., the highest possible accuracy attainable on this data
17sample), we included the subject themselves in this majority vote-based prediction.
Since each of the 10 participant groups viewed a unique trial set, these groups provided 10 independent
replications of the experiment. Models were compared to each other and to the lower bound of the noise
ceiling by a Wilcoxon signed-rank test using these 10 independent accuracy outcomes as paired samples. For
each analysis, the false discovery rate across multiple comparisons was controlled by the Benjamini-Hochberg
procedure [63].
In Fig. 4, we instead measure model-human consistency in a more continuous way, comparing the sentence
probability ratio in a model to the graded Likert ratings provided by humans; see Supplementary Section 6.3
for full details.
4.11 Selecting trials for model evaluation
All of the randomly sampled natural-sentence pairs (Fig. 1a) were evaluated for each of the candidate models.
Controversial sentence pairs (either natural, Fig. 1b or synthetic, Fig. 3) were included in a model’s evaluation
set only if they were formed to target that model specifically. The overall summary analysis (Fig. 4) evaluated
all models on all available sentence pairs.
4.12 Comparison to pseudo-log-likelihood acceptability measures
Wang & Cho [64] proposed an alternative approach for computing sentence probabilities in bidirectional
(BERT-like) models, using a pseudo-log-likelihood measure that simply sums the log-probability of each token
conditioned on all of the other tokens in the sentence. While this measure does not reflect a true probability
distribution [65], it is positively correlated with human acceptability judgments for several bidirectional models
[13, 66]. To directly compare this existing approach to our novel method for computing probabilities, we again
used the method of controversial sentence pairs to identify the approach most aligned with human judgments.
For each bidirectional model (BERT, RoBERTa, and ELECTRA), we created two copies of the model, each us-
ing a different approach for computing sentence probabilities. We synthesized 40 sentence pairs to maximally
differentiate between the two copies of each model, with each copy assigning a higher probability to a different
sentence in the pair. Subsequently, we tested 30 human participants, presenting each participant with all 120
sentence pairs. Model-human consistency was quantified as in the main experiment.
4.13 Data and code availability
The experimental stimuli, detailed behavioral testing results, sentence optimization code, and code for repro-
ducing all analyses and figures are available at github.com/dpmlab/contstimlang [67].
Acknowledgments
This material is based upon work partially supported by the National Science Foundation under Grant No.
1948004 to NK. This publication was made possible with the support of the Charles H. Revson Foundation to
TG. The statements made and views expressed, however, are solely the responsibility of the authors.
18Author Contributions
T.G., M.S., N.K., and C.B. designed the study. M.S. implemented the computational models and T.G. imple-
mented the sentence pair optimization procedures. M.S. conducted the behavioral experiments. T.G. and M.S.
analyzed the experiments’ results. T.G., M.S., N.K., and C.B. wrote the paper.
Competing Interests
The authors declare no competing interests.
References
1. Rumelhart, D. E., Hinton, G. E. & Williams, R. J. Learning representations by back-propagating errors.
Nature 323, 533–536. doi: 10.1038/323533a0 (1986).
2. Hochreiter, S. & Schmidhuber, J. Long Short-Term Memory. Neural Computation 9, 1735–1780. doi: 10.
1162/neco.1997.9.8.1735 (1997).
3. Devlin, J., Chang, M., Lee, K. & Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers
for Language Understanding inProceedings of the 2019 Conference of the North American Chapter of
the Association for Computational Linguistics: Human Language Technologies (Minneapolis, MN, USA,
2019), 4171–4186. doi: 10.18653/v1/n19-1423 .
4. Liu, Y ., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L. & Stoyanov,
V . RoBERTa: A Robustly Optimized BERT Pretraining Approach. Preprint at https://arxiv.org/
abs/1907.11692 (2019).
5. Conneau, A. & Lample, G. Cross-lingual Language Model Pretraining inAdvances in Neural Information
Processing Systems 32 (Vancouver, BC, Canada, 2019). URL:proceedings.neurips.cc/paper/
2019/file/c04c19c2c2474dbf5f7ac4372c5b9af1-Paper.pdf .
6. Clark, K., Luong, M., Le, Q. V . & Manning, C. D. ELECTRA: Pre-training Text Encoders as Discrimina-
tors Rather Than Generators in8th International Conference on Learning Representations, ICLR 2020
(Online, 2020). URL:openreview.net/forum?id=r1xMH1BtvB .
7. Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I., et al. Language models are un-
supervised multitask learners 2019. URL:cdn.openai.com/better- language- models/
language_models_are_unsupervised_multitask_learners.pdf .
8. Goodkind, A. & Bicknell, K. Predictive power of word surprisal for reading times is a linear function of
language model quality inProceedings of the 8th Workshop on Cognitive Modeling and Computational
Linguistics (CMCL 2018) (Salt Lake City, Utah, 2018), 10–18. doi: 10.18653/v1/W18-0102 .
9. Shain, C., Blank, I. A., van Schijndel, M., Schuler, W. & Fedorenko, E. fMRI reveals language-specific
predictive coding during naturalistic sentence comprehension. Neuropsychologia 138, 107307. doi: 10.
1016/j.neuropsychologia.2019.107307 (2020).
10. Broderick, M. P., Anderson, A. J., Di Liberto, G. M., Crosse, M. J. & Lalor, E. C. Electrophysiologi-
cal correlates of semantic dissimilarity reflect the comprehension of natural, narrative speech. Current
Biology 28, 803–809. doi: 10.1016/j.cub.2018.01.080 (2018).
1911. Goldstein, A., Zada, Z., Buchnik, E., Schain, M., Price, A., Aubrey, B., Nastase, S. A., Feder, A., Emanuel,
D., Cohen, A., Jansen, A., Gazula, H., Choe, G., Rao, A., Kim, C., Casto, C., Fanda, L., Doyle, W.,
Friedman, D., Dugan, P., Melloni, L., Reichart, R., Devore, S., Flinker, A., Hasenfratz, L., Levy, O.,
Hassidim, A., Brenner, M., Matias, Y ., Norman, K. A., Devinsky, O. & Hasson, U. Shared computational
principles for language processing in humans and deep language models. Nature Neuroscience 25, 369–
380. doi: 10.1038/s41593-022-01026-4 (2022).
12. Lau, J. H., Clark, A. & Lappin, S. Grammaticality, Acceptability, and Probability: A Probabilistic View
of Linguistic Knowledge. Cognitive Science 41, 1202–1241. doi: 10.1111/cogs.12414 (2017).
13. Lau, J. H., Armendariz, C., Lappin, S., Purver, M. & Shu, C. How Furiously Can Colorless Green Ideas
Sleep? Sentence Acceptability in Context. Transactions of the Association for Computational Linguistics
8, 296–310. doi: 10.1162/tacl_a_00315 (2020).
14. Wang, A., Singh, A., Michael, J., Hill, F., Levy, O. & Bowman, S. R. GLUE: A Multi-Task Benchmark
and Analysis Platform for Natural Language Understanding in7th International Conference on Learning
Representations, ICLR 2019, (New Orleans, LA, USA, 2019). URL:openreview.net/forum?id=
rJ4km2R5t7 .
15. Wang, A., Pruksachatkun, Y ., Nangia, N., Singh, A., Michael, J., Hill, F., Levy, O. & Bowman, S. Su-
perGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems inAdvances
in Neural Information Processing Systems 32 (Vancouver, BC, Canada, 2019). URL:proceedings.
neurips.cc/paper/2019/file/4496bf24afe7fab6f046bf4923da8de6-Paper.pdf .
16. Warstadt, A., Parrish, A., Liu, H., Mohananey, A., Peng, W., Wang, S.-F. & Bowman, S. R. BLiMP: The
Benchmark of Linguistic Minimal Pairs for English. Transactions of the Association for Computational
Linguistics 8, 377–392. doi: 10.1162/tacl_a_00321 (2020).
17. Kiela, D., Bartolo, M., Nie, Y ., Kaushik, D., Geiger, A., Wu, Z., Vidgen, B., Prasad, G., Singh, A., Ring-
shia, P., Ma, Z., Thrush, T., Riedel, S., Waseem, Z., Stenetorp, P., Jia, R., Bansal, M., Potts, C. & Williams,
A.Dynabench: Rethinking Benchmarking in NLP inProceedings of the 2021 Conference of the North
American Chapter of the Association for Computational Linguistics: Human Language Technologies
(Online, 2021), 4110–4124. doi: 10.18653/v1/2021.naacl-main.324 .
18. Box, G. E. & Hill, W. J. Discrimination Among Mechanistic Models. Technometrics 9, 57–71. doi: 10.
1080/00401706.1967.10490441 (1967).
19. Golan, T., Raju, P. C. & Kriegeskorte, N. Controversial stimuli: Pitting neural networks against each other
as models of human cognition. Proceedings of the National Academy of Sciences 117, 29330–29337.
doi:10.1073/pnas.1912334117 (2020).
20. Cross, D. V . Sequential dependencies and regression in psychophysical judgments. Perception & Psy-
chophysics 14, 547–552. doi: 10.3758/BF03211196 (1973).
21. Foley, H. J., Cross, D. V . & O’reilly, J. A. Pervasiveness and magnitude of context effects: Evidence
for the relativity of absolute magnitude estimation. Perception & Psychophysics 48, 551–558. doi: 10.
3758/BF03211601 (1990).
22. Petzschner, F. H., Glasauer, S. & Stephan, K. E. A Bayesian perspective on magnitude estimation. Trends
in Cognitive Sciences 19, 285–293. doi: 10.1016/j.tics.2015.03.002 (2015).
23. Greenbaum, S. Contextual Influence on Acceptability Judgments. Linguistics 15. doi: 10.1515/ling.
1977.15.187.5 (1977).
2024. Sch ¨utze, C. T. & Sprouse, J. in (eds Podesva, R. J. & Sharma, D.) 27–50 (Cambridge University Press,
Cambridge, 2014). doi: 10.1017/CBO9781139013734.004 .
25. Sprouse, J. & Almeida, D. Design sensitivity and statistical power in acceptability judgment experiments.
Glossa 2, 14. doi: 10.5334/gjgl.236 (2017).
26. Lindsay, G. W. Convolutional Neural Networks as a Model of the Visual System: Past, Present, and
Future. Journal of Cognitive Neuroscience 33, 2017–2031. doi: 10.1162/jocn_a_01544 (2021).
27. Wehbe, L., Vaswani, A., Knight, K. & Mitchell, T. Aligning context-based statistical models of language
with brain activity during reading inProceedings of the 2014 Conference on Empirical Methods in Nat-
ural Language Processing (EMNLP) (Doha, Qatar, 2014), 233–243. doi: 10.3115/v1/D14-1030 .
28. Toneva, M. & Wehbe, L. Interpreting and improving natural-language processing (in machines) with
natural language-processing (in the brain) inAdvances in Neural Information Processing Systems 32
(Vancouver, BC, Canada, 2019). URL:proceedings . neurips . cc / paper / 2019 / file /
749a8e6c231831ef7756db230b4359c8-Paper.pdf .
29. Heilbron, M., Armeni, K., Schoffelen, J.-M., Hagoort, P. & De Lange, F. P. A hierarchy of linguistic
predictions during natural language comprehension. Proceedings of the National Academy of Sciences
119, e2201968119. doi: 10.1073/pnas.2201968119 (2022).
30. Jain, S., V o, V ., Mahto, S., LeBel, A., Turek, J. S. & Huth, A. Interpretable multi-timescale models for
predicting fMRI responses to continuous natural speech inAdvances in Neural Information Processing
Systems 33 (Online, 2020), 13738–13749. URL:proceedings.neurips.cc/paper_files/
paper/2020/file/9e9a30b74c49d07d8150c8c83b1ccf07-Paper.pdf .
31. Lyu, B., Marslen-Wilson, W. D., Fang, Y . & Tyler, L. K. Finding structure in time: Humans, machines,
and language. bioRxiv. Preprint at https://www.biorxiv.org/content/10.1101/2021.
10.25.465687v2 (2021).
32. Schrimpf, M., Blank, I. A., Tuckute, G., Kauf, C., Hosseini, E. A., Kanwisher, N., Tenenbaum, J. B. &
Fedorenko, E. The neural architecture of language: Integrative modeling converges on predictive pro-
cessing. Proceedings of the National Academy of Sciences 118, e2105646118. doi: 10.1073/pnas.
2105646118 (2021).
33. Wilcox, E., Vani, P. & Levy, R. A Targeted Assessment of Incremental Processing in Neural Language
Models and Humans inProceedings of the 59th Annual Meeting of the Association for Computational
Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1:
Long Papers) (Online, 2021), 939–952. doi: 10.18653/v1/2021.acl-long.76 .
34. Caucheteux, C. & King, J.-R. Brains and algorithms partially converge in natural language processing.
Communications Biology 5, 134. doi: 10.1038/s42003-022-03036-1 (2022).
35. Arehalli, S., Dillon, B. & Linzen, T. Syntactic Surprisal From Neural Models Predicts, But Underesti-
mates, Human Processing Difficulty From Syntactic Ambiguities inProceedings of the 26th Conference
on Computational Natural Language Learning (CoNLL) (Abu Dhabi, United Arab Emirates (Hybrid),
2022), 301–313. doi: 10.18653/v1/2022.conll-1.20 .
36. Merkx, D. & Frank, S. L. Human Sentence Processing: Recurrence or Attention? Proceedings of the
Workshop on Cognitive Modeling and Computational Linguistics. doi:10.18653/v1/2021.cmcl-
1.2.URL:dx.doi.org/10.18653/v1/2021.cmcl-1.2 (2021).
2137. Michaelov, J. A., Bardolph, M. D., Coulson, S. & Bergen, B. K. Different kinds of cognitive plausibility:
why are transformers better than RNNs at predicting N400 amplitude? inProceedings of the Annual Meet-
ing of the Cognitive Science Society 43 (2021). URL:escholarship.org/uc/item/9z06m20f .
38. Rakocevic, L. I. Synthesizing controversial sentences for testing the brain-predictivity of language models
https://hdl.handle.net/1721.1/130713 . PhD thesis (Massachusetts Institute of Technol-
ogy, 2021).
39. Goodman, N. D. & Frank, M. C. Pragmatic Language Interpretation as Probabilistic Inference. Trends in
Cognitive Sciences 20, 818–829. doi: 10.1016/j.tics.2016.08.005 (2016).
40. Howell, S. R., Jankowicz, D. & Becker, S. A model of grounded language acquisition: Sensorimotor
features improve lexical and grammatical learning. Journal of Memory and Language 53, 258–276.
doi:https://doi.org/10.1016/j.jml.2005.03.002 (2005).
41. Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I. & Fergus, R. Intriguing
properties of neural networks Preprint at http://arxiv.org/abs/1312.6199 . 2013.
42. Goodfellow, I. J., Shlens, J. & Szegedy, C. Explaining and Harnessing Adversarial Examples in3rd
International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015,
Conference Track Proceedings (2015). URL:arxiv.org/abs/1412.6572 .
43. Zhang, W. E., Sheng, Q. Z., Alhazmi, A. & Li, C. Adversarial attacks on deep-learning models in natural
language processing: A survey. ACM Transactions on Intelligent Systems and Technology (TIST) 11, 1–
41. doi: 10.1145/3374217 (2020).
44. Liang, B., Li, H., Su, M., Bian, P., Li, X. & Shi, W. Deep Text Classification Can be Fooled inProceedings
of the Twenty-Seventh International Joint Conference on Artificial Intelligence, IJCAI-18 (Stockholm,
Sweden, 2018), 4208–4215. doi: 10.24963/ijcai.2018/585 .
45. Ebrahimi, J., Rao, A., Lowd, D. & Dou, D. HotFlip: White-Box Adversarial Examples for Text Classifica-
tioninProceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume
2: Short Papers) (Melbourne, Australia, 2018), 31–36. doi: 10.18653/v1/P18-2006 .
46. Abdou, M., Ravishankar, V ., Barrett, M., Belinkov, Y ., Elliott, D. & Søgaard, A. The Sensitivity of Lan-
guage Models and Humans to Winograd Schema Perturbations inProceedings of the 58th Annual Meet-
ing of the Association for Computational Linguistics (Online, 2020), 7590–7604. doi: 10.18653/v1/
2020.acl-main.679 .
47. Alzantot, M., Sharma, Y ., Elgohary, A., Ho, B.-J., Srivastava, M. & Chang, K.-W. Generating Natural
Language Adversarial Examples inProceedings of the 2018 Conference on Empirical Methods in Natural
Language Processing (Brussels, Belgium, 2018), 2890–2896. doi: 10.18653/v1/D18-1316 .
48. Ribeiro, M. T., Singh, S. & Guestrin, C. Semantically Equivalent Adversarial Rules for Debugging NLP
models inProceedings of the 56th Annual Meeting of the Association for Computational Linguistics
(Volume 1: Long Papers) (Melbourne, Australia, 2018), 856–865. doi: 10.18653/v1/P18-1079 .
49. Ren, S., Deng, Y ., He, K. & Che, W. Generating Natural Language Adversarial Examples through Prob-
ability Weighted Word Saliency inProceedings of the 57th Annual Meeting of the Association for Com-
putational Linguistics (Florence, Italy, 2019), 1085–1097. doi: 10.18653/v1/P19-1103 .
50. Morris, J., Lifland, E., Lanchantin, J., Ji, Y . & Qi, Y . Reevaluating Adversarial Examples in Natural
Language inFindings of the Association for Computational Linguistics: EMNLP 2020 (Online, 2020),
3829–3839. doi: 10.18653/v1/2020.findings-emnlp.341 .
2251. Wallace, E., Rodriguez, P., Feng, S., Yamada, I. & Boyd-Graber, J. Trick Me If You Can: Human-in-the-
Loop Generation of Adversarial Examples for Question Answering. Transactions of the Association for
Computational Linguistics 7, 387–401. doi: 10.1162/tacl_a_00279 (2019).
52. Perez, E., Huang, S., Song, F., Cai, T., Ring, R., Aslanides, J., Glaese, A., McAleese, N. & Irving, G.
Red Teaming Language Models with Language Models inProceedings of the 2022 Conference on Em-
pirical Methods in Natural Language Processing (Abu Dhabi, United Arab Emirates, 2022), 3419–3448.
doi:10.18653/v1/2022.emnlp-main.225 .
53. Gibson, E. Linguistic complexity: locality of syntactic dependencies. Cognition 68, 1–76. doi: 10.1016/
S0010-0277(98)00034-1 (1998).
54. Watt, W. C. The indiscreteness with which impenetrables are penetrated. Lingua 37, 95–128. doi: 10.
1016/0024-3841(75)90046-7 (1975).
55. Sch ¨utze, C. T. The empirical base of linguistics. Grammaticality judgments and linguistic methodology
Classics in Linguistics 2.doi:10.17169/langsci.b89.100 (Language Science Press, Berlin,
2016).
56. Bird, S., Klein, E. & Loper, E. Natural language processing with Python: analyzing text with the natural
language toolkit (”O’Reilly Media, Inc.”, Sebastopol, CA, USA, 2009).
57. Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein,
N., Antiga, L., Desmaison, A., Kopf, A., Yang, E., DeVito, Z., Raison, M., Tejani, A., Chilamkurthy,
S., Steiner, B., Fang, L., Bai, J. & Chintala, S. PyTorch: An Imperative Style, High-Performance Deep
Learning Library inAdvances in Neural Information Processing Systems 32 (Vancouver, BC, Canada,
2019), 8024–8035. URL:papers.neurips.cc/paper/9015-pytorch-an-imperative-
style-high-performance-deep-learning-library.pdf .
58. Wolf, T., Debut, L., Sanh, V ., Chaumond, J., Delangue, C., Moi, A., Cistac, P., Rault, T., Louf, R., Fun-
towicz, M., Davison, J., Shleifer, S., von Platen, P., Ma, C., Jernite, Y ., Plu, J., Xu, C., Scao, T. L., Gugger,
S., Drame, M., Lhoest, Q. & Rush, A. M. Transformers: State-of-the-Art Natural Language Processing
inProceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System
Demonstrations (Online, 2020), 38–45. doi: 10.18653/v1/2020.emnlp-demos.6 .
59. Yamakoshi, T., Griffiths, T. & Hawkins, R. Probing BERT’s priors with serial reproduction chains in
Findings of the Association for Computational Linguistics: ACL 2022 (Dublin, Ireland, 2022), 3977–
3992. doi: 10.18653/v1/2022.findings-acl.314 .
60. Chestnut, S. Perplexity Accessed: 2022-09-23. 2019. URL:drive.google.com/uc?export=
download&id=1gSNfGQ6LPxlNctMVwUKrQpUA7OLZ83PW .
61. Van Heuven, W. J. B., Mandera, P., Keuleers, E. & Brysbaert, M. Subtlex-UK: A New and Improved Word
Frequency Database for British English. Quarterly Journal of Experimental Psychology 67, 1176–1190.
doi:10.1080/17470218.2013.850521 (2014).
62. Wang, Z. & Simoncelli, E. P. Maximum differentiation (MAD) competition: A methodology for compar-
ing computational models of perceptual quantities. Journal of Vision 8, 8–8. doi: 10.1167/8.12.8
(2008).
63. Benjamini, Y . & Hochberg, Y . Controlling the False Discovery Rate: A Practical and Powerful Approach
to Multiple Testing. Journal of the Royal Statistical Society: Series B (Methodological) 57, 289–300.
doi:10.1111/j.2517-6161.1995.tb02031.x (1995).
2364. Wang, A. & Cho, K. BERT has a Mouth, and It Must Speak: BERT as a Markov Random Field Language
Model inProceedings of the Workshop on Methods for Optimizing and Evaluating Neural Language
Generation (Minneapolis, Minnesota, 2019), 30–36. doi: 10.18653/v1/W19-2304 .
65. Cho, K. BERT has a Mouth and must Speak, but it is not an MRF kyunghyuncho.me/bert-has-
a-mouth-and-must-speak-but-it-is-not-an-mrf/ . Accessed: 2022-09-28. 2019.
66. Salazar, J., Liang, D., Nguyen, T. Q. & Kirchhoff, K. Masked Language Model Scoring inProceedings
of the 58th Annual Meeting of the Association for Computational Linguistics (Online, 2020), 2699–2712.
doi:10.18653/v1/2020.acl-main.240 .
67. Golan, T., Siegelman, M., Kriegeskorte, N. & Baldassano, C. Code and data for ”Testing the limits of
natural language models for predicting human language judgments” version 1.2.2. 2023. doi: 10.5281/
zenodo.8147166 .
68. Shannon, C. E. A mathematical theory of communication. The Bell System Technical Journal 27, 379–
423. doi: 10.1002/j.1538-7305.1948.tb01338.x (1948).
69. Irvine, A., Langfus, J. & Callison-Burch, C. The American Local News Corpus inProceedings of the Ninth
International Conference on Language Resources and Evaluation (LREC’14) (Reykjavik, Iceland, 2014),
1305–1308. URL:www.lrec-conf.org/proceedings/lrec2014/pdf/914_Paper.pdf .
70. Kneser, R. & Ney, H. Improved backing-off for m-gram language modeling in1995 international confer-
ence on acoustics, speech, and signal processing 1 (1995), 181–184. doi: 10.1109/ICASSP.1995.
479394 .
71. Gurobi Optimization, LLC. Gurobi Optimizer Reference Manual 2021. URL:www.gurobi.com .
72. Woodbury, M. A. Rank Correlation when There are Equal Variates. The Annals of Mathematical Statistics
11, 358–362. URL:www.jstor.org/stable/2235684 (1940).
73. Sch ¨utt, H. H., Kipnis, A. D., Diedrichsen, J. & Kriegeskorte, N. Statistical inference on representational
geometries. eLife 12 (eds Serences, J. T. & Behrens, T. E.) e82566. doi: 10.7554/eLife.82566
(2023).
74. Nili, H., Wingfield, C., Walther, A., Su, L., Marslen-Wilson, W. & Kriegeskorte, N. A Toolbox for Rep-
resentational Similarity Analysis. PLOS Computational Biology 10, 1–11. doi: 10.1371/journal.
pcbi.1003553 (2014).
75. Pennington, J., Socher, R. & Manning, C. GloVe: Global Vectors for Word Representation inProceedings
of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) (Doha, Qatar,
2014), 1532–1543. doi: 10.3115/v1/D14-1162 .
76. Frank, S. L. & Willems, R. M. Word predictability and semantic similarity show distinct patterns of
brain activity during language comprehension. Language, Cognition and Neuroscience 32, 1192–1203.
doi:10.1080/23273798.2017.1323109 (2017).
5 Extended Data
24Extended Data Figure 1: An example of one experimental trial, as presented to the participants . The participant must
choose one sentence while providing their confidence rating on a 3-point scale.
GPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-grammodel 2
GPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-grammodel 10.93 0.88 0.86 0.83 0.87 0.76 0.82 0.78
0.93 0.90 0.83 0.88 0.89 0.80 0.84 0.82
0.88 0.90 0.87 0.84 0.90 0.83 0.90 0.86
0.86 0.83 0.87 0.84 0.83 0.77 0.83 0.77
0.83 0.88 0.84 0.84 0.86 0.81 0.79 0.77
0.87 0.89 0.90 0.83 0.86 0.87 0.91 0.84
0.76 0.80 0.83 0.77 0.81 0.87 0.84 0.89
0.82 0.84 0.90 0.83 0.79 0.91 0.84 0.91
0.78 0.82 0.86 0.77 0.77 0.84 0.89 0.91
0.00 0.25 0.50 0.75 1.00
between model agreement rate
(proportion of sentence pairs)
Extended Data Figure 2: Between-model agreement rate
on the probability ranking of the 90 randomly sampled
and paired natural sentence pairs evaluated in the exper-
iment . Each cell represents the proportion of sentence pairs
for which two models make congruent probability ranking
(i.e., both models assign a higher probability to sentence 1,
or both models assign a higher probability to sentence 2).
25GPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-grammodel 2
GPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-grammodel 10.51 0.53 0.78 0.83 0.66 0.84 0.79 0.88
0.49 0.60 0.65 0.78 0.60 0.80 0.76 0.86
0.47 0.40 0.63 0.75 0.57 0.79 0.64 0.59
0.22 0.35 0.37 0.58 0.30 0.44 0.60 0.54
0.17 0.22 0.25 0.42 0.47 0.49 0.42 0.55
0.34 0.40 0.43 0.70 0.53 0.84 0.62 0.70
0.16 0.20 0.21 0.56 0.51 0.16 0.49 0.62
0.21 0.24 0.36 0.40 0.58 0.38 0.51 0.63
0.12 0.14 0.41 0.46 0.45 0.30 0.38 0.37
0.00 0.25 0.50 0.75 1.00
human choice aligned with model 1
(proportion of trials)
GPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-grammodel 2
GPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-grammodel 10.55 0.61 0.79 0.84 0.85 0.86 0.76 0.92
0.45 0.36 0.60 0.89 0.81 0.96 0.94 0.95
0.39 0.64 0.49 0.88 0.89 0.95 0.98 0.95
0.21 0.40 0.51 0.71 0.73 0.86 0.78 0.90
0.16 0.11 0.12 0.29 0.51 0.70 0.72 0.91
0.15 0.19 0.11 0.27 0.49 0.79 0.65 0.71
0.14 0.04 0.05 0.14 0.30 0.21 0.32 0.55
0.24 0.06 0.02 0.22 0.28 0.35 0.68 0.67
0.08 0.05 0.05 0.10 0.09 0.29 0.45 0.33
0.00 0.25 0.50 0.75 1.00
human choice aligned with model 1
(proportion of trials)aNatural controversial sentences bSynthetic controversial sentencesExtended Data Figure 3: Pairwise model comparison of model-human consistency. For each pair of models (represented
as one cell in the matrices above), the only trials considered were those in which the stimuli were either selected (a) or
synthesized (b) to contrast the predictions of the two models. For these trials, the two models always made controversial
predictions (i.e., one sentence is preferred by the first model and the other sentence is preferred by the second model). The
matrices above depict the proportion of trials in which the binarized human judgments aligned with the row model (“model
1”). For example, GPT-2 (top-row) was always more aligned (green hues) with the human choices than its rival models. In
contrast, 2-gram (bottom-row) was always less aligned (purple hues) with the human choices than its rival models.
GPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-grammodels assigned as mreject
GPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-grammodels assigned as maccept0.21 0.24 0.43 0.23 0.29 0.36 0.17 0.22
0.21 0.16 0.29 0.51 0.29 0.23 0.18 0.39
0.17 0.20 0.27 0.34 0.27 0.40 0.27 0.20
0.14 0.18 0.27 0.44 0.23 0.25 0.33 0.20
0.07 0.07 0.11 0.16 0.17 0.13 0.08 0.22
0.03 0.01 0.07 0.04 0.17 0.06 0.16 0.04
0.05 0.03 0.05 0.03 0.03 0.01 0.04 0.06
0.04 0.03 0.02 0.08 0.05 0.04 0.06 0.04
0.01 0.02 0.01 0.02 0.04 0.01 0.07 0.04
0.00 0.25 0.50 0.75 1.00
humans choice aligned with maccept
(proportion of trials)
Extended Data Figure 4: Pairwise model analysis of human re-
sponse for natural vs. synthetic sentence pairs. In each optimiza-
tion condition, a synthetic sentence swas formed by modifying a nat-
ural sentence nso the synthetic sentence would be “rejected” by one
model ( mreject , columns), minimizing p(s|mreject ), and would be
“accepted” by another model ( maccept , rows), satisfying the constraint
p(s|maccept )≥p(n|maccept ). Each cell above summarizes model-
human agreement in trials resulting from one such optimization condi-
tion. The color of each cell denotes the proportion of trials in which
humans judged a synthetic sentence to be more likely than its natural
counterpart and hence aligned with maccept . For example, the top-right
cell depicts human judgments for sentence pairs formed to minimize
the probability assigned to the synthetic sentence by the simple 2-gram
model while ensuring that GPT-2 would judge the synthetic sentence
to be at least as likely as the natural sentence; humans favored the syn-
thetic sentence in only 22 out the 100 sentence pairs in this condition.
261.0
0.5
0.0 0.5 1.0
ordinal correlation between human ratings and models'
sentence pair probability log-ratio (signed-rank cosine similarity)RoBERTa
RoBERTa (PLL)
ELECTRA
ELECTRA (PLL)
BERT
BERT (PLL)
aRandomly sampled natural-sentence pairs
1.0
0.5
0.0 0.5 1.0
ordinal correlation between human ratings and models'
sentence pair probability log-ratio (signed-rank cosine similarity)RoBERTa
RoBERTa (PLL)
ELECTRA
ELECTRA (PLL)
BERT
BERT (PLL)
bSynthetic controversial sentence pairsExtended Data Figure 5: Human consistency of bidirectional transformers: approximate log-likelihood versus pseudo-
log-likelihood (PLL). Each dot in the plots above depicts the ordinal correlation between the judgments of one participant
and the predictions of one model. (a)The performance of BERT, RoBERTa, and ELECTRA in predicting the human judg-
ments of randomly sampled natural sentence pairs in the main experiment, using two different likelihood measures: our novel
approximate likelihood method (i.e., averaging multiple conditional probability chains, see Methods) and pseudo-likelihood
(PLL, summating the probability of each word given all of the other words [64]). For each model, we statistically com-
pared the two likelihood measures to each other and to the noise ceiling using a two-sided Wilcoxon signed-rank test across
the participants. False discovery rate was controlled at q <0.05for the 9 comparisons. When predicting human pref-
erences of natural sentences, the pseudo-log-likelihood measure is at least as accurate as our proposed approximate
log-likelihood measure. (b) Results from a follow-up experiment, in which we synthesized synthetic sentence pairs for each
of the model pairs, pitting the two alternative likelihood measures against each other. Statistical testing was conducted in the
same fashion as in panel a. These results indicate that for each of the three bidirectional language models, the approximate
log-likelihood measure is considerably and significantly ( q <0.05) more human-consistent than the pseudo-likelihood mea-
sure. Synthetic controversial sentence pairs uncover a dramatic failure mode of the pseudo-log-likelihood measure,
which remains covert when the evaluation is limited to randomly-sampled natural sentences. See Extended Data Table
2 for synthetic sentence pair examples.
270 25% 50% 75% 100%
human-choice prediction accuracyGPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-gram
Extended Data Figure 6: Model prediction accuracy for pairs of
natural and synthetic sentences, evaluating each model across
all of the sentence pairs in which it was targeted to rate the syn-
thetic sentence to be less probable than the natural sentence.
The data binning applied here is complementary to the one used in
Fig. 3b, where each model was evaluated across all of the sentence
pairs in which it was targeted to rate the synthetic sentence to be
at least as probable as the natural sentence. Unlike Fig. 3b, where
all of the models performed poorly, here no models were found to
be significantly below the lower bound on the noise ceiling; typ-
ically, when a sentence was optimized to decrease its probability
under any model (despite the sentence probability not decreasing
under a second model), humans agreed that the sentence became
less probable.
sentence log probability (model 1) log probability (model 2) # human choices
n: I always cover for him and make excuses. logp(n|GPT-2 ) =−36.46 log p(n|2-gram ) =−106.95 10
s: We either wish for it or ourselves do. logp(s|GPT-2 ) =−36.15 logp(s|2-gram ) =−122.28 0
n: This is why I will never understand boys. logp(n|RoBERTa ) =−46.88 log p(n|2-gram ) =−103.11 10
s: This is why I will never kiss boys. logp(s|RoBERTa ) =−46.75 logp(s|2-gram ) =−107.91 0
n: One of the ones I did required it. logp(n|ELECTRA ) =−35.97 log p(n|LSTM ) =−40.89 10
s: Many of the years I did done so. logp(s|ELECTRA ) =−35.77 logp(s|LSTM ) =−46.25 0
n: There were no guns in the Bronze Age. logp(n|BERT ) =−48.48 log p(n|ELECTRA ) =−30.40 10
s: There is rich finds from the Bronze Age. logp(s|BERT ) =−48.46 logp(s|ELECTRA ) =−44.34 0
n: You did a great job on cleaning them. logp(n|XLM ) =−40.38 log p(n|RNN) =−43.47 10
s: She did a great job at do me. logp(s|XLM ) =−39.89 logp(s|RNN) =−61.03 0
n: This logic has always seemed flawed to me. logp(n|LSTM ) =−39.77 log p(n|RNN) =−45.92 10
s: His cell has always seemed instinctively to me. logp(s|LSTM ) =−38.89 logp(s|RNN) =−62.81 0
s: Stand near the cafe and sip your coffee. logp(s|RNN) =−65.55 log p(s|ELECTRA ) =−34.46 10
n: Sit at the front and break your neck. logp(n|RNN) =−44.18 logp(n|ELECTRA ) =−34.65 0
n: Most of my jobs have been like this. logp(n|3-gram ) =−80.72 log p(n|LSTM ) =−35.07 10
s: One of my boyfriend have been like this. logp(s|3-gram ) =−80.63 logp(s|LSTM ) =−41.44 0
n: They even mentioned that I offer white flowers. logp(n|2-gram ) =−113.38 log p(n|BERT ) =−62.81 10
s: But even fancied that would logically contradictory philosophies. logp(s|2-gram ) =−113.24 logp(s|BERT ) =−117.98 0
Extended Data Table 1: Examples of pairs of synthetic and natural sentences that maximally contributed to each
model’s prediction error. For each model (double row, “model 1”), the table shows results for two sentences on which the
model failed severely. In each case, the failing model 1 prefers synthetic sentence s(higher log probability bolded), while the
model it was pitted against (“model 2”) and all 10 human subjects presented with that sentence pair prefer natural sentence
n. (When more than one sentence pair induced an equal maximal error in a model, the example included in the table was
chosen at random.)
28sentence pseudo-log-likelihood (PLL) approximate log probability # human choices
s1: I found so many in things and called. logp(s1|BERT (PLL) ) =−55.14 log p(s1|BERT ) =−55.89 30
s2: Khrushchev schizophrenic so far
disproportionately goldfish fished alone. logp(s2|BERT (PLL) ) =−22.84 logp(s2|BERT ) =−162.31 0
s1: Figures out if you are on the lead. logp(s1|BERT (PLL) ) =−38.11 log p(s1|BERT ) =−51.27 30
s2: Neighbours unsatisfactory indistinguishable
misinterpreting schizophrenic on homecoming
cheerleading. logp(s2|BERT (PLL) ) =−16.43 logp(s2|BERT ) =−258.91 0
s1: I just say this and not the point. logp(s1|ELECTRA (PLL) ) =−34.41 log p(s1|ELECTRA ) =−33.80 30
s2: Glastonbury reliably mobilize disenfranchised
homosexuals underestimate unhealthy skeptics. logp(s2|ELECTRA (PLL) ) =−11.81 logp(s2|ELECTRA ) =−162.62 0
s1: And diplomacy is more people to the place. logp(s1|ELECTRA (PLL) ) =−62.81 log p(s1|ELECTRA ) =−47.33 30
s2: Brezhnev ingenuity disembarking Acapulco
methamphetamine arthropods unaccompanied
Khrushchev. logp(s2|ELECTRA (PLL) ) =−34.00 logp(s2|ELECTRA ) =−230.97 0
s1: Sometimes what looks and feels real to you. logp(s1|RoBERTa (PLL) ) =−36.58 log p(s1|RoBERTa ) =−51.61 30
s2: Buying something breathes or crawls
aesthetically to decorate. logp(s2|RoBERTa (PLL) ) =−9.78 logp(s2|RoBERTa ) =−110.27 0
s1: In most other high priority packages were affected. logp(s1|RoBERTa (PLL) ) =−71.13 log p(s1|RoBERTa ) =−61.60 30
s2: Stravinsky cupboard nanny contented burglar
babysitting unsupervised bathtub. logp(s2|RoBERTa (PLL) ) =−21.86 logp(s2|RoBERTa ) =−164.70 0
Extended Data Table 2: Examples of controversial synthetic-sentence pairs that maximally contributed to the pre-
diction error of bidirectional transformers using pseudo-log-likelihood (PLL). For each bidirectional model, the table
displays two sentence pairs on which the model failed severely when its prediction was based on pseudo-log-likelihood
(PLL) estimates [64]. In each of these sentence pairs, the PLL estimate favors sentence s2(higher PLL bolded), while the
approximate log-likelihood estimate and most of the human subjects presented with that sentence pair preferred sentence
s1. (When more than one sentence pair induced an equal maximal error in a model, the example included in the table was
chosen at random.) Sentences with long, multi-token words (e.g., “methamphetamine”) have high PLL estimates since
each of their tokens is well predicted by the others tokens. And yet, the entire sentence is improbable according to
human judgments and approximate log-probability estimates based on proper conditional probability chains.
296 Supplementary Methods
6.1 Language models
N-gram models. N-gram models [68], the simplest language model class, are trained by counting the
number of occurrences of all unique phrases of length N words in large text corpora. N-gram models make
predictions about upcoming words by using empirical conditional probabilities in the training corpus. We
tested both 2-gram and 3-gram variants. In 2-gram models, all unique two-word phrases are counted, and
each upcoming word probability (probability of w2conditioned on previous word w1) is determined by
dividing the count of 2-gram w1, w2by the count of unigram (word) w1. In 3-gram models, all unique three-
word phrases are counted, and upcoming word probabilities (probability of w3conditioned on previous
words w1andw2) are determined by dividing the count of 3-gram w1, w2, w3by the count of 2-gram
w1, w2. In both such models, sentence probabilities can be computed as the product of all unidirectional
word transition probabilities in a given sentence. We trained both the 2-gram and 3-gram models on a large
corpus composed of text from four sources: 1. public comments from the social media website Reddit
(reddit .com) acquired using the public API at pushshift .io, 2. articles from Wikipedia, 3. English
books and poetry available for free at Project Gutenberg ( gutenberg .org), and 4. articles compiled in
the American Local News Corpus [69]. The n-gram probability estimates were regularized by means of
Kneser-Ney smoothing [70].
Recurrent neural network models. We also tested two recurrent neural network models, including a
simple recurrent neural network (RNN) [1] and a more complex long short-term memory recurrent neural
network (LSTM) [2]. We trained both of these models on a next word prediction task using the same corpus
used to train the n-gram models. Both the RNN and LSTM had a 256-feature embedding size and a 512-
feature hidden state size, and were trained over 100 independent batches of text for 50 epochs with a learning
rate of .002. Both models’ training sets were tokenized into individual words and consisted of a vocabulary
of94,607unique tokens.
Transformer models. Similar to RNNs, transformers are designed to make predictions about sequential
inputs. However, transformers do not use a recurrent architecture, and have a number of more complex
architectural features. For example, unlike the fixed token embeddings in classic RNNs, transformers utilize
context-dependent embeddings that vary depending on a token’s position. Most transformers also contain
multiple attention heads in each layer of the model, which can help direct the model to relevant tokens in
complex ways. We tested five models with varying architectures and training procedures, including BERT
[3], RoBERTa [4], XLM [5], ELECTRA [6], and GPT-2 [7].
• We used the large version of BERT (bi-directonal encoder representations from transformers), con-
taining 24 encoding layers, 1024 hidden units in the feedforward network element of the model, and
16 attention heads. BERT is a bi-directional model trained to perform two different tasks: 1. a masked
language modeling (MLM) task, in which 15 percent of tokens are replaced with a special [MASK]
token and BERT must predict the masked word, and 2. next sentence prediction (NSP), in which
BERT aims to predict the upcoming sentence in the training corpus given the current sentence.
• RoBERTa is also a bi-directional model that uses the same architecture as BERT. However, RoBERTa
was trained on exclusively the masked word prediction task (and not next sentence prediction), and
used a different optimization procedure (including longer training on a larger dataset). This makes
empirical comparisons between BERT and RoBERTa particularly interesting, because they differ only
in training procedure and not architecture.
30• XLM is a cross-lingual bi-directional model which, too, shares BERT’s original architecture. XLM is
trained on three different tasks: 1. the same MLM task used in both BERT and RoBERTa, 2. a causal
language modeling task where upcoming words are predicted from left to right, and 3. a translation
modeling task. On this task, each training example consists of the same text in two languages, and the
model performs a masked language modeling task using context from one language to predict tokens
of another. Such a task can help the XLM model become robust to idiosyncrasies of one particular
language that may not convey much linguistic information.
• The ELECTRA model uses a training approach that involves two transformer models: a generator
and a discriminator. While the generator performs a masked language modeling task similar to other
transformers, the discriminator simultaneously tries to figure out which masked tokens were replaced
by the generator. This task may be more efficient than pure masked token prediction, because it uses
information from all input tokens rather than only the masked subset.
• GPT-2, the second iteration of GPT OpenAI’s GPT model, is the only unidirectional transformer
model that we tested. We used the pretrained GPT-2-xl version, with 48 encoding layers and 25
attention heads in each layer. Because GPT-2 is unidirectional it was trained only on the causal
language modeling task, in which tokens are predicted from left to right.
6.2 Selection of controversial natural-sentence pairs
We evaluated 231,725eight-word sentences sampled from Reddit. Reddit comments were scraped from
across the entire website and all unique eight-word sentences were saved. These sentences were subse-
quently filtered to exclude blatant spelling errors, inappropriate language, and individual words that were
not included in the corpus used to train the n-gram and recurrent neural network models in our experiment.
We estimated logp(s|m)for each natural sentence sand each model mas described above. We
then rank-transformed the sentence probabilities separately for each model, assigning the fractional rank
r(s|m) = 0 to the least probable sentence according to model mandr(s|m) = 1 to the most probable
one. This step eliminated differences between models in terms of probability calibration.
Next, we aimed to filter this corpus for controversial sentences. To prune the candidate sentences, we
eliminated any sentence sfor which no pair of models m1,m2held(r(s|m1)<0.5)and(r(s|m2)≥0.5),
where r(s|m1)is the fractional rank assigned for sentence sby model m. This step ensured that all of the
remaining sentences had a below-median probability according to one model and above-median probability
according to another, for at least one pair of models. We also excluded sentences in which any word (except
for prepositions) appeared more than once. After this pruning, 85,749candidate sentences remained, from
which85749
2
≈3.67×109possible sentence pairs can be formed.
We aimed to select 360 controversial sentence pairs, devoting 10 sentence pairs to each of the 36
models pairs. First, we defined two 360-long integer vectors m1andm2, specifying for each of the
360 yet unselected sentence pairs which model pair they contrast. We then selected 360 sentence pairs
s1
1, s2
1
,
s1
2, s2
2
...,
s1
360, s2
360
by solving the following minimization problem:
{(s1
j∗, s2
j∗)|j= 1,2, ..360}= argmin
s1,s2X
j
r(s1
j|m1
j) +r(s2
j|m2
j)
(5)
subject to ∀jr(s1
j|m2
j)≥0.5 (5a)
∀jr(s2
j|m1
j)≥0.5 (5b)
All 720 sentences are unique. (5c)
31To achieve this, we used integer linear programming (ILP) as implemented by Gurobi [71]. We repre-
sented sentence allocation as a sparse binary tensor Sof dimensions 85,749 ×360×2 (sentences, trials,
pair members) and the fractional sentence probabilities ranks as a matrix Rof dimensions 85,749 ×9 (sen-
tences, models). This enabled us to express and solve the selection problem in Eq. 5 as a standard ILP
problem:
S∗= argmin
SX
i,jSi,j,1Ri,m1
j+Si,j,2Ri,m2
j(6)
subject to Si,j,1Ri,m2
j≥0.5 (6a)
Si,j,2Ri,m1
j≥0.5 (6b)
∀iX
j,kSi,j,k≤1(each sentence iis used only once in the experiment) (6c)
∀jP
iSi,j,1= 1
∀jP
iSi,j,2= 1)
(each trial jis allocated exactly one sentence pair) (6d)
Sis binary (6e)
6.3 Evaluation of model-human consistency: Correlating model log-probability
ratios to human Likert ratings
For every model mand experimental trial i, we evaluated the log probability ratio for the trial’s two sen-
tences:
LR(s1
i, s2
i|m) = logp(s2
i|m)
p(s1
i|m)(7)
The human Likert ratings were recoded to be symmetrical around zero, mapping the six ratings appear-
ing in Extended Data Fig. 1 to (−2.5,−1.5,−0.5,+0.5,+1.5,+2.5). We then sought to correlate the model
log-ratios and with the zero-centered human Likert ratings, quantifying how well the model log-ratios were
associated with human sentence-likeliness judgments. To allow for an ordinal (not necessarily linear) asso-
ciation between the log-ratios and Likert ratings, we rank-transformed both measures (ranking within each
model or each human) while retaining the sign of the values.
For each participant h:
r(s1
i, s2
i|h) = sign( y0(s1
i, s2
i|h))·R y0(s1
i, s2
i|h) 
, (8)
where y0(s1
i, s2
i|h))is the zero-centered Likert rating provided by subject hfor trial iandR(·)is rank
transform using random tie-breaking.
For each model m:
r(s1
i, s2
i|m) = sign( LR(s1
i, s2
i|m))·R LR(s1
i, s2
i|m) 
, (9)
A valid correlation measure of the model ranks and human ranks must be invariant to whether one
sentence was presented on the left ( s1) and the other on the right ( s2), or vice versa. Changing the sentence
order within a trial would flip the signs of both the log-ratio and the zero-centered Likert rating. Therefore,
the required correlation measure must be invariant to such coordinated sign flips, but not to flipping the sign
of just one of the measures. Since cosine similarity maintains such invariance, we introduced signed-rank
32cosine similarity , an ordinal analog of cosine similarity, substituting the raw data points for signed ranks (as
defined in Eq. 8-9):
SCSR=P
ir(s1
i, s2
i|m)r(s1
i, s2
i|h)qP
ir(s1
i, s2
i|m)2qP
ir(s1
i, s2
i|h)2. (10)
To eliminate the noise contributed by random tie-breaking, we used a closed-form expression of the
expected value of Eq. 10 over different random tie-breaking draws:
E(SCSR) =P
iE
r(s1
i, s2
i|m)
E
r(s1
i, s2
i|h)
pPn
k=1k2pPn
k=1k2=P
i¯r(s1
i, s2
i|m)¯r(s1
i, s2
i|h)Pn
k=1k2, (11)
where ¯r(·)denotes signed rank with average-rank assigned to ties instead of random tie-breaking, and n
denotes the number of evaluated sentence pairs. The expected value of the product in the numerator is
equal to the product of expected values of the factors since the random tie-breaking within each factor is
independent. The vector norms (the factors in the denominator) are constant since given no zero ratings,
each signed-rank rating vector always includes one of each rank 1ton(where nis the number of sentence
pairs considered), and the signs are eliminated by squaring. This derivation follows a classical result for
Spearman’s ρ[72] (see [73], appendix C, for a modern treatment). We empirically confirmed that averaging
SCSR as defined in Eq. 10 across a large number of random tie-breaking draws converges to E(SCSR)as
defined in Eq. 11. This latter expression (whose computation requires no actual random tie-breaking) was
used to quantify the correlation between each participant and model.
For each participant, the lower bound on the noise ceiling was calculated by replacing the model-derived
predictions with an across-participants average of the nine other participants’ signed-rank rating vectors. The
lower bound plotted in main text Fig. 4 is an across-subject average of this estimate. An upper bound on the
noise ceiling was calculated as a dot product between the participant’s expected signed-rank rating vector
(¯ r/pPk2) and a normalized, across-participants average of the expected signed-rank rating vectors of all
10 participants.
Inference was conducted in the same fashion as that employed for the binarized judgments (Wilcoxon
signed-rank tests across the 10 subject groups, controlling for false discovery rate).
7 Supplementary Results
7.1 Randomly sampled natural-sentence pairs fail to adjudicate among mod-
els
As a baseline, we created 90 pairs of natural sentence pairs by randomly sampling from a corpus of 8-word
sentences appearing on Reddit. Evaluating the sentence probabilities assigned to the sentences by the dif-
ferent models, we found that models tended to agree on which of the two sentences was more probable
(Extended Data Fig. 2). The between-model agreement rate ranged from 75.6% of the sentence pairs for
GPT-2 vs. RNN to 93.3% for GPT-2 vs. RoBERTa, with an average agreement between models of 84.5%.
Main text Fig. 1a (left-hand panel) provides a detailed graphical depiction of the relationship between sen-
tence probability ranks for one model pair (GPT-2 and RoBERTa).
We divided these 90 pairs into 10 sets of nine sentences and presented each set to a separate group of 10
subjects. To evaluate model-human alignment, we computed the proportion of trials where the model and the
participant agreed on which sentence was more probable. All of the nine language models performed above
33chance (50% accuracy) in predicting the human choices for the randomly sampled natural sentence pairs
(main text Fig. 1a, right-hand panel). Since we presented each group of 10 participants with a unique set of
sentence pairs, we could statistically test between-model differences while accounting for both participants
and sentence pairs as random factors by means of a simple two-sided Wilcoxon signed-rank test conducted
across the 10 participant groups. For the set of randomly sampled natural-sentence pairs, this test yielded
no significant prediction accuracy differences between the candidate models (controlling for false discovery
rate for all 36 model pairs at q<.05). This result is unsurprising considering the high level of between-
model agreement on the sentence probability ranking within each of these sentence pairs.
To obtain an estimate of the noise ceiling [74] (i.e., the best possible prediction accuracy for this dataset),
we predicted each participant’s choices by the majority vote of the nine other participants who were pre-
sented with the same trials. This measurement provided a lower bound on the noise ceiling. Including the
participant’s own choice in the prediction yields an upper bound, since no set of predictions can be more
human-aligned on average given the between-subject variability. For the randomly sampled natural sen-
tences, none of the models were found to be significantly less accurate than the lower bound on the noise
ceiling (controlling the false discovery rate for all nine models at q<.05). In other words, the 900 trials
of randomly sampled and paired natural sentences provided no statistical evidence that any of the language
models are human-inconsistent.
7.2 Comparing the accuracy of unnormalized and normalized sentence prob-
ability estimates
Previous studies (e.g., [13]) have found that normalizing language model sentence probability estimates by
the sentences’ token counts can result in greater alignment with human acceptability judgments. While we
deliberately used unnormalized sentence log probability when designing the experiment, we evaluated the
prediction accuracy of each model under such normalizations through a control analysis.
Rather than predicting human judgments based only on the relative log probabilities of the two sen-
tences, we instead used cross-validated logistic regression to predict human judgments using a combination
of unnormalized log probability differences (“LP”) and two measures from Lau and colleagues [13] that
incorporate information about the token counts in each sentence. The “MeanLP” measure normalized each
sentence’s log probability by its token count T, whereas the “PenLP” measure divided each sentence’s log
probability by a dampened version of its token count,
(5 +T)/(5 + 1)0.8. For models trained on whole
words (LSTM, RNN, 3-gram, and 2-gram), we used character count instead of token count.
For each language model m, we fitted a separate logistic regression to predict the individual binarized
sentence choices across the entire main experiment dataset by weighing the three predictors “LP,” “MeanLP,”
and “PenLP.” We did not include an intercept due to the symmetry of the prediction task (the presentation
of sentences as sentence 1 or 2 was randomized). We cross-validated the logistic regression’s accuracy by
leaving out one sentence pair at a time, using data from all conditions of the experiment.
Taking token count into consideration led to minor improvements in the prediction accuracy of most
models (an average improvement of 0.95%), but this adjustment did not change the hierarchy of the models
in terms of their human consistency (Supplementary Fig. S1). We hypothesize that the greater disparities
between unnormalized and normalized probability measures, as observed by Lau and colleagues [13] com-
pared to those found in our study, may be attributed to their experiment involving sentences of markedly
different lengths.
347.3 Models differ in their sensitivity to low-level linguistic features
While the controversial sentences presented in this study were synthesized without consideration for partic-
ular linguistic features, we performed a post hoc analysis to explore the contribution of different features
to model and human preferences (Supplementary Fig. S3). For each controversial synthetic sentence pair,
we computed the average log-transformed word frequency for each sentence (extracted from the publicly
available subtlex database [61]). We also computed the average pairwise correlation between semantic
GloVe vector representations [75] of all eight words, based on neuroimaging research showing that there
are specific neural signatures evoked by dissimilarity in semantic vectors [10, 76]. We performed paired
sample t-tests across sentence pairs between the linguistic feature preferences for models vs. humans, and
found that GPT-2, LSTM, RNN, 3-gram, and 2-gram models were significantly more likely (vs. humans)
to prefer sentences with low GloVe correlations, while ELECTRA was significantly more likely to prefer
high GloVe correlations (controlling the false discovery rate for all nine models at q<.05). For word fre-
quency, the RNN, 3-gram, and 2-gram models were significantly biased (vs. humans) to prefer sentences
with low-frequency words, while ELECTRA and XLM showed a significant bias for high-frequency words.
These results indicate that even strong models like GPT-2 and ELECTRA can exhibit subtle misalignments
with humans in their response to simple linguistic features, when evaluated on sentences synthesized to be
controversial.
358 Supplementary Figures
0 25% 50% 75% 100%
human-choice prediction accuracyGPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-gram
a
0 25% 50% 75% 100%
human-choice prediction accuracyGPT-2
RoBERTa
ELECTRA
BERT
XLM
LSTM
RNN
3-gram
2-gram
bUnnormalized sentence probability estimates
Normalized sentence probability estimates
Supplementary Fig. S1: The predictivity of normalized and unnormalized log-probability measures.
(a)Predicting human judgments from all conditions using only unnormalized log probability differences
(equivalent to Fig. 4 in the main text, except using binarized accuracy as a dependent measure). (b)Binarized
accuracy of the logistic regression optimally combining LP, MeanLP, and PenLP for each language model.
Relative model performance is nearly identical in these two analyses, indicating that tokenization differences
across models did not play a large confounding role in our main results.
36Supplementary Fig. S2: The task instructions provided to the participants at the beginning of the experimen-
tal session.
37abSupplementary Fig. S3: Linguistic feature values for synthetic sentence pairs. (a) GloVe correlation
values of the preferred and rejected sentence for each synthetic sentence pair. Each panel depicts preferences
for both humans (red) and a specific model (black), for sentence pairs that this model was involved in
synthesizing. Black sub-panel outlines indicate significant differences between the preferences of models
and humans on that particular set of sentence pairs, according to a paired sample t-test (controlling for false
discovery rate across all nine models at q<.05). (b)Same as (a), but for average log-transformed word
frequency.
389 Supplementary Tables
modelaccepted sentence
has more tokensequal
token-countsrejected sentence
has more tokens p-value
GPT-2 24 13 3 <0.0001
RoBERTa 6 18 16 0.0656
ELECTRA 12 21 7 0.3593
BERT 4 8 28 <0.0001
XLM 2 16 22 <0.0001
Supplementary Table S1: Token count control analysis. For each transformer model, we considered syn-
thetic controversial sentence pairs where the other targeted model was also a transformer (a total of 40
sentence pairs per model). For each such pair, we evaluated the token count of the synthetic sentence to
which the model assigned a higher probability (“accepted sentence”) and the token count of the synthetic
sentence to which the model assigned a lower probability (“rejected sentence”). For each model, this table
presents the number of sentence pairs in which the accepted sentence had a higher token count, both sen-
tences had an equal number of tokens, and the rejected sentence had a higher token count. We compared the
prevalence of higher token counts in accepted and rejected sentences using a binomial test ( H0:p= 0.5)
controlled for False Discovery Rate across five comparisons.
GPT-2 assigned significantly more tokens to accepted sentences, whereas BERT and XLM assigned sig-
nificantly more tokens to rejected sentences. For RoBeRTa and ELECTRA, no significant difference was
found. Note that since the controversial sentences are driven by relative model response properties, a sig-
nificant difference for a particular model does not necessarily indicate that token count biases the model’s
sentence probability estimates. For example, GPT-2’s apparent preference for sentences with a greater token
count might reflect biases of the alternative models pitted against GPT-2. These models might prefer shorter
sentences that exhibit undetected grammatical or semantic violations over longer but felicitous sentences.
Overall, these results indicate that while certain models’ probability estimates might be biased by to-
kenization, lower sentence probabilities were not systematically confounded by higher token counts.
39