A comment on "Neurophysiological dynamics of phrase-structure building during sentence processing" by Nelson et al (2017), Proceedings of the National Academy of Sciences USA 114(18), E3669-E3678.
Deep Dive into A dependency look at the reality of constituency.
A comment on “Neurophysiological dynamics of phrase-structure building during sentence processing” by Nelson et al (2017), Proceedings of the National Academy of Sciences USA 114(18), E3669-E3678.
104
Glottometrics 40, 2018, 104-106
A Dependency Look at the Reality of Constituency
Xinying Chen1,
Carlos Gómez-Rodríguez2,
Ramon Ferrer-i-Cancho3
Abstract. A comment on “Neurophysiological dynamics of phrase-structure building during sentence
processing” by Nelson et al (2017), Proceedings of the National Academy of Sciences USA 114(18),
E3669-E3678.
Recently, Nelson et al. (2017) have addressed the fundamental problem of the neurophysic-
ological support for complex syntactic operations of theoretical computational models. They
interpret their compelling results as supporting the neural reality of phrase structure. Such a
conclusion opens various questions.
First, constituency is not the only possible reality for the syntactic structure of sentences.
An alternative is dependency, where the structure of a sentence is defined by word pairwise
dependencies (Fig. 1). From that perspective, phrase structure is regarded as an epiphenomenon
of word-word dependencies and constituency (in a classical sense as that of X-bar theory) has
been argued to not exist (Mel’čuk, 2011). Furthermore, constituency may not be universal and
thus its suitability may depend on the language (Evans & Levinson, 2009). Dependency is a
stronger alternative for its simplicity, its close relationship with merge (Osborne, Putnam, &
Gross, 2011), its compatibility with recent cognitive observations (Gómez-Rodríguez, 2016), its
contribution to the cost of individual words even in isolation (Lester & Moscoso del Prado
Martin, 2016) and its success over phrase structure in computational linguistics, where it has
become predominant (Kübler, McDonald, & Nivre, 2009).
Second, the authors admit that a parser of the sentence might transiently conclude that
“ten sad students”… is a phrase consistently with a transient decrease in activity (1st paragraph of
p. 4). Unfortunately, their parser does not account for that as shown by the counts in Fig. 2 A of
Nelson et al. (2017). In contrast, a standard dependency parser would, because at that point it
would close the dependencies opened by “ten” and “sad” (Fig. 1). This raises the question of
whether the conclusions depend on the choice X-bar and particular parser as a model of phrase
structure. The conclusions by Nelson et al. (2017) may suffer from circularity, namely the
positive support for a particular X-bar model could be due to the fact that the source was a toy
artificial X-bar grammar. Future analyses would benefit from the use of natural sentences,
sentences with realistic probabilities that are also longer and more complex (sentence length does
not exceed 10 in Nelson et al. (2017)).
Third, dependency shows the limits of comparing phrase structure models against n-gram
models with n = 2, because only about 50% of adjacent words are linked (Liu, 2008; Ferrer-i-
1 Foreign Languages Research Center, School of Foreign Studies, Xi’an Jiaotong University, No.28
Xianning West Road, 710049 Xi’an, Shaanxi, P.R. China.
2 Universidade da Coruña. FASTPARSE Lab, LyS Research Group. Departamento de Computación.
Facultade de Informática, Elviña 15071 A Coruña, Spain
3 Complexity & Quantitative Linguistics Lab, LARCA Research Group, Departament de Ciències de la
Computació, Universitat Politècnica de Catalunya, Campus Nord, Edifici Omega, Jordi Girona Salgado 1-
3, 08034 Barcelona, Catalonia (Spain). Corresponding author, rferrericancho@cs.upc.edu.
A Dependency Look at the Reality of Constituency
105
Cancho, 2004), thus a bigram model misses 50% of the dependencies. Bigrams are a weak
baseline, as the common practice in computational linguistics is using at least smoothed trigram
models, and often 5-gram models, to obtain meaningful predictions (Jozefowicz, Vinyals,
Schuster, Shazeer, & Wu, 2016). A higher-order lexical n-gram model would strengthen the
current results. The authors also employ more sophisticated n-gram models. One is an unbounded
model based on part-of-speech categories, implying a dramatic loss of information with respect
to the original words which might explain its poor performance. The other is a syntactic n-gram,
but not enough information is provided about its definition and implementation. Regardless,
since the model is obtained from a corpus derived from a toy grammar and lexicon, its prob-
abilities are likely to be unrealistic and thus it is problematic.
In sum, dependency offers a better approach to the syntactic complexity of languages and
merge. n-gram models of higher complexity should be the subject of future research involving
realistic sentences.
Figure 1: Syntactic dependency structure of the sentence in Fig 2 A of Nelson et al. (2017)
according to Universal Dependencies (McDonald et al., 2013).
Acknowledgements
X.C. is supported by the Social Science Fund of Shaanxi State (2015K001). C.G.R is funded by
the European Research Counc
…(Full text truncated)…
This content is AI-processed based on ArXiv data.