A dependency look at the reality of constituency

Reading time: 4 minute
...

📝 Original Info

  • Title: A dependency look at the reality of constituency
  • ArXiv ID: 1708.07722
  • Date: 2020-09-24
  • Authors: ** - Xinying Chen (시안교통대학 외국어연구센터) - Carlos Gómez‑Rodríguez (스페인 코루냐 대학교 FASTPARSE Lab) - Ramón Ferrer‑i‑Cancho (바르셀로나 폴리테크닉 대학교 복합언어학 연구소) **

📝 Abstract

A comment on "Neurophysiological dynamics of phrase-structure building during sentence processing" by Nelson et al (2017), Proceedings of the National Academy of Sciences USA 114(18), E3669-E3678.

💡 Deep Analysis

📄 Full Content

104 Glottometrics 40, 2018, 104-106

A Dependency Look at the Reality of Constituency

Xinying Chen1,
Carlos Gómez-Rodríguez2,
Ramon Ferrer-i-Cancho3

Abstract. A comment on “Neurophysiological dynamics of phrase-structure building during sentence processing” by Nelson et al (2017), Proceedings of the National Academy of Sciences USA 114(18), E3669-E3678.

Recently, Nelson et al. (2017) have addressed the fundamental problem of the neurophysic- ological support for complex syntactic operations of theoretical computational models. They interpret their compelling results as supporting the neural reality of phrase structure. Such a conclusion opens various questions.

First, constituency is not the only possible reality for the syntactic structure of sentences. An alternative is dependency, where the structure of a sentence is defined by word pairwise dependencies (Fig. 1). From that perspective, phrase structure is regarded as an epiphenomenon of word-word dependencies and constituency (in a classical sense as that of X-bar theory) has been argued to not exist (Mel’čuk, 2011). Furthermore, constituency may not be universal and thus its suitability may depend on the language (Evans & Levinson, 2009). Dependency is a stronger alternative for its simplicity, its close relationship with merge (Osborne, Putnam, & Gross, 2011), its compatibility with recent cognitive observations (Gómez-Rodríguez, 2016), its contribution to the cost of individual words even in isolation (Lester & Moscoso del Prado Martin, 2016) and its success over phrase structure in computational linguistics, where it has become predominant (Kübler, McDonald, & Nivre, 2009).

Second, the authors admit that a parser of the sentence might transiently conclude that “ten sad students”… is a phrase consistently with a transient decrease in activity (1st paragraph of p. 4). Unfortunately, their parser does not account for that as shown by the counts in Fig. 2 A of Nelson et al. (2017). In contrast, a standard dependency parser would, because at that point it would close the dependencies opened by “ten” and “sad” (Fig. 1). This raises the question of whether the conclusions depend on the choice X-bar and particular parser as a model of phrase structure. The conclusions by Nelson et al. (2017) may suffer from circularity, namely the positive support for a particular X-bar model could be due to the fact that the source was a toy artificial X-bar grammar. Future analyses would benefit from the use of natural sentences, sentences with realistic probabilities that are also longer and more complex (sentence length does not exceed 10 in Nelson et al. (2017)).

Third, dependency shows the limits of comparing phrase structure models against n-gram models with n = 2, because only about 50% of adjacent words are linked (Liu, 2008; Ferrer-i-

1 Foreign Languages Research Center, School of Foreign Studies, Xi’an Jiaotong University, No.28 Xianning West Road, 710049 Xi’an, Shaanxi, P.R. China. 2 Universidade da Coruña. FASTPARSE Lab, LyS Research Group. Departamento de Computación. Facultade de Informática, Elviña 15071 A Coruña, Spain 3 Complexity & Quantitative Linguistics Lab, LARCA Research Group, Departament de Ciències de la Computació, Universitat Politècnica de Catalunya, Campus Nord, Edifici Omega, Jordi Girona Salgado 1- 3, 08034 Barcelona, Catalonia (Spain). Corresponding author, rferrericancho@cs.upc.edu. A Dependency Look at the Reality of Constituency

105 Cancho, 2004), thus a bigram model misses 50% of the dependencies. Bigrams are a weak baseline, as the common practice in computational linguistics is using at least smoothed trigram models, and often 5-gram models, to obtain meaningful predictions (Jozefowicz, Vinyals, Schuster, Shazeer, & Wu, 2016). A higher-order lexical n-gram model would strengthen the current results. The authors also employ more sophisticated n-gram models. One is an unbounded model based on part-of-speech categories, implying a dramatic loss of information with respect to the original words which might explain its poor performance. The other is a syntactic n-gram, but not enough information is provided about its definition and implementation. Regardless, since the model is obtained from a corpus derived from a toy grammar and lexicon, its prob- abilities are likely to be unrealistic and thus it is problematic.

In sum, dependency offers a better approach to the syntactic complexity of languages and merge. n-gram models of higher complexity should be the subject of future research involving realistic sentences.

Figure 1: Syntactic dependency structure of the sentence in Fig 2 A of Nelson et al. (2017) according to Universal Dependencies (McDonald et al., 2013).

Acknowledgements

X.C. is supported by the Social Science Fund of Shaanxi State (2015K001). C.G.R is funded by the European Research Counc

Reference

This content is AI-processed based on open access ArXiv data.

Start searching

Enter keywords to search articles

↑↓
ESC
⌘K Shortcut