Syntactic Structures by Chomsky (Summary)
07 Oct 2018This blog is a summary of the ideas outlined in Chomsky’s Syntactic Structures.
Table of Contents
Language
Any language L
is considered to be a set of sentences. Each sentence is finite in length and is constructed out of a finite set of elements (words). A sentence can be a sequence of phonemes or letters in an alphabet.
Grammar
Grammar has no relation to semantic meaning of a sentence. A sentence can be grammatically correct and not have any meaning. Since the probability of grammatically correct and incorrect sentences occurring in a corpus of text is highly dependent on the corpus, we cannot leverage statistics to find out if a sentence is grammatically correct or not. (Although we know that this is not true anymore as we can build robust language models that care capable of probabilistically generating grammatically correct sentences).
Conclusion: Grammar is autonomous and independent of meaning, and that probabilistic models give no particular insight into some of the basic problems of syntactic structure.
Phonemes and Morphemes
Phonemes are smallest elements in pronunciation. Morphemes are smallest elements in words.
Markov Processes for language generation
Chomsky talks of a finite state machine (finite state grammar) that can be used for generation of sentences. He argues that since since English is a non-finite state language, we cannot use a finite state machine to generate English sentences.
-
Finite state language
a[3]b[4]
is finite. -
Infinite state language
a*b*
is infinite.Similarly, the following demonstrates how English is also infinite.
Suppose \(S_{1}\), \(S_{2}\) and \(S_{3}\) are declarative sentences, we can construct the following sentences using grammar rules.
If \(S_{1}\), then \(S_{2}\).
Either \(S_{3}\) or \(S_{4}\).
Each of the above sentence is also declarative and hence can be expanded infinitely.
Limitations of Phrase structure description
-
If there are two sentences of the form
Z + X + W
andZ + Y + W
, we can use the conjunctionand
and construct a new sentence of the formZ-X+ and +Y-W
(Only ifX
andY
are constituents)For eg,
the scene - of the movie - was in Chicago
the scene - of the play - was in Chicago
the scene - of the movie and of the play - was in Chicago
The above assumption does not hold true in the following case where
X
andY
are not constituentsthe - liner sailed down the - river
the - tugboat chugged up the - river
the - liner sailed down the and the tugboat chugged up the - river
The limitation in this case is that unless
X
andY
are constituents, we cannot apply this grammar rule. But it is not possible to incorporate this rule in any phrase structure grammar. Why? Because we need to know the actual form of the 2 sentences and ifX
andY
are constituents (for which we need to know their values). -
Auxilary Verbs
-
Active passive relation
“John admires sincerity” <-> “sincerity admires John”
“sincerity frightens John” <-> “John frightens sincerity”
These examples demonstrate that even though these sentences are not violating any grammar rules, they are semantically incorrect and hence pose limitations to the assumption that grammar does not depend on semantics. These limitations can be overcome by having extra rules (which cannot be incorporated within the grammar itself) over the grammar.
Grammatical Transformation
-
In order to overcome the above limitations, we can assume the grammar to have 3 levels of rules:
-
Phrase Structure
-
Transformational Structure
-
Morphophonemics
Given a sentence, we first go through the phrase structure grammar to build a tree and get to the leaf nodes (terminal words).
We then run through the transformational structural rules (both optional and obligatory rules) to transform the set of terminal nodes.
We then apply the morphophonemics to arrive at the final sentence.
-
Procedure for formalizing computational linguistics
Of the 3 approaches outlined in the diagram above, Chomsky proposes to choose the one in which we can come to a reasonable solution to, hence evaluation. He says, generating (discovery) of the correct grammar given a data corpus is not feasible (we know this is not true now, because we can build language models given a huge corpus).
Explanation power of linguistics
Reason to have a separate representation for morphemes and phonemes is demonstrated here. The phoneme sequence /eneym/ can be understood ambiguously as both “a name” or a “an aim”.
If our grammar is only a single level system dealing with only phonemes, we have no way of representing this ambiguity. Therefore we need a second morphological layer.
Syntax and Semantics
The three levels of parsing of grammar gives us the ability to perform the following representations:
-
Represent sentences that can be understood in more than one way ambiguously.
-
Represent two sentences that are understood in a similar manner similarly on the transformational level.
Chomsky argues that there cannot be any relation between semantics and grammar i.e the burden of proof lies on the linguist claiming that the grammar should be dependent on semantics.
Assertions supporting dependence of grammar on meaning:
-
Two utterances are phonemically distinct if and only if they differ in meanings
This is refuted because of synonyms (utterance tokens that are phonemically different but means the same thing) and homonyms (utterance tokens that are phonemically identical but differ in meaning)
-
Morphemes are the smallest elements that have meanings
This is refuted because morpheme such as gl- in “gleam”, “glimmer” and “glow” does not carry any meaning in itself.
-
Grammatical sentences are those that have semantic significance
-
NP - VP
-> actor-action“the fighting stopped” has no actor-action relation
-
Verb - NP
-> action-goal or action-object of action“I missed the train” has no action-goal relation
-
Active sentence and corresponding passive sentence are synonymous
“everyone in the room knows at least two languages” is not synonymous to “at least two languages are known by everyone in the room”