0% found this document useful (0 votes)
6 views13 pages

Third Unit

The document discusses movement phenomena in language, focusing on how elements in sentences can be rearranged to form questions, including yes/no and wh-questions. It explains concepts like subject-auxiliary inversion, do-support, and transformational grammar, as well as how context-free grammars can be extended to handle interrogative sentences. Additionally, it introduces Augmented Transition Networks (ATNs) for natural language parsing and the concept of gap threading in syntax.

Uploaded by

gngpavani
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
6 views13 pages

Third Unit

The document discusses movement phenomena in language, focusing on how elements in sentences can be rearranged to form questions, including yes/no and wh-questions. It explains concepts like subject-auxiliary inversion, do-support, and transformational grammar, as well as how context-free grammars can be extended to handle interrogative sentences. Additionally, it introduces Augmented Transition Networks (ATNs) for natural language parsing and the concept of gap threading in syntax.

Uploaded by

gngpavani
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 13

🗣️Movement Phenomena in Language

Movement phenomena refer to structural changes in sentences where certain elements (like
words or phrases) appear to be "moved" from their original positions to form questions or
other constructions.

1. ✅ Yes/No Questions and Subject-Auxiliary Inversion

 Basic Observation:
Yes/no questions are similar to their corresponding assertions, except the subject
noun phrase (NP) and the first auxiliary verb are swapped.
 Examples:
o Jack is giving Sue a back rub. → Is Jack giving Sue a back rub?
o He will run in the marathon. → Will he run in the marathon?
 If there is no auxiliary, insert "do/does/did" (known as do-support):
o John went to the store. → Did John go to the store?
o Henry goes to school every day. → Does Henry go to school every day?
 📌 This swapping is called Subject-Auxiliary Inversion.

2. ✅ Wh-Questions and Wh-Movement

 Wh-questions ask about specific information (e.g., who, what, where, how, why).
 To form a wh-question:
1. Replace the questioned element with a wh-phrase.
2. Move the wh-phrase to the beginning of the sentence.
3. Apply subject-auxiliary inversion (unless the subject is being questioned).
 Examples:
Assertion: The fat man will put the book in the corner.

o How will the fat man put the book in the corner?
o What will the fat man angrily put in the corner?
o Where will the fat man angrily put the book?
o In what corner will the fat man angrily put the book?
 🔁 Still follows the subject-auxiliary inversion rule, and do-support is used if no
auxiliary is present:
o I found a bookcase.
→ Did I find a bookcase?
→ What did I find?
o

3. ✅ Terminology from Linguistics

 Movement: Refers to rearranging sentence elements (like moving auxiliaries or wh-


phrases).
 Subject-Auxiliary Inversion: Swapping the subject and the auxiliary verb.
 Do-Support: Inserting a "do" form when no auxiliary is present.
 Wh-Movement: Moving a wh-phrase to the start of the sentence.

4. ✅ Transformational Grammar (TG)

 Introduced by Noam Chomsky.


 TG proposes two levels of sentence structure:
o Deep Structure: The basic sentence (e.g., The cat will scratch John).
o Surface Structure: The final sentence we speak/write after transformations
(e.g., Will the cat scratch John?).
 Transformations: Rules that convert the deep structure to the surface structure (e.g.,
moving auxiliary, inserting wh-words).

Summary Table:

Assertion Sentence Yes/No Question Wh-Question Example

Is Jack giving Sue a back


Jack is giving Sue a back rub. Who is Jack giving a back rub to?
rub?

He will run in the marathon next Will he run in the When will he run in the
year. marathon? marathon?

John went to the store. Did John go to the store? Where did John go?

I found a bookcase. Did I find a bookcase? What did I find?


Assertion Sentence Yes/No Question Wh-Question Example

Handling Questions in Context-Free Grammars

Definition: A context-free grammar consists of a set of rules or productions, each of which


expresses theways that symbols of the language can be grouped and ordered together, and a
lexicon ofwords and symbols.

Context Free Grammar (CFG) - Formal Definition

Context-free grammar G is a 4-tuple.

G = (V, T, S, P)

These parameters are as follows;

V – Set of variables (also called asNon-terminal symbols)

T – Set of terminal symbols (lexicon)

o The symbols that refer to words in a language are calledterminal symbols.

o Lexicon is a set of rules that introduce these symbols

. S – Designated start symbol (one of the non-terminals, S∈ V)

 P – Set of productions (also called asrules).

o Each rule in P is of the formA → s, where

o A is a non-terminal (variable) symbol

.o Each rule can have only one non-terminal symbol on the left hand side of therule.

o s is a sequence of terminals and non-terminals. It is from (T U V)*, infinite setof strings.

🧠
2. Sample CFG and Derivation
Grammar Example:
S → NP VP
NP → Det Noun | NP PP
VP → Verb NP
PP → Pre NP
Det → 'a' | 'the'
Noun → 'cake' | 'child' | 'fork'
Verb → 'ate'
Pre → 'with'

Sample Sentence Derivation:

S → NP VP

→ Det Noun VP

→ the Noun VP

→ the child VP

→ the child Verb NP

→ the child ate NP

→ the child ate Det Noun

→ the child ate a Noun

→ the child ate a cake

✅ This shows that "the child ate a cake" is generated by the grammar and hence is
grammatically correct.

❓ 3. Extending CFGs to Handle Questions


CFGs as described above typically handle declarative sentences (statements).
But natural languages also contain interrogative sentences (questions), such as:

 Yes/No Questions: Did the child eat the cake?


 Wh-Questions: What did the child eat?

Goal:
➡️Extend the grammar minimally to support questions, while reusing as much of the

original grammar as possible.

✅ 4. Yes/No Questions
🛠 Mechanism: Subject–Auxiliary Inversion

In yes/no questions, the auxiliary verb moves in front of the subject noun phrase (NP).

 Statement: The child has eaten the cake.


 Question: Has the child eaten the cake?

📐 Extended Rule:

We add a rule like this:

S[+inv] → (AUX AGR ?a SUBCAT ?v) (NP AGR ?a) (VP VFORM ?v)

Where:

 AUX: Auxiliary verb


 NP: Subject noun phrase
 VP: Verb phrase
 AGR: Ensures subject-verb agreement
 VFORM: Ensures correct verb form follows the auxiliary

This single rule, when added, enables the grammar to produce valid yes/no questions by just
changing the order of subject and auxiliary. All other grammar components remain
unchanged.
❓ 5. Wh-Questions
These are more complex than yes/no questions because they involve gaps — missing
constituents in the sentence.

🔎 Example:

 Statement: You put the book in the shelf.


 Wh-question: What did you put the book in __?

Here, "what" replaces a noun phrase and leaves behind a gap.

🧩 Features Needed:

5.1 GAP Feature:

 Keeps track of missing elements (like the object NP in the example).


 Propagates through the sentence structure to indicate where something is missing.

5.2 WH Feature:

 Identifies wh-words (who, what, where, etc.) and their corresponding phrase types.
 These words serve as fillers for the gaps left in the sentence.

🧮 6. Categories of WH-Words
WH Word Phrase Type Replaced Example

who, whom, what Noun Phrase (NP) Who ate the cake?

what, which Determiner in NP What book did he steal?

where, when Prepositional Phrase (PP) Where did you put the book?
WH Word Phrase Type Replaced Example

how Adverbial Phrase (AdvP) How quickly did he run?

whose Possessive NP Whose book did you find?

🔄 7. How to Extend CFG for Wh-Questions


Step-by-step:

1. Introduce GAP feature to non-terminal rules (e.g., NP, PP) to allow missing
elements.
2. Add WH feature to identify wh-phrases and restrict them to appropriate roles.
3. Prevent gap propagation into the subject position in inverted clauses (e.g.,
auxiliary questions).
4. Allow fronted wh-phrases to act as fillers for the gap in the sentence body.

5.

🔚 8. Summary: CFG Extension Strategy


Aspect Declarative CFG Extended CFG for Questions

Sentence Form S → NP VP S[+inv] → AUX NP VP

Question Type Not handled Yes/No, Wh-questions

Auxiliary Handling None AUX comes before subject

Gap Tracking Not needed GAP feature in rules

WH-Phrase Handling Not needed WH feature for fronting

Grammar Reuse Full Full (with minimal additions)


Aspect Declarative CFG Extended CFG for Questions

Here’s a brief and clear set of notes on Augmented Transition Networks (ATNs) and hold
mechanisms based on your content:

Augmented Transition Networks (ATNs) – Brief Notes

What are ATNs?

 ATNs are a type of transition network used in natural language parsing.


 They extend finite state machines (FSMs) by:
o Allowing recursive procedures (networks can call themselves or others).
o Using registers (features) to store and manipulate data during parsing.

Why ATNs are Needed

 Traditional FSMs can only handle regular languages, not enough for complex
natural language constructs.
 Natural language involves hierarchical and context-sensitive structures.
 ATNs are designed to handle context-free and some context-sensitive aspects of
language.

Key Features of ATNs

 Registers (features): Store info like parts of speech, agreement, etc.


 Recursive calls: Networks can call sub-networks for complex constituents.
 Actions: Executed when arcs are followed, used to:
o Set registers (e.g., DET := *)
o Copy features (e.g., AGR := AGR*)
 Tests: Conditions to check if a transition is valid (e.g., agreement checks).
 Constituent building: When a network finishes, its registers form a constituent
structure.

Hold Mechanism

 ATNs use a special variable "*" to hold the result of following an arc.
 Example:
o If a word like "the" is accepted via a lexical arc, it's stored in "*".
o Then an action like DET := * stores it in the DET register.
o Similarly, AGR := AGR* copies the agreement info from * into the current
network.
Simple NP (Noun Phrase) ATN Example

 Lexical arc (e.g., for a determiner):


o Actions:
1. DET := * – Store the matched word.
2. AGR := AGR* – Copy agreement features.
 Tests on arcs enforce rules like subject-verb agreement.

ATN – Tests, Push Arcs & Agreement (Compact Notes)

 Arc Tests:
o An arc is followed only if its test passes.
o Example: Arc 2 checks if AGR of the current network and the noun in "*" have
a non-empty intersection.
 Push Arcs:
o Used to call sub-networks (e.g., NP).
o The result is stored in "*".
o Example: SUBJ := * assigns the parsed NP to the subject register.
 Agreement Check (e.g., Subject-Verb):
o Test compares AGR of SUBJ (noun phrase) with AGR of the verb (in "*").
o Arc proceeds only if their features match (i.e., agree).
 Ensures grammatical consistency, such as subject-verb agreement.
Consider an example. A trace of a parse of the sentence "The dog saw Jack"

Here's a concise and structured summary of the new points you've provided:
ATN – Parse Flow Summary

 Parsing starts in the S network.


 Arc 4 triggers a call to the NP network.
 NP network parses the phrase "The dog":
o Checks number agreement.
o Builds NP with AGR = plural.
 On returning (via pop arc), arc 4 is completed.
 NP is stored in SUBJ register.
 Arc 3 is then followed:
o Checks if SUBJ.AGR agrees with verb’s AGR.
 Another NP, "Jack", is accepted in a similar call to the NP network.

1. Gap Threading as a Parsing Concept

If you're referring to syntax or parsing, “gap threading” might relate to movement/gap


phenomena in natural language syntax — for example, in wh-movement (e.g., What did you
say that John bought ___?). In such sentences, parsers need to thread the gap left by a
displaced element (like "what") back to its original syntactic position.

In this context, gap threading is the process of:

 Keeping track of the "gap" (i.e., the missing constituent).


 Ensuring grammatical dependencies (e.g., subject/object roles) are maintained during
parsing.

It's common in transformational grammar and unification-based grammar frameworks


like HPSG or LFG, where gaps are handled using features that “thread” through syntactic
trees.

2. Multithreading in NLP (Gap + Threading?)

If by “gap threading” you're referring to parallelism or multithreading in NLP tasks, then it


could mean:

 Using threading to fill gaps in data, e.g., during coreference resolution, ellipsis
recovery, or missing data reconstruction.
 Using multithreading (in code) to speed up NLP tasks, like tokenization, tagging,
parsing, etc., especially for large corpora.

3. Text Gap Filling


In tasks like cloze tests, gap threading might refer to connecting gaps (blanks) to potential
fillers using context — in models like masked language models (e.g., BERT). Here,
threading refers metaphorically to connecting context before and after the gap.

You might also like