NLP - PPT - CH 1
NLP - PPT - CH 1
Processing
Pushpak Bhattacharya
Aditya Joshi
Chapter 1
Introduction
• Understand the relationship between linguistics, probability, and data in the context
of Natural Language Processing (NLP)
• This work explores the field of natural language processing (NLP): computational
techniques that process languages used by humans (i.e., ‘natural’ languages)
• One of the key reasons for the advancement of the human race is language
• In the case of written texts, symbols are the codes that carry meaning and, hence,
convey ideas
• Scripts of languages fall into three categories depending on how they treat vowels
• For the ‘abjad’ category, as in Arabic, vowels are mostly dropped completely
• They are used by people who may not be able to produce conventional sounds
in a language
• The former arises primarily due to multiple meanings of lexical units, specifically
words
• The latter arises from how phrases are linked to other sentential units
• The ambiguity arises due to a possible meaning of ‘love’ as a zero score in tennis and
the word ‘nothing’
• There are five words in this sentence, with the intended meaning as ‘it is reported by
the Government of Maharashtra that COVID-19 cases have increased’
v. Multimodal NLP (e.g., emotion recognition based on facial expressions and spoken
words)
vi. Sentiment and emotion analysis: (e.g., the recent ChatGPT tool)
• While structure brings a sense of determinism to language, it often misses the point
that language is not deterministic or fossilized
i. No grammar, as far as we know, can capture all and only the phenomena of the
language
ii. Given the left-hand side (LHS) of a production rule, multiple right-hand sides (RHS)
are possible
• As a result, new and unexpected language phenomena are witnessed, while some
fade away
“Use a dataset of textual examples as the ‘training data’ (in the sense of past data), to
learn a model that makes predictions on the ‘test data’ (in the sense of future data)”
ii. NLP must choose among possible meanings by assigning numerical scores to them
i. Rules many times proved woefully inadequate to capture the full spectrum and
complexity of language phenomena
• In general, disambiguation in NLP at any layer amounts to choosing the best among
possible labels, strings, trees, or graphs
• Deep learning introduced a continuum in which similar linguistic units were close to
each other because:
i. It could handle data sparsity
ii. it could learn vectors that are similar to each other
ii. Learning to solve particular NLP tasks. DNN for NLP relies on layers of neural
networks learning representations of text
• This book takes a historical perspective of NLP via the three generations when
describing the foundations and applications of NLP
• Several problems in NLP such as POS tagging, parsing, natural language inference,
sentiment analysis, question-answering is discussed