vnd.openxmlformats-officedocument.wordprocessingml.document&rendition=1
vnd.openxmlformats-officedocument.wordprocessingml.document&rendition=1
QUESTION BANK
Q No Question Marks Section Unit
1 Discuss in detail about Natural Language Processing and discuss with 12 Section-I 1
some applications.
2 Calculate the minimum edit distance between two words: “small” and 12 Section-I 1
“smell” using dynamic programming algorithm. Consider cost for
insertion, deletion and substitution is 1, 1, 2.
3 Differentiate non-word and real-word spelling correction? What is 12 Section-I 1
perplexity? Estimate the perplexity of the corpus based on unigram
language model: “the man is a thief but the man is a good man”
4 Find the Maximum Likelihood Estimate and how is it used Language 12 Section-I 1
Model? Given the Corpus
5 Discuss about morphology and morphemes. What is bounded and free 12 Section-I 1
morphemes explain with example. What is stemming and how it is
different from lemmatization?
6 Explain about Evaluation of Language Models and Basic Smoothing 12 Section-I 1
7 Explain about Noisy Channel Model for Spelling Correction-Gram 12 Section-I 1
Language Models
8 Develop a comprehensive text processing pipeline that includes 12 Section-I 1
tokenization, stemming, normalization, and spelling correction.
9 Apply different smoothing techniques to a language model and analyze 12 Section-I 1
their impact on performance.
10 Design an algorithm to correct spelling errors in a given text document. 12 Section-I 1
11 Distinguish between inflectional and derivational morphology explains 12 Section-II 2
with example. What is morphological analysis explain with example?
12 Explain in detail about POS tagging with example. 12 Section-II 2
13 What is POS tagging? Find the POS tag for the phrase “the light book”, 12 Section-II 2
using Viterbi algorithm in Hidden markov tagging model with the
following information.
14 Discuss the difference of real words and non-words? What is FSA and 12 Section-II 2
how inflections in words can be represented using FSA, explain with
example.
15 Classifying the problems of Hidden Markov model to predict the POS 12 Section-II 2
tags for a given sentence or phrase? Explain how Baum Welch
algorithm learns the parameters transition matrix, observation matrix
and initial state distribution.
16 State the smoothing in language model. What are the advantages 12 Section-II 2
smoothing? Find the Good turing smoothing for the following
sentence: “he is he is good man”
17 Explain the different categories of affixes in morphology with 12 Section-II 2
examples. What are the differences between content and functional
morphemes? What is the difference between regular and irregular forms
of verbs and nouns respectively?
18 Established why maximum entropy model is better than hidden Markov 12 Section-II 2
model. How POS tagging is achieved in maximum entropy model.
What is beam search explain in detail.
19 How is the uniformity maintained in maximum entropy model? Write 12 Section-II 2
the maximum entropy model principles.
20 Discuss in detail about Conditional Random Fields. 12 Section-II 2
21 Discuss about syntax, parsing and constituency. 12 Section-III 3
22 Differenciate between top down and bottom up parsing. Apply CYK 12 Section-III 3
algorithm to parse the sentence “a pilot likes flying planes” with given
grammar