Total No. of Questions: 8] SEAT No.
8
23
PB2258 [6263]-96 [Total No. of Pages :3
ic-
tat
B.E.(Computer Engineering)
7s
NATURAL LANGUAGE PROCESSING
8:5
(2019 Pattern) (Semester-VIII) (Elective-V) (410252 A)
02 91
3:4
Time : 2½ Hours] [Max. Marks : 70
0
41
7/0 13
Instructions to the candidates:
1) Answer Q1 or Q2, Q3 or Q4, Q5 or Q6, Q7 or Q8.
0
5/2
2) Neat diagrams must be drawn wherever necessary.
.23 GP
3) Figures to the right side indicate full marks.
4) Assume Suitable data if necessary.
E
81
8
C
23
ic-
Q1) a) What are generative models of language, and how do they differ from
16
tat
discriminative models? Provide an example of a generative model and
8.2
7s
describe how it can be used in NLP. [9]
.24
8:5
91
49
3:4
b) Define Latent Dirichlet Allocation (LDA) and explain how it is used for
30
41
topic modeling in text data. Discuss the key components of LDA,
01
02
including topics, documents, and word distributions. [9]
5/2
GP
7/0
OR
CE
81
8
23
.23
Q2) a) Describe the concept of contextualized representations, such as those
ic-
16
generated by BERT, and how they are used in natual language
tat
8.2
7s
processing. Discuss the advantages and disadvantages of contextualized
.24
8:5
representations. [10]
91
49
3:4
30
b) Suppose you have a text corpus of 10,000 words, and you want to build
41
a bigram model from this corpus. The vocabulary size of the corpus is
01
02
5,000. After counting the bigrams in the corpus, you found that the bigram
5/2
GP
“the cat” appears 50 times, while the unigram “the” appears 1000 times
7/0
and the unigram “cat” appears 100 times. Using the add-k smoothing
CE
81
method with k=0.5, what is the probability of the sentence “the cat sat on
.23
the mat”? [8]
16
8.2
.24
P.T.O.
49
8
Q3) a) Describe the Vector Space Model (VSM) for information retrieval. How
23
does VSM represent documents and queries, and how are similarities
ic-
calculated? Discuss the strengths and weaknesses of VSM. [9]
tat
7s
b) Discuss the different methods used for evaluating NER systems. What
8:5
02 91
are common metrics for measuring NER system performance, and how
3:4
0
can the results be analyzed to improve the system? [9]
41
7/0 13
OR
0
5/2
.23 GP
Q4) a) Define Cross-Lingual Information Retrieval (CLIR) and discuss the
E
81
challenges involved in retrieving information from languages different from
8
C
23
the query language. How do machine translation techniques assist in CLIR?
ic-
[9]
16
tat
8.2
7s
b) Explain the importance of entity extraction in NLP. How does entity
.24
8:5
91
extraction differ from named entity recognition, and provide examples of
49
3:4
real-world applications where entity extraction is crucial. [9
30
41
01
02
5/2
GP
7/0
Q5) a) Describe Walker’s algorithm for word sense disambiguation. How does
CE
it differ from other disambiguation techniques like Lesk’s Algorithm, and
81
8
23
what are the scenarios where it can be most effective? [9]
.23
ic-
16
tat
b) Compare the Indo Word Net with the traditional WordNet. What are the
8.2
7s
key differences and advantages of IndoWordNet, especially in the
.24
8:5
context of Indian languages? [8]
91
49
3:4
30
41
OR
01
02
Q6) a) Compare and constrast the natural Language Toolkit (NLTK), spaCy,
5/2
GP
and TexBlob. what are their main features and in what use cases are they
7/0
most suitable? [9]
CE
81
.23
b) What is the significance of PropBank and VerbNet in linguistic resources?
16
Provide examples of how these resoures can be used to extract semantic
8.2
information from text. [8]
.24
49
[6263]-96 2
8
23
Q7) a) Define natural language generation and its role in NLP. How does NLG
ic-
differ from text-to-speech synthesis, and what are the applications of
tat
NLG in data reporting and storytelling? [9]
7s
b) Discuss the challenges in cross-lingual translation and provide examples
8:5
of how it is beneficial in real-world applications. [8]
02 91
3:4
0
41
7/0 13 OR
Q8) a) Explain the key principles of rule-based machine translation. How do
0
5/2
.23 GP
rule-based techniques differ from statistical approaches in machine
translation? provide an example of a rule-based translation. [9]
E
81
8
b) Discuss the key components of a conversational agent, such as chatbots
C
23
or virtual assistants. How do natural language generation and
ic-
16
understanding play a role in creating effective conversational agents? [8]
tat
8.2
7s
.24
8:5
91
49
3:4
30
41
01
02
5/2
GP
7/0
CE
81
8
23
.23
ic-
16
tat
8.2
7s
.24
8:5
91
49
3:4
30
41
01
02
5/2
GP
7/0
CE
81
.23
16
8.2
.24
49
[6263]-96 3