NLP
NLP
a) Natural Logic Processing b) Natural Language Processing c) Natural Learning Processing d) Non-
Language Processing
6. What does a virtual assistant like Siri or Alexa primarily rely on?
Answer: c) NLP
8. Which technique in NLP is used to remove unnecessary words like “and” or “the”?
9. What is the process of breaking text into individual words or terms called?
Answer: b) Tokenization
10. Which is a challenge for NLP when dealing with human language?
a) Identifying numbers b) Processing unstructured data c) Parsing simple sentences d) None of the
above
Answer: b) Tokenization
12. In which step of text normalization are words reduced to their root form?
Answer: a) Lemmatization
16. Which type of chatbot uses large databases and can learn from more interactions?
Answer: c) Smart-bot
d) Speech recognition
18. Which AI model helps extract features from text using word frequency?
19. What is the primary step in converting text to numerical format in NLP?
a) Model performs perfectly b) Model does not fit well with the data c) Model fits too well to the data
d) Model generalizes well
Answer: c) TFIDF
22. In text normalization, which step involves reducing words to their base forms?
Answer: a) Stemming
23. What kind of words are typically removed during stopword removal in NLP?
a) Important keywords b) Common and unimportant words c) Rare words d) Numerical data
d) Identifying colors
Answer: b) Alexa
d) Analyzing sentiment
30. Which type of NLP bot adapts based on interactions and improves over time?
a) It can only handle predefined conversations b) It requires a large database c) It learns from user
interactions d) It uses machine learning
a) Removing prefixes from words b) Converting words to their base or dictionary form
34. What do virtual assistants like Google Assistant use for speech recognition?
Answer: c) NLP
d) A speech output
36. Which process removes words like “is” or “the” from text in NLP?
39. What is a common issue with NLP when dealing with human languages?
a) Words with multiple meanings b) Predicting stock prices c) Recognizing colors d) Detecting shapes
41. Which chatbot type typically connects users to human agents for complex queries?
Answer: a) Script-bot
43. Which of the following tools is used to reduce a word to its base form?
Answer: b) Lemmatizer
a) By using images b) By analyzing text data c) By calculating numerical values d) By tracking user
behavior
45. What is the primary challenge of translating human language into computer language?
a) A large collection of text data b) A summary of documents c) A chatbot model d) A set of stopwords
d) To tokenize text
Assertion-Reasoning Questions:
Assertion (A): NLP can be used to automatically summarize large amounts of text data.
Reason (R): NLP algorithms analyze text to extract important information while removing redundant or
irrelevant data.
Correct Answer: a) Both A and R are true, and R is the correct explanation of A.
Assertion (A): Stopwords are removed during NLP processing.
Reason (R): Stopwords are frequently used words that add significant value to the meaning of a
sentence.
Assertion (A): Lemmatization is a technique used to reduce words to their root forms.
Reason (R): Lemmatization always results in words that are meaningful, unlike stemming.
Correct Answer: a) Both A and R are true, and R is the correct explanation of A.
Assertion (A): Text classification can be used to categorize documents into predefined categories.
Correct Answer: b) Both A and R are true, but R is not the correct explanation of A.
Assertion (A): NLP chatbots can replace human customer service agents completely.
Reason (R): NLP-based chatbots use predefined scripts that limit their flexibility in conversations.
Case Study 1:
A company is using sentiment analysis to understand customer opinions about its new smartphone
model. They collect social media posts and customer reviews to analyze the data. They want to know
if the overall sentiment is positive, negative, or neutral, and they are also interested in identifying
specific features customers mention.
Questions:
How would the company identify specific features mentioned by customers in the reviews?
Answer: By using text classification or named entity recognition to extract key terms related to specific
features.
Answer: It helps the company understand customer opinions, improve the product, and respond to
issues or concerns in a timely manner.
Why might the company also want to remove stopwords in this analysis?
Answer: Stopwords are commonly used words that don’t add significant meaning to the analysis, so
removing them can help focus on more relevant terms.
How could the company improve the accuracy of their sentiment analysis over time?
Answer: By training the model on a larger dataset, using advanced algorithms like TFIDF or neural
networks, and continuously updating the dataset with new customer feedback.
Case Study 2:
A mental health app aims to provide users with a chatbot that interacts with them when they are
feeling stressed. The chatbot uses Cognitive Behavioral Therapy (CBT) techniques to guide
conversations and provide helpful suggestions.
Questions:
Answer: The chatbot uses Natural Language Processing for conversational interaction and therapy-based
suggestions.
How can NLP help the chatbot understand user inputs during a conversation?
Answer: NLP can process and analyze the text input, detect emotions or stress-related keywords, and
respond with appropriate therapeutic advice.
Answer: Text normalization helps the chatbot process unstructured input by breaking it down into
simpler forms, such as tokens or base forms of words, allowing for more accurate responses.
What challenge might the chatbot face when interpreting user emotions?
Answer: Multiple meanings of words and ambiguous expressions could lead to misunderstandings of
the user’s emotional state.
How can the chatbot be improved to better serve its users over time?
Answer: By incorporating machine learning techniques, the chatbot can learn from user interactions,
adapt responses, and improve its understanding of complex emotions and contexts.
Case Study 3:
A retail company is using an NLP-based virtual assistant to help customers on its website. The
assistant helps customers find products, provides recommendations, and answers questions about
the store’s return policy.
Questions:
What is the primary function of the virtual assistant in this case study?
Answer: The virtual assistant helps customers with product recommendations and answers frequently
asked questions, such as those related to return policies.
How does the virtual assistant understand customer queries about different products?
Answer: The assistant uses NLP techniques like text classification and named entity recognition to
identify keywords and understand customer intents.
What benefit does the company gain by using this virtual assistant?
Answer: The company provides a seamless shopping experience, reduces the workload on human
customer service agents, and improves customer satisfaction.
How could the virtual assistant improve product recommendations over time?
Answer: By analyzing past interactions and customer preferences, the assistant can use machine
learning algorithms to offer more personalized product suggestions.
What challenge might the virtual assistant face when interacting with customers?
Answer: It may struggle with ambiguous queries or complex, multi-part questions that require a deeper
understanding beyond predefined scripts.
Case Study 4:
Questions:
What NLP functionality is most important for the healthcare chatbot in this scenario?
Answer: The chatbot uses NLP for understanding patient queries, recognizing symptoms, and scheduling
tasks based on natural language input.
How could the chatbot ensure it gives accurate responses to symptom-related queries?
Answer: By using medical-specific databases and NLP models trained on healthcare-related data, the
chatbot can provide reliable advice and suggestions.
Answer: It provides convenient access to healthcare services, reduces wait times, and helps patients
manage their health effectively through timely reminders and assistance.
Answer: The chatbot may misinterpret symptoms or fail to account for complex medical conditions,
leading to incorrect recommendations.
How can the chatbot be improved to better assist patients over time?
Answer: Continuous updates to its medical knowledge base, as well as integration with patient medical
histories, can help the chatbot provide more accurate and personalized responses.
Case Study 5:
An educational platform is developing an NLP-powered system that helps students write better
essays. The system provides feedback on grammar, style, and sentence structure, and suggests
improvements based on common writing errors.
Questions:
How does the NLP system assist students in improving their writing?
Answer: The system uses NLP techniques like grammar checking, sentence segmentation, and syntax
analysis to identify errors and suggest improvements in writing.
What advantage does the system offer over traditional grammar checking tools?
Answer: The NLP system can provide more advanced feedback, such as style improvement and context-
based suggestions, going beyond simple grammar checks.
Answer: Sentence segmentation helps the system break down essays into individual sentences for easier
error detection and analysis of sentence structure.
What challenge might the system face when offering feedback on complex essays?
Answer: It may struggle with understanding the deeper meaning or context of sentences, leading to
inaccurate or irrelevant suggestions.
Answer: By incorporating machine learning and deeper contextual understanding, the system can offer
more tailored and accurate feedback based on the writing style and complexity of the essay.
Question-Answer:
What is Natural Language Processing (NLP)?
NLP is a sub-field of AI focused on enabling computers to understand and process human languages. It
involves programming computers to analyze and process large amounts of natural language data.
The three primary domains of AI discussed are Data Science, Computer Vision, and Natural Language
Processing (NLP).
A Script-bot works around a predefined script and is limited in functionality, whereas a Smart-bot uses
AI, learns from data, and is more flexible and capable of handling various tasks.
Sentiment analysis helps in identifying emotions and opinions expressed in text, enabling companies to
understand customer feedback and sentiment towards products or services.
The ‘Bag of Words’ model converts text into a collection of unique words from the corpus and counts
their frequency, disregarding the order in which they appear.
Text Normalization simplifies complex text into a form that machines can process, involving steps such
as tokenization, removing stopwords, and stemming.
TFIDF measures the importance of a word in a document relative to a collection of documents. Term
Frequency (TF) measures how often a word appears, while Inverse Document Frequency (IDF) reduces
the weight of common words across multiple documents.
Stopwords are common words that don’t add much meaning to the text (e.g., “and”, “the”). They are
removed to help the machine focus on more relevant terms.
Automatic Summarization can be used to extract key information from large texts, such as summarizing
news articles while avoiding redundancy and maximizing content diversity.
How does Cognitive Behavioral Therapy (CBT) relate to NLP-based projects in the document?
NLP can be used to create chatbots that apply primitive CBT techniques, helping users manage stress
and emotional well-being by interacting with them and suggesting when professional help may be
needed.
Computers struggle with complexities such as the arrangement of words (syntax), multiple meanings of
words, and understanding the context or emotional subtext behind statements.
The two types are Script-bots, which follow predefined scripts, and Smart-bots, which are AI-powered
and can learn from data to handle more complex interactions.
Stemming is used to reduce words to their root forms to simplify the data. However, it may produce
non-meaningful words (e.g., “studies” becomes “studi”).
Lemmatization also reduces words to their base form but ensures the resulting words are meaningful,
whereas stemming may produce non-meaningful root forms.
Data Acquisition involves collecting relevant conversational data, such as through surveys or interviews,
to enable machines to interpret language and emotions effectively in an NLP project.