Mcs1009 Ml Answer Key Part a and b
Mcs1009 Ml Answer Key Part a and b
QA102 (2 Marks):
Conditional probability is the probability of an event occurring given that another event
has already occurred. Example: P(A|B) = P(A ∩ B) / P(B).
QA103 (2 Marks):
Supervised learning uses labeled data to predict outcomes, while unsupervised learning
finds hidden patterns in unlabeled data.
QA104 (2 Marks):
Entropy measures the impurity or randomness in data. In information theory, it's used
to determine information gain.
QA105 (2 Marks):
Advantages: (1) Automates decision-making, (2) Learns from data. Challenges: (1)
Requires large data, (2) Risk of bias.
QA106 (2 Marks):
Precision = TP / (TP + FP), Recall = TP / (TP + FN). These are important in evaluating
classification performance, especially for imbalanced datasets.
QA107 (2 Marks):
Euclidean distance helps in measuring the closeness of data points in clustering and
classification tasks like KNN.
QA108 (2 Marks):
Z-score standardizes data by subtracting the mean and dividing by the standard
deviation. Helps in outlier detection.
QA201 (2 Marks):
Discriminative models learn the boundary between classes, e.g., Logistic Regression.
Generative models model data distribution, e.g., Naive Bayes.
QA202 (2 Marks):
The objective of Linear Regression is to model the relationship between a dependent
variable and one or more independent variables.
PART B (5 x 13 = 65 Marks)
QB101 (a) (13 Marks):
Bayes’ Theorem:
P(A|B) = [P(B|A) * P(A)] / P(B)
In ML, it's used in probabilistic classifiers like Naïve Bayes.
Example: Email spam filtering based on word probabilities.