Exam2021-2022 (Jan C)
Exam2021-2022 (Jan C)
2022-01-13, 17:00–20:00∗
Question 1 Given a hypothesis space L of binary classifiers. Given that the space contains
a finite amount n of classifiers, prove that V C(L) log2 (n)
Question 2 The first image is a binary dataset. The other images1 show decision surfaces
(coloured by certainty) for different classifiers trained on these data. Connect the classifiers:
Linear SVM Naive Bayes k-NN Random forest Decision tree
Question 3 Given the data in the table2 below, perform one step of decision tree learning,
and answer the following questions. Treat missing values as seen in the lectures.
1. What is the dataset’s class entropy?
2. What is the information gain of the chosen test?
3. What class would be predicted for an input (pp,yes,?)?
1
Question 4 Given the following pair of clauses.
Question 5 We have a binary dataset of 1600 instances with a class skew of 0.0625. On
this dataset, classifier has a positive precision of 0.8 and a TPR of 0.4.
Question 6 We trained two classifiers on the same binary dataset. Their decision bound-
aries are shown.4
1. Which model will have better accuracy on the training data? (2-3 sentences)
2. Which model will have better accuracy if we use a bigger set of testing data? (2-3
sentences)
3
Transcription note: Indeed, “balanced accuracy” was never seen in class, so try to think about what it
could be and why this dataset lends itself to balancing. We don’t make the rules :/.
4
Transcription note: The actual figure doesn’t seem to be on Google. It was basically two images of the
same twin point cloud (of about 1000 points altogether), where one image had a decision boundary with
long vertical and horizontal stretches, and the other had a decision boundary that also ran horizontally or
vertically, but more fine-grained, sometimes surrounding a single point.
2
Question 7 Execute the PC algorithm for a Bayesian network of variables A, B, C, D using
the below graphical model as an oracle for (un)conditional independences.
1. Give all the sets Sxy produced by the algorithm.
2. Give the final model.
C F D
Question 8 hAngelika’s question of death based on the 231-slide lecture.i You are given
a table of 20 artworks, along with their artist, and whether the artworks are expensive or
cheap. You are also given which of those artists is famous.5 You are also given some prior
percentages: of 320 artworks and 100 artists, we know P (fame) = 2%, P (exp | fame) = 40%,
P (exp | not fame) = 10%. Given that expensiveness follows from fame,
1. Draw a par-factor graph over all artworks and all artists.
2. Estimate the probabilities for famous, expensive given famous and expensive given
non-famous, using data from both the new table as well as the prior data.
3. Draw a par-factor graph for the situation where expensive artworks imply fame.
4. Can you estimate the probabilities for expensive, famous given expensive and famous
given non-expensive, or are we missing prior percentages to get there?
Question 9 Answer the following questions with 1
5
A4 text at most per answer.
1. What is learnt in model-based reinforcement learning?
2. What is learnt in model-free reinforcement learning?
3. What is off-policy learning?
4. What is the difference between passive and active model-free reinforcement learning?
5
Transcription note: The exact data don’t matter, so the reader is asked to come up with their own.