ML CH 1
ML CH 1
TECHNOLOGY
FACULTY OF ENGINEERING & TECHNOLOGY
PARUL UNIVERSITY
❑ The need for machine learning is increasing day by day. The reason behind the need for
machine learning is that it is capable of doing tasks that are too complex for a person to
implement directly. As a human, we have some limitations as we cannot access the huge
amount of data manually, so for this, we need some computer systems and here comes the
machine learning to make things easy for us.
❑ We can train machine learning algorithms by providing them the huge amount of data and let
them explore the data, construct the models, and predict the required output automatically.
The performance of the machine learning algorithm depends on the amount of data, and it
can be determined by the cost function. With the help of machine learning, we can save both
time and money.
Need for Machine Learning
❑ Web search: ranking page based on what you are most likely to click on.
❑ Computational biology: rational design drugs in the computer based on past
experiments.
❑ Finance: decide who to send what credit card offers to. Evaluation of risk on credit
offers. How to decide where to invest money.
❑ E-commerce: Predicting customer churn. Whether or not a transaction is fraudulent
❑ Space exploration: space probes and radio astronomy.
Applications of Machine Learning
• The main goal of the supervised learning technique is to map the input variable(x) with
the output variable(y). Some real-world applications of supervised learning are Risk
Assessment, Fraud Detection, Spam filtering, etc.
Learning Paradigms in Machine Learning
2) Unsupervised Learning
• Unsupervised learning is a learning method in which a machine learns without
any supervision.
• The training is provided to the machine with the set of data that has not been
labeled, classified, or categorized, and the algorithm needs to act on that data
without any supervision. The goal of unsupervised learning is to restructure
the input data into new features or a group of objects with similar patterns.
• In unsupervised learning, we don't have a predetermined result. The machine
tries to find useful insights from the huge amount of data.
Learning Paradigms in Machine Learning
Advantages of Unsupervised Learning
• Unsupervised learning is used for more complex tasks as compared to supervised
learning because, in unsupervised learning, we don't have labeled input data.
• Unsupervised learning is preferable as it is easy to get unlabeled data in comparison to
labeled data.
Disadvantages of Unsupervised Learning
• Unsupervised learning is intrinsically more difficult than supervised learning as it does
not have corresponding output.
• The result of the unsupervised learning algorithm might be less accurate as input data
is not labeled, and algorithms do not know the exact output in advance.
Learning Paradigms in Machine Learning
3) Reinforcement Learning
Reinforcement learning is a feedback-based learning method, in which a learning
agent gets a reward for each right action and gets a penalty for each wrong
action. The agent learns automatically with these feedbacks and improves its
performance. In reinforcement learning, the agent interacts with the environment
and explores it. The goal of an agent is to get the most reward points, and hence,
it improves its performance.
The robotic dog, which automatically learns the movement of his arms, is an
example of Reinforcement learning.
Learning Paradigms in Machine Learning
Advantages and Disadvantages of Reinforcement Learning
Advantages
• It helps in solving complex real-world problems which are difficult to be
solved by general techniques.
• The learning model of RL is similar to the learning of human beings; hence
most accurate results can be found.
• Helps in achieving long term results.
Disadvantage
• RL algorithms are not preferred for simple problems.
• RL algorithms require huge data and computations.
• Too much reinforcement learning can lead to an overload of states which can
weaken the results.
PAC - Learning
• PAC stand for Probably approximately correct
• Probably approximately correct (PAC) learning is a framework for
mathematical analysis of machine learning algorithm.
• In the other words PAC learning is a theoretical framework for analyzing the
generalization performance of machine learning algorithms.
Goal: With High Probability ("Probably"),the select hypothesis will have lower
error ("Approximately Correct"")
In the PAC model ,we specify two small parameter ,ε (epsilon) and δ (delta) and
require that with probability at least (1-δ ) a system learn a concept with error at
most ε.
PAC - Learning
ε and δ parameters:
❑ ε gives an upper bound on the error in the accuracy with which h
approximated(Accuracy : 1-ε )
❑ δ gives the probability of failure in the achieving this accuracy
(Confidence : 1-δ)
PAC - Learning
❑ A good learner will learn with high probability and close approximation
to the target concept
❑ With high probability, the selected hypothesis will have lower the error
("Approximately Correct") with the parameter ε and δ
PAC - Learning
PAC – Learning Example
PAC – Learning Example
r
PAC – Learning Example
PAC – Learning Example
PAC – Learning Example
PAC – Learning Example
PAC – Learning Example
PAC – Learning Example
PAC – Learning Example
PAC – Learning Example
PAC – Learning Example
PAC – Learning Example
PAC – Learning Example
Version Space
• Subset of hypothesis H consistent with training example ( D ) .
H= hypothesis Consistent