0% found this document useful (0 votes)
43 views9 pages

DL UNIT 1 (AB22) Continution

The history of machine learning spans several decades, beginning with early algorithms in the 1950s and 1960s, evolving through rule-based systems and a renaissance in the 1980s and 1990s with neural networks. The 2000s and 2010s saw the rise of deep learning fueled by big data, leading to modern applications in various fields. Key milestones include the development of foundational algorithms, significant breakthroughs in deep learning, and the emergence of frameworks that have shaped the current landscape of machine learning.

Uploaded by

venkata rajesh
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
43 views9 pages

DL UNIT 1 (AB22) Continution

The history of machine learning spans several decades, beginning with early algorithms in the 1950s and 1960s, evolving through rule-based systems and a renaissance in the 1980s and 1990s with neural networks. The 2000s and 2010s saw the rise of deep learning fueled by big data, leading to modern applications in various fields. Key milestones include the development of foundational algorithms, significant breakthroughs in deep learning, and the emergence of frameworks that have shaped the current landscape of machine learning.

Uploaded by

venkata rajesh
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 9

Q) summarize the history of machine learning?

Early Beginnings (1950s-1960s)

- The term "Machine Learning" was coined by Arthur Samuel in 1959

- The first machine learning algorithms were developed, including decision trees and linear regression

- The field was heavily influenced by the development of computer science and artificial intelligence

Rule-Based Systems (1970s-1980s)

- Machine learning focused on rule-based systems, which used hand-coded rules to make decisions

- Expert systems were developed, which mimicked human decision-making processes

- The field was dominated by symbolic reasoning and knowledge representation

Machine Learning Renaissance (1980s-1990s)

- The introduction of backpropagation and neural networks led to a resurgence in machine learning
research

- The development of support vector machines (SVMs) and other kernel-based methods

- The field began to shift towards statistical and probabilistic approaches

Big Data and Deep Learning (2000s-2010s)

- The availability of large datasets and computational power led to the development of deep learning
algorithms

- The introduction of convolutional neural networks (CNNs) and recurrent neural networks (RNNs)

- The field began to focus on large-scale machine learning and big data analytics

Modern Machine Learning (2010s-present)

- The development of transfer learning and pre-trained models

- The introduction of attention mechanisms and transformer models

- The field has expanded to include applications in natural language processing, computer vision, and
reinforcement learning

- The rise of deep learning frameworks such as TensorFlow, PyTorch, and Keras

Key Milestones

- 1951: The first computer program that could learn, called the "Logical Theorist," was developed by
Allen Newell and Herbert Simon

- 1965: The first neural network, called the "Perceptron," was developed by Frank Rosenblatt
- 1986: The backpropagation algorithm was introduced by David Rumelhart, Geoffrey Hinton, and
Ronald Williams

- 2011: The ImageNet Large Scale Visual Recognition Challenge (ILSVRC) was launched, which led to
significant advances in deep learning for computer vision

- 2014: The introduction of the AlexNet deep neural network, which won the ILSVRC competition and
sparked a wave of interest in deep learning

Notable Researchers

- Alan Turing: Developed the concept of the universal Turing machine and proposed the Turing test for
artificial intelligence

- Marvin Minsky: Developed the first neural network simulator and wrote the book "Perceptrons" with
Seymour Papert

- David Rumelhart: Introduced the backpropagation algorithm and developed the concept of distributed
representations

- Yann LeCun: Developed the convolutional neural network (CNN) and is a key figure in the development
of deep learning

- Andrew Ng: Developed the Stanford AI Lab (SAIL) and is a key figure in the development of deep
learning and AI education.

…………………………….. END…………

Q) overfitting and underfitting?

Overfitting

Overfitting occurs when a machine learning model is too complex and learns the noise in the training
data, rather than the underlying patterns. This results in a model that performs well on the training
data, but poorly on new, unseen data.

Underfitting

Underfitting occurs when a machine learning model is too simple and fails to capture the underlying
patterns in the training data. This results in a model that performs poorly on both the training data and
new, unseen data.

Key differences

- Overfitting: Model is too complex, learns noise in training data, and performs poorly on new data.

- Underfitting: Model is too simple, fails to capture underlying patterns, and performs poorly on both
training and new data.

Causes of overfitting

1. Model complexity: Models with too many parameters or layers can overfit the training data.

2. Noise in training data: Noisy or erroneous data can cause the model to learn the noise rather than the
underlying patterns.
3. Small training dataset: A small training dataset can lead to overfitting, as the model may not have
enough data to learn from.

Causes of underfitting

1. Model simplicity: Models that are too simple may not have enough capacity to learn the underlying
patterns in the data.

2. Insufficient training data: A small or incomplete training dataset can lead to underfitting, as the model
may not have enough data to learn from.

3. Poor feature engineering: Failure to extract relevant features from the data can lead to underfitting.

Solutions to overfitting

1. Regularization: Adding a penalty term to the loss function to discourage large weights.

2. Dropout: Randomly dropping out neurons during training to prevent over-reliance on any one neuron.

3. Early stopping: Stopping training when the model's performance on the validation set starts to
degrade.

4. Data augmentation: Increasing the size of the training dataset by adding noise or variations to the
existing data.

Solutions to underfitting

1. Increase model complexity: Adding more layers or parameters to the model to increase its capacity to
learn.

2. Collect more data: Gathering more data to increase the size of the training dataset.

3. Feature engineering: Extracting relevant features from the data to improve the model's performance.

4. Try different algorithms: Experimenting with different machine learning algorithms to find one that
better fits the data.

Real-world examples

- Overfitting: A model that predicts stock prices based on historical data, but performs poorly on new,
unseen data.

- Underfitting: A model that predicts customer churn based on a single feature, but fails to capture the
underlying patterns in the data.

Key takeaways

- Overfitting and underfitting are two common problems in machine learning that can affect a model's
performance.

- Understanding the causes of overfitting and underfitting can help you develop strategies to prevent or
address these issues.

- Regularization, dropout, and early stopping can help prevent overfitting, while increasing model
complexity, collecting more data, and feature engineering can help address underfitting.
…………………………… END…………………..

Q) supervised and unsupervised machine learning algorithms?

Supervised Machine Learning Algorithms

Supervised machine learning algorithms are trained on labeled data, where the correct output is already
known. The goal of these algorithms is to learn a mapping between input data and the corresponding
output labels, so that the algorithm can make predictions on new, unseen data.

Examples of Supervised Machine Learning Algorithms

1. Linear Regression: Linear regression is a linear approach to modeling the relationship between a
dependent variable and one or more independent variables.

2. Logistic Regression: Logistic regression is a linear approach to modeling the probability of a binary
outcome based on one or more predictor variables.

3. Decision Trees: Decision trees are a type of supervised learning algorithm that uses a tree-like model
to classify data or make predictions.

4. Random Forests: Random forests are an ensemble learning method that combines multiple decision
trees to improve the accuracy and robustness of predictions.

5. Support Vector Machines (SVMs): SVMs are a type of supervised learning algorithm that uses a
hyperplane to separate data into different classes.

Unsupervised Machine Learning Algorithms

Unsupervised machine learning algorithms are trained on unlabeled data, and the goal is to discover
patterns, relationships, or groupings in the data.

Examples of Unsupervised Machine Learning Algorithms

1. K-Means Clustering: K-means clustering is a type of unsupervised learning algorithm that groups
similar data points into clusters based on their features.

2. Hierarchical Clustering: Hierarchical clustering is a type of unsupervised learning algorithm that builds
a hierarchy of clusters by merging or splitting existing clusters.

3. Principal Component Analysis (PCA): PCA is a type of unsupervised learning algorithm that reduces the
dimensionality of high-dimensional data by transforming it into a new set of orthogonal features.

4. t-Distributed Stochastic Neighbor Embedding (t-SNE): t-SNE is a type of unsupervised learning


algorithm that maps high-dimensional data to a lower-dimensional space in a way that preserves the
local structure of the data.

5. Autoencoders: Autoencoders are a type of unsupervised learning algorithm that learns to compress
and reconstruct data, often for dimensionality reduction or generative modeling.

Key Differences

1. Labeled vs. Unlabeled Data: Supervised learning algorithms require labeled data, while unsupervised
learning algorithms do not.
2. Prediction vs. Pattern Discovery: Supervised learning algorithms are designed to make predictions,
while unsupervised learning algorithms are designed to discover patterns or relationships in the data.

3. Evaluation Metrics: Supervised learning algorithms are typically evaluated using metrics such as
accuracy, precision, and recall, while unsupervised learning algorithms are typically evaluated using
metrics such as clustering quality or dimensionality reduction.

Real-World Applications

1. Image Classification: Supervised learning algorithms can be used for image classification, such as
classifying images as either "cats" or "dogs".

2. Customer Segmentation: Unsupervised learning algorithms can be used for customer segmentation,
such as grouping customers based on their demographics and behavior.

3. Anomaly Detection: Unsupervised learning algorithms can be used for anomaly detection, such as
identifying unusual patterns in network traffic or transaction data.

4. Recommendation Systems: Supervised learning algorithms can be used for recommendation systems,
such as recommending products to users based on their past purchases and ratings.

5. Natural Language Processing: Supervised learning algorithms can be used for natural language
processing, such as sentiment analysis or text classification.

……………………… END………………..

Q) brief history and evolution of artificial intelligence?

Early Beginnings (1950s-1960s)

1. Dartmouth Summer Research Project (1956): The term "Artificial Intelligence" was coined by John
McCarthy, Marvin Minsky, Nathaniel Rochester, and Claude Shannon.

2. First AI Program (1956): The Logical Theorist, developed by Allen Newell and Herbert Simon, was the
first AI program.

3. ELIZA (1966): Joseph Weizenbaum developed ELIZA, a natural language processing (NLP) program that
could simulate a conversation.

Rule-Based Expert Systems (1970s-1980s)

1. MYCIN (1976): Edward Feigenbaum and his team developed MYCIN, a rule-based expert system for
diagnosing bacterial infections.

2. Expert Systems (1980s): Rule-based expert systems became popular, with applications in areas like
medicine, finance, and engineering.

Machine Learning and Neural Networks (1980s-1990s)

1. Backpropagation (1986): David Rumelhart, Geoffrey Hinton, and Ronald Williams introduced
backpropagation, a key algorithm for training neural networks.

2. Neural Networks (1990s): Neural networks became popular, with applications in areas like image
recognition, speech recognition, and NLP.
AI Winter (1980s-1990s)

1. Lack of Funding (1980s): AI research faced a decline in funding, leading to a period known as the "AI
winter."

2. Criticism and Skepticism (1990s): AI research was criticized for its lack of progress, leading to
skepticism about its potential.

Resurgence of AI (2000s-2010s)

1. Big Data and Compute Power (2000s): The availability of large datasets and increased compute power
enabled the development of more complex AI models.

2. Deep Learning (2010s): Deep learning techniques, such as convolutional neural networks (CNNs) and
recurrent neural networks (RNNs), became popular.

3. AI Applications (2010s): AI applications, such as virtual assistants, self-driving cars, and personalized
recommendations, became increasingly common.

Current State of AI (2020s)

1. Increased Adoption (2020s): AI is being adopted across industries, with applications in areas like
healthcare, finance, and education.

2. Advances in NLP and Computer Vision (2020s): NLP and computer vision have seen significant
advances, with applications in areas like chatbots, sentiment analysis, and image recognition.

3. Ethics and Responsibility (2020s): There is a growing focus on AI ethics and responsibility, with
concerns about bias, transparency, and accountability.

Future of AI

1. Increased Autonomy (2020s): AI systems are expected to become more autonomous, with the ability
to make decisions without human intervention.

2. Hybrid Approaches (2020s): Hybrid approaches, combining symbolic and connectionist AI, are
expected to become more popular.

3. Human-AI Collaboration (2020s): Human-AI collaboration is expected to become more prevalent, with
AI systems augmenting human capabilities.

Key Milestones

1. 1950: Alan Turing proposes the Turing Test, a measure of a machine's ability to exhibit intelligent
behavior.

2. 1956: The Dartmouth Summer Research Project on Artificial Intelligence is established.

3. 1965: The first AI program, Logical Theorist, is developed.

4. 1980: Expert systems become popular.

5. 2011: IBM's Watson system wins Jeopardy!, demonstrating the power of AI in natural language
processing.
6. 2014: Google acquires DeepMind, a leading AI research organization.

7. 2016: AlphaGo, a computer program developed by Google DeepMind, defeats a human world
champion in Go.

Notable Researchers

1. Alan Turing: Proposed the Turing Test and made significant contributions to the development of
computer science and AI.

2. Marvin Minsky: Developed the first neural network simulator and made significant contributions to
the development of AI.

3. John McCarthy: Coined the term "Artificial Intelligence" and made significant contributions to the
development of AI.

4. Geoffrey Hinton: Developed the backpropagation algorithm and made significant contributions to the
development of deep learning.

5. Yann LeCun: Developed the convolutional neural network (CNN) and made significant contributions to
the development of deep learning.

………………. END……………..

Q)Bias-Variance Trade-Off?

What is Bias?

In general, a machine learning model analyses the data, find patterns in it and make
predictions. While training, the model learns these patterns in the dataset and applies them to
test data for prediction. While making predictions, a difference occurs between prediction
values made by the model and actual values/expected values, and this difference is known as
bias errors or Errors due to bias. It can be defined as an inability of machine learning algorithms
such as Linear Regression to capture the true relationship between the data points. Each
algorithm begins with some amount of bias because bias occurs from assumptions in the
model, which makes the target function simple to learn. A model has either:

o Low Bias: A low bias model will make fewer assumptions about the form of the target
function.
o High Bias: A model with a high bias makes more assumptions, and the model becomes
unable to capture the important features of our dataset. A high bias model also cannot
perform well on new data.
Generally, a linear algorithm has a high bias, as it makes them learn fast. The simpler the
algorithm, the higher the bias it has likely to be introduced. Whereas a nonlinear algorithm
often has low bias.

Some examples of machine learning algorithms with low bias are Decision Trees, k-Nearest
Neighbours and Support Vector Machines. At the same time, an algorithm with high bias
is Linear Regression, Linear Discriminant Analysis and Logistic Regression.
Ways to reduce High Bias:
High bias mainly occurs due to a much simple model. Below are some ways to reduce the high
bias:

o Increase the input features as the model is underfitted.


o Decrease the regularization term.
o Use more complex models, such as including some polynomial features.

What is a Variance Error?

The variance would specify the amount of variation in the prediction if the different training
data was used. In simple words, variance tells that how much a random variable is different
from its expected value. Ideally, a model should not vary too much from one training dataset to
another, which means the algorithm should be good in understanding the hidden mapping
between inputs and output variables. Variance errors are either of low variance or high
variance.

Low variance means there is a small variation in the prediction of the target function with
changes in the training data set. At the same time, High variance shows a large variation in the
prediction of the target function with changes in the training dataset.

A model that shows high variance learns a lot and perform well with the training dataset, and
does not generalize well with the unseen dataset. As a result, such a model gives good results
with the training dataset but shows high error rates on the test dataset.

Since, with high variance, the model learns too much from the dataset, it leads to overfitting of
the model. A model with high variance has the below problems:

o A high variance model leads to overfitting.


o Increase model complexities.
Usually, nonlinear algorithms have a lot of flexibility to fit the model, have high variance.

Some examples of machine learning algorithms with low variance are, Linear Regression,
Logistic Regression, and Linear discriminant analysis. At the same time, algorithms with high
variance are decision tree, Support Vector Machine, and K-nearest neighbours.

Ways to Reduce High Variance:

o Reduce the input features or number of parameters as a model is overfitted.


o Do not use a much complex model.
o Increase the training data.
o Increase the Regularization term.

Bias-Variance Trade-Off

While building the machine learning model, it is really important to take care of bias and
variance in order to avoid overfitting and underfitting in the model. If the model is very simple
with fewer parameters, it may have low variance and high bias. Whereas, if the model has a
large number of parameters, it will have high variance and low bias. So, it is required to make a
balance between bias and variance errors, and this balance between the bias error and variance
error is known as the Bias-Variance trade-off.

For an accurate prediction of the model, algorithms need a low variance and low bias. But this
is not possible because bias and variance are related to each other:

o If we decrease the variance, it will increase the bias.


o If we decrease the bias, it will increase the variance.
Bias-Variance trade-off is a central issue in supervised learning. Ideally, we need a model that
accurately captures the regularities in training data and simultaneously generalizes well with
the unseen dataset. Unfortunately, doing this is not possible simultaneously. Because a high
variance algorithm may perform well with training data, but it may lead to overfitting to noisy
data. Whereas, high bias algorithm generates a much simple model that may not even capture
important regularities in the data. So, we need to find a sweet spot between bias and variance
to make an optimal model.

Hence, the Bias-Variance trade-off is about finding the sweet spot to make a balance between
bias and variance errors.

……………………………… end…………..

You might also like