Probability notes
Probability notes
Probability theory is the branch of mathematics that deals with the analysis of random phenomena. It
provides a framework for quantifying uncertainty, predicting outcomes, and modeling random events.
The key concepts in probability theory are:
Probability: A number between 0 and 1 that quantifies the likelihood of an event occurring.
2. Basic Definitions
Sample Space (S): The set of all possible outcomes of an experiment. For example, when rolling a die, the
sample space is
,
6
S={1,2,3,4,5,6}.
Event (A): A subset of the sample space. For example, getting an even number when rolling a die is the
event
A={2,4,6}.
b. Probability
A, denoted as
0≤P(A)≤1
P(S)=1, where
A and
B,
𝐴
∪
P(A∪B)=P(A)+P(B).
3. Probability Rules
If two events,
A and
B, are mutually exclusive (i.e., they cannot happen simultaneously), the probability that either event
occurs is:
∪
𝐵
P(A∪B)=P(A)+P(B)
For events that may not be mutually exclusive, the probability of the union is given by:
)
+
P(A∪B)=P(A)+P(B)−P(A∩B)
Where
If events
A and
𝐵
B are independent (the occurrence of one does not affect the probability of the other), the probability
that both events occur is:
P(A∩B)=P(A)⋅P(B)
d. Conditional Probability
𝐴
A given
B, denoted as
P(A∣B):
𝐵
)
P(A∣B)=
P(B)
P(A∩B)
>
P(B)>0.
4. Types of Probability
a. Theoretical Probability
Theoretical probability is based on the assumption that all outcomes are equally likely. For example,
when rolling a fair six-sided die, the probability of rolling a 3 is:
rolling a 3
P(rolling a 3)=
6
P(A)=
c. Subjective Probability
Subjective probability is based on personal judgment, intuition, or experience rather than statistical
analysis or theoretical models.
a. Random Variables
A random variable is a function that assigns a real number to each outcome of a random experiment.
There are two types:
Discrete Random Variable: Takes on a finite or countably infinite number of values (e.g., the number of
heads in 3 coin flips).
Continuous Random Variable: Takes on an uncountably infinite number of values within a range (e.g., the
height of a person).
b. Probability Distribution
The probability distribution of a random variable gives the probability of each possible outcome. It can
be:
Discrete Probability Distribution: For a discrete random variable, the probability distribution is given by a
probability mass function (PMF).
Continuous Probability Distribution: For a continuous random variable, the probability distribution is
given by a probability density function (PDF).
a. Binomial Distribution
The binomial distribution models the number of successes in a fixed number of independent Bernoulli
trials (each with two possible outcomes, success or failure). The probability mass function is given by:
(
1
P(X=k)=(
)p
(1−p)
n−k
Where:
b. Normal Distribution
The normal distribution (or Gaussian distribution) is a continuous probability distribution that is
symmetric around its mean. The probability density function is:
𝑓
f(x)=
2πσ
2
1
2σ
(x−μ)
Where
σ is the standard deviation. The normal distribution is important in many fields due to the Central Limit
Theorem.
c. Poisson Distribution
The Poisson distribution models the number of events occurring in a fixed interval of time or space,
given a known average rate. Its probability mass function is:
𝑋
=
P(X=k)=
k!
−λ
Where
X, denoted as
E(X), is the long-term average or the "center" of the distribution. For a discrete random variable:
E(X)=
∑
x⋅P(X=x)
E(X)=∫
−∞
xf(x)dx
b. Variance
𝑋
X, denoted as
Var
Var(X), measures the spread or dispersion of the distribution. For a discrete random variable:
Var
Var(X)=E(X
2
)−(E(X))
Law of Large Numbers (LLN) states that as the number of trials in an experiment increases, the empirical
probability (relative frequency) of an event will converge to the theoretical probability. In other words,
with a large enough sample size, the average of observed outcomes will get closer to the expected value.
The Central Limit Theorem (CLT) states that for a sufficiently large sample size, the distribution of the
sum (or average) of independent and identically distributed random variables will approach a normal
distribution, regardless of the original distribution.
Statistics: Probability theory underlies statistical methods such as hypothesis testing, confidence
intervals, and regression analysis.
Machine Learning: Many algorithms in machine learning, like Naive Bayes and Bayesian inference, are
based on probability theory.
11. Conclusion
Probability theory is essential for modeling uncertainty, making predictions, and understanding random
processes. It forms the foundation for various fields, from statistics to machine learning and decision
theory. Mastery of probability theory helps in interpreting data, analyzing risks, and solving real-world
problems.