Basic Definitions
-
Probability - Study of uncertainty around anything.
-
Random Experiment - An experiment whose set of all possible outcomes is known, but the outcome of any trial is unknown until the trial is conducted.
-
Sample Space - Set of all possible outcomes of a random experiment.
-
Event - A set of outcomes of the random experiment. A subset of the sample space.
-
Trial - One repetition of a random experiment.
Types of Events
-
Independent Events - Set of events where occurrence of one event has no affect on occurrence of another.
-
Dependent Events - Set of events where occurrence of one event has some effect on occurrence of another.
-
Mutually Exclusive Events - Set of events where occurrence of one event guarantees that other events in the sample space won’t occur.
Conditional Probability
- Conditional Probability - Probability of one event occurring given that another has occurred.
- Total Probability Theorem - The probability of one event is the sum of intersections of that event and partitions of the sample space.
- Bayes Theorem - It tells us how knowing the and are related.
Here -
- Prior means probability of the hypothesis before seeing the data.
- Likelihood means how likely is the data if the hypothesis is true.
- Evidence means the probability of the data.
- Posterior - means probability of the hypothesis after seeing the data.
Random Variable
-
Random Variable - A real number associated to every outcome of a random experiment.
-
Range of a random variable - The set of values a random variable can hold.
-
Discrete Random Variable - Random variable whose range is a finite or countably infinite set.
-
Continuous Random Variable - Random variable whose range is an uncountably infinite set.
Probability Distributions
-
Probability Distribution - Describes the probability of a random variable holding a value.
-
Probability Mass Function - A function which describes how probability is distributed over the values of a continuous random variable.
-
Cumulative Distribution - A function which describes the probability of the random variable holding any value less than some . It holds the accumulated probabilities till .
-
Independent and Identically Distributed - Events are i.i.d. if occurrence of one doesn’t affect the occurrence of another and the probabilities of both events is the same.
General distributions -
- Bernoulli Distribution - The probability distribution where success has a probability of and failure has a probability of .
- Binomial Distribution - The probability of achieving successes upon performing i.i.d. Bernoulli trials.
- Geometric Distribution - The probability of achieving the first success after failures upon performing i.i.d. Bernoulli trials.
- Poisson Distribution - The probability of a number of events occurring in a fixed interval given the average rate of events.
- Equally Likely Distribution - Probability of a discrete R.V. where the probabilities of each event occurring is equal.
- Uniform Distribution - A probability distribution with a constant probability density over a fixed interval.
- Exponential Distribution - A probability distribution which models the time until the first event occurs.
- Normal/Gaussian Distribution - A probability distribution in which values are symmetrically distributed about the mean, with most observations clustering near the mean and fewer occurring as we move away from it.
Expected Values and Variance
- Expected Value - The average of all outcomes of a random variable if the experiment is conducted a large number of times.
- The expected values for the popular distributions are -
| Distribution | Expected Value |
|---|---|
| Bernoulli | |
| Binomial | |
| Geometric | |
| Poisson | |
| Uniform | |
| Exponential | |
| Gaussian |
-
- Mean - The expected value of a random variable.
- Variance - The spread of the values of a random variable around the mean.
The variance for the popular distributions are -
| Distribution | Expected Value |
|---|---|
| Bernoulli | |
| Binomial | |
| Geometric | |
| Poisson | |
| Uniform | |
| Exponential | |
| Gaussian |
Joint Distribution
-
Joint Distribution - The joint distribution of multiple random variables is the probability associated to all possible permutation of values the random variables hold simultaneously.
-
Marginal Distribution - The probability distribution of one random variable holding a fixed value over all possible values for the rest of the random variable is called the margin distribution for that random variable.
Discrete Marginals -
Continuous Marginals -
-
Expected Value - only when are independent.
-
Variance - only when are independent.
-
Joint Cumulative Distribution - The joint cumulative distribution function tells the probability of a set of random variables holding values less than or equal to some threshold.
-
Joint Moments -
Covariance
-
Covariance -
-
Correlation -
If and are independent, and are uncorrelated. But if and are uncorrelated, that doesn’t mean and are independent.
But if and are two Uncorrelated Gaussian Random Variables, both are always independent.
Joint Conditional Probability
Using this we can rewrite the Bayes Theorem w.r.t joint probability.
Theorems
- Markov Inequality - For any non-negative random variable with a finite and ,
Markov Inequality gives an exaggerated estimate of the probabilities, especially for tails. At times the upper bound provided by the Markov Inequality can be greater than 1 too.
- Chebyshev’s Inequality - For any real valued random variable with mean and variance , we say that
- We can apply Chebyshev’s inequality for Normal Distributions and try to find the probability of some value lying beyond some distance away from the mean . For such a case,
- Central Limit Theorem - The sample mean and the sample variance converge to the population mean and variance, respectively, as the sample size increases.