Probability
- Basic Concepts
- Conditional Probability
- Random Variables
- Probability Distributions
- Variance of Random Variables
- Central Limit Theorem
- Law of Large Numbers
Concept | Definition | Characteristics/Rules | Examples | Use Cases |
---|---|---|---|---|
Experiment | A process producing uncertain but well-defined outcomes | Outcomes known in advance, actual outcome uncertain, repeatable | Flipping a coin, rolling a die, drawing a card, clicking an ad | Defines the starting point of probability analysis |
Outcome | A single possible result of an experiment | Represents one element from the sample space | Heads, Tails, rolling a "3", Ace of Spades | Used to build sample spaces |
Sample Space (S) | The set of all possible outcomes of an experiment | Finite (discrete) or infinite/continuous | for a coin | Framework for defining events |
Event (E) | A subset of the sample space | Can contain one or multiple outcomes | "Even die roll" = {2,4,6}; "at least one Head" = {HH,HT,TH} | Basis for probability calculations |
Probability of an Event | Numerical measure of likelihood (0 to 1) | impossible, certain | Probability of rolling 3 on die = 1/6 | Quantifies uncertainty |
Classical Probability | A priori assumption of equal likelihood | Rolling a 3 on a fair die = 1/6 | Used in games of chance, theoretical problems | |
Empirical Probability | Based on observed data/frequency | Ad clicks: 150/1000 = 0.15 | Foundation of data-driven analysis | |
Subjective Probability | Based on judgment or intuition | Not derived from calculation | Analyst predicts 70% chance of product success | Used in business decisions with scarce data |
Complementary Events | Opposite of an event () | Rule: | Click rate 0.15 → No Click 0.85 | Helps compute probabilities indirectly |
Mutually Exclusive Events | Events that cannot occur together | Rule: | Even vs odd die roll outcomes | Useful in disjoint scenarios |
Non-Mutually Exclusive Events | Events that can overlap | Rule: | Rolling even OR >4 → 2/3 | Key in overlapping categories or risks |
Concept | Formula | Example | Use Cases |
---|---|---|---|
Conditional Probability | Probability of event A given event B: | From a 52-card deck: Given the card is a Face Card, probability it's a King = 1/3 | Used in click-through rates, churn prediction, and medical diagnosis |
Independent Events | One event does not affect the other. | Two coin flips: probability of heads on first and second flip = | Simplifies modeling; many tests assume independence |
Dependent Events | One event affects the probability of the other. | Drawing two Kings in a row without replacement: | Critical for sequential data, customer behavior, anomaly chains |
Bayes' Theorem | Updates probability of A given evidence B: | Medical test: 1% prevalence, 95% true positive, 10% false positive. | Core for A/B testing, spam filtering, fraud detection, Naive Bayes models |
Concept | Definition | Properties | Examples | Use Cases |
---|---|---|---|---|
Random Variable | A function mapping outcomes of a random experiment to real numbers | Represents the numerical result of an unpredictable event | Number of heads in flips, height of a person, number of customers in a store | Forms foundation of modeling uncertainty and statistical inference |
Discrete Random Variable | Takes on finite or countably infinite distinct values (often integers) | Possible outcomes are countable | Coin flips, number of defective items, die roll, customer arrivals | Modeled with discrete distributions (e.g., Binomial, Poisson) |
Continuous Random Variable | Takes on values from any interval within the real line | Outcomes come from measurement, uncountably infinite | Height, weight, time to complete a task, daily sales revenue | Modeled with continuous distributions (e.g., Normal, Exponential) |
Probability Distribution | Describes how probabilities are assigned over possible values | Defines likelihood structure of a random variable | PMF for discrete, PDF for continuous, CDF for both | Core tool to compute likelihood, support inference, and fit models |
PMF (Probability Mass Function) | Assigns a probability to each possible value of a discrete random variable | Number of heads in two coin flips: PMF = | Essential for modeling counts and categorical outcomes | |
PDF (Probability Density Function) | Describes density for continuous random variables; probability is area under curve | Heights modeled by Normal distribution | Basis for calculating probabilities of ranges in continuous data | |
CDF (Cumulative Distribution Function) | Gives probability that . Works for both discrete and continuous variables |
| Coin flip example: | Used for quantiles, percentiles, range probabilities, and model fitting |
Type | Distribution | Parameters | PMF/PDF | Mean | Variance | Use Cases |
---|---|---|---|---|---|---|
Discrete | Bernoulli | Binary outcomes, click/no-click, churn models | ||||
Binomial | A/B testing, quality control, survey responses | |||||
Poisson | Rare events, call arrivals, web traffic, defects count | |||||
Geometric | Reliability, marketing conversion attempts, first defect detection | |||||
Continuous | Uniform | Random number generation, baseline models | ||||
Normal | Natural phenomena, CLT, parametric statistical tests, ML | |||||
Standard Normal | 0 | 1 | Z-scores, standardization, hypothesis testing | |||
Exponential | Reliability engineering, waiting times, customer lifetime | |||||
Chi-squared | (df) | Sum of squares of standard normals | Goodness-of-fit, independence tests, variance CI | |||
t-distribution | (df) | Symmetric, bell-shaped, heavier tails | 0 | t-tests, CI for mean, small samples | ||
F-distribution | Ratio of scaled chi-squared variates | Varies | ANOVA, regression model significance, variance comparison |
Aspect | Expected Value () | Variance () |
---|---|---|
Definition | The theoretical average (mean) of a random variable; a measure of central tendency | The average squared deviation from the mean; a measure of spread or variability |
Notation | or | or |
Discrete Formula | or | |
Continuous Formula | or | |
Units | Same as the random variable | Squared units of . Standard deviation () restores original units |
Properties |
|
|
Interpretation | Long-run average or "center" of the distribution | Degree of dispersion around the mean; how "spread out" values are |
Examples |
|
|
Use Cases | Decision making, expected returns, risk assessment, fairness in probability games, model evaluation | Risk measurement, quality control, hypothesis testing, error analysis (MSE), variability comparison between processes |
Aspect | Summary |
---|---|
Definition | States that the sampling distribution of the sample mean (or sum) of a large number of independent, identically distributed random variables approaches a normal distribution, regardless of the original population distribution |
Conditions |
|
Key Properties |
|
Importance |
|
Applications |
|
Examples |
|
Limitations |
|
Conceptual Visualization |
|
Concept | Definition | Key Properties | Applications |
---|---|---|---|
Law of Large Numbers (LLN) | As sample size increases, the sample mean converges to the population mean |
|
|
Weak Law of Large Numbers | Sample average converges to expected value in probability: |
|
|
Strong Law of Large Numbers | Sample average converges to expected value almost surely: |
|
|
Applications in Data Analysis | Explains why we can trust sample statistics with large samples |
|
|