▶ Table of Contents
Probability distributions can seem intimidating. I am no expert but I made this note to understand them.
Here is the web calculator for these distributions.
I. Foundations: Discrete vs. Continuous
Remember, the key difference is: these distributions are all discrete. This means they deal with things you count (whole numbers: 0, 1, 2, 3…), not things you measure (like height or temperature, which can be any value within a range).
II. Five Discrete Distributions: Explained
Let’s explore each distribution, focusing on what it counts, when to use it, and how to identify it.
A. Binomial Distribution
What it counts: The number of "successes" in a fixed number of independent trials, where each trial has only two possible outcomes ("success" or "failure"). Key parameters: n (number of trials), p (probability of success on a single trial).When to use it:
How to identify: Look for a fixed number of trials ( n ), two outcomes per trial (success/failure), independent trials, and a constant probability of success (p ). |
B. Discrete Uniform Distribution
What it counts: Each outcome in a finite set of possibilities has an equal chance of occurring. Key parameter: n (or range a to b ): The number of possible outcomes.When to use it:
How to identify: The key is "equally likely" or "uniform probability" for all outcomes. |
C. Geometric Distribution
What it counts: The number of trials needed to get the first success. Key parameter: p (probability of success on a single trial).When to use it:
How to identify: You're counting trials until the first success, trials are independent, the probability of success ( p ) is constant, and the number of trials is not fixed. |
D. Negative Binomial Distribution
What it counts: The number of trials needed to get a fixed number of successes. It's a generalization of the Geometric distribution. Key parameters: r (number of successes), p (probability of success on a single trial).When to use it:
How to identify: You're counting trials until a specific number of successes ( r ), trials are independent, the probability of success (p ) is constant, and the number of trials is not fixed. (The Geometric distribution is a special case where r = 1 .) |
E. Poisson Distribution
What it counts: The number of events occurring in a fixed interval of time or space, when the events happen independently and at a constant average rate. Key parameter: λ (lambda): The average number of events in the given interval.When to use it:
How to identify: You're counting events within a specific interval, events are independent, there's a known average rate ( λ ), and you're not dealing with "trials" in the same way as the other distributions. |
III. A Quick Reference
Distribution | What are you counting? | Fixed # of Trials? | Until First Success? | Constant Rate? | Key Parameters |
---|---|---|---|---|---|
Binomial | Number of successes in a fixed number of trials | Yes | No | No | n , p |
Discrete Uniform | Any one outcome from a set of equally likely possibilities | No (usually) | No | No | n (or a , b ) |
Geometric | Number of trials until the first success | No | Yes | No | p |
Negative Binomial | Number of trials until a fixed number of successes | No | No | No | r , p |
Poisson | Number of events in a fixed interval (time/space) | No | No | Yes | λ |
IV. Example
Llet’s use one scenario and ask five different questions, each corresponding to a different distribution. This will highlight how the question determines the appropriate distribution.
Scenario: Imagine you are a quality control inspector at a factory that makes widgets. Widgets are produced continuously.
Here are five different questions you might ask, and which distribution applies:
Binomial: You inspect a batch of 20 widgets (fixed number of trials). Each widget is either defective (failure) or not defective (success). You want to know the probability that exactly 3 of the 20 are defective. This is a binomial problem because you have a fixed number of trials (20), two outcomes per trial (defective/not defective), independent trials (one widget’s defectiveness doesn’t affect another), and a constant probability of a widget being defective.
Discrete Uniform: You randomly select one widget from the production line. You’re interested in the identification number stamped on the widget, which ranges from 1 to 100. Each number is equally likely. What’s the probability you pick widget #42? This is a discrete uniform problem because each ID number has an equal chance of being selected.
Geometric: You’re waiting for the first defective widget to come off the production line. You want to know the probability that the first defective widget is the 5th one you inspect. This is a geometric problem because you’re counting trials until the first success (finding a defective widget).
Negative Binomial: You want to find five defective widgets to analyze. You want to know the probability that you’ll need to inspect exactly 12 widgets to find your fifth defective one. This is a negative binomial problem because you’re counting trials until a fixed number of successes (five defective widgets).
Poisson: On average, 10 defective widgets are produced per hour. You want to know the probability that exactly 15 defective widgets will be produced in the next hour. This is a Poisson problem because you’re counting events (defective widgets) within a fixed interval (one hour), and you have a known average rate (10 per hour). The defects occur independently.
V. The Math Behind the Distributions
Now, let’s look at the mathematical formulas for each distribution: the Probability Mass Function (PMF), the Mean (Expected Value), and the Variance. We’ll also provide a worked example for each.
A. Functions of Discrete Random Variables
In practice, we often construct new random variables by performing arithmetic operations on other random variables. Let $X$ be a random variable, and let $c$ and $d$ be constants. If Y is a new random variable, and $Y = cX + d$, then $$E(Y) = cE(X) + d$$ $$V(Y) = c^2V(X)$$
B. Discrete Uniform Distribution
B. Discrete Uniform Distribution
Notation: $X \sim DU(a, b)$ (X follows a Discrete Uniform distribution with parameters a and b)
Probability Mass Function (PMF): $$f(x) = P(X = x) = \frac{1}{b - a + 1}, \quad x = a, a+1, …, b$$ This means each value between a and b (inclusive) has an equal probability of occurring.
Mean (Expected Value): $$E(X) = \mu = \frac{a + b}{2}$$
Variance: $$Var(X) = \sigma^2 = \frac{(b - a + 1)^2 - 1}{12}$$
Example: Rolling a fair six-sided die. Here, $a = 1$ and $b = 6$.
- $P(X = 3) = \frac{1}{6 - 1 + 1} = \frac{1}{6}$ (The probability of rolling a 3 is 1/6)
- $E(X) = \frac{1 + 6}{2} = 3.5$
- $Var(X) = \frac{(6 - 1 + 1)^2 - 1}{12} = \frac{36 - 1}{12} = \frac{35}{12} \approx 2.92$
C. Binomial Distribution
Notation: $X \sim Bin(n, p)$ (X follows a Binomial distribution with parameters n and p)
Probability Mass Function (PMF): $$f(x) = P(X = x) = \binom{n}{x} p^x (1 - p)^{n - x}, \quad x = 0, 1, …, n$$ where $\binom{n}{x} = \frac{n!}{x!(n-x)!}$ is the binomial coefficient, representing the number of ways to choose x successes from n trials.
Mean (Expected Value): $$E(X) = \mu = np$$
Variance: $$Var(X) = \sigma^2 = np(1 - p)$$
Example: Flipping a fair coin 5 times (n=5, p=0.5). What’s the probability of getting exactly 2 heads (x=2)?
$P(X = 2) = \binom{5}{2} (0.5)^2 (1 - 0.5)^{5 - 2} = \frac{5!}{2!3!} (0.25)(0.125) = 10 \times 0.25 \times 0.125 = 0.3125$
The binomial coefficient $\binom{5}{2} = \frac{5!}{2!3!} = \frac{5\times4\times3\times2\times1}{(2\times1)(3\times2\times1)} = \frac{120}{12} = 10$
$E(X) = 5 \times 0.5 = 2.5$
$Var(X) = 5 \times 0.5 \times (1 - 0.5) = 1.25$
D. Negative Binomial Distribution
- Notation: $X \sim NB(r, p)$
- Probability Mass Function (pmf): $$f(x) = P(X=x) = \binom{x-1}{r-1}p^{r}(1-p)^{x-r} \quad x=r, r+1, r+2,…$$
- Mean: $$\mu = E(X) = \frac{r}{p}$$
- Variance: $$\sigma^{2} = V(X) = \frac{r(1-p)}{p^{2}}$$
E. Geometric Distribution
Notation: $X \sim Geom(p)$ (X follows a Geometric distribution with parameter p)
Probability Mass Function (PMF): $$f(x) = P(X = x) = p(1 - p)^{x - 1}, \quad x = 1, 2, …$$
Mean (Expected Value): $$E(X) = \mu = \frac{1}{p}$$
Variance: $$Var(X) = \sigma^2 = \frac{1 - p}{p^2}$$
Example: Rolling a die until you get a 6 (p = 1/6). What’s the probability it takes exactly 3 rolls?
$P(X = 3) = (\frac{1}{6})(1 - \frac{1}{6})^{3 - 1} = (\frac{1}{6})(\frac{5}{6})^2 = \frac{25}{216} \approx 0.116$
$E(X) = \frac{1}{1/6} = 6$
$Var(X) = \frac{1 - 1/6}{(1/6)^2} = \frac{5/6}{1/36} = 30$
F. Poisson Distribution
Notation: $X \sim Poisson(\lambda)$ (X follows a Poisson distribution with parameter λ) Note: Often, the parameter is given as $\lambda T$, where $\lambda$ is the rate and $T$ is the length of the interval. We use $\lambda T$ in formulas for consistency.
Probability Mass Function (PMF): $$f(x) = P(X = x) = \frac{e^{-\lambda T} (\lambda T)^x}{x!}, \quad x = 0, 1, 2, …$$
Mean (Expected Value): $$E(X) = \mu = \lambda T$$
Variance: $$Var(X) = \sigma^2 = \lambda T$$
Example: Customers arrive at a store at an average rate of 5 per hour ($\lambda = 5$). What’s the probability that exactly 2 customers arrive in a given hour (T = 1)?
$P(X = 2) = \frac{e^{-5 \times 1} (5 \times 1)^2}{2!} = \frac{e^{-5} \times 25}{2} \approx 0.084$
$E(X) = 5 \times 1 = 5$
$Var(X) = 5 \times 1 = 5$