100% FREE Updated: Mar 2026 Probability Theory Random Variables and Distributions

Elementary Distributions

Comprehensive study notes on Elementary Distributions for CMI M.Sc. and Ph.D. Computer Science preparation. This chapter covers key concepts, formulas, and examples needed for your exam.

Elementary Distributions

This chapter lays the foundation for understanding probabilistic phenomena by introducing elementary discrete and continuous distributions. Mastery of these distributions is critical for comprehending advanced probabilistic models and is frequently assessed in examinations through both direct application and theoretical analysis.

Chapter Contents

|

| Topic |

|---|-------| | 1 | Discrete Distributions | | 2 | Continuous Distributions |

We begin with Discrete Distributions.

Part 1: Discrete Distributions

Discrete distributions model random variables that can take on a finite or countably infinite number of values. We use them to analyze outcomes such as the number of successes in trials or the count of events in a specific interval.

---

Core Concepts

1. Bernoulli Distribution

The Bernoulli distribution describes the outcome of a single trial with exactly two possible results: "success" (with probability pp) or "failure" (with probability 1βˆ’p1-p).

πŸ“ Bernoulli Probability Mass Function (PMF)

Let XX be a Bernoulli random variable. Its PMF is:

P(X=x)=px(1βˆ’p)1βˆ’xforΒ x∈{0,1}P(X=x) = p^x (1-p)^{1-x} \quad \text{for } x \in \{0, 1\}

Where:
pp = probability of success (0≀p≀10 \le p \le 1)
x=1x=1 for success, x=0x=0 for failure

Expected Value: E[X]=pE[X] = p
Variance: Var⁑(X)=p(1βˆ’p)\operatorname{Var}(X) = p(1-p)

Worked Example:

Consider a component that functions correctly with a probability of 0.90.9. We want to find the probability that a single randomly selected component functions correctly.

Step 1: Define the random variable and parameters.

> Let X=1X=1 if the component functions correctly, and X=0X=0 otherwise. This is a Bernoulli trial.
> The probability of success is p=0.9p = 0.9.

Step 2: Apply the Bernoulli PMF for x=1x=1.

>

P(X=1)=p1(1βˆ’p)1βˆ’1=p1(1βˆ’p)0=pP(X=1) = p^1 (1-p)^{1-1} = p^1 (1-p)^0 = p

>
P(X=1)=0.9P(X=1) = 0.9

Answer: The probability that a single component functions correctly is 0.90.9.

:::question type="MCQ" question="A quality control inspector checks a product for defects. The probability of a product being defective is 0.050.05. What is the probability that a randomly selected product is not defective?" options=["0.05","0.95","0.50","0.10"] answer="0.95" hint="Identify success and failure, and their probabilities." solution="Let X=1X=1 if the product is defective (success) and X=0X=0 if it is not defective (failure).
The probability of success is p=0.05p = 0.05.
We want to find the probability that the product is not defective, which corresponds to X=0X=0.

Step 1: Identify the probability of failure.
>

P(X=0)=(1βˆ’p)P(X=0) = (1-p)

>
P(X=0)=1βˆ’0.05P(X=0) = 1 - 0.05

>
P(X=0)=0.95P(X=0) = 0.95

Answer: The probability that a randomly selected product is not defective is 0.950.95."
:::

---

2. Binomial Distribution

The Binomial distribution models the number of successes in a fixed number of independent Bernoulli trials. Each trial has the same probability of success, pp.

πŸ“ Binomial Probability Mass Function (PMF)

Let XX be a Binomial random variable. Its PMF is:

P(X=k)=(nk)pk(1βˆ’p)nβˆ’kforΒ k∈{0,1,…,n}P(X=k) = \binom{n}{k} p^k (1-p)^{n-k} \quad \text{for } k \in \{0, 1, \ldots, n\}

Where:
nn = number of trials
kk = number of successes
pp = probability of success on a single trial (0≀p≀10 \le p \le 1)
(nk)=n!k!(nβˆ’k)!\binom{n}{k} = \frac{n!}{k!(n-k)!} is the binomial coefficient.

Expected Value: E[X]=npE[X] = np
Variance: Var⁑(X)=np(1βˆ’p)\operatorname{Var}(X) = np(1-p)

Worked Example:

A fair coin is tossed 10 times. We want to find the probability of getting exactly 7 heads.

Step 1: Define the random variable and parameters.

> Let XX be the number of heads in 10 tosses. This follows a Binomial distribution.
> Number of trials, n=10n = 10.
> Probability of success (getting a head), p=0.5p = 0.5.
> Number of desired successes, k=7k = 7.

Step 2: Apply the Binomial PMF.

>

P(X=7)=(107)(0.5)7(1βˆ’0.5)10βˆ’7P(X=7) = \binom{10}{7} (0.5)^7 (1-0.5)^{10-7}

>
P(X=7)=(107)(0.5)7(0.5)3P(X=7) = \binom{10}{7} (0.5)^7 (0.5)^3

>
P(X=7)=10!7!3!(0.5)10P(X=7) = \frac{10!}{7!3!} (0.5)^{10}

>
P(X=7)=10Γ—9Γ—83Γ—2Γ—1(0.5)10P(X=7) = \frac{10 \times 9 \times 8}{3 \times 2 \times 1} (0.5)^{10}

>
P(X=7)=120Γ—0.0009765625P(X=7) = 120 \times 0.0009765625

>
P(X=7)=0.1171875P(X=7) = 0.1171875

Answer: The probability of getting exactly 7 heads in 10 tosses is approximately 0.1170.117.

:::question type="NAT" question="A manufacturing process produces 5% defective items. If a random sample of 20 items is selected, what is the expected number of defective items in the sample?" answer="1.0" hint="Recall the formula for the expected value of a Binomial distribution." solution="Let XX be the number of defective items in a sample of 20. This follows a Binomial distribution.

Step 1: Identify the parameters.
> Number of trials, n=20n = 20.
> Probability of success (item being defective), p=0.05p = 0.05.

Step 2: Apply the formula for the expected value of a Binomial distribution.
>

E[X]=npE[X] = np

>
E[X]=20Γ—0.05E[X] = 20 \times 0.05

>
E[X]=1E[X] = 1

Answer: The expected number of defective items is 1.01.0."
:::

---

3. Geometric Distribution

The Geometric distribution models the number of Bernoulli trials required to achieve the first success. This includes the success itself.

πŸ“ Geometric Probability Mass Function (PMF)

Let XX be a Geometric random variable. Its PMF is:

P(X=k)=(1βˆ’p)kβˆ’1pforΒ k∈{1,2,3,…}P(X=k) = (1-p)^{k-1} p \quad \text{for } k \in \{1, 2, 3, \ldots\}

Where:
kk = number of trials until the first success (including the success)
pp = probability of success on a single trial (0<p≀10 < p \le 1)

Expected Value: E[X]=1pE[X] = \frac{1}{p}
Variance: Var⁑(X)=1βˆ’pp2\operatorname{Var}(X) = \frac{1-p}{p^2}

Worked Example:

A quality control process involves testing items until a non-defective item is found. The probability of an item being non-defective is 0.80.8. We want to find the probability that the first non-defective item is found on the 3rd trial.

Step 1: Define the random variable and parameters.

> Let XX be the number of trials until the first non-defective item is found. This follows a Geometric distribution.
> Probability of success (non-defective item), p=0.8p = 0.8.
> Number of desired trials, k=3k = 3.

Step 2: Apply the Geometric PMF.

>

P(X=3)=(1βˆ’p)3βˆ’1pP(X=3) = (1-p)^{3-1} p

>
P(X=3)=(1βˆ’0.8)2(0.8)P(X=3) = (1-0.8)^{2} (0.8)

>
P(X=3)=(0.2)2(0.8)P(X=3) = (0.2)^2 (0.8)

>
P(X=3)=0.04Γ—0.8P(X=3) = 0.04 \times 0.8

>
P(X=3)=0.032P(X=3) = 0.032

Answer: The probability that the first non-defective item is found on the 3rd trial is 0.0320.032.

:::question type="MCQ" question="A basketball player makes a free throw with a probability of 0.70.7. What is the probability that the player's first successful free throw occurs on their 4th attempt?" options=["0.7","0.021","0.0063","0.0189"] answer="0.0189" hint="The first success occurs on the kk-th trial means kβˆ’1k-1 failures followed by 1 success." solution="Let XX be the number of attempts until the first successful free throw. This follows a Geometric distribution.

Step 1: Identify the parameters.
> Probability of success, p=0.7p = 0.7.
> Number of desired trials, k=4k = 4.

Step 2: Apply the Geometric PMF.
>

P(X=k)=(1βˆ’p)kβˆ’1pP(X=k) = (1-p)^{k-1} p

>
P(X=4)=(1βˆ’0.7)4βˆ’1(0.7)P(X=4) = (1-0.7)^{4-1} (0.7)

>
P(X=4)=(0.3)3(0.7)P(X=4) = (0.3)^3 (0.7)

>
P(X=4)=0.027Γ—0.7P(X=4) = 0.027 \times 0.7

>
P(X=4)=0.0189P(X=4) = 0.0189

Answer: The probability that the player's first successful free throw occurs on their 4th attempt is 0.01890.0189."
:::

---

4. Negative Binomial Distribution

The Negative Binomial distribution generalizes the Geometric distribution. It models the number of Bernoulli trials required to achieve a specified number of successes, rr.

πŸ“ Negative Binomial Probability Mass Function (PMF)

Let XX be a Negative Binomial random variable. Its PMF is:

P(X=k)=(kβˆ’1rβˆ’1)pr(1βˆ’p)kβˆ’rforΒ k∈{r,r+1,r+2,…}P(X=k) = \binom{k-1}{r-1} p^r (1-p)^{k-r} \quad \text{for } k \in \{r, r+1, r+2, \ldots\}

Where:
kk = total number of trials until rr-th success (including the rr-th success)
rr = number of desired successes
pp = probability of success on a single trial (0<p≀10 < p \le 1)

Expected Value: E[X]=rpE[X] = \frac{r}{p}
Variance: Var⁑(X)=r(1βˆ’p)p2\operatorname{Var}(X) = \frac{r(1-p)}{p^2}

Worked Example:

A software company is hiring developers, and each candidate has a 0.20.2 probability of passing the coding interview, independently. We want to find the probability that the 5th successful hire occurs on the 10th interview.

Step 1: Define the random variable and parameters.

> Let XX be the number of interviews until the 5th successful hire. This follows a Negative Binomial distribution.
> Number of desired successes, r=5r = 5.
> Probability of success (passing interview), p=0.2p = 0.2.
> Total number of trials, k=10k = 10.

Step 2: Apply the Negative Binomial PMF.

>

P(X=10)=(10βˆ’15βˆ’1)(0.2)5(1βˆ’0.2)10βˆ’5P(X=10) = \binom{10-1}{5-1} (0.2)^5 (1-0.2)^{10-5}

>
P(X=10)=(94)(0.2)5(0.8)5P(X=10) = \binom{9}{4} (0.2)^5 (0.8)^5

>
P(X=10)=9!4!5!(0.00032)(0.32768)P(X=10) = \frac{9!}{4!5!} (0.00032) (0.32768)

>
P(X=10)=9Γ—8Γ—7Γ—64Γ—3Γ—2Γ—1(0.00032)(0.32768)P(X=10) = \frac{9 \times 8 \times 7 \times 6}{4 \times 3 \times 2 \times 1} (0.00032) (0.32768)

>
P(X=10)=126Γ—0.0001048576P(X=10) = 126 \times 0.0001048576

>
P(X=10)β‰ˆ0.01321P(X=10) \approx 0.01321

Answer: The probability that the 5th successful hire occurs on the 10th interview is approximately 0.01320.0132.

:::question type="MCQ" question="A machine produces items, and the probability of an item being non-defective is 0.90.9. We are interested in finding the 3rd non-defective item. What is the expected number of items we need to examine until we find the 3rd non-defective item?" options=["3.33","2.7","3","10"] answer="3.33" hint="Use the expected value formula for the Negative Binomial distribution." solution="Let XX be the number of items examined until the 3rd non-defective item is found. This follows a Negative Binomial distribution.

Step 1: Identify the parameters.
> Number of desired successes, r=3r = 3.
> Probability of success (non-defective item), p=0.9p = 0.9.

Step 2: Apply the formula for the expected value of a Negative Binomial distribution.
>

E[X]=rpE[X] = \frac{r}{p}

>
E[X]=30.9E[X] = \frac{3}{0.9}

>
E[X]=3.333…E[X] = 3.333\ldots

Answer: The expected number of items we need to examine is approximately 3.333.33."
:::

---

5. Poisson Distribution

The Poisson distribution models the number of events occurring in a fixed interval of time or space, given that these events occur with a known constant mean rate and independently of the time since the last event. It is often used for rare events.

πŸ“ Poisson Probability Mass Function (PMF)

Let XX be a Poisson random variable. Its PMF is:

P(X=k)=eβˆ’Ξ»Ξ»kk!forΒ k∈{0,1,2,…}P(X=k) = \frac{e^{-\lambda} \lambda^k}{k!} \quad \text{for } k \in \{0, 1, 2, \ldots\}

Where:
kk = number of events
Ξ»\lambda = average rate of events in the given interval (Ξ»>0\lambda > 0)
ee = Euler's number (approximately 2.718282.71828)

Expected Value: E[X]=Ξ»E[X] = \lambda
Variance: Var⁑(X)=λ\operatorname{Var}(X) = \lambda

Worked Example:

A call center receives an average of 44 calls per hour. We want to find the probability that the call center receives exactly 22 calls in a specific hour.

Step 1: Define the random variable and parameters.

> Let XX be the number of calls received in an hour. This follows a Poisson distribution.
> The average rate of calls, Ξ»=4\lambda = 4.
> Number of desired calls, k=2k = 2.

Step 2: Apply the Poisson PMF.

>

P(X=2)=eβˆ’4422!P(X=2) = \frac{e^{-4} 4^2}{2!}

>
P(X=2)=eβˆ’4Γ—162P(X=2) = \frac{e^{-4} \times 16}{2}

>
P(X=2)=8eβˆ’4P(X=2) = 8 e^{-4}

>
P(X=2)β‰ˆ8Γ—0.0183156P(X=2) \approx 8 \times 0.0183156

>
P(X=2)β‰ˆ0.14652P(X=2) \approx 0.14652

Answer: The probability that the call center receives exactly 2 calls in an hour is approximately 0.14650.1465.

:::question type="NAT" question="The number of typos on a page of a certain book follows a Poisson distribution with an average of 0.50.5 typos per page. What is the probability that a randomly selected page has no typos? (Round to 4 decimal places)" answer="0.6065" hint="Use the Poisson PMF with k=0k=0." solution="Let XX be the number of typos on a page. This follows a Poisson distribution.

Step 1: Identify the parameters.
> Average rate of typos, Ξ»=0.5\lambda = 0.5.
> Number of desired typos, k=0k = 0.

Step 2: Apply the Poisson PMF.
>

P(X=k)=eβˆ’Ξ»Ξ»kk!P(X=k) = \frac{e^{-\lambda} \lambda^k}{k!}

>
P(X=0)=eβˆ’0.5(0.5)00!P(X=0) = \frac{e^{-0.5} (0.5)^0}{0!}

>
P(X=0)=eβˆ’0.5Γ—11P(X=0) = \frac{e^{-0.5} \times 1}{1}

>
P(X=0)=eβˆ’0.5P(X=0) = e^{-0.5}

>
P(X=0)β‰ˆ0.6065306597…P(X=0) \approx 0.6065306597 \ldots

>
P(X=0)β‰ˆ0.6065P(X=0) \approx 0.6065

Answer: The probability that a randomly selected page has no typos is approximately 0.60650.6065."
:::

---

6. Hypergeometric Distribution

The Hypergeometric distribution models the number of successes in a sample drawn without replacement from a finite population containing a known number of successes and failures.

πŸ“ Hypergeometric Probability Mass Function (PMF)

Let XX be a Hypergeometric random variable. Its PMF is:

P(X=k)=(Kk)(Nβˆ’Knβˆ’k)(Nn)P(X=k) = \frac{\binom{K}{k} \binom{N-K}{n-k}}{\binom{N}{n}}

Where:
NN = total number of items in the population
KK = total number of "success" items in the population
nn = number of items drawn in the sample
kk = number of "success" items in the sample (number of desired successes)
max⁑(0,nβˆ’(Nβˆ’K))≀k≀min⁑(n,K)\max(0, n-(N-K)) \le k \le \min(n, K)

Expected Value: E[X]=nKNE[X] = n \frac{K}{N}
Variance: Var⁑(X)=nKNNβˆ’KNNβˆ’nNβˆ’1\operatorname{Var}(X) = n \frac{K}{N} \frac{N-K}{N} \frac{N-n}{N-1}

Worked Example:

A box contains 1010 items, of which 33 are defective. If a sample of 44 items is drawn randomly without replacement, we want to find the probability that exactly 11 of the sampled items is defective.

Step 1: Define the random variable and parameters.

> Let XX be the number of defective items in the sample. This follows a Hypergeometric distribution.
> Total population size, N=10N = 10.
> Total number of defective items in population, K=3K = 3.
> Sample size, n=4n = 4.
> Number of desired defective items in sample, k=1k = 1.

Step 2: Apply the Hypergeometric PMF.

>

P(X=1)=(31)(10βˆ’34βˆ’1)(104)P(X=1) = \frac{\binom{3}{1} \binom{10-3}{4-1}}{\binom{10}{4}}

>
P(X=1)=(31)(73)(104)P(X=1) = \frac{\binom{3}{1} \binom{7}{3}}{\binom{10}{4}}

>
P(X=1)=3Γ—7!3!4!10!4!6!P(X=1) = \frac{3 \times \frac{7!}{3!4!}}{\frac{10!}{4!6!}}

>
P(X=1)=3Γ—7Γ—6Γ—53Γ—2Γ—110Γ—9Γ—8Γ—74Γ—3Γ—2Γ—1P(X=1) = \frac{3 \times \frac{7 \times 6 \times 5}{3 \times 2 \times 1}}{\frac{10 \times 9 \times 8 \times 7}{4 \times 3 \times 2 \times 1}}

>
P(X=1)=3Γ—35210P(X=1) = \frac{3 \times 35}{210}

>
P(X=1)=105210P(X=1) = \frac{105}{210}

>
P(X=1)=0.5P(X=1) = 0.5

Answer: The probability that exactly 1 of the sampled items is defective is 0.50.5.

:::question type="MCQ" question="A deck of 52 cards contains 4 aces. If 5 cards are drawn randomly without replacement, what is the probability that exactly 2 of them are aces?" options=["(42)(483)(525)\frac{\binom{4}{2}\binom{48}{3}}{\binom{52}{5}}","(42)(485)(525)\frac{\binom{4}{2}\binom{48}{5}}{\binom{52}{5}}","(42)(523)(525)\frac{\binom{4}{2}\binom{52}{3}}{\binom{52}{5}}","(42)(483)(523)\frac{\binom{4}{2}\binom{48}{3}}{\binom{52}{3}}"] answer="(42)(483)(525)\frac{\binom{4}{2}\binom{48}{3}}{\binom{52}{5}}" hint="Identify the total population, number of successes in population, sample size, and number of successes in sample." solution="Let XX be the number of aces in the sample of 5 cards. This follows a Hypergeometric distribution.

Step 1: Identify the parameters.
> Total population size, N=52N = 52 (total cards).
> Total number of 'success' items (aces) in population, K=4K = 4.
> Sample size, n=5n = 5 (cards drawn).
> Number of desired 'success' items (aces) in sample, k=2k = 2.

Step 2: Apply the Hypergeometric PMF.
>

P(X=k)=(Kk)(Nβˆ’Knβˆ’k)(Nn)P(X=k) = \frac{\binom{K}{k} \binom{N-K}{n-k}}{\binom{N}{n}}

>
P(X=2)=(42)(52βˆ’45βˆ’2)(525)P(X=2) = \frac{\binom{4}{2} \binom{52-4}{5-2}}{\binom{52}{5}}

>
P(X=2)=(42)(483)(525)P(X=2) = \frac{\binom{4}{2} \binom{48}{3}}{\binom{52}{5}}

Answer: The correct expression for the probability is (42)(483)(525)\frac{\binom{4}{2}\binom{48}{3}}{\binom{52}{5}}."
:::

---

7. Discrete Uniform Distribution

The Discrete Uniform distribution assigns equal probability to each outcome in a finite set of possible values.

πŸ“ Discrete Uniform Probability Mass Function (PMF)

Let XX be a Discrete Uniform random variable. Its PMF is:

P(X=x)=1nforΒ x∈{x1,x2,…,xn}P(X=x) = \frac{1}{n} \quad \text{for } x \in \{x_1, x_2, \ldots, x_n\}

Where:
nn = total number of possible outcomes
xix_i = each distinct outcome

If the outcomes are integers from aa to bb (inclusive), then n=bβˆ’a+1n = b-a+1.

P(X=x)=1bβˆ’a+1forΒ x∈{a,a+1,…,b}P(X=x) = \frac{1}{b-a+1} \quad \text{for } x \in \{a, a+1, \ldots, b\}

Expected Value: E[X]=a+b2E[X] = \frac{a+b}{2}
Variance: Var⁑(X)=(bβˆ’a+1)2βˆ’112=(bβˆ’a)(bβˆ’a+2)12\operatorname{Var}(X) = \frac{(b-a+1)^2 - 1}{12} = \frac{(b-a)(b-a+2)}{12}

Worked Example:

A fair six-sided die is rolled. We want to find the probability of rolling a 4.

Step 1: Define the random variable and parameters.

> Let XX be the outcome of the die roll. This follows a Discrete Uniform distribution.
> The possible outcomes are {1,2,3,4,5,6}\{1, 2, 3, 4, 5, 6\}.
> Number of possible outcomes, n=6n = 6.
> Desired outcome, x=4x = 4.

Step 2: Apply the Discrete Uniform PMF.

>

P(X=4)=1nP(X=4) = \frac{1}{n}

>
P(X=4)=16P(X=4) = \frac{1}{6}

Answer: The probability of rolling a 4 is 16\frac{1}{6}.

:::question type="NAT" question="A random number generator produces integers from 1 to 100 (inclusive), with each integer having an equal probability of being generated. What is the expected value of the generated number?" answer="50.5" hint="Identify the range of outcomes and use the expected value formula for a Discrete Uniform distribution." solution="Let XX be the generated integer. This follows a Discrete Uniform distribution.

Step 1: Identify the parameters.
> The range of outcomes is from a=1a=1 to b=100b=100.
> Total number of outcomes, n=bβˆ’a+1=100βˆ’1+1=100n = b-a+1 = 100-1+1 = 100.

Step 2: Apply the formula for the expected value of a Discrete Uniform distribution.
>

E[X]=a+b2E[X] = \frac{a+b}{2}

>
E[X]=1+1002E[X] = \frac{1+100}{2}

>
E[X]=1012E[X] = \frac{101}{2}

>
E[X]=50.5E[X] = 50.5

Answer: The expected value of the generated number is 50.550.5."
:::

---

Advanced Applications

Worked Example:

An online server experiences failures at an average rate of 0.50.5 failures per day.

  • What is the probability of having exactly 2 failures in a 3-day period?

  • If the server is critical, what is the probability that it runs for 5 days without any failures?
  • Step 1: Identify the distribution and parameters for part 1.

    > The number of failures in a fixed period follows a Poisson distribution.
    > The average rate of failures per day is 0.50.5.
    > For a 3-day period, the new average rate Ξ»\lambda is 0.5Β failures/dayΓ—3Β days=1.50.5 \text{ failures/day} \times 3 \text{ days} = 1.5.
    > We want k=2k=2 failures.

    Step 2: Calculate the probability for part 1.

    >

    P(X=2)=eβˆ’1.5(1.5)22!P(X=2) = \frac{e^{-1.5} (1.5)^2}{2!}

    >
    P(X=2)=eβˆ’1.5Γ—2.252P(X=2) = \frac{e^{-1.5} \times 2.25}{2}

    >
    P(X=2)=1.125Γ—eβˆ’1.5P(X=2) = 1.125 \times e^{-1.5}

    >
    P(X=2)β‰ˆ1.125Γ—0.22313P(X=2) \approx 1.125 \times 0.22313

    >
    P(X=2)β‰ˆ0.2510P(X=2) \approx 0.2510

    Step 3: Identify the distribution and parameters for part 2.

    > For the server to run for 5 days without any failures, this means 0 failures in a 5-day period. This is still a Poisson distribution.
    > The average rate of failures per day is 0.50.5.
    > For a 5-day period, the new average rate Ξ»\lambda is 0.5Β failures/dayΓ—5Β days=2.50.5 \text{ failures/day} \times 5 \text{ days} = 2.5.
    > We want k=0k=0 failures.

    Step 4: Calculate the probability for part 2.

    >

    P(X=0)=eβˆ’2.5(2.5)00!P(X=0) = \frac{e^{-2.5} (2.5)^0}{0!}

    >
    P(X=0)=eβˆ’2.5P(X=0) = e^{-2.5}

    >
    P(X=0)β‰ˆ0.08208P(X=0) \approx 0.08208

    Answer:

  • The probability of exactly 2 failures in a 3-day period is approximately 0.25100.2510.

  • The probability that the server runs for 5 days without any failures is approximately 0.08210.0821.
  • :::question type="MSQ" question="A company sends out marketing emails, and historically, 20% of the recipients open the email. If 10 emails are sent to different recipients, which of the following statements are correct? (Select all that apply)" options=["The probability that exactly 3 emails are opened is (103)(0.2)3(0.8)7\binom{10}{3}(0.2)^3(0.8)^7.","The expected number of opened emails is 2.","The probability that the first opened email is the 5th one sent is (0.8)4(0.2)(0.8)^4(0.2).","The variance of the number of opened emails is 2."] answer="The probability that exactly 3 emails are opened is \binom{10}{3}(0.2)^3(0.8)^7}.,The expected number of opened emails is 2.,The probability that the first opened email is the 5th one sent is (0.8)4(0.2)(0.8)^4(0.2)." hint="Analyze each statement based on Binomial and Geometric distributions." solution="Let XX be the number of opened emails out of 10. This is a Binomial distribution with n=10n=10 and p=0.2p=0.2.

    Statement 1: The probability that exactly 3 emails are opened is (103)(0.2)3(0.8)7\binom{10}{3}(0.2)^3(0.8)^7.
    > This is directly from the Binomial PMF: P(X=k)=(nk)pk(1βˆ’p)nβˆ’kP(X=k) = \binom{n}{k}p^k(1-p)^{n-k}.
    > For k=3k=3, P(X=3)=(103)(0.2)3(0.8)10βˆ’3=(103)(0.2)3(0.8)7P(X=3) = \binom{10}{3}(0.2)^3(0.8)^{10-3} = \binom{10}{3}(0.2)^3(0.8)^7.
    > This statement is Correct.

    Statement 2: The expected number of opened emails is 2.
    > For a Binomial distribution, E[X]=npE[X] = np.
    > E[X]=10Γ—0.2=2E[X] = 10 \times 0.2 = 2.
    > This statement is Correct.

    Statement 3: The probability that the first opened email is the 5th one sent is (0.8)4(0.2)(0.8)^4(0.2).
    > Let YY be the number of emails sent until the first one is opened. This is a Geometric distribution with p=0.2p=0.2.
    > P(Y=k)=(1βˆ’p)kβˆ’1pP(Y=k) = (1-p)^{k-1}p.
    > For k=5k=5, P(Y=5)=(1βˆ’0.2)5βˆ’1(0.2)=(0.8)4(0.2)P(Y=5) = (1-0.2)^{5-1}(0.2) = (0.8)^4(0.2).
    > This statement is Correct.

    Statement 4: The variance of the number of opened emails is 2.
    > For a Binomial distribution, Var⁑(X)=np(1βˆ’p)\operatorname{Var}(X) = np(1-p).
    > Var⁑(X)=10Γ—0.2Γ—(1βˆ’0.2)=10Γ—0.2Γ—0.8=1.6\operatorname{Var}(X) = 10 \times 0.2 \times (1-0.2) = 10 \times 0.2 \times 0.8 = 1.6.
    > The statement says the variance is 2, which is incorrect.
    > This statement is Incorrect.

    Answer: The correct options are 'The probability that exactly 3 emails are opened is (103)(0.2)3(0.8)7\binom{10}{3}(0.2)^3(0.8)^7.', 'The expected number of opened emails is 2.', 'The probability that the first opened email is the 5th one sent is (0.8)4(0.2)(0.8)^4(0.2).' "
    :::

    ---

    Problem-Solving Strategies

    πŸ’‘ Identifying the Correct Distribution

    When faced with a problem involving discrete random variables, carefully analyze the problem context to select the appropriate distribution:

      • Bernoulli: Single trial, two outcomes (success/failure).

      • Binomial: Fixed number of independent trials, counting successes. Key phrases: "out of nn trials," "number of successes."

      • Geometric: Number of trials until the first success. Key phrases: "first success on the kk-th trial," "how many attempts until."

      • Negative Binomial: Number of trials until the rr-th success. Key phrases: "rr-th success on the kk-th trial."

      • Poisson: Number of events in a fixed interval (time/space), average rate given. Key phrases: "average number of events per unit," "number of occurrences."

      • Hypergeometric: Sampling without replacement from a finite population with two categories. Key phrases: "drawn from a batch," "without replacement."

      • Discrete Uniform: Each outcome in a finite set is equally likely. Key phrases: "fair die," "random integer from aa to bb."

    ---

    Common Mistakes

    ⚠️ Watch Out

    ❌ Confusing Binomial and Hypergeometric:
    Students often use the Binomial distribution for sampling without replacement.
    βœ… Correct approach: If sampling is without replacement from a finite population, use the Hypergeometric distribution. If sampling is with replacement or from an infinite population (or large enough to approximate), use the Binomial.

    ❌ Incorrectly identifying kk for Geometric vs. Negative Binomial:
    For Geometric, kk is the total trials including the first success. For Negative Binomial, kk is the total trials including the rr-th success.
    βœ… Correct approach: Read carefully whether the question asks for the number of trials before the rr-th success or the total trials up to and including the rr-th success. The standard formulas define kk as total trials.

    ❌ Misinterpreting λ\lambda for Poisson distribution:
    The parameter Ξ»\lambda must correspond to the given interval in the question. If the average is per hour and the question asks about a 3-hour period, Ξ»\lambda must be scaled.
    βœ… Correct approach: Always adjust Ξ»\lambda to match the time or space unit specified in the probability question. For example, if average is 2 events/hour, then for a 0.5-hour interval, Ξ»=2Γ—0.5=1\lambda = 2 \times 0.5 = 1.

    ---

    Practice Questions

    :::question type="NAT" question="A particular type of integrated circuit has a failure rate of 1 in 100 during the first 1000 hours of operation. If a batch of 50 such circuits is tested, what is the variance of the number of circuits that fail within the first 1000 hours?" answer="0.495" hint="Identify the distribution and its parameters. Recall the variance formula." solution="Let XX be the number of circuits that fail in the batch of 50. This follows a Binomial distribution.

    Step 1: Identify the parameters.
    > Number of trials (circuits), n=50n = 50.
    > Probability of success (a circuit failing), p=1/100=0.01p = 1/100 = 0.01.

    Step 2: Apply the formula for the variance of a Binomial distribution.
    >

    Var⁑(X)=np(1βˆ’p)\operatorname{Var}(X) = np(1-p)

    >
    Var⁑(X)=50Γ—0.01Γ—(1βˆ’0.01)\operatorname{Var}(X) = 50 \times 0.01 \times (1-0.01)

    >
    Var⁑(X)=0.5Γ—0.99\operatorname{Var}(X) = 0.5 \times 0.99

    >
    Var⁑(X)=0.495\operatorname{Var}(X) = 0.495

    Answer: The variance of the number of failing circuits is 0.4950.495."
    :::

    :::question type="MCQ" question="A biased coin has a probability of landing heads as 0.60.6. What is the probability that the 3rd head occurs on the 5th toss?" options=["(53)(0.6)3(0.4)2\binom{5}{3}(0.6)^3(0.4)^2","(42)(0.6)3(0.4)2\binom{4}{2}(0.6)^3(0.4)^2","(0.4)2(0.6)(0.4)^2(0.6)","(52)(0.6)3(0.4)2\binom{5}{2}(0.6)^3(0.4)^2"] answer="(42)(0.6)3(0.4)2\binom{4}{2}(0.6)^3(0.4)^2" hint="This involves a specific number of successes on a specific trial number." solution="Let XX be the number of tosses until the 3rd head. This follows a Negative Binomial distribution.

    Step 1: Identify the parameters.
    > Number of desired successes, r=3r = 3.
    > Probability of success (heads), p=0.6p = 0.6.
    > Total number of trials, k=5k = 5.

    Step 2: Apply the Negative Binomial PMF.
    >

    P(X=k)=(kβˆ’1rβˆ’1)pr(1βˆ’p)kβˆ’rP(X=k) = \binom{k-1}{r-1} p^r (1-p)^{k-r}

    >
    P(X=5)=(5βˆ’13βˆ’1)(0.6)3(1βˆ’0.6)5βˆ’3P(X=5) = \binom{5-1}{3-1} (0.6)^3 (1-0.6)^{5-3}

    >
    P(X=5)=(42)(0.6)3(0.4)2P(X=5) = \binom{4}{2} (0.6)^3 (0.4)^2

    Answer: The correct option is '(42)(0.6)3(0.4)2\binom{4}{2}(0.6)^3(0.4)^2'."
    :::

    :::question type="MSQ" question="Which of the following scenarios can be appropriately modeled by a Poisson distribution? (Select all that apply)" options=["The number of defective items in a sample of 1000 taken from a large production line with a known defect rate.","The number of cars passing a specific point on a highway in a 5-minute interval, given the average traffic flow.","The number of customers arriving at a store between 9 AM and 10 AM, given an average arrival rate.","The number of heads obtained when flipping a coin 20 times."] answer="The number of cars passing a specific point on a highway in a 5-minute interval, given the average traffic flow.,The number of customers arriving at a store between 9 AM and 10 AM, given an average arrival rate." hint="Poisson models events in an interval with an average rate." solution="Option 1: The number of defective items in a sample of 1000 taken from a large production line with a known defect rate.
    > This is a Binomial distribution (n=1000n=1000, p=p= defect rate). It can be approximated by Poisson if nn is large and pp is small, but it's fundamentally Binomial. So, not appropriately modeled as its primary distribution.

    Option 2: The number of cars passing a specific point on a highway in a 5-minute interval, given the average traffic flow.
    > This is a classic Poisson scenario: counting events (cars) in a fixed interval (5 minutes) with an average rate.
    > This statement is Correct.

    Option 3: The number of customers arriving at a store between 9 AM and 10 AM, given an average arrival rate.
    > Another classic Poisson scenario: counting events (customer arrivals) in a fixed time interval (1 hour) with an average rate.
    > This statement is Correct.

    Option 4: The number of heads obtained when flipping a coin 20 times.
    > This is a Binomial distribution (n=20n=20, p=0.5p=0.5).
    > This statement is Incorrect.

    Answer: The correct options are 'The number of cars passing a specific point on a highway in a 5-minute interval, given the average traffic flow.', 'The number of customers arriving at a store between 9 AM and 10 AM, given an average arrival rate.' "
    :::

    :::question type="NAT" question="A box contains 8 red balls and 4 blue balls. If 3 balls are drawn randomly without replacement, what is the probability that all 3 balls are red? (Round to 4 decimal places)" answer="0.2545" hint="This is sampling without replacement from a finite population." solution="Let XX be the number of red balls drawn. This follows a Hypergeometric distribution.

    Step 1: Identify the parameters.
    > Total population size, N=8Β (red)+4Β (blue)=12N = 8 \text{ (red)} + 4 \text{ (blue)} = 12.
    > Total number of 'success' items (red balls) in population, K=8K = 8.
    > Sample size, n=3n = 3.
    > Number of desired 'success' items (red balls) in sample, k=3k = 3.

    Step 2: Apply the Hypergeometric PMF.
    >

    P(X=k)=(Kk)(Nβˆ’Knβˆ’k)(Nn)P(X=k) = \frac{\binom{K}{k} \binom{N-K}{n-k}}{\binom{N}{n}}

    >
    P(X=3)=(83)(12βˆ’83βˆ’3)(123)P(X=3) = \frac{\binom{8}{3} \binom{12-8}{3-3}}{\binom{12}{3}}

    >
    P(X=3)=(83)(40)(123)P(X=3) = \frac{\binom{8}{3} \binom{4}{0}}{\binom{12}{3}}

    >
    P(X=3)=8!3!5!Γ—112!3!9!P(X=3) = \frac{\frac{8!}{3!5!} \times 1}{\frac{12!}{3!9!}}

    >
    P(X=3)=8Γ—7Γ—63Γ—2Γ—112Γ—11Γ—103Γ—2Γ—1P(X=3) = \frac{\frac{8 \times 7 \times 6}{3 \times 2 \times 1}}{ \frac{12 \times 11 \times 10}{3 \times 2 \times 1}}

    >
    P(X=3)=56220P(X=3) = \frac{56}{220}

    >
    P(X=3)β‰ˆ0.254545…P(X=3) \approx 0.254545 \ldots

    >
    P(X=3)β‰ˆ0.2545P(X=3) \approx 0.2545

    Answer: The probability that all 3 balls are red is approximately 0.25450.2545."
    :::

    :::question type="MCQ" question="A computer program generates a random integer XX between 1 and 10 (inclusive), such that each integer has an equal probability of being chosen. What is the variance of XX?" options=["2.5","8.25","10","9.1667"] answer="8.25" hint="This is a Discrete Uniform distribution. Use the variance formula for integers aa to bb." solution="Let XX be the random integer generated. This follows a Discrete Uniform distribution.

    Step 1: Identify the parameters.
    > The range of outcomes is from a=1a=1 to b=10b=10.
    > Total number of outcomes, n=bβˆ’a+1=10βˆ’1+1=10n = b-a+1 = 10-1+1 = 10.

    Step 2: Apply the formula for the variance of a Discrete Uniform distribution.
    >

    Var⁑(X)=(bβˆ’a+1)2βˆ’112\operatorname{Var}(X) = \frac{(b-a+1)^2 - 1}{12}

    >
    Var⁑(X)=(10βˆ’1+1)2βˆ’112\operatorname{Var}(X) = \frac{(10-1+1)^2 - 1}{12}

    >
    Var⁑(X)=(10)2βˆ’112\operatorname{Var}(X) = \frac{(10)^2 - 1}{12}

    >
    Var⁑(X)=100βˆ’112\operatorname{Var}(X) = \frac{100 - 1}{12}

    >
    Var⁑(X)=9912\operatorname{Var}(X) = \frac{99}{12}

    >
    Var⁑(X)=8.25\operatorname{Var}(X) = 8.25

    Answer: The variance of XX is 8.258.25."
    :::

    ---

    Summary

    ❗ Key Formulas & Takeaways

    |

    | Formula/Concept | Expression | Expected Value | Variance |

    |---|----------------|------------|----------------|----------| | 1 | Bernoulli | P(X=x)=px(1βˆ’p)1βˆ’xP(X=x) = p^x (1-p)^{1-x} | pp | p(1βˆ’p)p(1-p) | | 2 | Binomial | P(X=k)=(nk)pk(1βˆ’p)nβˆ’kP(X=k) = \binom{n}{k} p^k (1-p)^{n-k} | npnp | np(1βˆ’p)np(1-p) | | 3 | Geometric | P(X=k)=(1βˆ’p)kβˆ’1pP(X=k) = (1-p)^{k-1} p | 1/p1/p | (1βˆ’p)/p2(1-p)/p^2 | | 4 | Negative Binomial | P(X=k)=(kβˆ’1rβˆ’1)pr(1βˆ’p)kβˆ’rP(X=k) = \binom{k-1}{r-1} p^r (1-p)^{k-r} | r/pr/p | r(1βˆ’p)/p2r(1-p)/p^2 | | 5 | Poisson | P(X=k)=eβˆ’Ξ»Ξ»kk!P(X=k) = \frac{e^{-\lambda} \lambda^k}{k!} | Ξ»\lambda | Ξ»\lambda | | 6 | Hypergeometric | P(X=k)=(Kk)(Nβˆ’Knβˆ’k)(Nn)P(X=k) = \frac{\binom{K}{k} \binom{N-K}{n-k}}{\binom{N}{n}} | nK/NnK/N | nKNNβˆ’KNNβˆ’nNβˆ’1n \frac{K}{N} \frac{N-K}{N} \frac{N-n}{N-1} | | 7 | Discrete Uniform | P(X=x)=1bβˆ’a+1P(X=x) = \frac{1}{b-a+1} | (a+b)/2(a+b)/2 | (bβˆ’a+1)2βˆ’112\frac{(b-a+1)^2 - 1}{12} |

    ---

    What's Next?

    πŸ’‘ Continue Learning

    This topic connects to:

      • Continuous Distributions: Understanding discrete distributions is foundational for studying continuous counterparts like the Exponential, Normal, and Uniform distributions.

      • Central Limit Theorem: The Binomial distribution, under certain conditions, can be approximated by the Normal distribution, which is a key concept in the Central Limit Theorem.

      • Moment Generating Functions (MGFs): MGFs provide a powerful tool to derive expected values and variances for these distributions, and to prove their properties.

      • Stochastic Processes: Discrete distributions are building blocks for modeling discrete-time stochastic processes, such as Markov chains.

    ---

    πŸ’‘ Next Up

    Proceeding to Continuous Distributions.

    ---

    Part 2: Continuous Distributions

    Continuous distributions model random variables that can take any value within a given range, providing a framework for analyzing probabilities of events over continuous scales. We apply these distributions to solve problems involving quantities such as time, distance, or measurement errors.

    ---

    Core Concepts

    1. Probability Density Function (PDF) and Cumulative Distribution Function (CDF)

    The Probability Density Function (PDF), fX(x)f_X(x), describes the relative likelihood for a continuous random variable XX to take on a given value xx. The Cumulative Distribution Function (CDF), FX(x)F_X(x), gives the probability that XX will take a value less than or equal to xx.

    πŸ“ PDF and CDF Relationship

    For a continuous random variable XX:

    • FX(x)=βˆ«βˆ’βˆžxfX(t) dtF_X(x) = \int_{-\infty}^{x} f_X(t) \, dt

    • fX(x)=ddxFX(x)f_X(x) = \frac{d}{dx} F_X(x) (where FX(x)F_X(x) is differentiable)

    Where:
    fX(x)β‰₯0f_X(x) \geq 0 for all xx
    βˆ«βˆ’βˆžβˆžfX(x) dx=1\int_{-\infty}^{\infty} f_X(x) \, dx = 1 (total probability is 1)
    0≀FX(x)≀10 \leq F_X(x) \leq 1 for all xx
    FX(x)F_X(x) is non-decreasing
    * lim⁑xβ†’βˆ’βˆžFX(x)=0\lim_{x \to -\infty} F_X(x) = 0 and lim⁑xβ†’βˆžFX(x)=1\lim_{x \to \infty} F_X(x) = 1
    When to use: To define the probability characteristics of continuous random variables.

    Worked Example:
    Consider a random variable XX with CDF given by:

    FX(x)={0x<0x20≀x<11xβ‰₯1F_X(x) = \begin{cases} 0 & x < 0 \\ x^2 & 0 \leq x < 1 \\ 1 & x \geq 1 \end{cases}

    Find the PDF, fX(x)f_X(x).

    Step 1: Differentiate FX(x)F_X(x) with respect to xx for the interval where it is non-constant.

    >

    fX(x)=ddx(x2)=2xforΒ 0≀x<1f_X(x) = \frac{d}{dx} (x^2) = 2x \quad \text{for } 0 \leq x < 1

    Step 2: Combine with the piecewise definition.

    >

    fX(x)={2x0≀x<10otherwisef_X(x) = \begin{cases} 2x & 0 \leq x < 1 \\ 0 & \text{otherwise} \end{cases}

    Answer: The PDF is fX(x)=2xf_X(x) = 2x for 0≀x<10 \leq x < 1 and 00 otherwise.

    :::question type="MCQ" question="A continuous random variable XX has a PDF fX(x)=12eβˆ’x/2f_X(x) = \frac{1}{2}e^{-x/2} for xβ‰₯0x \geq 0 and 00 otherwise. What is P(X>2)P(X > 2)?" options=["eβˆ’1e^{-1}","eβˆ’2e^{-2}","1βˆ’eβˆ’11 - e^{-1}","1βˆ’eβˆ’21 - e^{-2}"] answer="eβˆ’1e^{-1}" hint="Use the CDF or direct integration of the PDF." solution="Step 1: Calculate the probability by integrating the PDF from 22 to ∞\infty.

    >

    P(X>2)=∫2∞12eβˆ’x/2 dxP(X > 2) = \int_{2}^{\infty} \frac{1}{2}e^{-x/2} \, dx

    Step 2: Perform the integration.

    >

    P(X>2)=[βˆ’eβˆ’x/2]2∞P(X > 2) = \left[ -e^{-x/2} \right]_{2}^{\infty}

    >
    P(X>2)=lim⁑bβ†’βˆž(βˆ’eβˆ’b/2)βˆ’(βˆ’eβˆ’2/2)P(X > 2) = \lim_{b \to \infty} (-e^{-b/2}) - (-e^{-2/2})

    >
    P(X>2)=0βˆ’(βˆ’eβˆ’1)P(X > 2) = 0 - (-e^{-1})

    >
    P(X>2)=eβˆ’1P(X > 2) = e^{-1}

    "
    :::

    ---

    2. Expectation and Variance for Continuous Random Variables

    The expectation (mean) E⁑[X]\operatorname{E}[X] represents the average value of a continuous random variable XX. The variance Var⁑(X)\operatorname{Var}(X) measures the spread or dispersion of the values of XX around its mean.

    πŸ“ Expectation and Variance

    For a continuous random variable XX with PDF fX(x)f_X(x):

    • Expectation: E⁑[X]=βˆ«βˆ’βˆžβˆžxfX(x) dx\operatorname{E}[X] = \int_{-\infty}^{\infty} x f_X(x) \, dx

    • Expectation of a function g(X)g(X): E⁑[g(X)]=βˆ«βˆ’βˆžβˆžg(x)fX(x) dx\operatorname{E}[g(X)] = \int_{-\infty}^{\infty} g(x) f_X(x) \, dx

    • Variance: Var⁑(X)=E⁑[(Xβˆ’E⁑[X])2]=E⁑[X2]βˆ’(E⁑[X])2\operatorname{Var}(X) = \operatorname{E}[(X - \operatorname{E}[X])^2] = \operatorname{E}[X^2] - (\operatorname{E}[X])^2

    • Second Moment: E⁑[X2]=βˆ«βˆ’βˆžβˆžx2fX(x) dx\operatorname{E}[X^2] = \int_{-\infty}^{\infty} x^2 f_X(x) \, dx

    When to use: To characterize the central tendency and variability of a continuous distribution.

    Worked Example:
    A continuous random variable XX has the PDF fX(x)=3x2f_X(x) = 3x^2 for 0≀x≀10 \leq x \leq 1 and 00 otherwise. Find E⁑[X]\operatorname{E}[X] and Var⁑(X)\operatorname{Var}(X).

    Step 1: Calculate E⁑[X]\operatorname{E}[X].

    >

    E⁑[X]=∫01x(3x2) dx=∫013x3 dx\operatorname{E}[X] = \int_{0}^{1} x (3x^2) \, dx = \int_{0}^{1} 3x^3 \, dx

    >
    E⁑[X]=[3x44]01=3(1)44βˆ’3(0)44=34\operatorname{E}[X] = \left[ \frac{3x^4}{4} \right]_{0}^{1} = \frac{3(1)^4}{4} - \frac{3(0)^4}{4} = \frac{3}{4}

    Step 2: Calculate E⁑[X2]\operatorname{E}[X^2].

    >

    E⁑[X2]=∫01x2(3x2) dx=∫013x4 dx\operatorname{E}[X^2] = \int_{0}^{1} x^2 (3x^2) \, dx = \int_{0}^{1} 3x^4 \, dx

    >
    E⁑[X2]=[3x55]01=3(1)55βˆ’3(0)55=35\operatorname{E}[X^2] = \left[ \frac{3x^5}{5} \right]_{0}^{1} = \frac{3(1)^5}{5} - \frac{3(0)^5}{5} = \frac{3}{5}

    Step 3: Calculate Var⁑(X)\operatorname{Var}(X).

    >

    Var⁑(X)=E⁑[X2]βˆ’(E⁑[X])2=35βˆ’(34)2\operatorname{Var}(X) = \operatorname{E}[X^2] - (\operatorname{E}[X])^2 = \frac{3}{5} - \left(\frac{3}{4}\right)^2

    >
    Var⁑(X)=35βˆ’916=48βˆ’4580=380\operatorname{Var}(X) = \frac{3}{5} - \frac{9}{16} = \frac{48 - 45}{80} = \frac{3}{80}

    Answer: E⁑[X]=34\operatorname{E}[X] = \frac{3}{4} and Var⁑(X)=380\operatorname{Var}(X) = \frac{3}{80}.

    :::question type="NAT" question="Let XX be a continuous random variable with PDF fX(x)=c(1βˆ’x2)f_X(x) = c(1-x^2) for 0≀x≀10 \leq x \leq 1 and 00 otherwise. Find E⁑[X]\operatorname{E}[X]. Round your answer to two decimal places." answer="0.38" hint="First, find the constant cc by integrating the PDF over its domain and setting it to 1. Then calculate E⁑[X]\operatorname{E}[X] using the formula." solution="Step 1: Find the constant cc. The total probability must be 1.

    >

    ∫01c(1βˆ’x2) dx=1\int_{0}^{1} c(1-x^2) \, dx = 1

    >
    c[xβˆ’x33]01=1c \left[ x - \frac{x^3}{3} \right]_{0}^{1} = 1

    >
    c((1βˆ’13)βˆ’(0βˆ’0))=1c \left( (1 - \frac{1}{3}) - (0 - 0) \right) = 1

    >
    c(23)=1β€…β€ŠβŸΉβ€…β€Šc=32c \left( \frac{2}{3} \right) = 1 \implies c = \frac{3}{2}

    Step 2: Calculate E⁑[X]\operatorname{E}[X] using the determined PDF fX(x)=32(1βˆ’x2)f_X(x) = \frac{3}{2}(1-x^2).

    >

    E⁑[X]=∫01xβ‹…32(1βˆ’x2) dx\operatorname{E}[X] = \int_{0}^{1} x \cdot \frac{3}{2}(1-x^2) \, dx

    >
    E⁑[X]=32∫01(xβˆ’x3) dx\operatorname{E}[X] = \frac{3}{2} \int_{0}^{1} (x - x^3) \, dx

    >
    E⁑[X]=32[x22βˆ’x44]01\operatorname{E}[X] = \frac{3}{2} \left[ \frac{x^2}{2} - \frac{x^4}{4} \right]_{0}^{1}

    >
    E⁑[X]=32((122βˆ’144)βˆ’(022βˆ’044))\operatorname{E}[X] = \frac{3}{2} \left( \left(\frac{1^2}{2} - \frac{1^4}{4}\right) - \left(\frac{0^2}{2} - \frac{0^4}{4}\right) \right)

    >
    E⁑[X]=32(12βˆ’14)\operatorname{E}[X] = \frac{3}{2} \left( \frac{1}{2} - \frac{1}{4} \right)

    >
    E⁑[X]=32(14)=38\operatorname{E}[X] = \frac{3}{2} \left( \frac{1}{4} \right) = \frac{3}{8}

    Step 3: Convert to two decimal places.

    >

    E⁑[X]=38=0.375β‰ˆ0.38\operatorname{E}[X] = \frac{3}{8} = 0.375 \approx 0.38

    "
    :::

    ---

    3. Uniform Distribution

    A continuous random variable XX has a Uniform distribution over the interval [a,b][a, b] if its PDF is constant within this interval and zero elsewhere. This implies that all values within the interval are equally likely.

    πŸ“ Uniform Distribution U(a,b)U(a, b)

    PDF:

    fX(x)={1bβˆ’aa≀x≀b0otherwisef_X(x) = \begin{cases} \frac{1}{b-a} & a \leq x \leq b \\ 0 & \text{otherwise} \end{cases}

    CDF:
    FX(x)={0x<axβˆ’abβˆ’aa≀x<b1xβ‰₯bF_X(x) = \begin{cases} 0 & x < a \\ \frac{x-a}{b-a} & a \leq x < b \\ 1 & x \geq b \end{cases}

    Mean: E⁑[X]=a+b2\operatorname{E}[X] = \frac{a+b}{2}
    Variance: Var⁑(X)=(bβˆ’a)212\operatorname{Var}(X) = \frac{(b-a)^2}{12}

    When to use: When all outcomes within a specified range are equally probable.

    Worked Example:
    A bus arrives at a stop at a random time between 10:00 AM and 10:15 AM. Let XX be the arrival time in minutes past 10:00 AM. Find the probability that the bus arrives between 10:05 AM and 10:10 AM.

    Step 1: Define the distribution parameters.
    The arrival time XX is uniformly distributed over the interval [0,15][0, 15] minutes. So, a=0a=0 and b=15b=15.

    Step 2: Write the PDF.

    >

    fX(x)=115βˆ’0=115forΒ 0≀x≀15f_X(x) = \frac{1}{15-0} = \frac{1}{15} \quad \text{for } 0 \leq x \leq 15

    Step 3: Calculate the probability P(5<X<10)P(5 < X < 10).

    >

    P(5<X<10)=∫510115 dxP(5 < X < 10) = \int_{5}^{10} \frac{1}{15} \, dx

    >
    P(5<X<10)=115[x]510P(5 < X < 10) = \frac{1}{15} [x]_{5}^{10}

    >
    P(5<X<10)=115(10βˆ’5)=515=13P(5 < X < 10) = \frac{1}{15} (10 - 5) = \frac{5}{15} = \frac{1}{3}

    Answer: The probability that the bus arrives between 10:05 AM and 10:10 AM is 13\frac{1}{3}.

    :::question type="MCQ" question="The lifespan of a certain electronic component, in hours, is uniformly distributed between 500 and 1500. What is the expected lifespan of the component?" options=["750 hours","1000 hours","1250 hours","1500 hours"] answer="1000 hours" hint="For a uniform distribution U(a,b)U(a, b), the expected value is simply the midpoint of the interval." solution="Step 1: Identify the parameters of the uniform distribution.
    The lifespan XX is uniformly distributed over [500,1500][500, 1500]. So, a=500a=500 and b=1500b=1500.

    Step 2: Apply the formula for the mean of a uniform distribution.

    >

    E⁑[X]=a+b2\operatorname{E}[X] = \frac{a+b}{2}

    >
    E⁑[X]=500+15002\operatorname{E}[X] = \frac{500+1500}{2}

    >
    E⁑[X]=20002=1000\operatorname{E}[X] = \frac{2000}{2} = 1000

    Answer: The expected lifespan of the component is 1000 hours."
    :::

    ---

    4. Exponential Distribution

    The Exponential distribution models the time until an event occurs in a Poisson process, where events occur continuously and independently at a constant average rate. It exhibits the memoryless property.

    πŸ“ Exponential Distribution Exp⁑(Ξ»)\operatorname{Exp}(\lambda)

    PDF:

    fX(x)={Ξ»eβˆ’Ξ»xxβ‰₯00x<0f_X(x) = \begin{cases} \lambda e^{-\lambda x} & x \geq 0 \\ 0 & x < 0 \end{cases}

    CDF:
    FX(x)={1βˆ’eβˆ’Ξ»xxβ‰₯00x<0F_X(x) = \begin{cases} 1 - e^{-\lambda x} & x \geq 0 \\ 0 & x < 0 \end{cases}

    Mean: E⁑[X]=1λ\operatorname{E}[X] = \frac{1}{\lambda}
    Variance: Var⁑(X)=1λ2\operatorname{Var}(X) = \frac{1}{\lambda^2}
    Memoryless Property: P(X>s+t∣X>s)=P(X>t)P(X > s+t \mid X > s) = P(X > t) for s,tβ‰₯0s, t \geq 0.

    When to use: For modeling waiting times, lifetimes of components, or inter-arrival times.

    Worked Example:
    The time (in minutes) a customer spends waiting for service at a bank follows an exponential distribution with an average waiting time of 5 minutes. What is the probability that a customer waits more than 10 minutes?

    Step 1: Determine the rate parameter Ξ»\lambda.
    The average waiting time is E⁑[X]=5\operatorname{E}[X] = 5 minutes. For an exponential distribution, E⁑[X]=1λ\operatorname{E}[X] = \frac{1}{\lambda}.

    >

    5=1Ξ»β€…β€ŠβŸΉβ€…β€ŠΞ»=155 = \frac{1}{\lambda} \implies \lambda = \frac{1}{5}

    Step 2: Use the CDF or direct integration to find P(X>10)P(X > 10).

    >

    P(X>10)=1βˆ’FX(10)=1βˆ’(1βˆ’eβˆ’Ξ»β‹…10)=eβˆ’Ξ»β‹…10P(X > 10) = 1 - F_X(10) = 1 - (1 - e^{-\lambda \cdot 10}) = e^{-\lambda \cdot 10}

    >
    P(X>10)=eβˆ’(1/5)β‹…10=eβˆ’2P(X > 10) = e^{-(1/5) \cdot 10} = e^{-2}

    Answer: The probability that a customer waits more than 10 minutes is eβˆ’2e^{-2}.

    :::question type="MCQ" question="The lifetime of a light bulb follows an exponential distribution with a mean of 800 hours. Given that a light bulb has already lasted 600 hours, what is the probability that it will last for at least another 200 hours?" options=["eβˆ’1/4e^{-1/4}","eβˆ’1/2e^{-1/2}","eβˆ’1e^{-1}","1βˆ’eβˆ’1/41 - e^{-1/4}"] answer="eβˆ’1/4e^{-1/4}" hint="This question tests the memoryless property of the exponential distribution." solution="Step 1: Determine the rate parameter Ξ»\lambda.
    The mean lifetime is E⁑[X]=800\operatorname{E}[X] = 800 hours. So, 1Ξ»=800β€…β€ŠβŸΉβ€…β€ŠΞ»=1800\frac{1}{\lambda} = 800 \implies \lambda = \frac{1}{800}.

    Step 2: Apply the memoryless property.
    The memoryless property states that P(X>s+t∣X>s)=P(X>t)P(X > s+t \mid X > s) = P(X > t).
    Here, s=600s = 600 hours (already lasted) and t=200t = 200 hours (another 200 hours).
    We need to find P(X>600+200∣X>600)=P(X>200)P(X > 600+200 \mid X > 600) = P(X > 200).

    Step 3: Calculate P(X>200)P(X > 200).

    >

    P(X>200)=eβˆ’Ξ»β‹…200P(X > 200) = e^{-\lambda \cdot 200}

    >
    P(X>200)=eβˆ’(1/800)β‹…200=eβˆ’200/800=eβˆ’1/4P(X > 200) = e^{-(1/800) \cdot 200} = e^{-200/800} = e^{-1/4}

    Answer: The probability that it will last for at least another 200 hours is eβˆ’1/4e^{-1/4}.
    "
    :::

    ---

    5. Normal (Gaussian) Distribution

    The Normal distribution is a symmetric, bell-shaped distribution characterized by its mean ΞΌ\mu and variance Οƒ2\sigma^2. It is fundamental in statistics due to the Central Limit Theorem.

    πŸ“ Normal Distribution N(ΞΌ,Οƒ2)N(\mu, \sigma^2)

    PDF:

    fX(x)=1Οƒ2Ο€eβˆ’(xβˆ’ΞΌ)2/(2Οƒ2)forΒ βˆ’βˆž<x<∞f_X(x) = \frac{1}{\sigma \sqrt{2\pi}} e^{-(x-\mu)^2 / (2\sigma^2)} \quad \text{for } -\infty < x < \infty

    CDF: FX(x)=Ξ¦(xβˆ’ΞΌΟƒ)F_X(x) = \Phi\left(\frac{x-\mu}{\sigma}\right), where Ξ¦(z)\Phi(z) is the CDF of the Standard Normal Distribution N(0,1)N(0,1). The CDF has no closed-form expression and is typically found using tables or software.
    Mean: E⁑[X]=μ\operatorname{E}[X] = \mu
    Variance: Var⁑(X)=Οƒ2\operatorname{Var}(X) = \sigma^2
    Standardization: If X∼N(ΞΌ,Οƒ2)X \sim N(\mu, \sigma^2), then Z=Xβˆ’ΞΌΟƒβˆΌN(0,1)Z = \frac{X-\mu}{\sigma} \sim N(0,1).

    When to use: For modeling natural phenomena, measurement errors, or sums of many independent random variables.

    Worked Example:
    Suppose the scores on a standardized test are normally distributed with a mean of 70 and a standard deviation of 10. What is the probability that a randomly selected student scores between 60 and 85? (Use Ξ¦(1.5)β‰ˆ0.9332\Phi(1.5) \approx 0.9332 and Ξ¦(βˆ’1)β‰ˆ0.1587\Phi(-1) \approx 0.1587).

    Step 1: Define the distribution parameters.
    X∼N(70,102)X \sim N(70, 10^2), so ΞΌ=70\mu = 70 and Οƒ=10\sigma = 10.

    Step 2: Standardize the values x1=60x_1 = 60 and x2=85x_2 = 85 to ZZ-scores.

    >

    Z1=60βˆ’7010=βˆ’1010=βˆ’1Z_1 = \frac{60 - 70}{10} = \frac{-10}{10} = -1

    >
    Z2=85βˆ’7010=1510=1.5Z_2 = \frac{85 - 70}{10} = \frac{15}{10} = 1.5

    Step 3: Calculate the probability using the standard normal CDF.

    >

    P(60<X<85)=P(βˆ’1<Z<1.5)P(60 < X < 85) = P(-1 < Z < 1.5)

    >
    P(βˆ’1<Z<1.5)=Ξ¦(1.5)βˆ’Ξ¦(βˆ’1)P(-1 < Z < 1.5) = \Phi(1.5) - \Phi(-1)

    >
    P(βˆ’1<Z<1.5)β‰ˆ0.9332βˆ’0.1587P(-1 < Z < 1.5) \approx 0.9332 - 0.1587

    >
    P(βˆ’1<Z<1.5)β‰ˆ0.7745P(-1 < Z < 1.5) \approx 0.7745

    Answer: The probability that a student scores between 60 and 85 is approximately 0.77450.7745.

    :::question type="MCQ" question="The weights of adult males in a city are normally distributed with a mean of 75 kg and a standard deviation of 5 kg. What percentage of adult males weigh between 70 kg and 80 kg? (Use the empirical rule or Ξ¦(1)β‰ˆ0.8413\Phi(1) \approx 0.8413)" options=["68.26%","95.45%","99.73%","50%"] answer="68.26%" hint="Recognize that the interval [70,80][70, 80] is within one standard deviation of the mean. Recall the empirical rule for normal distributions." solution="Step 1: Identify the distribution parameters.
    X∼N(75,52)X \sim N(75, 5^2), so ΞΌ=75\mu = 75 and Οƒ=5\sigma = 5.

    Step 2: Observe the given interval.
    The interval is [70,80][70, 80].
    70=75βˆ’5=ΞΌβˆ’Οƒ70 = 75 - 5 = \mu - \sigma
    80=75+5=ΞΌ+Οƒ80 = 75 + 5 = \mu + \sigma
    We are looking for the probability P(ΞΌβˆ’Οƒ<X<ΞΌ+Οƒ)P(\mu - \sigma < X < \mu + \sigma).

    Step 3: Apply the empirical rule.
    For a normal distribution, approximately 68.26% of data falls within one standard deviation of the mean.

    Alternatively, using Z-scores:
    Z1=70βˆ’755=βˆ’1Z_1 = \frac{70-75}{5} = -1
    Z2=80βˆ’755=1Z_2 = \frac{80-75}{5} = 1
    P(70<X<80)=P(βˆ’1<Z<1)=Ξ¦(1)βˆ’Ξ¦(βˆ’1)P(70 < X < 80) = P(-1 < Z < 1) = \Phi(1) - \Phi(-1)
    Using symmetry, Ξ¦(βˆ’1)=1βˆ’Ξ¦(1)\Phi(-1) = 1 - \Phi(1).
    P(βˆ’1<Z<1)=Ξ¦(1)βˆ’(1βˆ’Ξ¦(1))=2Ξ¦(1)βˆ’1P(-1 < Z < 1) = \Phi(1) - (1 - \Phi(1)) = 2\Phi(1) - 1
    Given Ξ¦(1)β‰ˆ0.8413\Phi(1) \approx 0.8413:
    P(βˆ’1<Z<1)β‰ˆ2(0.8413)βˆ’1=1.6826βˆ’1=0.6826P(-1 < Z < 1) \approx 2(0.8413) - 1 = 1.6826 - 1 = 0.6826.

    Answer: Approximately 68.26% of adult males weigh between 70 kg and 80 kg."
    :::

    ---

    6. Gamma Distribution

    The Gamma distribution is a versatile distribution used to model waiting times or the sum of independent exponentially distributed random variables. It is characterized by two parameters: shape (Ξ±\alpha) and rate (Ξ²\beta or scale ΞΈ=1/Ξ²\theta = 1/\beta).

    πŸ“ Gamma Distribution Gamma⁑(Ξ±,Ξ²)\operatorname{Gamma}(\alpha, \beta)

    PDF:

    fX(x)={Ξ²Ξ±xΞ±βˆ’1eβˆ’Ξ²xΞ“(Ξ±)xβ‰₯00x<0f_X(x) = \begin{cases} \frac{\beta^\alpha x^{\alpha-1} e^{-\beta x}}{\Gamma(\alpha)} & x \geq 0 \\ 0 & x < 0 \end{cases}

    Where:
    Ξ“(Ξ±)=∫0∞tΞ±βˆ’1eβˆ’t dt\Gamma(\alpha) = \int_{0}^{\infty} t^{\alpha-1} e^{-t} \, dt is the Gamma function.
    For integer Ξ±\alpha, Ξ“(Ξ±)=(Ξ±βˆ’1)!\Gamma(\alpha) = (\alpha-1)!.
    Mean: E⁑[X]=αβ\operatorname{E}[X] = \frac{\alpha}{\beta}
    Variance: Var⁑(X)=αβ2\operatorname{Var}(X) = \frac{\alpha}{\beta^2}
    Relationship to Exponential: If X1,…,XnX_1, \dots, X_n are i.i.d. Exp⁑(Ξ²)\operatorname{Exp}(\beta), then βˆ‘i=1nXi∼Gamma⁑(n,Ξ²)\sum_{i=1}^n X_i \sim \operatorname{Gamma}(n, \beta). (Here Ξ±=n\alpha=n)

    When to use: For modeling waiting times until nn events occur, or in reliability analysis.

    Worked Example:
    Suppose the waiting time for the first customer at a store is exponentially distributed with a mean of 2 minutes. What is the mean waiting time for the 3rd customer, assuming waiting times are independent?

    Step 1: Identify the parameters for a single exponential waiting time.
    The mean waiting time for the first customer (an exponential random variable) is 2 minutes. So for Exp⁑(Ξ²)\operatorname{Exp}(\beta), E⁑[X]=1Ξ²=2β€…β€ŠβŸΉβ€…β€ŠΞ²=12\operatorname{E}[X] = \frac{1}{\beta} = 2 \implies \beta = \frac{1}{2}.

    Step 2: Relate to the Gamma distribution.
    The waiting time for the 3rd customer is the sum of 3 independent exponential random variables, each with rate Ξ²=1/2\beta = 1/2. This sum follows a Gamma distribution with shape parameter Ξ±=3\alpha = 3 and rate parameter Ξ²=1/2\beta = 1/2.
    So, X∼Gamma⁑(3,1/2)X \sim \operatorname{Gamma}(3, 1/2).

    Step 3: Calculate the mean of the Gamma distribution.

    >

    E⁑[X]=αβ\operatorname{E}[X] = \frac{\alpha}{\beta}

    >
    E⁑[X]=31/2=6\operatorname{E}[X] = \frac{3}{1/2} = 6

    Answer: The mean waiting time for the 3rd customer is 6 minutes.

    :::question type="MCQ" question="A system has three identical components, each with a lifetime that follows an exponential distribution with a mean of 100 hours. If the system fails when all three components fail (they operate in parallel), what is the expected time until system failure? Assume component failures are independent." options=["100 hours","200 hours","300 hours","Not enough information"] answer="300 hours" hint="The sum of independent exponential random variables follows a Gamma distribution. Identify the parameters for the Gamma distribution." solution="Step 1: Determine the rate parameter Ξ²\beta for each component's exponential lifetime.
    Mean lifetime E⁑[Xi]=100\operatorname{E}[X_i] = 100 hours. For Exp⁑(β)\operatorname{Exp}(\beta), E⁑[Xi]=1β\operatorname{E}[X_i] = \frac{1}{\beta}.
    So, 1Ξ²=100β€…β€ŠβŸΉβ€…β€ŠΞ²=1100\frac{1}{\beta} = 100 \implies \beta = \frac{1}{100}.

    Step 2: Recognize that the system failure time is the sum of three independent exponential random variables.
    Let X1,X2,X3X_1, X_2, X_3 be the lifetimes of the three components. The system fails when all three fail, meaning the system lifetime is Y=X1+X2+X3Y = X_1 + X_2 + X_3.
    Since Xi∼Exp⁑(1/100)X_i \sim \operatorname{Exp}(1/100) independently, their sum YY follows a Gamma distribution with shape parameter α=3\alpha = 3 and rate parameter β=1/100\beta = 1/100.
    So, Y∼Gamma⁑(3,1/100)Y \sim \operatorname{Gamma}(3, 1/100).

    Step 3: Calculate the expected value of the Gamma distribution.

    >

    E⁑[Y]=αβ\operatorname{E}[Y] = \frac{\alpha}{\beta}

    >
    E⁑[Y]=31/100=3Γ—100=300\operatorname{E}[Y] = \frac{3}{1/100} = 3 \times 100 = 300

    Answer: The expected time until system failure is 300 hours."
    :::

    ---

    7. Beta Distribution

    The Beta distribution is defined on the interval [0,1][0, 1] and is widely used to model probabilities or proportions. It is characterized by two positive shape parameters, Ξ±\alpha and Ξ²\beta.

    πŸ“ Beta Distribution Beta⁑(Ξ±,Ξ²)\operatorname{Beta}(\alpha, \beta)

    PDF:

    fX(x)={xΞ±βˆ’1(1βˆ’x)Ξ²βˆ’1B(Ξ±,Ξ²)0≀x≀10otherwisef_X(x) = \begin{cases} \frac{x^{\alpha-1}(1-x)^{\beta-1}}{B(\alpha, \beta)} & 0 \leq x \leq 1 \\ 0 & \text{otherwise} \end{cases}

    Where:
    * B(Ξ±,Ξ²)=Ξ“(Ξ±)Ξ“(Ξ²)Ξ“(Ξ±+Ξ²)B(\alpha, \beta) = \frac{\Gamma(\alpha)\Gamma(\beta)}{\Gamma(\alpha+\beta)} is the Beta function.
    Mean: E⁑[X]=αα+β\operatorname{E}[X] = \frac{\alpha}{\alpha+\beta}
    Variance: Var⁑(X)=αβ(α+β)2(α+β+1)\operatorname{Var}(X) = \frac{\alpha\beta}{(\alpha+\beta)^2(\alpha+\beta+1)}

    When to use: For modeling probabilities, proportions, or values constrained between 0 and 1.

    Worked Example:
    A random variable XX representing the proportion of time a machine is operational follows a Beta distribution with parameters Ξ±=2\alpha=2 and Ξ²=3\beta=3. What is the expected proportion of time the machine is operational?

    Step 1: Identify the parameters of the Beta distribution.
    X∼Beta⁑(2,3)X \sim \operatorname{Beta}(2, 3), so α=2\alpha = 2 and β=3\beta = 3.

    Step 2: Apply the formula for the mean of a Beta distribution.

    >

    E⁑[X]=αα+β\operatorname{E}[X] = \frac{\alpha}{\alpha+\beta}

    >
    E⁑[X]=22+3=25\operatorname{E}[X] = \frac{2}{2+3} = \frac{2}{5}

    Answer: The expected proportion of time the machine is operational is 25\frac{2}{5} or 0.40.4.

    :::question type="MCQ" question="The proportion of defective items produced by a manufacturing process is modeled by a Beta distribution with parameters Ξ±=1\alpha=1 and Ξ²=1\beta=1. What is the PDF of this proportion?" options=["f(x)=1f(x) = 1 for 0≀x≀10 \leq x \leq 1","f(x)=xf(x) = x for 0≀x≀10 \leq x \leq 1","f(x)=1βˆ’xf(x) = 1-x for 0≀x≀10 \leq x \leq 1","f(x)=2x(1βˆ’x)f(x) = 2x(1-x) for 0≀x≀10 \leq x \leq 1"] answer="f(x)=1f(x) = 1 for 0≀x≀10 \leq x \leq 1" hint="Recall the definition of the Beta function B(Ξ±,Ξ²)B(\alpha, \beta) and simplify the PDF for the given parameters. What distribution is this equivalent to?" solution="Step 1: Identify the parameters of the Beta distribution.
    Ξ±=1\alpha = 1 and Ξ²=1\beta = 1.

    Step 2: Write down the PDF formula for Beta⁑(α,β)\operatorname{Beta}(\alpha, \beta).

    >

    fX(x)=xΞ±βˆ’1(1βˆ’x)Ξ²βˆ’1B(Ξ±,Ξ²)forΒ 0≀x≀1f_X(x) = \frac{x^{\alpha-1}(1-x)^{\beta-1}}{B(\alpha, \beta)} \quad \text{for } 0 \leq x \leq 1

    Step 3: Substitute Ξ±=1\alpha=1 and Ξ²=1\beta=1 into the PDF.

    >

    fX(x)=x1βˆ’1(1βˆ’x)1βˆ’1B(1,1)=x0(1βˆ’x)0B(1,1)=1B(1,1)f_X(x) = \frac{x^{1-1}(1-x)^{1-1}}{B(1, 1)} = \frac{x^0(1-x)^0}{B(1, 1)} = \frac{1}{B(1, 1)}

    Step 4: Calculate the Beta function B(1,1)B(1, 1).
    B(1,1)=Ξ“(1)Ξ“(1)Ξ“(1+1)=(1βˆ’1)!(1βˆ’1)!(2βˆ’1)!=0!β‹…0!1!=1β‹…11=1B(1, 1) = \frac{\Gamma(1)\Gamma(1)}{\Gamma(1+1)} = \frac{(1-1)!(1-1)!}{(2-1)!} = \frac{0! \cdot 0!}{1!} = \frac{1 \cdot 1}{1} = 1.

    Step 5: Substitute B(1,1)B(1, 1) back into the PDF.

    >

    fX(x)=11=1forΒ 0≀x≀1f_X(x) = \frac{1}{1} = 1 \quad \text{for } 0 \leq x \leq 1

    This is the PDF of a Uniform distribution U(0,1)U(0,1).

    Answer: The PDF is f(x)=1f(x) = 1 for 0≀x≀10 \leq x \leq 1."
    :::

    ---

    8. Lognormal Distribution

    A random variable XX is Lognormally distributed if its logarithm, ln⁑(X)\ln(X), is normally distributed. This distribution is suitable for modeling variables that are positively skewed and bounded below by zero, such as financial asset prices or income.

    πŸ“ Lognormal Distribution LN(ΞΌ,Οƒ2)LN(\mu, \sigma^2)

    If Y=ln⁑(X)Y = \ln(X) is normally distributed with mean ΞΌ\mu and variance Οƒ2\sigma^2 (i.e., Y∼N(ΞΌ,Οƒ2)Y \sim N(\mu, \sigma^2)), then XX is lognormally distributed.
    PDF:

    fX(x)=1xΟƒ2Ο€eβˆ’(ln⁑xβˆ’ΞΌ)2/(2Οƒ2)forΒ x>0f_X(x) = \frac{1}{x \sigma \sqrt{2\pi}} e^{-(\ln x - \mu)^2 / (2\sigma^2)} \quad \text{for } x > 0

    Mean: E⁑[X]=eΞΌ+Οƒ2/2\operatorname{E}[X] = e^{\mu + \sigma^2/2}
    Variance: Var⁑(X)=(eΟƒ2βˆ’1)e2ΞΌ+Οƒ2\operatorname{Var}(X) = (e^{\sigma^2} - 1)e^{2\mu + \sigma^2}

    When to use: For modeling variables that are products of many independent positive random variables, or naturally positive, skewed data.

    Worked Example:
    The price of a stock, XX, is lognormally distributed. If ln⁑(X)\ln(X) is normally distributed with mean ΞΌ=0.05\mu=0.05 and standard deviation Οƒ=0.10\sigma=0.10, what is the expected price of the stock?

    Step 1: Identify the parameters of the underlying normal distribution.
    For ln⁑(X)∼N(ΞΌ,Οƒ2)\ln(X) \sim N(\mu, \sigma^2), we have ΞΌ=0.05\mu = 0.05 and Οƒ=0.10\sigma = 0.10.
    Thus, Οƒ2=(0.10)2=0.01\sigma^2 = (0.10)^2 = 0.01.

    Step 2: Apply the formula for the mean of a Lognormal distribution.

    >

    E⁑[X]=eΞΌ+Οƒ2/2\operatorname{E}[X] = e^{\mu + \sigma^2/2}

    >
    E⁑[X]=e0.05+0.01/2\operatorname{E}[X] = e^{0.05 + 0.01/2}

    >
    E⁑[X]=e0.05+0.005\operatorname{E}[X] = e^{0.05 + 0.005}

    >
    E⁑[X]=e0.055\operatorname{E}[X] = e^{0.055}

    Answer: The expected price of the stock is e0.055e^{0.055}. (Numerical value is approx. 1.05651.0565).

    :::question type="MCQ" question="A random variable XX follows a Lognormal distribution such that ln⁑(X)∼N(2,0.25)\ln(X) \sim N(2, 0.25). Which of the following statements is true regarding XX?" options=["E⁑[X]=e2\operatorname{E}[X] = e^2","The median of XX is e2e^2","The variance of XX is 0.250.25","XX can take negative values"] answer="The median of XX is e2e^2" hint="Recall the properties of the Lognormal distribution, particularly how its median relates to the mean of the underlying normal distribution." solution="Step 1: Identify the parameters of the underlying normal distribution.
    For ln⁑(X)∼N(ΞΌ,Οƒ2)\ln(X) \sim N(\mu, \sigma^2), we have ΞΌ=2\mu = 2 and Οƒ2=0.25\sigma^2 = 0.25.

    Step 2: Evaluate each option.
    * E⁑[X]=e2\operatorname{E}[X] = e^2: The mean of a Lognormal distribution is eΞΌ+Οƒ2/2e^{\mu + \sigma^2/2}. Here, E⁑[X]=e2+0.25/2=e2.125\operatorname{E}[X] = e^{2 + 0.25/2} = e^{2.125}. So, this statement is false.
    * The median of XX is e2e^2: For a Lognormal distribution, the median is eΞΌe^\mu. Since ΞΌ=2\mu=2, the median is e2e^2. This statement is true.
    * The variance of XX is 0.250.25: The variance of XX is (eΟƒ2βˆ’1)e2ΞΌ+Οƒ2(e^{\sigma^2} - 1)e^{2\mu + \sigma^2}. This is not 0.250.25. The variance of ln⁑(X)\ln(X) is 0.250.25. So, this statement is false.
    * XX can take negative values: The Lognormal distribution is defined for x>0x > 0. So, this statement is false.

    Answer: The median of XX is e2e^2"
    :::

    ---

    9. Weibull Distribution

    The Weibull distribution is a flexible distribution commonly used in reliability engineering to model lifetimes of components. It can model decreasing, constant, or increasing failure rates depending on its shape parameter.

    πŸ“ Weibull Distribution W(k,Ξ»)W(k, \lambda)

    PDF:

    fX(x)={kΞ»(xΞ»)kβˆ’1eβˆ’(x/Ξ»)kxβ‰₯00x<0f_X(x) = \begin{cases} \frac{k}{\lambda} \left(\frac{x}{\lambda}\right)^{k-1} e^{-(x/\lambda)^k} & x \geq 0 \\ 0 & x < 0 \end{cases}

    Where:
    k>0k > 0 is the shape parameter.
    Ξ»>0\lambda > 0 is the scale parameter.
    Mean: E⁑[X]=λΓ(1+1/k)\operatorname{E}[X] = \lambda \Gamma(1 + 1/k)
    Variance: Var⁑(X)=Ξ»2[Ξ“(1+2/k)βˆ’(Ξ“(1+1/k))2]\operatorname{Var}(X) = \lambda^2 \left[ \Gamma(1 + 2/k) - (\Gamma(1 + 1/k))^2 \right]
    Relationship to Exponential: If k=1k=1, the Weibull distribution reduces to the Exponential distribution with rate 1/Ξ»1/\lambda.

    When to use: For modeling material fatigue, wind speed, or reliability of systems.

    Worked Example:
    The lifetime of a certain type of bearing (in years) follows a Weibull distribution with shape parameter k=1k=1 and scale parameter Ξ»=5\lambda=5. What is the mean lifetime of these bearings?

    Step 1: Identify the parameters of the Weibull distribution.
    k=1k=1 and Ξ»=5\lambda=5.

    Step 2: Apply the formula for the mean of a Weibull distribution.

    >

    E⁑[X]=λΓ(1+1/k)\operatorname{E}[X] = \lambda \Gamma(1 + 1/k)

    >
    E⁑[X]=5Ξ“(1+1/1)\operatorname{E}[X] = 5 \Gamma(1 + 1/1)

    >
    E⁑[X]=5Ξ“(2)\operatorname{E}[X] = 5 \Gamma(2)

    Step 3: Evaluate the Gamma function.
    Recall that for integer nn, Ξ“(n)=(nβˆ’1)!\Gamma(n) = (n-1)!. So, Ξ“(2)=(2βˆ’1)!=1!=1\Gamma(2) = (2-1)! = 1! = 1.

    >

    E⁑[X]=5β‹…1=5\operatorname{E}[X] = 5 \cdot 1 = 5

    Answer: The mean lifetime of the bearings is 5 years. (Note: Since k=1k=1, this is an exponential distribution with mean Ξ»=5\lambda=5).

    :::question type="MCQ" question="A device's time to failure (in months) is modeled by a Weibull distribution with parameters k=2k=2 and Ξ»=10\lambda=10. What is the initial failure rate trend of this device?" options=["Constant failure rate","Increasing failure rate","Decreasing failure rate","Cannot be determined without more information"] answer="Increasing failure rate" hint="The shape parameter kk determines the failure rate trend for a Weibull distribution. Consider the cases k<1k<1, k=1k=1, and k>1k>1." solution="Step 1: Identify the shape parameter kk.
    Here, k=2k=2.

    Step 2: Relate the shape parameter to the failure rate trend.
    * If k<1k < 1, the failure rate is decreasing over time. This indicates 'infant mortality' or devices that improve with age.
    * If k=1k = 1, the failure rate is constant (equivalent to an exponential distribution). This implies random failures, independent of age.
    * If k>1k > 1, the failure rate is increasing over time. This indicates 'wear-out' or devices that are more likely to fail as they age.

    Step 3: Conclude based on k=2k=2.
    Since k=2>1k=2 > 1, the device exhibits an increasing failure rate.

    Answer: Increasing failure rate"
    :::

    ---

    10. Cauchy Distribution

    The Cauchy distribution is a peculiar continuous distribution known for its heavy tails and the fact that its mean and variance are undefined. It is a classic example of a distribution for which the Law of Large Numbers does not apply.

    πŸ“ Cauchy Distribution Cauchy⁑(x0,Ξ³)\operatorname{Cauchy}(x_0, \gamma)

    PDF:

    fX(x)=1πγ[1+(xβˆ’x0Ξ³)2]forΒ βˆ’βˆž<x<∞f_X(x) = \frac{1}{\pi \gamma \left[ 1 + \left(\frac{x - x_0}{\gamma}\right)^2 \right]} \quad \text{for } -\infty < x < \infty

    Where:
    x0x_0 is the location parameter (median and mode).
    Ξ³>0\gamma > 0 is the scale parameter.
    Mean: Undefined
    Variance: Undefined

    When to use: In physics (e.g., resonance phenomena), or as a counterexample in probability theory due to its undefined moments.

    Worked Example:
    A random variable XX follows a Cauchy distribution with location parameter x0=0x_0=0 and scale parameter Ξ³=1\gamma=1. What is the median of this distribution?

    Step 1: Identify the location parameter.
    For a Cauchy distribution, the location parameter x0x_0 is also its median. Here, x0=0x_0=0.

    Step 2: State the median.
    The median of the Cauchy distribution is x0x_0.

    >

    Median(X)=0\text{Median}(X) = 0

    Answer: The median of the distribution is 0.

    :::question type="MCQ" question="Which of the following statements is true about a random variable XX following a Cauchy distribution?" options=["Its mean is always zero.","Its variance is finite.","The Central Limit Theorem applies to sums of Cauchy random variables.","Its PDF is symmetric around its location parameter."] answer="Its PDF is symmetric around its location parameter." hint="Recall the unique properties of the Cauchy distribution, especially regarding its moments and symmetry." solution="Step 1: Evaluate each statement based on the properties of the Cauchy distribution.
    * Its mean is always zero. False. The mean of a Cauchy distribution is undefined, not necessarily zero.
    * Its variance is finite. False. The variance of a Cauchy distribution is undefined.
    * The Central Limit Theorem applies to sums of Cauchy random variables. False. The Central Limit Theorem requires finite variance for the sum to converge to a normal distribution. For Cauchy random variables, the average of nn i.i.d. Cauchy variables is itself a Cauchy variable, not a normal one.
    * Its PDF is symmetric around its location parameter. True. The PDF fX(x)=1πγ[1+(xβˆ’x0Ξ³)2]f_X(x) = \frac{1}{\pi \gamma \left[ 1 + \left(\frac{x - x_0}{\gamma}\right)^2 \right]} is clearly symmetric around x0x_0, as replacing (xβˆ’x0)(x-x_0) with βˆ’(xβˆ’x0)-(x-x_0) does not change the value of the PDF.

    Answer: Its PDF is symmetric around its location parameter."
    :::

    ---

    Advanced Applications

    We apply the concepts of continuous distributions to solve problems involving transformations of random variables and finding probabilities in more complex scenarios.

    Worked Example:
    Let XX be a continuous random variable with PDF fX(x)=2xf_X(x) = 2x for 0≀x≀10 \leq x \leq 1 and 00 otherwise. Find the PDF of Y=X2Y = X^2.

    Step 1: Find the CDF of XX.
    For 0≀x≀10 \leq x \leq 1:

    >

    FX(x)=∫0x2t dt=[t2]0x=x2F_X(x) = \int_{0}^{x} 2t \, dt = [t^2]_{0}^{x} = x^2

    So, FX(x)=x2F_X(x) = x^2 for 0≀x≀10 \leq x \leq 1.

    Step 2: Find the CDF of YY.
    Since Y=X2Y = X^2, for 0≀y≀10 \leq y \leq 1, we have X=YX = \sqrt{Y}.
    The CDF of YY, FY(y)F_Y(y), is P(Y≀y)P(Y \leq y).

    >

    FY(y)=P(X2≀y)=P(X≀y)F_Y(y) = P(X^2 \leq y) = P(X \leq \sqrt{y})

    >
    FY(y)=FX(y)=(y)2=yforΒ 0≀y≀1F_Y(y) = F_X(\sqrt{y}) = (\sqrt{y})^2 = y \quad \text{for } 0 \leq y \leq 1

    And FY(y)=0F_Y(y) = 0 for y<0y < 0, FY(y)=1F_Y(y) = 1 for yβ‰₯1y \geq 1.

    Step 3: Find the PDF of YY by differentiating FY(y)F_Y(y).

    >

    fY(y)=ddyFY(y)=ddy(y)=1forΒ 0≀y≀1f_Y(y) = \frac{d}{dy} F_Y(y) = \frac{d}{dy} (y) = 1 \quad \text{for } 0 \leq y \leq 1

    >
    fY(y)={10≀y≀10otherwisef_Y(y) = \begin{cases} 1 & 0 \leq y \leq 1 \\ 0 & \text{otherwise} \end{cases}

    Answer: The PDF of Y=X2Y = X^2 is fY(y)=1f_Y(y) = 1 for 0≀y≀10 \leq y \leq 1 and 00 otherwise, which is a Uniform distribution U(0,1)U(0,1).

    :::question type="NAT" question="Let XX be an exponential random variable with parameter Ξ»=1\lambda = 1. Find the median of XX. Round your answer to two decimal places." answer="0.69" hint="The median mm is the value such that P(X≀m)=0.5P(X \leq m) = 0.5. Use the CDF of the exponential distribution." solution="Step 1: Write the CDF for an exponential distribution with Ξ»=1\lambda=1.

    >

    FX(x)=1βˆ’eβˆ’Ξ»x=1βˆ’eβˆ’xforΒ xβ‰₯0F_X(x) = 1 - e^{-\lambda x} = 1 - e^{-x} \quad \text{for } x \geq 0

    Step 2: Set FX(m)=0.5F_X(m) = 0.5 and solve for mm.

    >

    1βˆ’eβˆ’m=0.51 - e^{-m} = 0.5

    >
    eβˆ’m=0.5e^{-m} = 0.5

    >
    βˆ’m=ln⁑(0.5)-m = \ln(0.5)

    >
    m=βˆ’ln⁑(0.5)=ln⁑(1/0.5)=ln⁑(2)m = -\ln(0.5) = \ln(1/0.5) = \ln(2)

    Step 3: Calculate the numerical value and round to two decimal places.

    >

    m=ln⁑(2)β‰ˆ0.693147...m = \ln(2) \approx 0.693147...

    >
    mβ‰ˆ0.69m \approx 0.69

    Answer: 0.69"
    :::

    ---

    Problem-Solving Strategies

    πŸ’‘ CDF Method for Transformations

    When finding the PDF of a transformed variable Y=g(X)Y=g(X):

    • Find the CDF of YY: FY(y)=P(Y≀y)=P(g(X)≀y)F_Y(y) = P(Y \leq y) = P(g(X) \leq y).

    • Express P(g(X)≀y)P(g(X) \leq y) in terms of the CDF of XX, FX(x)F_X(x). This often involves solving for XX in terms of yy.

    • Differentiate FY(y)F_Y(y) with respect to yy to get fY(y)f_Y(y).

    πŸ’‘ Using Z-scores for Normal Distribution

    For any problem involving probabilities or percentiles with a Normal distribution N(ΞΌ,Οƒ2)N(\mu, \sigma^2), always standardize the values to ZZ-scores using Z=Xβˆ’ΞΌΟƒZ = \frac{X-\mu}{\sigma}. This allows the use of standard normal tables or Ξ¦(z)\Phi(z) values.

    ---

    Common Mistakes

    ⚠️ PDF vs. Probability

    ❌ Thinking fX(x)f_X(x) directly gives P(X=x)P(X=x).
    βœ… For continuous random variables, P(X=x)=0P(X=x) = 0. The PDF fX(x)f_X(x) represents the probability density at xx, not the probability. Probabilities are found by integrating the PDF over an interval.

    ⚠️ Memoryless Property Misuse

    ❌ Applying the memoryless property to non-exponential distributions.
    βœ… The memoryless property P(X>s+t∣X>s)=P(X>t)P(X > s+t \mid X > s) = P(X > t) is unique to the exponential distribution (and geometric for discrete). Do not assume it for other continuous distributions.

    ⚠️ Undefined Moments for Cauchy

    ❌ Attempting to calculate the mean or variance of a Cauchy distributed variable.
    βœ… The mean and variance of a Cauchy distribution are undefined. This is a key characteristic.

    ---

    Practice Questions

    :::question type="MCQ" question="The amount of time (in hours) a student spends studying for an exam is a continuous random variable XX with PDF fX(x)=kx(2βˆ’x)f_X(x) = kx(2-x) for 0≀x≀20 \leq x \leq 2 and 00 otherwise. What is the value of kk?" options=["1/21/2","3/43/4","3/83/8","1/41/4"] answer="3/43/4" hint="The total probability over the domain of the PDF must integrate to 1." solution="Step 1: Set the integral of the PDF over its domain equal to 1.

    >

    ∫02kx(2βˆ’x) dx=1\int_{0}^{2} kx(2-x) \, dx = 1

    >
    k∫02(2xβˆ’x2) dx=1k \int_{0}^{2} (2x - x^2) \, dx = 1

    Step 2: Evaluate the integral.

    >

    k[x2βˆ’x33]02=1k \left[ x^2 - \frac{x^3}{3} \right]_{0}^{2} = 1

    >
    k((22βˆ’233)βˆ’(02βˆ’033))=1k \left( \left(2^2 - \frac{2^3}{3}\right) - \left(0^2 - \frac{0^3}{3}\right) \right) = 1

    >
    k(4βˆ’83)=1k \left( 4 - \frac{8}{3} \right) = 1

    >
    k(12βˆ’83)=1k \left( \frac{12 - 8}{3} \right) = 1

    >
    k(43)=1k \left( \frac{4}{3} \right) = 1

    Step 3: Solve for kk.

    >

    k=34k = \frac{3}{4}

    "
    :::

    :::question type="NAT" question="The lifetime of a product (in years) is exponentially distributed with a mean of 4 years. What is the probability that the product lasts exactly 5 years? Round your answer to two decimal places." answer="0.00" hint="For a continuous random variable, the probability of it taking an exact value is zero." solution="Step 1: Understand the nature of continuous random variables.
    For any continuous random variable XX and any specific value xx, the probability P(X=x)P(X=x) is always 0. This is because the probability is defined as the area under the PDF curve, and the area under a single point is zero.

    Step 2: Apply this principle to the question.
    The question asks for the probability that the product lasts exactly 5 years. Since lifetime is a continuous variable, this probability is 0.

    >

    P(X=5)=0P(X=5) = 0

    Answer: 0.00"
    :::

    :::question type="MSQ" question="Let XX be a random variable with PDF fX(x)=1x2f_X(x) = \frac{1}{x^2} for xβ‰₯1x \geq 1 and 00 otherwise. Select ALL correct statements." options=["FX(x)=1βˆ’1/xF_X(x) = 1 - 1/x for xβ‰₯1x \geq 1","E⁑[X]\operatorname{E}[X] is finite.","Var⁑(X)\operatorname{Var}(X) is finite.","P(X>2)=1/2P(X > 2) = 1/2"] answer="FX(x)=1βˆ’1/xF_X(x) = 1 - 1/x for xβ‰₯1x \geq 1,P(X>2)=1/2P(X > 2) = 1/2" hint="Calculate the CDF, expectation, and variance. Remember to integrate carefully over the domain." solution="Step 1: Calculate the CDF FX(x)F_X(x).
    For xβ‰₯1x \geq 1:

    >

    FX(x)=∫1x1t2 dt=[βˆ’1t]1xF_X(x) = \int_{1}^{x} \frac{1}{t^2} \, dt = \left[ -\frac{1}{t} \right]_{1}^{x}

    >
    FX(x)=βˆ’1xβˆ’(βˆ’11)=1βˆ’1xF_X(x) = -\frac{1}{x} - \left(-\frac{1}{1}\right) = 1 - \frac{1}{x}

    So, FX(x)=1βˆ’1/xF_X(x) = 1 - 1/x for xβ‰₯1x \geq 1. This statement is correct.

    Step 2: Calculate E⁑[X]\operatorname{E}[X].

    >

    E⁑[X]=∫1∞xβ‹…1x2 dx=∫1∞1x dx\operatorname{E}[X] = \int_{1}^{\infty} x \cdot \frac{1}{x^2} \, dx = \int_{1}^{\infty} \frac{1}{x} \, dx

    >
    E⁑[X]=[ln⁑∣x∣]1∞=lim⁑bβ†’βˆž(ln⁑b)βˆ’ln⁑(1)=βˆžβˆ’0=∞\operatorname{E}[X] = \left[ \ln|x| \right]_{1}^{\infty} = \lim_{b \to \infty} (\ln b) - \ln(1) = \infty - 0 = \infty

    Since E⁑[X]\operatorname{E}[X] is infinite, it is not finite. This statement is incorrect.

    Step 3: Calculate Var⁑(X)\operatorname{Var}(X).
    Since E⁑[X]\operatorname{E}[X] is infinite, Var⁑(X)\operatorname{Var}(X) must also be infinite (or undefined).
    Var⁑(X)=E⁑[X2]βˆ’(E⁑[X])2\operatorname{Var}(X) = \operatorname{E}[X^2] - (\operatorname{E}[X])^2. If E⁑[X]\operatorname{E}[X] is infinite, E⁑[X2]\operatorname{E}[X^2] will also be infinite.
    >

    E⁑[X2]=∫1∞x2β‹…1x2 dx=∫1∞1 dx=[x]1∞=∞\operatorname{E}[X^2] = \int_{1}^{\infty} x^2 \cdot \frac{1}{x^2} \, dx = \int_{1}^{\infty} 1 \, dx = [x]_{1}^{\infty} = \infty

    So, Var⁑(X)\operatorname{Var}(X) is infinite. This statement is incorrect.

    Step 4: Calculate P(X>2)P(X > 2).
    Using the CDF:

    >

    P(X>2)=1βˆ’FX(2)=1βˆ’(1βˆ’12)=1βˆ’12=12P(X > 2) = 1 - F_X(2) = 1 - \left(1 - \frac{1}{2}\right) = 1 - \frac{1}{2} = \frac{1}{2}

    Alternatively, by integrating the PDF:

    >

    P(X>2)=∫2∞1x2 dx=[βˆ’1x]2∞P(X > 2) = \int_{2}^{\infty} \frac{1}{x^2} \, dx = \left[ -\frac{1}{x} \right]_{2}^{\infty}

    >
    P(X>2)=lim⁑bβ†’βˆž(βˆ’1b)βˆ’(βˆ’12)=0βˆ’(βˆ’12)=12P(X > 2) = \lim_{b \to \infty} \left(-\frac{1}{b}\right) - \left(-\frac{1}{2}\right) = 0 - \left(-\frac{1}{2}\right) = \frac{1}{2}

    This statement is correct.

    Answer: FX(x)=1βˆ’1/xF_X(x) = 1 - 1/x for xβ‰₯1x \geq 1,P(X>2)=1/2P(X > 2) = 1/2"
    :::

    :::question type="MCQ" question="If X∼N(5,4)X \sim N(5, 4), what is the distribution of Y=2X+3Y = 2X + 3?" options=["N(13,16)N(13, 16)","N(13,8)N(13, 8)","N(13,4)N(13, 4)","N(10,16)N(10, 16)"] answer="N(13,16)N(13, 16)" hint="Recall the properties of linear transformations of normal random variables: E⁑[aX+b]=aE⁑[X]+b\operatorname{E}[aX+b] = a\operatorname{E}[X]+b and Var⁑(aX+b)=a2Var⁑(X)\operatorname{Var}(aX+b) = a^2\operatorname{Var}(X)." solution="Step 1: Identify the mean and variance of XX.
    X∼N(5,4)X \sim N(5, 4), so E⁑[X]=5\operatorname{E}[X] = 5 and Var⁑(X)=4\operatorname{Var}(X) = 4.

    Step 2: Calculate the mean of YY.

    >

    E⁑[Y]=E⁑[2X+3]=2E⁑[X]+3\operatorname{E}[Y] = \operatorname{E}[2X + 3] = 2\operatorname{E}[X] + 3

    >
    E⁑[Y]=2(5)+3=10+3=13\operatorname{E}[Y] = 2(5) + 3 = 10 + 3 = 13

    Step 3: Calculate the variance of YY.

    >

    Var⁑(Y)=Var⁑(2X+3)=22Var⁑(X)\operatorname{Var}(Y) = \operatorname{Var}(2X + 3) = 2^2 \operatorname{Var}(X)

    >
    Var⁑(Y)=4β‹…4=16\operatorname{Var}(Y) = 4 \cdot 4 = 16

    Step 4: State the distribution of YY.
    A linear transformation of a normal random variable is also a normal random variable.
    So, Y∼N(13,16)Y \sim N(13, 16).

    Answer: N(13,16)N(13, 16)"
    :::

    :::question type="NAT" question="A random variable XX has a PDF fX(x)=12eβˆ’βˆ£x∣f_X(x) = \frac{1}{2}e^{-|x|} for βˆ’βˆž<x<∞-\infty < x < \infty. What is E⁑[∣X∣]\operatorname{E}[|X|]? Round your answer to two decimal places." answer="1.00" hint="Split the integral into two parts due to the absolute value in the PDF and the expectation formula." solution="Step 1: Write the formula for E⁑[∣X∣]\operatorname{E}[|X|].

    >

    E⁑[∣X∣]=βˆ«βˆ’βˆžβˆžβˆ£x∣fX(x) dx\operatorname{E}[|X|] = \int_{-\infty}^{\infty} |x| f_X(x) \, dx

    >
    E⁑[∣X∣]=βˆ«βˆ’βˆžβˆžβˆ£x∣12eβˆ’βˆ£xβˆ£β€‰dx\operatorname{E}[|X|] = \int_{-\infty}^{\infty} |x| \frac{1}{2}e^{-|x|} \, dx

    Step 2: Split the integral due to the absolute value.
    Since fX(x)f_X(x) is symmetric around 0, we can simplify the integral.

    >

    E⁑[∣X∣]=2∫0∞x12eβˆ’x dx\operatorname{E}[|X|] = 2 \int_{0}^{\infty} x \frac{1}{2}e^{-x} \, dx

    >
    E⁑[∣X∣]=∫0∞xeβˆ’x dx\operatorname{E}[|X|] = \int_{0}^{\infty} x e^{-x} \, dx

    Step 3: Evaluate the integral using integration by parts, or recognize it as the mean of an exponential distribution.
    The integral ∫0∞xeβˆ’x dx\int_{0}^{\infty} x e^{-x} \, dx is the mean of an exponential distribution with Ξ»=1\lambda = 1. The mean of Exp⁑(Ξ»)\operatorname{Exp}(\lambda) is 1/Ξ»1/\lambda.
    Here, Ξ»=1\lambda=1, so the integral value is 1/1=11/1 = 1.

    Alternatively, using integration by parts:
    Let u=xu=x, dv=eβˆ’xdxdv=e^{-x}dx. Then du=dxdu=dx, v=βˆ’eβˆ’xv=-e^{-x}.

    >

    ∫xeβˆ’x dx=βˆ’xeβˆ’xβˆ’βˆ«(βˆ’eβˆ’x) dx=βˆ’xeβˆ’xβˆ’eβˆ’x\int x e^{-x} \, dx = -xe^{-x} - \int (-e^{-x}) \, dx = -xe^{-x} - e^{-x}

    >
    ∫0∞xeβˆ’x dx=[βˆ’xeβˆ’xβˆ’eβˆ’x]0∞\int_{0}^{\infty} x e^{-x} \, dx = \left[ -xe^{-x} - e^{-x} \right]_{0}^{\infty}

    >
    =lim⁑bβ†’βˆž(βˆ’beβˆ’bβˆ’eβˆ’b)βˆ’(βˆ’0e0βˆ’e0)= \lim_{b \to \infty} (-be^{-b} - e^{-b}) - (-0e^0 - e^0)

    >
    =(0βˆ’0)βˆ’(0βˆ’1)=1= (0 - 0) - (0 - 1) = 1

    Answer: 1.00"
    :::

    ---

    Summary

    ❗ Key Formulas & Takeaways

    | Formula/Concept | Expression |
    |---|------------|
    | PDF properties | fX(x)β‰₯0f_X(x) \geq 0, ∫fX(x) dx=1\int f_X(x) \, dx = 1 |
    | CDF from PDF | FX(x)=βˆ«βˆ’βˆžxfX(t) dtF_X(x) = \int_{-\infty}^{x} f_X(t) \, dt |
    | Expectation | E⁑[X]=βˆ«βˆ’βˆžβˆžxfX(x) dx\operatorname{E}[X] = \int_{-\infty}^{\infty} x f_X(x) \, dx |
    | Variance | Var⁑(X)=E⁑[X2]βˆ’(E⁑[X])2\operatorname{Var}(X) = \operatorname{E}[X^2] - (\operatorname{E}[X])^2 |
    | Uniform U(a,b)U(a,b) Mean | E⁑[X]=(a+b)/2\operatorname{E}[X] = (a+b)/2 |
    | Uniform U(a,b)U(a,b) Variance | Var⁑(X)=(bβˆ’a)2/12\operatorname{Var}(X) = (b-a)^2/12 |
    | Exponential Exp⁑(λ)\operatorname{Exp}(\lambda) Mean | E⁑[X]=1/λ\operatorname{E}[X] = 1/\lambda |
    | Exponential Exp⁑(λ)\operatorname{Exp}(\lambda) Memoryless | P(X>s+t∣X>s)=P(X>t)P(X > s+t \mid X > s) = P(X > t) |
    | Normal N(ΞΌ,Οƒ2)N(\mu, \sigma^2) Standardization | Z=(Xβˆ’ΞΌ)/ΟƒβˆΌN(0,1)Z = (X-\mu)/\sigma \sim N(0,1) |
    | Gamma Gamma⁑(α,β)\operatorname{Gamma}(\alpha, \beta) Mean | E⁑[X]=α/β\operatorname{E}[X] = \alpha/\beta |
    | Beta Beta⁑(α,β)\operatorname{Beta}(\alpha, \beta) Mean | E⁑[X]=α/(α+β)\operatorname{E}[X] = \alpha/(\alpha+\beta) |
    | Lognormal LN(ΞΌ,Οƒ2)LN(\mu, \sigma^2) Mean | E⁑[X]=eΞΌ+Οƒ2/2\operatorname{E}[X] = e^{\mu + \sigma^2/2} |
    | Cauchy Distribution | Mean and Variance undefined |

    ---

    What's Next?

    πŸ’‘ Continue Learning

    This topic connects to:

      • Joint Distributions: Understanding how multiple continuous random variables interact.

      • Central Limit Theorem: The Normal distribution's role in the convergence of sums of random variables.

      • Statistical Inference: Using continuous distributions to construct confidence intervals and perform hypothesis tests.

      • Stochastic Processes: Applying exponential and gamma distributions in continuous-time models (e.g., queuing theory).

    ---

    Chapter Summary

    ❗ Elementary Distributions β€” Key Points

    Differentiate between discrete and continuous random variables, understanding their respective probability mass functions (PMF) and probability density functions (PDF), and cumulative distribution functions (CDF).
    Master the properties, parameters, and applications of common discrete distributions: Bernoulli, Binomial, Poisson, Geometric, and Hypergeometric.
    Understand the characteristics, parameters, and use cases of fundamental continuous distributions: Uniform, Exponential, and Normal (Gaussian).
    Calculate and interpret the expectation, variance, and higher moments for these elementary distributions.
    Recognize the conditions under which specific distributions (e.g., Poisson approximation to Binomial, Normal approximation to Binomial) can be applied.
    Grasp the significance of distribution parameters in determining shape, spread, and central tendency.

    ---

    Chapter Review Questions

    :::question type="MCQ" question="A call center receives calls at an average rate of 5 calls per hour. Assuming the number of calls follows a Poisson distribution, what is the probability of receiving exactly 3 calls in a 30-minute period?" options=["0.1404","0.2138","0.0821","0.2565"] answer="0.2138" hint="Adjust the Poisson rate parameter (Ξ»\lambda) to match the specified time period before applying the PMF." solution="For a 30-minute period, the average rate of calls is Ξ»=5Β calls/hourΓ—0.5Β hours=2.5\lambda = 5 \text{ calls/hour} \times 0.5 \text{ hours} = 2.5.
    The probability of receiving exactly kk calls in a Poisson distribution is given by P(X=k)=eβˆ’Ξ»Ξ»kk!P(X=k) = \frac{e^{-\lambda} \lambda^k}{k!}.
    For k=3k=3 and Ξ»=2.5\lambda=2.5:

    P(X=3)=eβˆ’2.5(2.5)33!=0.082085Γ—15.6256β‰ˆ0.2138P(X=3) = \frac{e^{-2.5} (2.5)^3}{3!} = \frac{0.082085 \times 15.625}{6} \approx 0.2138
    "
    :::

    :::question type="NAT" question="The lifetime of a certain electronic component (in hours) is exponentially distributed with a mean of 200 hours. If a component has already functioned for 150 hours, what is the probability that it will last for at least another 100 hours? (Provide answer to 4 decimal places.)" answer="0.6065" hint="Recall the memoryless property of the exponential distribution." solution="For an exponentially distributed random variable XX with mean ΞΌ\mu, the rate parameter is Ξ»=1/ΞΌ\lambda = 1/\mu. Here, ΞΌ=200\mu=200, so Ξ»=1/200\lambda = 1/200.
    The memoryless property states that P(X>t+s∣X>t)=P(X>s)P(X > t+s | X > t) = P(X > s).
    In this case, t=150t=150 hours (time already functioned) and s=100s=100 hours (additional time needed).
    So, P(X>150+100∣X>150)=P(X>100)P(X > 150+100 | X > 150) = P(X > 100).
    The CDF of an exponential distribution is F(x)=1βˆ’eβˆ’Ξ»xF(x) = 1 - e^{-\lambda x}, so P(X>x)=eβˆ’Ξ»xP(X > x) = e^{-\lambda x}.
    P(X>100)=eβˆ’(1/200)Γ—100=eβˆ’0.5β‰ˆ0.60653P(X > 100) = e^{-(1/200) \times 100} = e^{-0.5} \approx 0.60653.
    Rounding to 4 decimal places, the answer is 0.6065."
    :::

    :::question type="MCQ" question="Which of the following distributions describes the number of trials required to achieve the first success in a sequence of independent Bernoulli trials, where the probability of success remains constant for each trial?" options=["Binomial distribution","Poisson distribution","Geometric distribution","Hypergeometric distribution"] answer="Geometric distribution" hint="Consider the stopping condition for each distribution type." solution="The Geometric distribution models the number of independent Bernoulli trials needed to get the first success. The Binomial distribution models the number of successes in a fixed number of trials. The Poisson distribution models the number of events in a fixed interval of time or space. The Hypergeometric distribution models the number of successes in draws without replacement."
    :::

    :::question type="NAT" question="A discrete random variable XX has the following probability mass function: P(X=1)=0.3P(X=1) = 0.3, P(X=2)=0.5P(X=2) = 0.5, P(X=3)=0.2P(X=3) = 0.2. What is the variance of XX?" answer="0.49" hint="The variance can be calculated as Var⁑(X)=E[X2]βˆ’(E[X])2\operatorname{Var}(X) = E[X^2] - (E[X])^2." solution="First, calculate the expected value E[X]E[X]:
    E[X]=(1Γ—0.3)+(2Γ—0.5)+(3Γ—0.2)=0.3+1.0+0.6=1.9E[X] = (1 \times 0.3) + (2 \times 0.5) + (3 \times 0.2) = 0.3 + 1.0 + 0.6 = 1.9.
    Next, calculate E[X2]E[X^2]:
    E[X2]=(12Γ—0.3)+(22Γ—0.5)+(32Γ—0.2)=(1Γ—0.3)+(4Γ—0.5)+(9Γ—0.2)=0.3+2.0+1.8=4.1E[X^2] = (1^2 \times 0.3) + (2^2 \times 0.5) + (3^2 \times 0.2) = (1 \times 0.3) + (4 \times 0.5) + (9 \times 0.2) = 0.3 + 2.0 + 1.8 = 4.1.
    Finally, calculate the variance:
    Var⁑(X)=E[X2]βˆ’(E[X])2=4.1βˆ’(1.9)2=4.1βˆ’3.61=0.49\operatorname{Var}(X) = E[X^2] - (E[X])^2 = 4.1 - (1.9)^2 = 4.1 - 3.61 = 0.49."
    :::

    ---

    What's Next?

    πŸ’‘ Continue Your CMI Journey

    Building on the foundation of elementary distributions, subsequent chapters will delve into multivariate distributions, exploring the relationships between multiple random variables. Understanding these basic distributions is also crucial for comprehending advanced topics such as transformations of random variables and the fundamental limit theorems of probability theory.

    🎯 Key Points to Remember

    • βœ“ Master the core concepts in Elementary Distributions before moving to advanced topics
    • βœ“ Practice with previous year questions to understand exam patterns
    • βœ“ Review short notes regularly for quick revision before exams

    Related Topics in Probability Theory

    More Resources

    Why Choose MastersUp?

    🎯

    AI-Powered Plans

    Personalized study schedules based on your exam date and learning pace

    πŸ“š

    15,000+ Questions

    Verified questions with detailed solutions from past papers

    πŸ“Š

    Smart Analytics

    Track your progress with subject-wise performance insights

    πŸ”–

    Bookmark & Revise

    Save important questions for quick revision before exams

    Start Your Free Preparation β†’

    No credit card required β€’ Free forever for basic features