100% FREE Updated: Apr 2026 Probability Random variables

Discrete random variables

Comprehensive study notes on Discrete random variables for CMI BS Hons preparation. This chapter covers key concepts, formulas, and examples needed for your exam.

Discrete random variables

This chapter establishes the foundational concepts of discrete random variables, encompassing their probability distributions, expectation, and variance. A thorough understanding of these principles is critical for subsequent advanced topics in probability and statistics, and is consistently evaluated in CMI examinations through both theoretical and applied problems.

---

Chapter Contents

|

| Topic |

|---|-------| | 1 | Probability distribution | | 2 | Expectation | | 3 | Variance at school level | | 4 | Simple modelling problems |

---

We begin with Probability distribution.

Part 1: Probability distribution

Probability Distribution

Overview

A probability distribution describes how probability is assigned to the possible values of a random variable. In CMI-style questions, this topic is not just about definitions. It includes checking whether a formula really defines a distribution, building distributions from experiments like coin tosses and die throws, and computing exact probabilities from repeated independent trials. ---

Learning Objectives

❗ By the End of This Topic

After studying this topic, you will be able to:

  • Define and use the probability mass function of a discrete random variable.

  • Check whether a given table or formula is a valid probability distribution.

  • Construct distributions from simple experiments.

  • Use Bernoulli and binomial ideas in repeated independent trials.

  • Compute probabilities such as β€œexactly kk”, β€œat least one”, and β€œat most kk”.

---

Core Idea

πŸ“– Discrete Random Variable

A discrete random variable XX is a variable that takes only finitely many or countably many values.

Its probability distribution is given by the function

p(x)=P(X=x)\qquad p(x) = P(X=x)

called the probability mass function or pmf.

---

Valid Probability Distribution

πŸ“ PMF Conditions

For a discrete random variable XX with pmf p(x)p(x), the following must hold:

  • p(x)β‰₯0\qquad p(x) \ge 0 for every possible value of xx

  • βˆ‘xp(x)=1\qquad \sum_x p(x) = 1

❗ Interpretation

The first condition says probabilities cannot be negative.

The second condition says the total probability over all possible values must be exactly 11.

---

Distribution Table

A discrete distribution is often written in tabular form. | xx | x1x_1 | x2x_2 | x3x_3 | β‹―\cdots | |---|---|---|---|---| | P(X=x)P(X=x) | p1p_1 | p2p_2 | p3p_3 | β‹―\cdots | where p1+p2+p3+β‹―=1\qquad p_1+p_2+p_3+\cdots = 1 ---

Cumulative Distribution Function

πŸ“ CDF

The cumulative distribution function of XX is

F(x)=P(X≀x)\qquad F(x)=P(X\le x)

For a discrete random variable, the cdf is a step function. Useful facts:
  • F(x)F(x) is nondecreasing
  • 0≀F(x)≀10 \le F(x) \le 1
  • as xβ†’βˆ’βˆžx\to -\infty, F(x)β†’0F(x)\to 0
  • as xβ†’βˆžx\to \infty, F(x)β†’1F(x)\to 1
---

Expectation and Variance

πŸ“ Expectation

If XX has pmf p(x)p(x), then

E[X]=βˆ‘xx p(x)\qquad E[X] = \sum_x x\,p(x)

πŸ“ Variance

The variance of XX is

Var⁑(X)=E[X2]βˆ’(E[X])2\qquad \operatorname{Var}(X)=E[X^2]-(E[X])^2

where

E[X2]=βˆ‘xx2p(x)\qquad E[X^2]=\sum_x x^2 p(x)

Expectation gives the average value in the long run; variance measures spread. ---

Bernoulli Distribution

πŸ“– Bernoulli Random Variable

A Bernoulli random variable takes only the values 00 and 11.

If

P(X=1)=p,P(X=0)=1βˆ’p\qquad P(X=1)=p,\qquad P(X=0)=1-p

then XX has Bernoulli distribution with parameter pp.

This models a single success-failure experiment. Examples:
  • success = β€œhead occurs”
  • success = β€œdie shows an even number”
  • success = β€œoutcome divisible by 33”
---

Binomial Distribution

πŸ“– Binomial Distribution

If an experiment is repeated nn times independently, and each trial has success probability pp, then the number of successes XX has binomial distribution:

P(X=k)=(nk)pk(1βˆ’p)nβˆ’k,k=0,1,2,…,n\qquad P(X=k)=\binom{n}{k}p^k(1-p)^{n-k},\qquad k=0,1,2,\dots,n

This is one of the most important discrete distributions for exam problems. ---

Standard Probability Forms

πŸ“ Exactly, At Least One, At Most

If X∼Bin⁑(n,p)X\sim \operatorname{Bin}(n,p), then:

    • Exactly kk successes:


P(X=k)=(nk)pk(1βˆ’p)nβˆ’k\qquad P(X=k)=\binom{n}{k}p^k(1-p)^{n-k}

    • At least one success:


P(Xβ‰₯1)=1βˆ’P(X=0)=1βˆ’(1βˆ’p)n\qquad P(X\ge 1)=1-P(X=0)=1-(1-p)^n

    • At most one success:


P(X≀1)=P(X=0)+P(X=1)\qquad P(X\le 1)=P(X=0)+P(X=1)

The complement method is especially important for β€œone or more” questions. ---

PYQ-Style Example 1

One or more of the first three throws is 44 Each throw is independent, and P(notΒ 4)=56\qquad P(\text{not }4)=\dfrac{5}{6} So $\qquad P(\text{at least one }4\text{ in first three throws}) =1-\left(\dfrac{5}{6}\right)^3$ $\qquad =1-\dfrac{125}{216} =\dfrac{91}{216}$ ---

PYQ-Style Example 2

Exactly two of the last four throws are divisible by 33 A die outcome is divisible by 33 if it is 33 or 66, so p=26=13\qquad p = \dfrac{2}{6}=\dfrac{1}{3} Let XX be the number of such throws among the last four. Then X∼Bin⁑(4,13)\qquad X\sim \operatorname{Bin}\left(4,\dfrac{1}{3}\right) Hence P(X=2)=(42)(13)2(23)2\qquad P(X=2)=\binom{4}{2}\left(\dfrac{1}{3}\right)^2\left(\dfrac{2}{3}\right)^2 $\qquad =6\cdot \dfrac{1}{9}\cdot \dfrac{4}{9} =\dfrac{24}{81} =\dfrac{8}{27}$ ---

Constructing a Distribution from an Experiment

πŸ’‘ Standard Method

To build the distribution of a random variable:

  • identify all possible values of the variable

  • compute the probability of each value

  • check that all probabilities are nonnegative

  • check that their sum is 11

Example A fair coin is tossed twice. Let XX be the number of heads. Possible values: 0,1,2\qquad 0,1,2 Probabilities:
  • P(X=0)=14\qquad P(X=0)=\dfrac{1}{4}
  • P(X=1)=24=12\qquad P(X=1)=\dfrac{2}{4}=\dfrac{1}{2}
  • P(X=2)=14\qquad P(X=2)=\dfrac{1}{4}
So the distribution is | xx | 00 | 11 | 22 | |---|---|---|---| | P(X=x)P(X=x) | 14\dfrac{1}{4} | 12\dfrac{1}{2} | 14\dfrac{1}{4} | ---

Common Mistakes

⚠️ Avoid These Errors
    • ❌ Forgetting that total probability must add to 11
βœ… Always check the sum
    • ❌ Using binomial formula without independence
βœ… Binomial model needs independent identical trials
    • ❌ Mixing up β€œat least one” with β€œexactly one”
βœ… Use complement for β€œat least one”
    • ❌ Ignoring the support of the random variable
βœ… First list all possible values clearly
    • ❌ Treating expectation as always one of the possible values
βœ… Expectation is an average, not necessarily an attained value
---

CMI Strategy

πŸ’‘ How to Attack Probability Distribution Questions

  • First decide whether the question is about a pmf, cdf, or event probability.

  • If a formula is given, check nonnegativity and total sum 11.

  • For repeated trials, test whether the setup is Bernoulli/binomial.

  • For β€œone or more”, use complement before expanding anything.

  • For β€œexactly kk”, use the binomial coefficient carefully.

  • Write the final answer in exact form whenever possible.

---

Practice Questions

:::question type="MCQ" question="Which of the following can be a valid probability mass function on the set {0,1,2}\{0,1,2\}?" options=["P(X=0)=0.2,Β P(X=1)=0.3,Β P(X=2)=0.4P(X=0)=0.2,\ P(X=1)=0.3,\ P(X=2)=0.4","P(X=0)=0.1,Β P(X=1)=0.4,Β P(X=2)=0.5P(X=0)=0.1,\ P(X=1)=0.4,\ P(X=2)=0.5","P(X=0)=βˆ’0.1,Β P(X=1)=0.6,Β P(X=2)=0.5P(X=0)=-0.1,\ P(X=1)=0.6,\ P(X=2)=0.5","P(X=0)=12,Β P(X=1)=12,Β P(X=2)=12P(X=0)=\dfrac{1}{2},\ P(X=1)=\dfrac{1}{2},\ P(X=2)=\dfrac{1}{2}"] answer="B" hint="Check nonnegativity and whether the probabilities add to 11." solution="A valid pmf must have all probabilities nonnegative and total sum 11. Option A sums to 0.90.9, so it is invalid. Option B has nonnegative probabilities and 0.1+0.4+0.5=1\qquad 0.1+0.4+0.5=1 so it is valid. Option C has a negative probability. Option D sums to 32\dfrac{3}{2}, so it is invalid. Hence the correct option is B\boxed{B}." ::: :::question type="NAT" question="A fair die is thrown 33 times. Find the probability that 44 appears in one or more of these throws." answer="\\dfrac{91}{216}" hint="Use the complement event 'no throw shows 44'." solution="The probability that a single throw does not show 44 is 56\qquad \dfrac{5}{6} So the probability that none of the three throws shows 44 is (56)3=125216\qquad \left(\dfrac{5}{6}\right)^3=\dfrac{125}{216} Therefore the probability that 44 appears in one or more throws is 1βˆ’125216=91216\qquad 1-\dfrac{125}{216}=\dfrac{91}{216} Hence the answer is 91216\boxed{\dfrac{91}{216}}." ::: :::question type="MSQ" question="Which of the following statements are true for a discrete random variable XX with pmf p(x)p(x)?" options=["p(x)β‰₯0p(x)\ge 0 for every possible value of xx","βˆ‘xp(x)=1\sum_x p(x)=1","F(x)=P(X≀x)F(x)=P(X\le x) is nondecreasing","E[X]E[X] must always be an integer"] answer="A,B,C" hint="Recall the basic definitions of pmf, cdf, and expectation." solution="1. True.
  • True.
  • True.
  • False. The expectation need not be an integer.
  • Hence the correct answer is A,B,C\boxed{A,B,C}." ::: :::question type="SUB" question="A fair die is thrown twice. Let XX be the number of throws whose outcome is divisible by 33. Find the probability distribution of XX." answer="P(X=0)=49,Β P(X=1)=49,Β P(X=2)=19P(X=0)=\dfrac{4}{9},\ P(X=1)=\dfrac{4}{9},\ P(X=2)=\dfrac{1}{9}" hint="An outcome is divisible by 33 if it is 33 or 66." solution="A die outcome is divisible by 33 if it is 33 or 66, so the success probability on one throw is p=26=13\qquad p=\dfrac{2}{6}=\dfrac{1}{3} Since there are two independent throws, the random variable XX, the number of divisible-by-33 outcomes, has binomial distribution: X∼Bin⁑(2,13)\qquad X\sim \operatorname{Bin}\left(2,\dfrac{1}{3}\right) Therefore: P(X=0)=(23)2=49\qquad P(X=0)=\left(\dfrac{2}{3}\right)^2=\dfrac{4}{9} P(X=1)=(21)(13)(23)=49\qquad P(X=1)=\binom{2}{1}\left(\dfrac{1}{3}\right)\left(\dfrac{2}{3}\right)=\dfrac{4}{9} P(X=2)=(13)2=19\qquad P(X=2)=\left(\dfrac{1}{3}\right)^2=\dfrac{1}{9} So the required distribution is P(X=0)=49,P(X=1)=49,P(X=2)=19\qquad P(X=0)=\dfrac{4}{9},\quad P(X=1)=\dfrac{4}{9},\quad P(X=2)=\dfrac{1}{9}." ::: ---

    Summary

    ❗ Key Takeaways for CMI

    • A valid discrete distribution must have nonnegative probabilities summing to 11.

    • The pmf is p(x)=P(X=x)p(x)=P(X=x), while the cdf is F(x)=P(X≀x)F(x)=P(X\le x).

    • Bernoulli distribution models one success-failure trial.

    • Binomial distribution models the number of successes in repeated independent identical trials.

    • β€œAt least one” is often easiest by complement.

    • Exact algebraic probability expressions are preferred over decimal approximations.

    ---

    πŸ’‘ Next Up

    Proceeding to Expectation.

    ---

    Part 2: Expectation

    Expectation

    Overview

    Expectation is the average or mean value of a random variable in the long run. In school-level and olympiad-style probability, expectation is one of the most powerful summary quantities because it often turns a complicated random process into a simple weighted sum. In exam problems, the main difficulty is modeling the random variable correctly before applying the formula. ---

    Learning Objectives

    ❗ By the End of This Topic

    After studying this topic, you will be able to:

    • Compute expectation for a discrete random variable.

    • Build a random variable from a probability problem.

    • Use linearity of expectation.

    • Interpret expectation even when the value itself need not be achievable.

    • Solve basic modelling questions involving expected counts and gains.

    ---

    Core Idea

    πŸ“– Expectation of a Discrete Random Variable

    If a discrete random variable XX takes values

    x1,x2,…,xn\qquad x_1,x_2,\dots,x_n

    with probabilities

    p1,p2,…,pn\qquad p_1,p_2,\dots,p_n,

    then the expectation of XX is

    E(X)=x1p1+x2p2+β‹―+xnpn\qquad E(X)=x_1p_1+x_2p_2+\cdots+x_np_n

    ❗ Interpretation

    Expectation is the weighted average of possible values of the random variable.

    ---

    Probability Distribution Conditions

    πŸ“ What a Valid Distribution Must Satisfy

    If XX takes values xix_i with probabilities pip_i, then:

    • piβ‰₯0\qquad p_i \ge 0

    • βˆ‘pi=1\qquad \sum p_i = 1

    Only after checking this should expectation be computed. ::: ---

    Linearity of Expectation

    πŸ“ Linearity

    For random variables XX and YY,

    E(X+Y)=E(X)+E(Y)\qquad E(X+Y)=E(X)+E(Y)

    and for a constant cc,

    E(cX)=cE(X)\qquad E(cX)=cE(X)

    This remains true whether or not XX and YY are independent. :::
    ❗ Why This Matters

    Linearity allows us to compute expectations of complicated quantities by breaking them into simpler parts.

    ---

    Expectation of Common Variables

    πŸ“ Useful Standard Values

    • If XX is Bernoulli with success probability pp, then


    E(X)=p\qquad E(X)=p

    • If a fair die is rolled and XX is the face value, then


    E(X)=1+2+3+4+5+66=72\qquad E(X)=\frac{1+2+3+4+5+6}{6}=\frac{7}{2}

    ---

    Minimal Worked Examples

    Example 1 A fair die is rolled. Let XX be the outcome. Then E(X)=1+2+3+4+5+66=216=72\qquad E(X)=\frac{1+2+3+4+5+6}{6}=\frac{21}{6}=\frac{7}{2} --- Example 2 A random variable XX takes values 0,1,20,1,2 with probabilities 14,Β 12,Β 14\qquad \frac{1}{4},\ \frac{1}{2},\ \frac{1}{4} Then E(X)=0β‹…14+1β‹…12+2β‹…14=0+12+12=1\qquad E(X)=0\cdot \frac{1}{4}+1\cdot \frac{1}{2}+2\cdot \frac{1}{4}=0+\frac{1}{2}+\frac{1}{2}=1 So E(X)=1\qquad E(X)=1 ::: ---

    Modelling Expectation

    πŸ’‘ How to Build the Random Variable

    In many questions, the main step is deciding what XX represents.

    Examples:

      • number of heads in repeated tosses

      • score obtained in a game

      • number of defective items chosen

      • profit or loss in a scheme

    Once XX is defined clearly, expectation usually becomes routine. ---

    Expected Value Need Not Be a Possible Outcome

    ⚠️ Important Concept

    Expectation need not be one of the actual values taken by the random variable.

    For example, in one roll of a fair die:

    E(X)=72\qquad E(X)=\frac{7}{2}

    but the die never actually shows 72\frac{7}{2}.

    ---

    Indicator Variable Idea

    πŸ“ Expected Count

    If II is an indicator variable that is 11 when an event occurs and 00 otherwise, then

    E(I)=P(I=1)\qquad E(I)=P(I=1)

    This is often used to count expected numbers of successes.

    ---

    CMI Strategy

    πŸ’‘ How to Attack Expectation Questions

    • Define the random variable clearly.

    • List its possible values and probabilities.

    • Check that probabilities sum to 11.

    • Compute the weighted average.

    • Use linearity when several pieces are involved.

    ---

    Common Mistakes

    ⚠️ Avoid These Errors
      • ❌ Averaging values without using probabilities
      • ❌ Forgetting that probabilities must sum to 11
      • ❌ Confusing expectation with most likely value
      • ❌ Assuming expectation must be one of the actual outcomes
    ---

    Practice Questions

    :::question type="MCQ" question="If a fair die is rolled once and XX is the outcome, then E(X)E(X) is" options=["33","72\dfrac{7}{2}","44","212\dfrac{21}{2}"] answer="B" hint="Use the average of all six equally likely values." solution="We have E(X)=1+2+3+4+5+66=216=72\qquad E(X)=\frac{1+2+3+4+5+6}{6}=\frac{21}{6}=\frac{7}{2} Hence the correct option is B\boxed{B}." ::: :::question type="NAT" question="A random variable XX takes values 1,2,31,2,3 with probabilities 12,13,16\dfrac{1}{2},\dfrac{1}{3},\dfrac{1}{6} respectively. Find E(X)E(X)." answer="5/3" hint="Compute the weighted average." solution="We compute E(X)=1β‹…12+2β‹…13+3β‹…16\qquad E(X)=1\cdot \frac{1}{2}+2\cdot \frac{1}{3}+3\cdot \frac{1}{6} =12+23+12=1+23=53\qquad = \frac{1}{2}+\frac{2}{3}+\frac{1}{2}=1+\frac{2}{3}=\frac{5}{3} So the answer is 53\boxed{\frac{5}{3}}." ::: :::question type="MSQ" question="Which of the following are true?" options=["Expectation is a weighted average","Expectation of a die roll need not be an integer","Probabilities in a distribution must sum to 11","Expectation is always the most likely outcome"] answer="A,B,C" hint="Check definition and interpretation carefully." solution="1. True. 2. True, for example a fair die has expectation 72\frac{7}{2}. 3. True. 4. False. Expectation need not be the most likely value. Hence the correct answer is A,B,C\boxed{A,B,C}." ::: :::question type="SUB" question="A fair coin is tossed three times. Let XX be the number of heads obtained. Find E(X)E(X)." answer="32\dfrac{3}{2}" hint="List the binomial probabilities or use linearity of expectation." solution="Let XX be the number of heads in three tosses. Using linearity of expectation, write X=I1+I2+I3\qquad X=I_1+I_2+I_3 where Ij=1I_j=1 if the jjth toss is a head and 00 otherwise. For each toss, E(Ij)=P(head)=12\qquad E(I_j)=P(\text{head})=\frac{1}{2} So E(X)=E(I1)+E(I2)+E(I3)=12+12+12=32\qquad E(X)=E(I_1)+E(I_2)+E(I_3)=\frac{1}{2}+\frac{1}{2}+\frac{1}{2}=\frac{3}{2} Hence the expected number of heads is 32\boxed{\frac{3}{2}}." ::: ---

    Summary

    ❗ Key Takeaways for CMI

    • Expectation is the weighted average of a random variable.

    • A correct model comes before the formula.

    • Linearity of expectation is one of the most useful tools in probability.

    • Expectation need not be a possible outcome.

    • Expected counts are often computed using indicator variables.

    ---

    πŸ’‘ Next Up

    Proceeding to Variance at school level.

    ---

    Part 3: Variance at school level

    Variance at School Level

    Overview

    Variance measures how spread out a random variable is around its mean. At school level, the important goals are to compute variance correctly, understand its meaning, and use the shortcut formula Var⁑(X)=E(X2)βˆ’[E(X)]2\qquad \operatorname{Var}(X)=E(X^2)-[E(X)]^2 In exam problems, the main trap is arithmetic: students often compute the mean correctly but forget to square it or mishandle the second moment. ---

    Learning Objectives

    ❗ By the End of This Topic

    After studying this topic, you will be able to:

    • Define variance of a discrete random variable.

    • Compute variance directly or using the shortcut formula.

    • Find standard deviation from variance.

    • Interpret variance as spread.

    • Avoid common errors in mean-and-square calculations.

    ---

    Core Idea

    πŸ“– Variance

    If XX is a random variable with mean

    ΞΌ=E(X)\qquad \mu = E(X),

    then the variance of XX is

    Var⁑(X)=E((Xβˆ’ΞΌ)2)\qquad \operatorname{Var}(X)=E\big((X-\mu)^2\big)

    This measures average squared deviation from the mean. ::: ---

    Shortcut Formula

    πŸ“ Most Useful Formula

    A very important identity is

    Var⁑(X)=E(X2)βˆ’[E(X)]2\qquad \operatorname{Var}(X)=E(X^2)-[E(X)]^2

    where E(X2)=βˆ‘xi2pi\qquad E(X^2)=\sum x_i^2 p_i if XX takes values xix_i with probabilities pip_i. ::: ---

    Standard Deviation

    πŸ“ Standard Deviation

    The standard deviation of XX is

    Οƒ=Var⁑(X)\qquad \sigma = \sqrt{\operatorname{Var}(X)}

    Variance has squared units, while standard deviation has the same unit as XX. ::: ---

    Basic Properties

    πŸ“ High-Value Properties

    • Var⁑(X)β‰₯0\qquad \operatorname{Var}(X)\ge 0

    • Var⁑(c)=0\qquad \operatorname{Var}(c)=0 for a constant random variable

    • Var⁑(aX)=a2Var⁑(X)\qquad \operatorname{Var}(aX)=a^2\operatorname{Var}(X)

    • Small variance means values are tightly clustered near the mean

    ---

    Minimal Worked Examples

    Example 1 Suppose XX takes values 00 and 11 with probabilities 12\frac{1}{2} each. Then E(X)=0β‹…12+1β‹…12=12\qquad E(X)=0\cdot \frac{1}{2}+1\cdot \frac{1}{2}=\frac{1}{2} Also, E(X2)=02β‹…12+12β‹…12=12\qquad E(X^2)=0^2\cdot \frac{1}{2}+1^2\cdot \frac{1}{2}=\frac{1}{2} So Var⁑(X)=12βˆ’(12)2=12βˆ’14=14\qquad \operatorname{Var}(X)=\frac{1}{2}-\left(\frac{1}{2}\right)^2=\frac{1}{2}-\frac{1}{4}=\frac{1}{4} --- Example 2 A fair die is rolled and XX is the outcome. We know E(X)=72\qquad E(X)=\frac{7}{2} Now E(X2)=12+22+32+42+52+626=916\qquad E(X^2)=\frac{1^2+2^2+3^2+4^2+5^2+6^2}{6}=\frac{91}{6} Hence Var⁑(X)=916βˆ’(72)2=916βˆ’494=3512\qquad \operatorname{Var}(X)=\frac{91}{6}-\left(\frac{7}{2}\right)^2=\frac{91}{6}-\frac{49}{4}=\frac{35}{12} ---

    Interpretation

    ❗ What Variance Tells You

    Variance does not tell you where the centre is; expectation already does that.

    Variance tells you how much the values typically fluctuate around the mean.

    A random variable with all mass at one point has zero variance. ---

    Direct vs Shortcut Computation

    πŸ’‘ Which Formula to Use?

    You may compute variance directly from

    E((Xβˆ’ΞΌ)2)\qquad E((X-\mu)^2)

    or by the shortcut

    E(X2)βˆ’[E(X)]2\qquad E(X^2)-[E(X)]^2

    The shortcut is usually faster in exam problems.

    ---

    Common Mistakes

    ⚠️ Avoid These Errors
      • ❌ Using Var⁑(X)=E(X2)βˆ’E(X)\operatorname{Var}(X)=E(X^2)-E(X)
      • ❌ Forgetting to square the mean
      • ❌ Computing E(X2)E(X^2) as [E(X)]2[E(X)]^2
      • ❌ Forgetting that variance cannot be negative
    ---

    CMI Strategy

    πŸ’‘ How to Attack Variance Questions

    • Find E(X)E(X) carefully.

    • Find E(X2)E(X^2) separately.

    • Use

    Var⁑(X)=E(X2)βˆ’[E(X)]2\qquad \operatorname{Var}(X)=E(X^2)-[E(X)]^2
    • Simplify only after both pieces are correct.

    • If asked, take square root at the end for standard deviation.

    ---

    Practice Questions

    :::question type="MCQ" question="Which of the following is equal to Var⁑(X)\operatorname{Var}(X)?" options=["E(X2)βˆ’E(X)E(X^2)-E(X)","E(X2)βˆ’[E(X)]2E(X^2)-[E(X)]^2","[E(X)]2βˆ’E(X2)[E(X)]^2-E(X^2)","E(X)E(X)"] answer="B" hint="Recall the shortcut formula." solution="The standard identity is Var⁑(X)=E(X2)βˆ’[E(X)]2\qquad \operatorname{Var}(X)=E(X^2)-[E(X)]^2 Hence the correct option is B\boxed{B}." ::: :::question type="NAT" question="A random variable XX takes values 00 and 22 with probabilities 12\dfrac{1}{2} each. Find Var⁑(X)\operatorname{Var}(X)." answer="1" hint="Compute E(X)E(X) and E(X2)E(X^2)." solution="We have E(X)=0β‹…12+2β‹…12=1\qquad E(X)=0\cdot \frac{1}{2}+2\cdot \frac{1}{2}=1 Also, E(X2)=02β‹…12+22β‹…12=2\qquad E(X^2)=0^2\cdot \frac{1}{2}+2^2\cdot \frac{1}{2}=2 Hence Var⁑(X)=E(X2)βˆ’[E(X)]2=2βˆ’12=1\qquad \operatorname{Var}(X)=E(X^2)-[E(X)]^2=2-1^2=1 So the answer is 1\boxed{1}." ::: :::question type="MSQ" question="Which of the following are always true?" options=["Variance is never negative","A constant random variable has variance 00","Standard deviation is the square root of variance","Variance always equals expectation"] answer="A,B,C" hint="Recall the definitions." solution="1. True. 2. True. 3. True. 4. False. Variance and expectation measure different things. Hence the correct answer is A,B,C\boxed{A,B,C}." ::: :::question type="SUB" question="A fair coin is tossed twice. Let XX be the number of heads obtained. Find E(X)E(X) and Var⁑(X)\operatorname{Var}(X)." answer="E(X)=1,Var⁑(X)=12E(X)=1,\\ \operatorname{Var}(X)=\dfrac{1}{2}" hint="Use the distribution of XX." solution="For two fair tosses, the random variable XX takes values 0,1,20,1,2 with probabilities 14,Β 12,Β 14\qquad \frac{1}{4},\ \frac{1}{2},\ \frac{1}{4} So E(X)=0β‹…14+1β‹…12+2β‹…14=1\qquad E(X)=0\cdot \frac{1}{4}+1\cdot \frac{1}{2}+2\cdot \frac{1}{4}=1 Also, E(X2)=02β‹…14+12β‹…12+22β‹…14=0+12+1=32\qquad E(X^2)=0^2\cdot \frac{1}{4}+1^2\cdot \frac{1}{2}+2^2\cdot \frac{1}{4}=0+\frac{1}{2}+1=\frac{3}{2} Hence Var⁑(X)=E(X2)βˆ’[E(X)]2=32βˆ’1=12\qquad \operatorname{Var}(X)=E(X^2)-[E(X)]^2=\frac{3}{2}-1=\frac{1}{2} Therefore E(X)=1,Β Var⁑(X)=12\qquad \boxed{E(X)=1,\ \operatorname{Var}(X)=\frac{1}{2}}." ::: ---

    Summary

    ❗ Key Takeaways for CMI

    • Variance measures spread around the mean.

    • The shortcut formula is

    Var⁑(X)=E(X2)βˆ’[E(X)]2\qquad \operatorname{Var}(X)=E(X^2)-[E(X)]^2
    • Variance is always nonnegative.

    • Standard deviation is the square root of variance.

    • Careful computation of E(X)E(X) and E(X2)E(X^2) is the core skill.

    ---

    πŸ’‘ Next Up

    Proceeding to Simple modelling problems.

    ---

    Part 4: Simple modelling problems

    Simple Modelling Problems

    Overview

    Simple modelling problems in probability ask you to convert a real situation into a random variable, event structure, or probability distribution. The mathematics is usually not hard once the model is correct β€” the real challenge is identifying what is random, what the possible outcomes are, and what assumptions are being made. ---

    Learning Objectives

    ❗ By the End of This Topic

    After studying this topic, you will be able to:

    • Translate a verbal situation into a probability model.

    • Identify outcomes, events, and random variables clearly.

    • Compute probabilities and expectations in basic applied settings.

    • Distinguish between model assumptions and conclusions.

    • Check whether a model is realistic and internally consistent.

    ---

    Core Idea

    πŸ“– Probability Model

    A probability model consists of:

    • a sample space,

    • a rule assigning probabilities,

    • the event or random variable of interest.

    In simple modelling problems, the main step is setting this up correctly. ::: ---

    Standard Questions in Modelling

    πŸ“ What to Ask First

    When reading a modelling problem, ask:

    • What are the possible outcomes?

    • Are the outcomes equally likely?

    • What is the random variable?

    • What probability or expectation is required?

    • Are there hidden assumptions such as independence?

    ---

    Common Modelling Situations

    πŸ“ Frequent School-Level Models

    • Tosses of coins

    • Rolls of dice

    • Drawing cards or balls

    • Defective / non-defective items

    • Success-failure trials

    • Gain-loss games

    ---

    Choosing the Right Random Variable

    πŸ’‘ Model the Quantity Asked

    If the question asks for:

      • number of successes, define that count

      • total score, define score sum

      • gain or loss, define profit variable

      • waiting time, define number of trials

    A good model often turns words into a simple table of values and probabilities. ::: ---

    Minimal Worked Examples

    Example 1 A fair coin is tossed twice. Let XX be the number of heads. The sample space is {HH,HT,TH,TT}\qquad \{HH,HT,TH,TT\} So XX takes values:
    • 00 for TTTT
    • 11 for HT,THHT,TH
    • 22 for HHHH
    Hence the distribution is: P(X=0)=14,P(X=1)=12,P(X=2)=14\qquad P(X=0)=\frac{1}{4},\quad P(X=1)=\frac{1}{2},\quad P(X=2)=\frac{1}{4} --- Example 2 A game pays:
    • 22 points for a head
    • 00 points for a tail
    A fair coin is tossed once. If XX is the score, then P(X=2)=12,P(X=0)=12\qquad P(X=2)=\frac{1}{2},\qquad P(X=0)=\frac{1}{2} So E(X)=2β‹…12+0β‹…12=1\qquad E(X)=2\cdot \frac{1}{2}+0\cdot \frac{1}{2}=1 ::: ---

    Independence in Modelling

    ❗ Use Independence Only When Justified

    If repeated trials are described as fair and separate, independence is usually intended.

    Examples:

      • repeated coin tosses

      • repeated fair die rolls


    But in sampling without replacement, outcomes are not independent.

    ---

    Common Modelling Errors

    ⚠️ Avoid These Errors
      • ❌ Assuming equally likely outcomes when they are not
      • ❌ Defining the wrong random variable
      • ❌ Forgetting restrictions such as β€œwithout replacement”
      • ❌ Mixing up event probability with expected value
    ---

    CMI Strategy

    πŸ’‘ How to Attack Modelling Questions

    • Write the experiment first.

    • Write the sample space or value table.

    • Define the random variable precisely.

    • Compute probabilities before jumping to expectation or variance.

    • Check whether independence or replacement is involved.

    ---

    Practice Questions

    :::question type="MCQ" question="A fair coin is tossed once. Let XX be the score, where X=1X=1 for Head and X=0X=0 for Tail. Then E(X)E(X) is" options=["00","12\dfrac{1}{2}","11","22"] answer="B" hint="This is a Bernoulli model with success probability 12\frac{1}{2}." solution="The score is 11 with probability 12\frac{1}{2} and 00 with probability 12\frac{1}{2}. Therefore E(X)=1β‹…12+0β‹…12=12\qquad E(X)=1\cdot \frac{1}{2}+0\cdot \frac{1}{2}=\frac{1}{2} So the correct option is B\boxed{B}." ::: :::question type="NAT" question="A fair die is rolled once. Let XX be the indicator of the event 'the outcome is even'. Find E(X)E(X)." answer="1/2" hint="For an indicator variable, expectation equals the probability of the event." solution="The event 'even' has probability 36=12\qquad \frac{3}{6}=\frac{1}{2} Since XX is the indicator of that event, E(X)=P(X=1)=12\qquad E(X)=P(X=1)=\frac{1}{2} Hence the answer is 12\boxed{\frac{1}{2}}." ::: :::question type="MSQ" question="Which of the following are important first steps in a modelling problem?" options=["Identify the random experiment","Define the random variable clearly","Decide whether outcomes are equally likely","Assume all events are independent without checking"] answer="A,B,C" hint="Think about model-building, not guesswork." solution="1. True. 2. True. 3. True. 4. False. Independence must be justified, not assumed automatically. Hence the correct answer is A,B,C\boxed{A,B,C}." ::: :::question type="SUB" question="A fair coin is tossed three times. Let XX be the number of tails obtained. Construct the probability distribution of XX." answer="P(X=0)=18,P(X=1)=38,P(X=2)=38,P(X=3)=18P(X=0)=\frac18,\\ P(X=1)=\frac38,\\ P(X=2)=\frac38,\\ P(X=3)=\frac18" hint="Count how many sequences have exactly 0,1,2,30,1,2,3 tails." solution="For three fair tosses, there are 23=82^3=8 equally likely outcomes. Let XX be the number of tails.
    • X=0X=0 only for HHHHHH, so
    P(X=0)=18\qquad P(X=0)=\frac{1}{8}
    • X=1X=1 for THH,HTH,HHTTHH,HTH,HHT, so
    P(X=1)=38\qquad P(X=1)=\frac{3}{8}
    • X=2X=2 for TTH,THT,HTTTTH,THT,HTT, so
    P(X=2)=38\qquad P(X=2)=\frac{3}{8}
    • X=3X=3 only for TTTTTT, so
    P(X=3)=18\qquad P(X=3)=\frac{1}{8} Hence the probability distribution is P(X=0)=18,Β P(X=1)=38,Β P(X=2)=38,Β P(X=3)=18\qquad \boxed{P(X=0)=\frac18,\ P(X=1)=\frac38,\ P(X=2)=\frac38,\ P(X=3)=\frac18}." ::: ---

    Summary

    ❗ Key Takeaways for CMI

    • A good probability model starts with the experiment and sample space.

    • The random variable must match the quantity asked.

    • Expectation and variance come only after the model is set correctly.

    • Independence and equal likelihood must not be assumed blindly.

    • Many modelling problems are easy once the setup is clean.

    ---

    Chapter Summary

    ❗ Discrete random variables β€” Key Points

    • A Discrete Random Variable (DRV) XX takes on a finite or countably infinite number of distinct values. Its behaviour is described by a Probability Mass Function (PMF) P⁑(X=x)\operatorname{P}(X=x), which satisfies P⁑(X=x)β‰₯0\operatorname{P}(X=x) \ge 0 for all xx and βˆ‘xP⁑(X=x)=1\sum_x \operatorname{P}(X=x) = 1.

    • The Expectation (or mean) of a DRV XX, denoted E⁑[X]\operatorname{E}[X], is the weighted average of its possible values: E⁑[X]=βˆ‘xxP⁑(X=x)\operatorname{E}[X] = \sum_x x \operatorname{P}(X=x).

    • Expectation is a linear operator: for constants a,ba, b, E⁑[aX+b]=aE⁑[X]+b\operatorname{E}[aX+b] = a\operatorname{E}[X]+b.

    • The Variance of a DRV XX, denoted Var⁑[X]\operatorname{Var}[X], quantifies the spread of its distribution around its mean: Var⁑[X]=E⁑[(Xβˆ’E⁑[X])2]=E⁑[X2]βˆ’(E⁑[X])2\operatorname{Var}[X] = \operatorname{E}[(X-\operatorname{E}[X])^2] = \operatorname{E}[X^2] - (\operatorname{E}[X])^2.

    • For constants a,ba, b, the variance property is Var⁑[aX+b]=a2Var⁑[X]\operatorname{Var}[aX+b] = a^2\operatorname{Var}[X]. The standard deviation is ΟƒX=Var⁑[X]\sigma_X = \sqrt{\operatorname{Var}[X]}.
    • Simple modelling problems involve defining a DRV based on a real-world scenario, constructing its PMF, and then using the PMF to calculate probabilities, expectation, and variance, interpreting these values in context.

    • A thorough understanding of these concepts is fundamental for analysing discrete data and forms the bedrock for more advanced topics in probability and statistics.

    ---

    Chapter Review Questions

    :::question type="MCQ" question="A discrete random variable XX has the following probability mass function:
    P⁑(X=1)=c\operatorname{P}(X=1) = c
    P⁑(X=2)=2c\operatorname{P}(X=2) = 2c
    P⁑(X=3)=3c\operatorname{P}(X=3) = 3c
    P⁑(X=4)=4c\operatorname{P}(X=4) = 4c
    What is the probability that XX is an even number?" options=["1/101/10","3/103/10","2/52/5","3/53/5"] answer="3/5" hint="First, determine the value of cc using the property that the sum of all probabilities must equal 1. Then, identify the outcomes for which XX is even and sum their probabilities." solution="The sum of all probabilities must equal 1:

    c+2c+3c+4c=1c + 2c + 3c + 4c = 1

    10c=1β€…β€ŠβŸΉβ€…β€Šc=11010c = 1 \implies c = \frac{1}{10}

    The probability that XX is an even number is P⁑(X=2)+P⁑(X=4)\operatorname{P}(X=2) + \operatorname{P}(X=4):
    P⁑(XΒ isΒ even)=2c+4c=6c=6Γ—110=610=35\operatorname{P}(X \text{ is even}) = 2c + 4c = 6c = 6 \times \frac{1}{10} = \frac{6}{10} = \frac{3}{5}
    "
    :::

    :::question type="NAT" question="A game involves rolling a fair four-sided die (numbered 1, 2, 3, 4). You win points equal to the number rolled, unless you roll a 4, in which case you lose 5 points. What is the expected number of points you will win?" answer="0.25" hint="Define a random variable for the points won. List all possible outcomes and their corresponding probabilities and points. Calculate the expectation using the formula E⁑[X]=βˆ‘xP⁑(X=x)\operatorname{E}[X] = \sum x \operatorname{P}(X=x)." solution="Let XX be the random variable representing the points won.
    The possible outcomes for the die roll are 1, 2, 3, 4, each with a probability of 1/41/4.
    The corresponding points are:
    If die roll is 1, points X=1X=1.
    If die roll is 2, points X=2X=2.
    If die roll is 3, points X=3X=3.
    If die roll is 4, points X=βˆ’5X=-5.

    The expectation E⁑[X]\operatorname{E}[X] is:

    E⁑[X]=(1Γ—14)+(2Γ—14)+(3Γ—14)+(βˆ’5Γ—14)\operatorname{E}[X] = (1 \times \frac{1}{4}) + (2 \times \frac{1}{4}) + (3 \times \frac{1}{4}) + (-5 \times \frac{1}{4})

    E⁑[X]=1+2+3βˆ’54=14=0.25\operatorname{E}[X] = \frac{1+2+3-5}{4} = \frac{1}{4} = 0.25
    "
    :::

    :::question type="MCQ" question="A random variable XX has E⁑[X]=3\operatorname{E}[X]=3 and Var⁑[X]=4\operatorname{Var}[X]=4. What is Var⁑[2Xβˆ’1]\operatorname{Var}[2X-1]?" options=["77","1111","1515","1616"] answer="16" hint="Recall the properties of variance for linear transformations: Var⁑[aX+b]=a2Var⁑[X]\operatorname{Var}[aX+b] = a^2\operatorname{Var}[X]." solution="Using the property of variance, Var⁑[aX+b]=a2Var⁑[X]\operatorname{Var}[aX+b] = a^2\operatorname{Var}[X].
    In this case, a=2a=2 and b=βˆ’1b=-1.

    Var⁑[2Xβˆ’1]=22Var⁑[X]\operatorname{Var}[2X-1] = 2^2 \operatorname{Var}[X]

    Var⁑[2Xβˆ’1]=4Γ—4=16\operatorname{Var}[2X-1] = 4 \times 4 = 16
    "
    :::

    :::question type="NAT" question="A box contains 3 red balls and 2 blue balls. Two balls are drawn randomly without replacement. Let YY be the number of red balls drawn. Calculate E⁑[Y]\operatorname{E}[Y]." answer="1.2" hint="First, determine the possible values for YY and their probabilities (PMF). Then, use the expectation formula." solution="Let YY be the number of red balls drawn. The possible values for YY are 0, 1, or 2.
    Total number of balls = 5. Number of ways to draw 2 balls from 5 is (52)=5Γ—42=10\binom{5}{2} = \frac{5 \times 4}{2} = 10.

    P⁑(Y=0)\operatorname{P}(Y=0): No red balls, meaning 2 blue balls are drawn.

    P⁑(Y=0)=(22)(52)=110\operatorname{P}(Y=0) = \frac{\binom{2}{2}}{\binom{5}{2}} = \frac{1}{10}

    P⁑(Y=1)\operatorname{P}(Y=1): One red ball and one blue ball are drawn.
    P⁑(Y=1)=(31)(21)(52)=3Γ—210=610\operatorname{P}(Y=1) = \frac{\binom{3}{1}\binom{2}{1}}{\binom{5}{2}} = \frac{3 \times 2}{10} = \frac{6}{10}

    P⁑(Y=2)\operatorname{P}(Y=2): Two red balls are drawn.
    P⁑(Y=2)=(32)(52)=310\operatorname{P}(Y=2) = \frac{\binom{3}{2}}{\binom{5}{2}} = \frac{3}{10}

    Check: 1/10+6/10+3/10=11/10 + 6/10 + 3/10 = 1. The PMF is correct.

    Now, calculate the expectation E⁑[Y]\operatorname{E}[Y]:

    E⁑[Y]=(0Γ—P⁑(Y=0))+(1Γ—P⁑(Y=1))+(2Γ—P⁑(Y=2))\operatorname{E}[Y] = (0 \times \operatorname{P}(Y=0)) + (1 \times \operatorname{P}(Y=1)) + (2 \times \operatorname{P}(Y=2))

    E⁑[Y]=(0Γ—110)+(1Γ—610)+(2Γ—310)\operatorname{E}[Y] = (0 \times \frac{1}{10}) + (1 \times \frac{6}{10}) + (2 \times \frac{3}{10})

    E⁑[Y]=0+610+610=1210=1.2\operatorname{E}[Y] = 0 + \frac{6}{10} + \frac{6}{10} = \frac{12}{10} = 1.2
    "
    :::

    ---

    What's Next?

    πŸ’‘ Continue Your CMI Journey

    This chapter has laid the essential groundwork for understanding discrete random variables, their distributions, and key summary statistics like expectation and variance. Building on this foundation, your CMI journey will next delve into Continuous Random Variables, where outcomes can take any value within an interval, necessitating the use of probability density functions and integral calculus. You will also encounter specific distributions like the Binomial, Poisson, and Geometric, which are vital for modelling various real-world phenomena involving discrete counts or trials. A robust understanding of discrete variables is indispensable for mastering more advanced topics in probability theory and statistical inference.

    🎯 Key Points to Remember

    • βœ“ Master the core concepts in Discrete random variables before moving to advanced topics
    • βœ“ Practice with previous year questions to understand exam patterns
    • βœ“ Review short notes regularly for quick revision before exams

    Related Topics in Probability

    More Resources

    Why Choose MastersUp?

    🎯

    AI-Powered Plans

    Personalized study schedules based on your exam date and learning pace

    πŸ“š

    15,000+ Questions

    Verified questions with detailed solutions from past papers

    πŸ“Š

    Smart Analytics

    Track your progress with subject-wise performance insights

    πŸ”–

    Bookmark & Revise

    Save important questions for quick revision before exams

    Start Your Free Preparation β†’

    No credit card required β€’ Free forever for basic features