AP Syllabus focus:
‘Essential Knowledge UNC-2.A.5: Elaborate on how the relative frequency of an outcome or event in simulated or empirical data is used to estimate the probability of that outcome or event. Include examples to illustrate this concept clearly.’
Estimating probabilities through simulation uses repeated random trials to approximate how often outcomes occur. This approach connects empirical patterns to theoretical probability, strengthening understanding of randomness and likelihood.
Estimating Probabilities Through Simulation
Estimating probabilities with simulation centers on the idea that relative frequency—the proportion of times an outcome occurs in repeated trials—serves as an approximation of its true probability. Simulation allows students to explore random processes when theoretical probability is difficult or impossible to compute, and it provides an accessible, data-driven route to understanding chance behavior.
Understanding Relative Frequency in Simulated Data
A random process is any situation in which outcomes are determined by chance, and simulated versions of these processes aim to mimic their real-world counterparts. When conducting a simulation, the key quantity of interest is the relative frequency, defined as the number of times a particular outcome occurs divided by the total number of simulated trials.
Relative Frequency: The proportion of trials in which a specific outcome or event occurs, calculated as (count of the outcome) ÷ (total number of trials).
Relative frequency provides a practical estimate of an event’s likelihood because it reflects how often the event appears under repeated random conditions. As the number of simulated trials increases, the relative frequency generally becomes more stable, offering a more precise approximation of the true probability.
Why Simulation Produces Useful Probability Estimates
Simulations help model contexts where theoretical reasoning is complex, or where empirical data from real-world trials may be unavailable. Because simulations represent random outcomes using chance-based mechanisms, such as random digit tables or computer random generators, patterns from the simulated data mirror the randomness present in actual processes.
Important reasons simulation improves probability estimation include:
Replicating randomness: Simulation tools ensure each trial imitates the behavior of a real random process.
Generating large datasets: Simulations can be run many times quickly, enabling more accurate relative frequency calculations.
Reducing obstacles: Some real-world trials may be too slow, costly, or impractical to repeat manually.
Each simulated trial contributes one outcome to the dataset, and these outcomes accumulate to reflect long-run tendencies associated with the event of interest.
Recording Outcomes in Simulation
Simulation requires careful and systematic recording of both individual outcomes and total counts so that relative frequency can be computed accurately. Clear documentation ensures the resulting probability estimates are reliable and reproducible.
A well-structured simulation typically includes:
A defined outcome of interest, such as a particular event or category.
A consistent mechanism for generating random results.
A system to tally every simulated outcome.
A final computation of relative frequency after all trials are completed.
These outcomes are often summarized in tables or graphs, making it easier to compare the likelihood of different outcomes.

Simulated probability distribution for the number of heads in four coin flips, based on 50 Monte Carlo trials. Each bar represents the estimated probability from relative frequency. The use of R code in the source introduces additional context, but only the displayed probability distribution is relevant here. Source.
These steps ensure alignment with the requirement that “all possible outcomes are associated with values determined by chance,” so that the simulation meaningfully models the random process.
Using Relative Frequency to Estimate Probability
An event’s probability represents the long-run proportion of times that event would occur over many repeated trials of the process. Simulation approximates this by using the relative frequency from repeated simulated trials as the best available estimate of the unknown probability.
EQUATION
= Event of interest, no units
= Count of trials where event E occurs
= Total number of simulated repetitions
Because simulations provide empirical evidence about how often events occur, the resulting relative frequency is interpreted as a reasonable stand-in for the true probability. Although the estimate may fluctuate with fewer trials, increasing the number of simulations narrows this variability and yields more dependable results.
Repeating the simulation many times lets us see how much the relative frequency of an event can vary from run to run, even when the underlying probability stays the same.

Dot plot of the proportion of heads across 100 simulations of 10,000 flips each. Each dot represents a relative-frequency estimate of the probability of heads, illustrating variability between simulations. The dotted reference lines and margin-of-error discussion in the source extend slightly beyond the syllabus requirements. Source.
Features of High-Quality Probability Simulations
To generate credible probability estimates, a simulation must accurately represent the underlying random process. This requires attention to several features:
Consistency: Each trial must use the same probability mechanism.
Independence: Outcomes of one trial must not influence later trials.
Accuracy in recording: Every outcome and the total number of trials must be tracked without omission.
Adequate number of trials: More repetitions generally lead to more stable and interpretable relative frequencies.
As the number of trials increases, simulated relative frequencies tend to stabilize near the true probability, even though small random fluctuations never completely disappear.

Running average of heads in 100 simulated coin tosses. The line becomes less erratic as more trials accumulate, approaching the theoretical probability 0.5 through stabilization of relative frequency. The Stata tool includes optional confidence-interval displays that are not needed for this subsubtopic. Source.
These characteristics ensure the simulation adheres closely to the behavior of the actual process and aligns with the AP Statistics emphasis on using empirical data to approximate unknown probabilities.
Connecting Simulation to Probability Reasoning
Simulation bridges intuition and formal probability by showing how random behavior accumulates over many trials. By observing how relative frequencies stabilize as more data are collected, students gain an appreciation for the empirical foundation of probability. This approach fosters deeper understanding of randomness, variability, and the logic underlying probability estimation, supporting further study in statistical inference and modeling.
FAQ
A simulation is valid when its structure reflects the essential features of the real random process. This includes using the correct probabilities, ensuring independence between trials, and representing all possible outcomes.
It must also avoid systematic bias. For example, a random number generator with uneven output would distort the results.
Finally, the simulation method should be repeatable so that others can produce similar results using the same procedure.
The number of trials depends on the level of precision needed. More trials reduce random fluctuation and produce a more stable relative frequency.
A practical approach is to run the simulation until additional trials no longer change the estimated probability meaningfully.
In applied settings, time or computing limits also influence the number of trials chosen.
Simulations rely on randomness, so each student’s set of trials represents a different sequence of random outcomes.
These differing sequences lead to natural variation in relative frequency, even when using the same model.
However, as the number of trials increases, the estimates from both students typically become more similar.
Simulation can show how frequently an outcome occurs under a given probability model.
If an outcome appears rarely in many simulated runs, it may be considered unusual in context.
This supports informal reasoning about extremity without applying formal hypothesis testing.
Signs of instability include large swings in the relative frequency as more trials are added.
Other indicators include:
• small total number of trials
• inconsistent estimates across repeated simulations
• reliance on a flawed or biased random mechanism
Unstable estimates suggest that more trials or an improved simulation design are needed.
Practice Questions
Question 1 (1–3 marks)
A student runs a simulation of 200 trials to estimate the probability that a randomly selected seed germinates. The seed germinated in 54 of the simulated trials.
Estimate the probability of germination based on the simulation, and explain briefly why this estimate is reasonable
Question 1
• 1 mark for correctly calculating the relative frequency:
54 ÷ 200 = 0.27 (or equivalent such as 27%).
• 1 mark for stating that this value is the estimated probability.
• 1 mark for explaining why this estimate is reasonable, e.g. because simulations approximate long-run relative frequencies or because repeated trials mimic the random process.
Question 2 (4–6 marks)
A researcher wants to estimate the probability that a particular email is classified as spam by an automated filter. To do this, the researcher uses a computer program to simulate 1,000 independent email classifications.
The simulation results show that 312 of the simulated emails were classified as spam.
(a) Use the simulation results to estimate the probability that an email is classified as spam.
(b) The researcher repeats the entire simulation process nine more times, each time running 1,000 simulated classifications. The proportions classified as spam across the ten simulations range from 0.296 to 0.324. Explain why the proportions vary between simulations, even though the underlying probability is the same.
(c) State why simulations with a larger number of trials generally produce more accurate estimates of the true probability.
Question 2
(a) (1 mark)
• Correct estimate of probability: 312 ÷ 1,000 = 0.312.
(b) (2–3 marks)
• 1 mark for stating that simulation results vary due to randomness or chance variation.
• 1 mark for explaining that each simulated set of 1,000 emails is a different sample of outcomes.
• Optional full explanation mark for noting that even when the true probability is fixed, relative frequencies fluctuate from sample to sample.
(c) (2 marks)
• 1 mark for stating that larger numbers of trials reduce variability in the relative frequency.
• 1 mark for explaining that more trials produce estimates that are closer to the true long-run probability.
