All Topics
mathematics-international-0607-core | cambridge-igcse
Responsive Image
2. Number
5. Transformations and Vectors
Calculating probability of events

Topic 2/3

left-arrow
left-arrow
archive-add download share

Your Flashcards are Ready!

15 Flashcards in this deck.

or
NavTopLeftBtn
NavTopRightBtn
3
Still Learning
I know
12

Calculating Probability of Events

Introduction

Probability is a fundamental concept in mathematics that quantifies the likelihood of events occurring. In the context of the Cambridge IGCSE Mathematics curriculum (0607 Core), understanding how to calculate probabilities is essential for solving real-world problems and advancing to more complex statistical analyses. This article delves into the methods and principles of calculating the probability of events, providing a comprehensive guide for students preparing for their examinations.

Key Concepts

Understanding Probability

Probability measures the chance that a particular event will occur. It is expressed as a number between 0 and 1, where 0 indicates impossibility and 1 indicates certainty. The probability of an event can also be represented as a percentage between 0% and 100%.

Basic Probability Formula

The fundamental formula for probability is:

$$P(A) = \frac{\text{Number of favorable outcomes for } A}{\text{Total number of possible outcomes}}$$

Where:

  • P(A) is the probability of event A occurring.
  • Number of favorable outcomes refers to outcomes that satisfy event A.
  • Total number of possible outcomes is the total number of all possible results in the experiment.

Example: Consider a fair six-sided die. What is the probability of rolling a 4?

Solution: There is only one favorable outcome (rolling a 4) out of six possible outcomes.

$$P(4) = \frac{1}{6} \approx 0.1667 \, \text{or} \, 16.67\%$$

Types of Events

Events can be categorized based on their outcomes:

  • Simple Events: Events with a single outcome. For example, drawing an Ace from a standard deck of cards.
  • Compound Events: Events with more than one outcome. For example, drawing a King or a Queen from a deck of cards.
  • Independent Events: Events where the outcome of one does not affect the outcome of another. For example, flipping a coin and rolling a die.
  • Dependent Events: Events where the outcome of one affects the outcome of another. For example, drawing two cards from a deck without replacement.

Mutually Exclusive Events

Two events are mutually exclusive if they cannot occur simultaneously. This means the occurrence of one event excludes the possibility of the other event occurring.

Formula:

$$P(A \text{ or } B) = P(A) + P(B)$$

Only applicable if A and B are mutually exclusive.

Example: Rolling a 2 or a 5 on a six-sided die.

Solution: Since a die cannot show both 2 and 5 at the same time:

$$P(2 \text{ or } 5) = P(2) + P(5) = \frac{1}{6} + \frac{1}{6} = \frac{2}{6} = \frac{1}{3} \approx 0.3333 \, \text{or} \, 33.33\%$$

Non-Mutually Exclusive Events

Events that can occur simultaneously are known as non-mutually exclusive events. In such cases, the probability of both events occurring together must be subtracted to avoid double-counting.

Formula:

$$P(A \text{ or } B) = P(A) + P(B) - P(A \text{ and } B)$$

Example: Drawing a card that is either a King or a Heart from a standard deck.

Solution: There are 4 Kings and 13 Hearts, but one King is also a Heart (King of Hearts).

$$P(\text{King or Heart}) = \frac{4}{52} + \frac{13}{52} - \frac{1}{52} = \frac{16}{52} = \frac{4}{13} \approx 0.3077 \, \text{or} \, 30.77\%$$

Complementary Events

The complement of an event A is the event that A does not occur. The probabilities of complementary events sum to 1.

Formula:

$$P(A') = 1 - P(A)$$

Example: If the probability of it raining tomorrow is 0.3, what is the probability that it will not rain?

Solution:

$$P(\text{No Rain}) = 1 - 0.3 = 0.7 \, \text{or} \, 70\%$$

Independent Events

When two events are independent, the occurrence of one does not affect the probability of the other.

Formula:

$$P(A \text{ and } B) = P(A) \times P(B)$$

Example: Tossing a fair coin and rolling a fair six-sided die.

Solution: Probability of getting Heads and a 4:

$$P(\text{Heads and } 4) = \frac{1}{2} \times \frac{1}{6} = \frac{1}{12} \approx 0.0833 \, \text{or} \, 8.33\%$$

Dependent Events

In dependent events, the outcome of one event affects the outcome of another.

Example: Drawing two cards from a deck without replacement.

Solution: Probability of drawing an Ace first and then another Ace:

$$P(\text{First Ace}) = \frac{4}{52} = \frac{1}{13}$$

After drawing one Ace, there are now 3 Aces left out of 51 cards.

$$P(\text{Second Ace}) = \frac{3}{51} = \frac{1}{17}$$

Thus:

$$P(\text{Two Aces}) = \frac{1}{13} \times \frac{1}{17} = \frac{1}{221} \approx 0.0045 \, \text{or} \, 0.45\%$$

Permutations and Combinations

Permutations and combinations are used to calculate the number of ways events can occur.

  • Permutations: Order matters. The number of permutations of n items taken r at a time is:
  • $$P(n, r) = \frac{n!}{(n - r)!}$$

  • Combinations: Order does not matter. The number of combinations of n items taken r at a time is:
  • $$C(n, r) = \frac{n!}{r!(n - r)!}$$

Example: How many ways can 3 students be selected from a group of 5?

Solution (Combinations):

$$C(5, 3) = \frac{5!}{3!(5 - 3)!} = \frac{120}{6 \times 2} = 10$$

Probability Distributions

A probability distribution assigns probabilities to each possible outcome of a random variable. For discrete random variables, this is often represented in a table.

Example: Probability distribution for rolling a six-sided die:

Outcome Probability
1 $\frac{1}{6}$
2 $\frac{1}{6}$
3 $\frac{1}{6}$
4 $\frac{1}{6}$
5 $\frac{1}{6}$
6 $\frac{1}{6}$

Expected Value

The expected value is the average outcome if an experiment is repeated many times. It is calculated by multiplying each outcome by its probability and summing the results.

Formula:

$$E(X) = \sum (x_i \times P(x_i))$$

Example: Expected value when rolling a six-sided die:

$$E(X) = 1 \times \frac{1}{6} + 2 \times \frac{1}{6} + 3 \times \frac{1}{6} + 4 \times \frac{1}{6} + 5 \times \frac{1}{6} + 6 \times \frac{1}{6} = \frac{21}{6} = 3.5$$

Law of Large Numbers

The Law of Large Numbers states that as the number of trials increases, the experimental probability of an event tends to get closer to the theoretical probability.

Example: If you flip a fair coin 1000 times, the proportion of heads should be close to 50%.

Conditional Probability

Conditional probability is the probability of an event occurring given that another event has already occurred.

Formula:

$$P(A|B) = \frac{P(A \text{ and } B)}{P(B)}$$

Example: In a deck of 52 cards, what is the probability of drawing an Ace given that the first card drawn was a King?

Solution: After drawing a King, there are still 4 Aces left out of 51 cards.

$$P(\text{Ace|King}) = \frac{4}{51} \approx 0.0784 \, \text{or} \, 7.84\%$$

Bayes' Theorem

Bayes' Theorem relates the conditional and marginal probabilities of events A and B.

Formula:

$$P(A|B) = \frac{P(B|A) \times P(A)}{P(B)}$$

Example: Suppose 5% of the population has a certain disease. A test for the disease has a 99% true positive rate and a 2% false positive rate. What is the probability that a person has the disease given that they tested positive?

Solution:

Let A be the event of having the disease, and B be testing positive.

$$P(A) = 0.05$$

$$P(B|A) = 0.99$$

$$P(B) = P(B|A)P(A) + P(B|A')P(A') = (0.99 \times 0.05) + (0.02 \times 0.95) = 0.0495 + 0.019 = 0.0685$$

Thus:

$$P(A|B) = \frac{0.99 \times 0.05}{0.0685} \approx \frac{0.0495}{0.0685} \approx 0.7226 \, \text{or} \, 72.26\%$$

This means there's a 72.26% chance that a person has the disease given a positive test result.

Probability Trees

Probability trees are diagrams that represent all possible outcomes of a sequence of events, making it easier to visualize and calculate probabilities of compound events.

Example: Tossing two coins.

Solution: The tree diagram would have two branches for the first toss (Heads or Tails) and for each of those, two branches for the second toss (Heads or Tails).

Possible outcomes: HH, HT, TH, TT.

Probability of each outcome: Each outcome has a probability of $\frac{1}{4}$.

Applications of Probability

Probability is widely used in various fields such as:

  • Finance: Assessing risk and return in investments.
  • Medicine: Determining the likelihood of diseases and treatment outcomes.
  • Engineering: Reliability testing and quality control.
  • Sports: Predicting outcomes of games and performances.
  • Daily Life: Making decisions based on uncertain outcomes, like weather forecasts.

Common Misconceptions

  • Gambler's Fallacy: The belief that past random events affect the probabilities in future independent events. For example, thinking that after flipping three heads in a row, tails are "due."
  • Confusing Probability with Possibility: Just because an event is possible doesn't mean it is equally probable as other events.
  • Overlooking Dependent Events: Ignoring how the outcome of one event affects another in dependent scenarios.

Advanced Concepts

Joint Probability

Joint probability refers to the probability of two events happening simultaneously. It is crucial in understanding the relationship between multiple events.

Formula:

$$P(A \text{ and } B)$$

Example: Probability of drawing a red card that is also a King from a standard deck.

Solution: There are two red Kings (King of Hearts and King of Diamonds).

$$P(\text{Red King}) = \frac{2}{52} = \frac{1}{26} \approx 0.0385 \, \text{or} \, 3.85\%$$

Conditional Probability with Multiple Conditions

Calculating the probability of an event given multiple conditions requires considering all given information and how it affects the sample space.

Example: In a class of 30 students, 12 play football, 15 play basketball, and 5 play both. What is the probability that a student plays football given that they play basketball?

Solution:

$$P(\text{Football|Basketball}) = \frac{P(\text{Football and Basketball})}{P(\text{Basketball})} = \frac{5}{15} = \frac{1}{3} \approx 0.3333 \, \text{or} \, 33.33\%$$

Bayesian Probability

Bayesian probability allows updating the probability of a hypothesis as more evidence becomes available. It's particularly useful in fields like medical diagnostics and machine learning.

Example: Revisiting the earlier example with disease testing, Bayesian probability helps refine the likelihood of disease as more test results become available.

Probability Distributions: Binomial and Normal

Dive deeper into specific probability distributions that model different types of data and phenomena.

Binomial Distribution

The binomial distribution models the number of successes in a fixed number of independent trials, each with the same probability of success.

Conditions:

  • Fixed number of trials (n).
  • Each trial has two possible outcomes: success or failure.
  • The probability of success (p) is constant for each trial.
  • Trials are independent.

Formula:

$$P(X = k) = C(n, k) p^k (1 - p)^{n - k}$$

Example: Probability of getting exactly 3 heads in 5 coin tosses.

Solution:

$$P(X = 3) = C(5, 3) \left(\frac{1}{2}\right)^3 \left(1 - \frac{1}{2}\right)^{5 - 3} = 10 \times \frac{1}{8} \times \frac{1}{4} = \frac{10}{32} = \frac{5}{16} \approx 0.3125 \, \text{or} \, 31.25\%$$

Normal Distribution

The normal distribution, also known as the Gaussian distribution, is a continuous probability distribution characterized by its bell-shaped curve, symmetric around the mean.

Properties:

  • Defined by two parameters: mean ($\mu$) and standard deviation ($\sigma$).
  • Approximately 68% of data lies within one standard deviation of the mean.
  • About 95% within two standard deviations, and 99.7% within three.

Standard Normal Distribution: A normal distribution with $\mu = 0$ and $\sigma = 1$.

Applications:

  • Modeling natural phenomena like heights and test scores.
  • Statistical inference and hypothesis testing.

Expected Value and Variance

While the expected value provides the average outcome, variance measures the spread of the probabilities around the expected value.

Variance Formula:

$$Var(X) = E[(X - \mu)^2] = \sum (x_i - \mu)^2 P(x_i)$$

Example: Rolling a six-sided die.

Solution:

First, calculate the expected value:

$$E(X) = 3.5$$

Then, calculate the variance:

$$Var(X) = (1 - 3.5)^2 \times \frac{1}{6} + (2 - 3.5)^2 \times \frac{1}{6} + \cdots + (6 - 3.5)^2 \times \frac{1}{6} = \frac{17.5}{6} \approx 2.9167$$

Standard Deviation:

$$\sigma = \sqrt{Var(X)} \approx \sqrt{2.9167} \approx 1.7078$$

Law of Total Probability

The Law of Total Probability relates marginal probabilities to conditional probabilities across different scenarios or partitions of the sample space.

Formula:

$$P(B) = P(B|A_1)P(A_1) + P(B|A_2)P(A_2) + \cdots + P(B|A_n)P(A_n)$$

Example: Suppose a factory has two machines producing widgets. Machine 1 produces 60% of the widgets with a defect rate of 2%, and Machine 2 produces 40% with a defect rate of 5%. What is the probability that a randomly selected widget is defective?

Solution:

$$P(\text{Defective}) = P(\text{Defective|Machine 1})P(\text{Machine 1}) + P(\text{Defective|Machine 2})P(\text{Machine 2})$$

$$= 0.02 \times 0.6 + 0.05 \times 0.4 = 0.012 + 0.02 = 0.032 \, \text{or} \, 3.2\%$$

Markov Chains

Markov Chains are mathematical systems that undergo transitions from one state to another based on certain probabilistic rules, with the property that the next state depends only on the current state and not on the sequence of events that preceded it.

Applications:

  • Predicting weather patterns.
  • Google’s PageRank algorithm.
  • Modeling queues in service systems.

Monte Carlo Simulations

Monte Carlo simulations use random sampling and statistical modeling to estimate mathematical functions and mimic the operation of complex systems.

Applications:

  • Financial forecasting.
  • Risk assessment.
  • Optimizing logistics and operations.

Interdisciplinary Connections

Probability theory intersects with various disciplines, enhancing its applicability and utility:

  • Physics: Quantum mechanics relies heavily on probabilistic models.
  • Biology: Genetics and evolutionary biology use probability to predict trait distributions.
  • Economics: Used in modeling market behaviors and in game theory.
  • Computer Science: Essential in algorithms, machine learning, and artificial intelligence.
  • Medicine: Critical in biostatistics and epidemiology for understanding disease spread and treatment efficacy.

Advanced Problem-Solving Techniques

To tackle complex probability problems, students should master various techniques:

  • Inclusion-Exclusion Principle: Used to calculate the probability of the union of multiple events.
  • Recursive Thinking: Breaking down problems into smaller, more manageable sub-problems.
  • Visualization: Utilizing diagrams like probability trees and Venn diagrams to conceptualize problems.
  • Algebraic Manipulation: Rearranging formulas and equations to isolate desired variables.

Common Advanced Problems

Problem 1: A box contains 5 red, 7 blue, and 8 green marbles. If two marbles are drawn at random without replacement, what is the probability that both marbles are blue?

Solution:

Total marbles: 20.

Probability of first marble being blue: $\frac{7}{20}$.

After drawing one blue marble, remaining marbles: 19, with 6 blue marbles.

Probability of second marble being blue: $\frac{6}{19}$.

Thus:

$$P(\text{Both Blue}) = \frac{7}{20} \times \frac{6}{19} = \frac{42}{380} = \frac{21}{190} \approx 0.1105 \, \text{or} \, 11.05\%$$

Problem 2: In a standard deck of cards, what is the probability of drawing a Queen or a card of Spades?

Solution:

Number of Queens: 4.

Number of Spades: 13.

Overlap (Queen of Spades): 1.

Using the formula for non-mutually exclusive events:

$$P(\text{Queen or Spade}) = P(\text{Queen}) + P(\text{Spade}) - P(\text{Queen of Spades})$$

$$= \frac{4}{52} + \frac{13}{52} - \frac{1}{52} = \frac{16}{52} = \frac{4}{13} \approx 0.3077 \, \text{or} \, 30.77\%$$

Problem 3: A fair die is rolled twice. What is the probability that the sum of the two rolls is greater than 9?

Solution:

Possible sums greater than 9: 10, 11, 12.

Number of outcomes for each sum:

  • 10: (4,6), (5,5), (6,4) → 3 outcomes.
  • 11: (5,6), (6,5) → 2 outcomes.
  • 12: (6,6) → 1 outcome.

Total favorable outcomes: 6.

Total possible outcomes when rolling a die twice: 36.

Thus:

$$P(\text{Sum} > 9) = \frac{6}{36} = \frac{1}{6} \approx 0.1667 \, \text{or} \, 16.67\%$$

Markov Chains in Detail

Markov Chains are powerful tools for modeling systems that transition from one state to another. Each state represents a possible condition, and transitions are governed by probabilities.

Key Components:

  • States: Distinct conditions or positions in the system.
  • Transition Probabilities: The probability of moving from one state to another.
  • Transition Matrix: A matrix representing all transition probabilities between states.

Example: Weather Prediction.

Consider a simple model with two states: Sunny (S) and Rainy (R).

  • Probability of it being Sunny tomorrow if it's Sunny today: 0.8
  • Probability of it being Rainy tomorrow if it's Sunny today: 0.2
  • Probability of it being Sunny tomorrow if it's Rainy today: 0.4
  • Probability of it being Rainy tomorrow if it's Rainy today: 0.6

The transition matrix (P) is:

Sunny Rainy
Sunny 0.8 0.2
Rainy 0.4 0.6

This matrix can be used to predict future weather patterns based on current conditions.

Monte Carlo Methods in Depth

Monte Carlo simulations rely on repeated random sampling to obtain numerical results. They are particularly useful for modeling complex systems where analytical solutions are difficult to derive.

Steps Involved:

  1. Define a domain of possible inputs.
  2. Generate inputs randomly from a probability distribution.
  3. Perform a deterministic computation using the inputs.
  4. Aggregate the results to form a probability distribution of the outcomes.

Example: Estimating the value of π using Monte Carlo simulation.

Solution:

  • Consider a square with side length 2, centered at the origin, and a quarter circle of radius 1 inside it.
  • Randomly generate points within the square.
  • Determine the fraction of points that fall inside the quarter circle.
  • Multiply this fraction by 4 to approximate π.

If 10,000 points are generated and 7,850 fall inside the quarter circle:

$$\pi \approx 4 \times \frac{7850}{10000} = 3.14$$

Advanced Probability Theorems

Understanding advanced theorems enhances problem-solving capabilities:

  • Central Limit Theorem: States that the distribution of the sum of a large number of independent, identically distributed variables tends toward a normal distribution, regardless of the original distribution.
  • Law of Large Numbers: As previously mentioned, the average of the results obtained from a large number of trials tends to be close to the expected value.
  • Chebyshev’s Inequality: Provides a bound on the probability that the value of a random variable deviates from its mean.

Bayesian Networks

Bayesian Networks are graphical models that represent the probabilistic relationships among a set of variables. They are used to model uncertainty and make predictions based on known data.

Components:

  • Nodes: Represent random variables.
  • Edges: Indicate conditional dependencies.

Applications:

  • Medical diagnosis.
  • Risk management.
  • Machine learning algorithms.

Stochastic Processes

Stochastic processes are collections of random variables representing the evolution of a system over time. They are used to model phenomena that evolve with inherent randomness.

Types:

  • Discrete-time stochastic processes: Observations at discrete time intervals.
  • Continuous-time stochastic processes: Observations continuously over time.

Example: Stock prices modeled as a stochastic process to predict future values based on past behavior.

Markov Decision Processes

Markov Decision Processes (MDPs) provide a mathematical framework for modeling decision-making in situations where outcomes are partly random and partly under the control of a decision-maker.

Components:

  • States: Represent all possible situations.
  • Actions: Choices available to the decision-maker.
  • Transition Probabilities: The probability of moving from one state to another after an action.
  • Rewards: Immediate gain or loss from taking an action in a state.

Applications:

  • Robotics and autonomous systems.
  • Economics and game theory.
  • Operations research and logistics.

Probability in Machine Learning

Probability theory underpins many machine learning algorithms, enabling models to handle uncertainty and make predictions based on data.

Applications:

  • Naive Bayes Classifier: Uses Bayes' Theorem for classification tasks.
  • Hidden Markov Models: Used in speech recognition and bioinformatics.
  • Probabilistic Graphical Models: Represent complex dependencies among variables.

Advanced Counting Techniques

Complex probability problems often require advanced counting methods:

  • Inclusion-Exclusion Principle: Helps in calculating the union of multiple overlapping sets.
  • Pigeonhole Principle: Ensures that certain conditions are met when objects are distributed into containers.
  • Multinomial Coefficients: Extend the concept of combinations to more than two categories.

Example: How many ways can 10 identical balls be distributed into 3 distinct boxes?

Solution: Using combinations with repetition:

$$C(n + k - 1, k - 1) = C(10 + 3 - 1, 3 - 1) = C(12, 2) = \frac{12 \times 11}{2 \times 1} = 66$$

Generating Functions

Generating functions encode sequences of numbers (like probabilities) into algebraic forms, facilitating the manipulation and extraction of sequence information.

Applications:

  • Solving recurrence relations.
  • Analyzing probability distributions.
  • Calculating moments and cumulants in statistics.

Random Variables

Random variables are functions that assign numerical values to each outcome in a sample space, enabling the application of probability to numerical data.

Types:

  • Discrete Random Variables: Take on a countable number of distinct values.
  • Continuous Random Variables: Take on an infinite number of possible values within a range.

Example: The number of heads in 10 coin tosses (discrete) vs. the height of students in a class (continuous).

Moment Generating Functions

Moment generating functions (MGFs) are used to uniquely determine the probability distribution of a random variable by encoding its moments (mean, variance, etc.).

Formula:

$$M_X(t) = E[e^{tX}]$$

Applications:

  • Calculating moments of random variables.
  • Proving the Central Limit Theorem.
  • Facilitating the analysis of sums of independent random variables.

Comparison Table

Concept Definition Application
Independent Events Events where the outcome of one does not affect the other. Coin tosses, rolling dice.
Dependent Events Events where the outcome of one affects the other. Drawing cards without replacement.
Mutually Exclusive Events Events that cannot occur simultaneously. Rolling a 2 or a 5 on a die.
Complementary Events Events where one is the opposite of the other. Rain vs. no rain.
Permutations Arrangements where order matters. Ordering books on a shelf.
Combinations Selections where order does not matter. Selecting team members.

Summary and Key Takeaways

  • Probability quantifies the likelihood of events, ranging from 0 to 1.
  • Understanding different types of events (mutually exclusive, independent) is crucial for accurate calculations.
  • Advanced concepts like Bayes' Theorem and probability distributions expand the application of probability in various fields.
  • Mastering counting techniques and probability theorems enhances problem-solving skills.
  • Interdisciplinary connections highlight the pervasive role of probability in everyday life and advanced scientific research.

Coming Soon!

coming soon
Examiner Tip
star

Tips

To excel in calculating probabilities, always start by clearly defining your events and determining if they are independent or dependent. Use the acronym PAC to remember: Probability of Anti-events Complementary. For mutually exclusive events, simply add their probabilities, but for non-mutually exclusive events, remember to subtract the intersection to avoid double-counting. Additionally, visualize complex problems with probability trees or Venn diagrams to better understand relationships between events. Practicing these strategies will enhance your problem-solving skills and boost your confidence for exams.

Did You Know
star

Did You Know

Did you know that the Birthday Paradox reveals that in a group of just 23 people, there's about a 50% chance that two individuals share the same birthday? This counterintuitive result highlights how probability can challenge our natural instincts. Additionally, in the realm of quantum mechanics, probability plays a crucial role in predicting the behavior of particles, where outcomes are inherently uncertain until measured. Moreover, probability theory is fundamental in genetics, helping scientists predict the likelihood of inheriting specific traits and understanding genetic variation within populations.

Common Mistakes
star

Common Mistakes

One common mistake students make is confusing independent events with mutually exclusive events. For example, assuming that rolling a 2 and a 3 on a die are mutually exclusive, when in reality, they are independent events. Another frequent error is incorrectly adding probabilities without considering overlap, such as calculating the probability of drawing a King or a Heart from a deck without subtracting the joint probability of drawing the King of Hearts. Lastly, students often misapply probability formulas to dependent events, overlooking how the outcome of one event affects the probability of another.

FAQ

What is the difference between probability and odds?
Probability measures the likelihood of an event occurring, expressed as a ratio of favorable outcomes to total outcomes. Odds, on the other hand, compare the number of favorable outcomes to the number of unfavorable ones.
How do you determine if two events are independent?
Two events are independent if the occurrence of one does not affect the probability of the other. Mathematically, P(A and B) = P(A) × P(B).
What is the complement of an event?
The complement of an event A is the event that A does not occur. It is calculated as P(A') = 1 - P(A).
How do you calculate the probability of compound events?
For compound events, determine if the events are independent or dependent. Use appropriate formulas: P(A and B) = P(A) × P(B) for independent events, and P(A and B) = P(A) × P(B|A) for dependent events.
Can probabilities be greater than 1?
No, probabilities range from 0 to 1. A probability of 0 means the event will not occur, while a probability of 1 means the event is certain to occur.
2. Number
5. Transformations and Vectors
Download PDF
Get PDF
Download PDF
PDF
Share
Share
Explore
Explore
How would you like to practise?
close