Spaces:
Running
Running
# /// script | |
# requires-python = ">=3.11" | |
# dependencies = [ | |
# "marimo", | |
# "matplotlib==3.10.0", | |
# "numpy==2.2.2", | |
# ] | |
# /// | |
import marimo | |
__generated_with = "0.11.2" | |
app = marimo.App(width="medium") | |
def _(): | |
import marimo as mo | |
return (mo,) | |
def _(mo): | |
mo.md( | |
r""" | |
# Axioms of Probability | |
Probability theory is built on three fundamental axioms, known as the [Kolmogorov axioms](https://en.wikipedia.org/wiki/Probability_axioms). These axioms form | |
the mathematical foundation for all of probability theory[<sup>1</sup>](https://chrispiech.github.io/probabilityForComputerScientists/en/part1/probability). | |
Let's explore each axiom and understand why they make intuitive sense: | |
""" | |
) | |
return | |
def _(mo): | |
mo.md( | |
r""" | |
## The Three Axioms | |
| Axiom | Mathematical Form | Meaning | | |
|-------|------------------|----------| | |
| **Axiom 1** | $0 \leq P(E) \leq 1$ | All probabilities are between 0 and 1 | | |
| **Axiom 2** | $P(S) = 1$ | The probability of the sample space is 1 | | |
| **Axiom 3** | $P(E \cup F) = P(E) + P(F)$ | For mutually exclusive events, probabilities add | | |
where the set $S$ is the sample space (all possible outcomes), and $E$ and $F$ are sets that represent events. The notation $P(E)$ denotes the probability of $E$, which you can interpret as the chance that something happens. $P(E) = 0$ means that the event cannot happen, while $P(E) = 1$ means the event will happen no matter what; $P(E) = 0.5$ means that $E$ has a 50% chance of happening. | |
For an example, when rolling a fair six-sided die once, the sample space $S$ is the set of die faces ${1, 2, 3, 4, 5, 6}$, and there are many possible events; we'll see some examples below. | |
""" | |
) | |
return | |
def _(mo): | |
mo.md( | |
r""" | |
## Understanding Through Examples | |
Let's explore these axioms using a simple experiment: rolling a fair six-sided die. | |
We'll use this to demonstrate why each axiom makes intuitive sense. | |
""" | |
) | |
return | |
def _(event): | |
event | |
return | |
def _(mo): | |
# Create an interactive widget to explore different events | |
event = mo.ui.dropdown( | |
options=[ | |
"Rolling an even number (2,4,6)", | |
"Rolling an odd number (1,3,5)", | |
"Rolling a prime number (2,3,5)", | |
"Rolling less than 4 (1,2,3)", | |
"Any possible roll (1,2,3,4,5,6)", | |
], | |
value="Rolling an even number (2,4,6)", | |
label="Select an event" | |
) | |
return (event,) | |
def _(event, mo, np, plt): | |
# Define the probabilities for each event | |
event_map = { | |
"Rolling an even number (2,4,6)": [2, 4, 6], | |
"Rolling an odd number (1,3,5)": [1, 3, 5], | |
"Rolling a prime number (2,3,5)": [2, 3, 5], | |
"Rolling less than 4 (1,2,3)": [1, 2, 3], | |
"Any possible roll (1,2,3,4,5,6)": [1, 2, 3, 4, 5, 6], | |
} | |
# Get outcomes directly from the event value | |
outcomes = event_map[event.value] | |
prob = len(outcomes) / 6 | |
# Visualize the probability | |
dice = np.arange(1, 7) | |
colors = ['#1f77b4' if d in outcomes else '#d3d3d3' for d in dice] | |
fig, ax = plt.subplots(figsize=(8, 2)) | |
ax.bar(dice, np.ones_like(dice), color=colors) | |
ax.set_xticks(dice) | |
ax.set_yticks([]) | |
ax.set_title(f"P(Event) = {prob:.2f}") | |
# Add explanation | |
explanation = mo.md(f""" | |
**Event**: {event.value} | |
**Probability**: {prob:.2f} | |
**Favorable outcomes**: {outcomes} | |
This example demonstrates: | |
- Axiom 1: The probability is between 0 and 1 | |
- Axiom 2: For the sample space, P(S) = 1 | |
- Axiom 3: The probability is the sum of individual outcome probabilities | |
""") | |
mo.hstack([plt.gcf(), explanation]) | |
return ax, colors, dice, event_map, explanation, fig, outcomes, prob | |
def _(mo): | |
mo.md( | |
r""" | |
## Why These Axioms Matter | |
These axioms are more than just rules - they provide the foundation for all of probability theory: | |
1. **Non-negativity** (Axiom 1) makes intuitive sense: you can't have a negative number of occurrences | |
in any experiment. | |
2. **Normalization** (Axiom 2) ensures that something must happen - the total probability must be 1. | |
3. **Additivity** (Axiom 3) lets us build complex probabilities from simple ones, but only for events | |
that can't happen together (mutually exclusive events). | |
From these simple rules, we can derive all the powerful tools of probability theory that are used in | |
statistics, machine learning, and other fields. | |
""" | |
) | |
return | |
def _(mo): | |
mo.md( | |
r""" | |
## π€ Test Your Understanding | |
Consider rolling two dice. Which of these statements follow from the axioms? | |
<details> | |
<summary>1. P(sum is 13) = 0</summary> | |
β Correct! This follows from Axiom 1. Since no combination of dice can sum to 13, | |
the probability must be non-negative but can be 0. | |
</details> | |
<details> | |
<summary>2. P(sum is 7) + P(sum is not 7) = 1</summary> | |
β Correct! This follows from Axioms 2 and 3. These events are mutually exclusive and cover | |
the entire sample space. | |
</details> | |
<details> | |
<summary>3. P(first die is 6 or second die is 6) = P(first die is 6) + P(second die is 6)</summary> | |
β Incorrect! This doesn't follow from Axiom 3 because the events are not mutually exclusive - | |
you could roll (6,6). | |
</details> | |
""" | |
) | |
return | |
def _(): | |
import numpy as np | |
import matplotlib.pyplot as plt | |
return np, plt | |
if __name__ == "__main__": | |
app.run() | |