Writing ml
ml 10 min read 25 January 2024

Probability as an Operating System for Better Decisions

Bayesian reasoning, belief updating, and calibrated uncertainty — how probabilistic thinking changes the way you interpret evidence and make decisions under uncertainty.

Most people treat probability as a tool for calculating lottery odds or understanding insurance pricing. It’s actually a framework for reasoning under uncertainty — one that applies everywhere from clinical diagnosis to model evaluation to everyday decisions.

If you learn to think probabilistically, you’ll be surprised how often it reframes problems that seemed intractable.

Probability as Belief

The classical definition of probability (frequency of outcomes in repeated experiments) is fine for coin flips. It breaks down for one-off events: “What is the probability that this drug trial succeeds?” “What is the probability that this startup exits within five years?”

The Bayesian interpretation is more general: probability represents a degree of belief, quantifying uncertainty about a proposition. It can be updated as evidence arrives. This makes probability a tool for reasoning, not just calculation.

A probability of 0.7 for a drug trial succeeding doesn’t mean “70% of similar trials succeed.” It means “given what I know now, I believe this trial is more likely to succeed than fail, with that relative confidence.”

Conditional Probability and Bayes’ Theorem

Conditional probability $P(A|B)$ is the probability of A given that B has occurred. It’s the most important concept in probabilistic reasoning.

Bayes’ Theorem connects conditional probabilities:

$$P(A|B) = \frac{P(B|A) \cdot P(A)}{P(B)}$$

In words: to update your belief in hypothesis A after observing evidence B, multiply your prior belief $P(A)$ by the likelihood $P(B|A)$ (how likely you’d see B if A were true), and normalize.

The named terms:

This is the fundamental process of rational belief updating.

A Diagnostic Example

Suppose a disease affects 1% of the population. A test is 99% accurate (both sensitivity and specificity). You test positive. What is the probability you actually have the disease?

Most people intuit around 99%. The actual answer:

$$P(\text{disease} | \text{positive}) = \frac{0.99 \times 0.01}{0.99 \times 0.01 + 0.01 \times 0.99} = 0.5$$

50%. Because the disease is rare, even a very accurate test generates many false positives among the large healthy population.

This is base rate neglect — one of the most common errors in probabilistic reasoning. The prior matters.

The Four Disciplines of Probabilistic Thinking

1. Specify Your Prior Explicitly

Before seeing data, state your prior beliefs clearly. This forces you to confront what you actually believe and why. Vague beliefs produce vague reasoning.

Priors come from domain knowledge, historical data, reference classes (“what’s the base rate for projects like this?”), and calibrated estimation. The prior doesn’t have to be precise — even a rough probability range is better than an implicit, unstated assumption.

2. Update on Evidence, Not Narrative

When new evidence arrives, update your beliefs systematically via Bayes’ rule. The mistake people make is narrative updating — changing beliefs based on how compelling a story sounds rather than how much the evidence actually shifts the probability.

Strong evidence that makes an unlikely hypothesis look likely deserves a large update. Weak or ambiguous evidence that “sounds” significant deserves a small update.

3. Maintain Calibration

A well-calibrated forecaster’s 70% confidence predictions are right 70% of the time. Most people are overconfident: their “80% sure” calls are right less than 80% of the time.

Calibration is a skill. It improves with feedback. Make probabilistic predictions, track outcomes, and adjust. This is why meteorologists are better calibrated than pundits — meteorologists get daily feedback.

4. Separate Uncertainty Types

Aleatory uncertainty: randomness that cannot be reduced with more data. A fair die has aleatory uncertainty; no amount of information tells you which face will land.

Epistemic uncertainty: uncertainty from lack of knowledge, reducible with more data. Whether a drug is effective is epistemically uncertain — more trials give you a clearer answer.

Conflating these leads to poor decisions. Epistemic uncertainty is worth reducing (run the experiment). Aleatory uncertainty isn’t (accept the distribution and design around it).

Comparing Probability Distributions

Once you move from single events to machine learning systems, you need to measure how different two probability distributions are. Several metrics do this, each with different properties.

KL Divergence

$$D_{KL}(p||q) = -\int p(x) \log\frac{q(x)}{p(x)}$$

The extra bits required to represent samples from $p$ using a code designed for $q$. Not symmetric: $D_{KL}(p||q) \neq D_{KL}(q||p)$. When $p$ and $q$ don’t overlap at all, KL divergence is infinite — a problem in training GANs.

In practice: cross-entropy loss in classification is equivalent to minimizing KL divergence between the predicted and true label distributions (the entropy term is constant).

Jensen-Shannon Divergence

$$D_{JS}(p||q) = \frac{1}{2}D_{KL}(p||\frac{p+q}{2}) + \frac{1}{2}D_{KL}(q||\frac{p+q}{2})$$

A symmetric, bounded (0 to 1 with base-2 log) version of KL. Handles non-overlapping distributions. Preferred for comparing distributions in GAN training.

Wasserstein Distance (Earth Mover’s Distance)

$$W(p,q) = \inf_{\gamma \sim \Pi(p,q)} \mathbb{E}_{(x,y) \sim \gamma}[||x - y||]$$

Measures the minimum “work” required to transform distribution $p$ into distribution $q$. Sensitive to distance between distributions, not just overlap. This is why WGAN training is more stable than original GAN: even when distributions don’t overlap, the gradient is informative.

When $p$ and $q$ are non-overlapping uniform distributions separated by distance $\theta$: KL divergence is infinite, JS is a constant $\log 2$, and Wasserstein is $\theta$. Only Wasserstein carries the useful signal.

Generative vs Discriminative Models

In ML, probabilistic thinking shows up in the fundamental distinction between model types:

Discriminative models learn $P(y|x)$ — the probability of the label given the input. Logistic regression, SVMs, most neural networks. They learn a decision boundary. Direct and efficient for prediction tasks.

Generative models learn the joint distribution $P(x, y) = P(x|y)P(y)$. Naive Bayes, VAEs, GANs. They learn what data looks like, not just how to classify it. Useful for generating new samples, anomaly detection (data that has low probability under the model is anomalous), and learning with limited labels.

Entropy as a Measure of Uncertainty

Entropy of a random variable $X$:

$$H(X) = -\sum_X P(X) \log P(X)$$

High entropy = high uncertainty. A uniform distribution over 10 classes has maximum entropy — you know nothing. A distribution that puts all probability mass on one class has zero entropy — you’re certain.

Entropy appears everywhere:

Understanding entropy gives you a unified vocabulary across all these domains.

The Practical Upshot

Probabilistic thinking changes how you handle evidence and decisions:

  1. When someone shows you evidence for a surprising claim, ask: how likely is this evidence under the null hypothesis? Strong evidence for an unlikely claim should shift your belief — but only proportionally to the evidence’s specificity.

  2. When making predictions, give ranges with confidence levels, not point estimates. “I’m 80% confident the project finishes between 6 and 9 weeks” is more honest and more actionable than “it’ll take 8 weeks.”

  3. When evaluating models, look at calibration curves — not just accuracy. A model that’s right 90% of the time but says “99% confident” when it’s wrong is worse than one that’s right 85% with accurate uncertainty estimates.

  4. When designing systems that depend on probabilistic model outputs (like fraud flagging or clinical decision support), design around the full distribution of predictions, not just the point estimate.

Probability doesn’t eliminate uncertainty. It makes uncertainty tractable.

probability bayesian-reasoning uncertainty decision-making statistics
← All articles

Lets collaborate!

Whether you need a quantitative researcher, an machine learning systems builder, or a technical advisor — I'm available for select consulting engagements.

Get in Touch →