← All Essays
◆ Decoded Decision Making ~13 min read

Decision Theory Decoded

Core Idea: Every decision is a bet under uncertainty. The question is how to make that bet well. Classical theory says: calculate expected value. Behavioral research says: humans don't do that—they feel losses twice as sharply as gains, warp probabilities, and anchor to reference points. Ecological rationality says: sometimes simple rules beat complex calculations. The practical synthesis is that good decision-making isn't about being perfectly rational. It's about knowing which tool to reach for—optimization, heuristics, or structured deliberation—based on the stakes, the uncertainty, and whether you can undo the choice if you're wrong.

In 1738, a Swiss mathematician named Daniel Bernoulli posed a puzzle that would take two and a half centuries to fully unravel. Imagine a coin-flipping game: heads pays you $2 on the first flip, $4 if the first heads comes on the second flip, $8 on the third, and so on—doubling each time. How much would you pay to play? The mathematically "correct" answer, using expected value, is infinity. The sum of each outcome's probability times its payoff diverges. But no sane person would pay more than about $20. Everyone who encounters this problem—the St. Petersburg paradox—feels, immediately and correctly, that something is wrong with the math. Not the arithmetic. The framework. The gap between what the formula says you should do and what every fiber of your judgment says you should do is the crack through which the entire field of decision theory poured.

Bernoulli's answer was that we don't value money linearly. The jump from $0 to $10,000 changes your life. The jump from $1,000,000 to $1,010,000 barely registers. We value the utility of money, not the money itself, and utility curves flatten as wealth grows. This single insight launched formal decision theory and set up the central question the field has been wrestling with ever since: what does it mean to decide well, and why do humans so reliably deviate from whatever answer the math provides?

Expected Value: The Foundation That Cracks

Let's start where every decision textbook starts: expected value. The idea is elemental. For any choice you face, list the possible outcomes, estimate their probabilities, estimate their payoffs, multiply each probability by its payoff, and add them up. The option with the highest expected value is the "rational" choice. Expected value is the backbone of insurance pricing, casino mathematics, and basic cost-benefit analysis. If you're betting on a fair coin, and heads pays $100 while tails costs $50, the expected value is positive: (0.5 × $100) + (0.5 × −$50) = $25. Take the bet.

Expected value works beautifully when three conditions hold: the probabilities are known, the payoffs scale linearly with your welfare, and you're making the decision many times—enough for the law of large numbers to smooth out bad luck. A casino can rely on expected value because it plays thousands of hands every night. The math that makes casinos profitable also makes EV-reasoning dangerous for the individual gambler sitting at the table, because she is not playing thousands of hands. She is playing one session, with finite money, and the variance can destroy her long before the expected value materializes.

This is the first crack: expected value ignores what you can afford to lose. A bet with positive expected value but a 10% chance of wiping out your savings is not a good bet, no matter what the math says. The expected value framework treats ruin as just another outcome to multiply by its probability. But ruin is different from a bad quarter. Ruin is absorbing—once you're bankrupt, you can't play again. The bet that maximizes expected value can simultaneously maximize the probability of ruin, and pretending those are the same thing has destroyed more fortunes than any other mathematical error in history.

Expected Utility: A Better Curve

Bernoulli's fix—later formalized by John von Neumann and Oskar Morgenstern in their 1944 masterpiece Theory of Games and Economic Behavior—was to replace value with utility. Instead of maximizing expected dollars, maximize expected utility, where utility is a concave function of wealth. "Concave" means the curve bends downward as wealth increases: each additional dollar adds less utility than the last. If your utility function is the logarithm of wealth (Bernoulli's original proposal), then the utility of $1,000 is much more than half the utility of $2,000—and the St. Petersburg paradox dissolves, because the infinite sum of expected values becomes a finite sum of expected utilities.

Expected utility theory (EU) gave economics a unified framework for understanding risk preferences. Risk aversion isn't irrational—it's the natural consequence of having a concave utility function, which in turn reflects the basic reality that the poorer you are, the more each dollar matters. A millionaire and a minimum-wage worker face the same coin flip, but the stakes are structurally different because their positions on the utility curve are different. EU elegantly captures this. It dominated economics, finance, and decision science for two centuries. It earned multiple Nobel Prizes. And then Daniel Kahneman and Amos Tversky broke it.

Prospect Theory: The Map of Human Irrationality

In a series of experiments through the 1970s, Kahneman and Tversky documented systematic patterns in how people actually make decisions under risk—patterns that no utility function, however creatively drawn, could explain. Their 1979 paper "Prospect Theory: An Analysis of Decision under Risk" is one of the most cited papers in the history of economics, and it redrew the map of human decision-making.

Three findings form the core. The first is reference dependence. People don't evaluate outcomes in terms of final wealth—they evaluate them as gains or losses relative to a reference point, usually wherever they currently stand. A salary of $80,000 feels wonderful if you earned $60,000 last year and terrible if you earned $100,000. Same wealth. Different reference point. Different experience. This seems obvious once stated, but it violates a foundational assumption of expected utility theory: that only final outcomes matter, not the path that got you there.

The second is loss aversion—the most robust finding in behavioral economics. Losses hurt roughly twice as much as equivalent gains please. Kahneman and Tversky estimated a loss aversion coefficient of approximately 2.0 to 2.5: the pain of losing $100 is psychologically equivalent to the pleasure of gaining $200 to $250. This is not risk aversion, which EU already explains. This is a separate asymmetry in how gains and losses are experienced. Loss aversion explains why people hold losing stocks too long (selling would crystallize the loss), why they reject favorable gambles that carry any chance of loss, why they anchor to the price they paid for something rather than its current market value, and why negotiations so often stall—both sides frame concessions as losses, which feel disproportionately painful.

The third is the S-shaped value function. For gains, the curve is concave—diminishing sensitivity, just as Bernoulli described. But for losses, the curve is convex: people become risk-seeking when all options involve losses. If you must choose between a certain loss of $500 and a 50/50 chance of losing $1,000 or nothing, most people gamble—even though the expected values are identical. The desire to avoid the certain pain of a definite loss drives people toward risk. This is why gamblers double down, why failing companies throw good money after bad, and why countries escalate wars they're losing rather than cut their losses. The value function is steeper for losses than gains and curved in opposite directions on either side of the reference point—an S-shape that generates a distinctive fourfold pattern of risk attitudes: risk-averse for likely gains, risk-seeking for unlikely gains (lottery tickets), risk-seeking for likely losses (doubling down), and risk-averse for unlikely losses (insurance against rare catastrophes).

On top of all this, Kahneman and Tversky showed that people don't process probabilities linearly. They overweight small probabilities (treating a 1% chance as if it were 5%) and underweight large ones (treating a 95% chance as if it were 80%). This probability distortion, layered onto the S-shaped value function and loss aversion, explains an astonishing range of real-world behavior that classical theory labels "irrational"—from the simultaneous purchase of lottery tickets and insurance to the equity premium puzzle (why stocks have historically returned so much more than bonds, given their actual risk profile).

When You Don't Even Know the Odds

Prospect theory improved dramatically on expected utility, but both frameworks assume you know the probabilities. In most of life's consequential decisions—career choices, business ventures, geopolitical strategies, whom to trust—you don't. The economist Frank Knight drew the crucial distinction in 1921: risk is uncertainty with known probability distributions (rolling dice, drawing from a shuffled deck). Uncertainty is uncertainty without them (launching a startup, predicting a war, choosing a life partner). Most of what matters is uncertainty, not risk.

In 1961, Daniel Ellsberg—yes, the same Daniel Ellsberg who would later leak the Pentagon Papers—published an experiment that revealed something striking about how humans respond to this distinction. Imagine two urns. Urn A contains 50 red balls and 50 black. Urn B contains 100 balls, some red and some black, in unknown proportions. You can bet on drawing a red ball from either urn. Most people choose Urn A. The known 50/50 risk feels safer than the unknown proportions of Urn B, even though there is no rational basis for assuming Urn B is worse—for all you know, it could be 99 red and 1 black.

This is ambiguity aversion: a preference for known risks over unknown ones that is independent of risk aversion. It cannot be explained by any utility function because it violates the Sure-Thing Principle—one of the foundational axioms of expected utility theory. Ellsberg showed that the axioms of rational choice, as formalized by Leonard Savage in 1954, are axioms that humans systematically and predictably violate when facing genuine uncertainty. Whether this makes humans irrational or reveals the axioms as inadequate descriptions of rationality depends on your philosophical commitments. What it demonstrates practically is that in domains of genuine uncertainty—entrepreneurship, strategy, novel technology, personal relationships—the entire apparatus of probability-weighted optimization hits a wall. You can't optimize what you can't quantify.

When Simple Rules Win

If optimization requires inputs that don't exist, what should you do instead? The German psychologist Gerd Gigerenzer spent three decades answering this question, and his answer is counterintuitive: use less information, not more. Gigerenzer and his Adaptive Behavior and Cognition (ABC) Research Group developed a research program around "fast and frugal heuristics"—simple decision rules that use minimal information and zero computation, yet match or outperform sophisticated optimization in uncertain environments.

Consider the recognition heuristic: if you recognize one option but not the other, infer the recognized option ranks higher on the criterion you care about. When American students were asked which of two German cities had a larger population—one they'd heard of and one they hadn't—they guessed correctly more often than German students who recognized both cities and tried to reason through multiple demographic cues. Less knowledge, filtered through a simple rule, outperformed more knowledge processed through complex reasoning. The additional information available to the German students introduced noise without proportional signal.

Or consider Herbert Simon's concept of satisficing, coined in 1956. Instead of evaluating all options to find the best one (optimizing), set an aspiration level and choose the first option that exceeds it. This sounds lazy. It is, in many environments, optimal. The mathematical "secretary problem"—a formalization of optimal stopping—proves that even with perfect information, the best strategy for sequential choices involves evaluating roughly 37% of candidates, setting your threshold at the best one seen so far, and then selecting the next candidate who exceeds that threshold. You don't evaluate everyone. You can't afford to. The cost of continued search eventually exceeds the expected improvement.

Perhaps the most striking result is the 1/N rule: when allocating resources across N options, simply divide equally. In a rigorous 2009 study, DeMiguel, Garlappi, and Uppal tested 1/N against fourteen optimized portfolio allocation strategies—including Harry Markowitz's Nobel Prize-winning mean-variance optimization—across multiple datasets. The naive 1/N rule outperformed the majority of them. The reason cuts to the heart of the optimization-versus-heuristics debate: optimal strategies require estimated inputs (expected returns, covariances), and estimation error in those inputs swamps the theoretical gains from optimization. The simple rule, by ignoring these estimates entirely, avoids the noise they introduce.

Gigerenzer's core principle—ecological rationality—states that the quality of a decision strategy depends on the match between the strategy and the environment. Complex strategies excel in well-defined, information-rich environments with stable statistical structure. Simple heuristics excel in uncertain, information-poor environments where overfitting is a greater danger than underfitting. The practical implication: don't use the same decision process for everything. Match the tool to the terrain.

The Pre-Mortem: Imagining Failure Before It Happens

The psychologist Gary Klein, who spent his career studying how experts make decisions in high-stakes, time-pressured environments—firefighters, military commanders, intensive-care nurses—developed a technique in 1998 that is, by a significant margin, the highest-impact practical intervention in the decision science literature. It's called the pre-mortem, and it works like this: before implementing a decision, the team imagines that it is one year in the future and the decision has failed spectacularly. Then, working individually and silently, each member writes down all the reasons the decision might have failed.

The power of this technique comes from a cognitive phenomenon called prospective hindsight: people generate approximately 30% more explanations for an outcome when told it has already occurred than when asked to predict whether it will occur. The pre-mortem exploits this asymmetry to surface risks and failure modes that standard risk assessment misses. It also circumvents two of the most dangerous failure modes in group decision-making: groupthink—where the desire for consensus suppresses dissent—and overconfidence—where the team systematically underestimates the probability of failure. By framing failure as a given and asking "why," rather than asking "could this fail?" (to which the answer is always an underconfident "probably not"), the pre-mortem grants psychological permission to voice concerns that would otherwise feel disloyal or pessimistic.

Paired with the pre-mortem is the reversibility heuristic. Jeff Bezos popularized the distinction between Type 1 decisions (irreversible, high-consequence—walking through a one-way door) and Type 2 decisions (reversible, low-consequence—walking through a two-way door). The practical rule: make Type 2 decisions fast, with minimal analysis, because the cost of being wrong is the cost of reversing course, which is low. Reserve extensive deliberation, pre-mortems, and scenario analysis for Type 1 decisions, where you can't undo the outcome. Most organizations systematically over-deliberate on Type 2 decisions (slowing everything down) and under-deliberate on Type 1 decisions (because the urgency that attaches to big decisions compresses the timeline). The correction is simple: classify first, then calibrate effort.

Decision Hygiene: The Invisible Error

In 2021, Daniel Kahneman, Olivier Sibony, and Cass Sunstein published Noise: A Flaw in Human Judgment, identifying a category of decision error that had been hiding in plain sight. Everyone knows about bias—systematic, directional error. If a judge sentences Black defendants more harshly than white defendants for the same crime, that's bias. But what if different judges sentence the same defendant, with the same facts, to wildly different terms? That's noise—unwanted variability in judgments that should be identical. And the research suggests that in most organizational decision-making, noise is at least as large a source of error as bias.

Studies have found that insurance underwriters assessing identical cases produced estimates that differed by an average of 55%. Forensic fingerprint examiners shown the same prints on separate occasions sometimes changed their conclusions. Judges' sentencing decisions correlated with irrelevant factors—the outdoor temperature, whether the local sports team won the night before, how close it was to lunch. None of this is bias in the traditional sense. It's scatter. It's the same judge making different decisions on the same case depending on when the case hits her desk.

Kahneman, Sibony, and Sunstein proposed decision hygiene—protocols designed to reduce noise without needing to diagnose specific biases. The key techniques: have multiple evaluators assess independently before any discussion, then aggregate their judgments (structurally equivalent to the wisdom-of-crowds effect). Decompose complex judgments into independent components—evaluate each factor separately, score them on predefined scales, then combine the scores. Use reference class forecasting: before analyzing the specifics of your situation, anchor on the base rate. What fraction of startups like yours succeed? What fraction of software projects finish on time? What fraction of mergers achieve their projected synergies? The outside view—the base rate from the reference class—is almost always more accurate than the inside view—your detailed narrative about why this case is special.

Decision hygiene is not glamorous. It doesn't feel like insight. But the evidence suggests it produces larger improvements in decision quality than almost any amount of individual debiasing, because it addresses the variability that individuals can't see in themselves.

The Meta-Decision: Knowing When to Stop Thinking

Every decision framework generates an ironic recursion: how do you decide how much time to spend deciding? Spend too little, and you miss important considerations. Spend too much, and you waste time that could be spent acting, suffer analysis paralysis, and may actually make things worse—research shows that after a certain point, additional information introduces more noise than signal.

The practical resolution is itself a heuristic. The value of additional deliberation equals the probability that more thinking will change your choice, multiplied by the difference in value between the option you'd choose with more analysis and the option you'd choose now. For most decisions, this product is small. Either the options are close enough in value that choosing "wrong" barely matters, or your intuition is already well-calibrated enough that analysis rarely changes the outcome. The decisions that warrant deep analysis are those that are irreversible, high-stakes, and where your intuition is poorly calibrated—which is to say, situations where the terrain is unfamiliar and the consequences are lasting.

Herbert Simon, who coined "satisficing" and won the Nobel Prize for his work on bounded rationality, put it best: the goal of decision-making isn't to be optimal. It's to be good enough, fast enough, with resources that are always finite. A theory of decision-making that ignores the cost of decision-making is like a theory of transportation that ignores the cost of fuel. The meta-decision—how much to analyze—is itself the most important decision in the sequence, because it determines how you'll spend the only truly non-renewable resource: your time and attention.

How This Was Decoded

This analysis synthesized three traditions that are too rarely integrated. The normative tradition—from Daniel Bernoulli through von Neumann and Morgenstern to Leonard Savage—defines what a perfectly rational agent would do and provides the mathematical backbone. The behavioral tradition—Daniel Kahneman and Amos Tversky's prospect theory, the heuristics-and-biases research program, and the noise framework developed with Olivier Sibony and Cass Sunstein—maps where real humans systematically deviate. The ecological rationality tradition—Gerd Gigerenzer's fast-and-frugal heuristics, Herbert Simon's satisficing, and the broader research on when simplicity outperforms optimization—provides the crucial corrective that deviation from classical rationality is often adaptive rather than defective. Frank Knight's risk-uncertainty distinction and Daniel Ellsberg's ambiguity aversion experiments frame the boundaries where formal optimization breaks down. Gary Klein's pre-mortem technique and the reversibility heuristic represent the highest-impact practical interventions from the applied decision science literature. The core decoded insight: there is no single "rational" way to decide. Rationality is ecological—it depends on the match between your strategy and your environment. The goal is not to be a perfect calculator. It is to be a good calibrator: knowing when to calculate, when to satisfice, when to deliberate deeply, and when to act.

Want the compressed, high-density version? Read the agent/research version →

You're reading the human-friendly version Switch to Agent/Research Version →