Gsd-skill-creator behavioral-economics

Systematic departures from rational choice theory and their implications for economic analysis and policy. Covers cognitive heuristics (anchoring, availability, representativeness), biases (loss aversion, status quo, overconfidence), prospect theory (reference dependence, probability weighting, diminishing sensitivity), nudge theory and choice architecture, and the integration of psychological findings into economic models. Use when analyzing decision-making under uncertainty, evaluating policy interventions that exploit behavioral patterns, or assessing where standard rational-agent models break down.

install
source · Clone the upstream repo
git clone https://github.com/Tibsfox/gsd-skill-creator
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/Tibsfox/gsd-skill-creator "$T" && mkdir -p ~/.claude/skills && cp -r "$T/examples/skills/economics/behavioral-economics" ~/.claude/skills/tibsfox-gsd-skill-creator-behavioral-economics && rm -rf "$T"
manifest: examples/skills/economics/behavioral-economics/SKILL.md
source content

Behavioral Economics

Standard economics assumes agents are rational: they have stable preferences, process information correctly, and maximize expected utility. Behavioral economics documents the systematic ways real humans depart from this ideal and builds models that incorporate these departures. The field was launched by Kahneman and Tversky's work on heuristics and biases in the 1970s and formalized by prospect theory (1979). It is not a rejection of economics but an enrichment -- the rational model is the benchmark from which behavioral findings are measured.

Agent affinity: varian (pedagogical exposition, connecting behavioral findings to standard theory), sen (welfare implications, capability approach as alternative to revealed preference), ostrom (behavioral foundations of cooperation in commons)

Concept IDs: econ-marginal-thinking, econ-trade-offs, econ-opportunity-cost, econ-market-failures

The Behavioral Economics Toolbox at a Glance

#TopicCore questionKey framework
1HeuristicsHow do people actually process information?Anchoring, availability, representativeness
2BiasesWhere do systematic errors arise?Loss aversion, status quo bias, overconfidence
3Prospect theoryHow do people evaluate risky outcomes?Reference dependence, probability weighting
4Intertemporal choiceWhy do people struggle with patience?Hyperbolic discounting, present bias
5Social preferencesAre people purely self-interested?Fairness, reciprocity, altruism
6Nudges and choice architectureCan policy exploit behavioral patterns?Default effects, framing, simplification
7Limits of behavioral economicsWhen does the rational model work fine?Markets, repeated interactions, high stakes

Topic 1 -- Heuristics

What they are. Heuristics are mental shortcuts that reduce complex problems to simpler judgments. They are often accurate and always efficient, but they can produce systematic errors (biases) in specific, predictable circumstances.

Anchoring. People's estimates are influenced by irrelevant starting points. Tversky and Kahneman (1974) asked subjects whether the percentage of African countries in the UN was higher or lower than a randomly generated number, then asked for their best estimate. Subjects anchored heavily on the random number. In economic contexts, anchoring affects negotiations (the first offer sets the anchor), pricing (suggested retail price), and financial forecasts.

Availability. People judge the probability of events by how easily examples come to mind. Plane crashes are vivid and memorable, so people overestimate aviation risk relative to driving risk -- even though driving is far more dangerous per mile. Availability distorts risk assessment, insurance demand, and policy priorities (resources flow to dramatic risks rather than statistical ones).

Representativeness. People judge probabilities by similarity to a prototype rather than by base rates. The "Linda problem" (Tversky and Kahneman, 1983): Linda is a former philosophy student concerned with social justice. Is she more likely to be (a) a bank teller, or (b) a bank teller and active in the feminist movement? Most people choose (b), which is a conjunction fallacy -- the conjunction of two events cannot be more probable than either event alone. In economic contexts, representativeness drives stereotyping in hiring, overreaction to salient narratives in financial markets, and the "hot hand" fallacy.

Worked heuristic example. An investor reads that a tech startup has tripled revenue for two consecutive years. Using representativeness, the investor judges the company as "the next Google" because the growth pattern matches the prototype of a breakout tech company. Base rate neglected: 90% of startups with early rapid growth fail within five years. The investor overweights the salient narrative and underweights the base rate, overpaying for the stock. This is a general pattern in financial markets -- stories move prices more than statistics, especially for individual investors. Professional investors are not immune but markets partially correct through arbitrage.

Base rate neglect. The general form of representativeness errors. A medical test with 99% sensitivity and 1% false positive rate sounds highly reliable. But if the disease prevalence (base rate) is 1 in 1,000, a positive result is more likely to be a false positive than a true positive: P(disease | positive) = 0.99 * 0.001 / (0.99 * 0.001 + 0.01 * 0.999) = approximately 9%. Doctors, judges, and jurors routinely make this error when evaluating diagnostic and forensic evidence. In economic contexts, base rate neglect leads to overconfident credit assessments, insurance mispricing, and investment mistakes.

The affect heuristic. People make judgments based on emotional reactions rather than deliberate analysis. If a technology feels exciting (AI, blockchain), people overestimate its benefits and underestimate its risks. If a technology feels scary (nuclear power, genetic modification), the opposite occurs. The affect heuristic explains asymmetric risk regulation: nuclear power is regulated far more stringently than coal power per unit of mortality risk because nuclear feels scarier.

Topic 2 -- Biases

Loss aversion. Losses loom larger than equivalent gains. Kahneman and Tversky estimated the loss aversion coefficient at approximately 2: losing $100 feels about twice as bad as gaining $100 feels good. This explains the endowment effect (people demand more to sell an object than they would pay to buy it), the disposition effect in finance (investors sell winners too early and hold losers too long), and resistance to policy changes (the prospect of losing existing benefits outweighs the prospect of gaining new ones).

Status quo bias. People disproportionately stick with the current state of affairs. This is partly loss aversion (change involves potential losses), partly the effort cost of switching, and partly a heuristic that the status quo has survived this long so it must be acceptable. Status quo bias explains low rates of organ donation in opt-in systems, inertia in retirement savings, and resistance to institutional reform.

Overconfidence. People systematically overestimate their knowledge and abilities. Calibration studies show that when people say they are 90% confident, they are right about 70-80% of the time. Overconfidence drives excess trading in financial markets (Barber and Odean, 2001), overoptimistic business plans, and insufficient precautionary behavior. The planning fallacy -- systematic underestimation of how long tasks will take -- is a special case.

Sunk cost fallacy. Rational agents ignore sunk costs -- what is spent is spent and should not affect future decisions. But people persist in failing endeavors because they have already invested time, money, or effort. "We've come this far" is not a valid economic argument for continuing, but it is a powerful psychological one. The Concorde fallacy (named after the supersonic jet that kept flying despite never being economically viable) illustrates the cost: billions were spent because governments could not bring themselves to write off earlier investment.

Worked bias example. A company has invested $10 million in a software project that is now clearly failing -- the technology is obsolete and a competitor has released a superior product. A rational analysis says: the $10 million is sunk. The question is whether future investment will generate returns exceeding future costs. If not, abandon the project. But managers feel the $10 million as a loss and continue investing to "justify" the sunk cost, eventually losing $20 million instead of $10 million. The bias is strongest when the same person who authorized the initial investment makes the continuation decision -- they are "throwing good money after bad" to avoid admitting an error. De-biasing strategies: separate the continuation decision from the initial decision (have a different person evaluate), use pre-commitment rules ("we will evaluate this project at month 6 with a fresh cost-benefit analysis"), and normalize failure ("killing a bad project is a success, not a failure").

Confirmation bias. People seek information that confirms existing beliefs and avoid information that contradicts them. In economic contexts, investors seek news that validates their holdings, managers seek data that supports their strategies, and policymakers seek evidence that justifies their positions. Confirmation bias is especially dangerous in financial markets, where it can sustain bubbles: investors who bought at high prices seek bullish analysis that confirms their purchase was wise, creating demand for optimistic forecasts.

Topic 3 -- Prospect Theory

The framework. Prospect theory (Kahneman and Tversky, 1979) is the most influential alternative to expected utility theory. It has three components:

  1. Reference dependence. People evaluate outcomes as gains or losses relative to a reference point, not as final wealth levels. The reference point is typically the status quo but can be an expectation, an aspiration, or a social comparison.

  2. Diminishing sensitivity. The value function is concave for gains and convex for losses -- the difference between $0 and $100 feels larger than the difference between $1,000 and $1,100. This produces risk aversion for gains and risk seeking for losses.

  3. Probability weighting. People overweight small probabilities and underweight large ones. This explains both lottery purchases (overweighting the tiny chance of a large gain) and insurance purchases (overweighting the small chance of a large loss) -- behaviors that are inconsistent under expected utility theory.

The value function. The value function is S-shaped: concave above the reference point (gains), convex below (losses), and steeper for losses than gains (loss aversion). This single function explains a wide range of anomalies: the endowment effect, risk-seeking in the domain of losses, the equity premium puzzle, and asymmetric reactions to good and bad news.

The isolation effect. People simplify choices by ignoring components that alternatives share. This can produce inconsistencies: the same final outcome is evaluated differently depending on how the problem is framed (whether identical components are included or stripped out).

Topic 4 -- Intertemporal Choice

Hyperbolic discounting. Standard economics assumes exponential discounting: a person who prefers $100 today over $110 tomorrow should also prefer $100 in 30 days over $110 in 31 days. But people often reverse their preference as the earlier date approaches -- they are patient when both options are in the future but impatient when the earlier option is now. This "present bias" is captured by hyperbolic (or quasi-hyperbolic) discount functions.

Self-control problems. Present bias creates a gap between what people plan to do and what they actually do. People plan to save, diet, and exercise -- then do not follow through. Sophisticated agents recognize their future self-control problem and seek commitment devices (automatic payroll deductions, gym contracts with penalties, deadline commitments). Naive agents do not anticipate the problem and are perpetually surprised by their own behavior.

Thaler's mental accounting. People compartmentalize money into mental accounts (rent money, vacation money, emergency fund) and apply different decision rules to each. This violates fungibility -- a dollar is a dollar regardless of which mental account it comes from -- but it serves as a self-control mechanism. Blowing the vacation budget feels different from raiding the emergency fund, even if the economic consequence is identical.

Worked intertemporal example. A worker is offered two retirement savings plans. Plan A: opt-in, 3% contribution rate, immediate paycheck reduction of $150/month. Plan B: automatic enrollment at 3%, must opt out to stop contributing. In Plan A, the loss of $150/month is salient and immediate; present bias makes the worker delay enrollment indefinitely ("I'll start next month"). In Plan B, the worker must actively choose to lose the employer match and future retirement income. Inertia and loss aversion keep the worker enrolled. The result: Plan B achieves 90%+ participation; Plan A achieves 40-60%. The economic content is identical but the behavioral framing changes the outcome dramatically. This is the empirical basis for Thaler's Save More Tomorrow program, which increased retirement savings rates by billions of dollars.

Beta-delta model. The quasi-hyperbolic discount function formalizes present bias: utility from future consumption at time t is weighted by beta * delta^t, where delta is the standard exponential discount factor and beta < 1 captures the extra discounting of all future periods relative to the present. If beta = 0.7 and delta = 0.99, a person is roughly patient about trade-offs between next year and the year after (both discounted by 0.99), but heavily discounts next month relative to today (by 0.7). The beta parameter captures self-control problems; the delta parameter captures genuine time preference.

Commitment devices. Recognizing their own present bias, sophisticated agents seek ways to constrain their future selves. Odysseus binding himself to the mast is the literary archetype. Modern examples: automatic payroll deductions, gym memberships with cancellation fees, software that blocks social media during work hours, and Stickk.com (commitment contracts with financial penalties for failure). The demand for commitment devices is itself evidence of present bias -- a perfectly patient person would not need them.

Topic 5 -- Social Preferences

Beyond self-interest. Standard economics assumes agents maximize their own payoff. Experimental evidence -- ultimatum games, dictator games, public goods games, trust games -- shows that people also care about fairness, reciprocity, and the outcomes of others.

Fairness. In ultimatum games, responders routinely reject offers they consider unfair (typically below 20-30% of the total), even though rejection means getting nothing. Proposers anticipate this and offer 40-50% on average. This cannot be explained by self-interest alone.

Reciprocity. People reward kind behavior and punish unkind behavior, even at personal cost. In trust games, investors send money to trustees, who often (but not always) return a fair share. In public goods games with punishment, cooperation rates rise dramatically when free riders can be sanctioned.

Ostrom's insight. The behavioral foundations of cooperation are critical for understanding common pool resource governance. People are not unconditionally cooperative or selfish -- they are conditional cooperators who will contribute as long as they believe others are contributing and free riders are sanctioned. Ostrom's institutional design principles create the conditions for conditional cooperation to sustain itself.

Topic 6 -- Nudges and Choice Architecture

The concept. Thaler and Sunstein (2008) defined a nudge as "any aspect of the choice architecture that alters people's behavior in a predictable way without forbidding any options or significantly changing their economic incentives." Nudges exploit behavioral patterns rather than fighting them.

Default effects. The most powerful nudge. Organ donation rates are 90%+ in opt-out countries and 15-30% in opt-in countries (Johnson and Goldstein, 2003). Automatic enrollment in retirement savings dramatically increases participation. The default works because of status quo bias and effort cost -- most people stick with whatever they were given.

Framing. How a choice is described affects decisions. "90% survival rate" elicits different choices than "10% mortality rate" even though they convey the same information. Positive framing of healthy food options, energy-saving framing of thermostat settings, and gain-framed health messages all exploit this pattern.

Simplification. Complex forms and procedures discourage participation even when people intend to participate. Simplifying tax filing (automatic returns), financial aid applications (pre-populated forms), and health insurance enrollment (fewer plan options) increases uptake. Complexity is a barrier that disproportionately affects low-income and low-education populations.

Libertarian paternalism. Nudges are described as "libertarian" because they preserve freedom of choice (people can always opt out) and "paternalist" because they steer people toward choices the designer believes are better for them. Critics argue that who defines "better" is a political question, not a psychological one, and that nudges can be manipulative when used by firms (dark patterns in web design) rather than benevolent planners.

Worked nudge example. A cafeteria wants to increase healthy food choices. Three nudge interventions, tested experimentally: (1) Place fruit at eye level and move desserts to a less convenient location. Result: fruit consumption up 25%, dessert consumption down 15%. (2) Use smaller plates for the buffet. Result: calorie intake down 20%. (3) Label healthy options with a green "daily pick" sticker. Result: healthy option selection up 30%. None of these interventions restrict choice -- every item remains available at the same price. They work by changing the default (what is easiest to reach), the reference point (smaller plate makes a normal portion look full), and the framing (social validation via the "daily pick" label). Total implementation cost: approximately $500 for signage and plate replacement. This cost-effectiveness is why behavioral insights teams in government and corporate settings have proliferated.

Sludge. Thaler coined the term "sludge" for friction that discourages beneficial action. Requiring complex paperwork to claim a tax rebate is sludge. Making cancellation of subscriptions difficult is sludge. Requiring in-person visits for government services that could be done online is sludge. Sludge disproportionately affects those with the least time, education, and bureaucratic literacy. "Sludge audits" -- systematically identifying and removing unnecessary friction -- are a low-cost, high-impact policy intervention.

Topic 7 -- Limits of Behavioral Economics

When rationality works. Markets with repeated interactions, competition, and clear feedback tend to discipline behavioral biases. Professional traders may be overconfident but markets correct persistent mispricing over time. Firms that ignore sunk costs outperform those that don't. The rational model is a better description of equilibrium outcomes in competitive markets than of individual decision-making in laboratory settings.

External validity concerns. Many behavioral findings come from laboratory experiments with small stakes and student subjects. Whether they scale to real-world decisions with large stakes and experienced actors is an ongoing debate. Some findings replicate robustly (loss aversion, default effects); others are fragile (ego depletion, some social priming effects).

The paternalism problem. If people's choices do not reflect their "true" preferences, whose preferences count? Behavioral economics provides tools for improving decision-making but does not resolve the philosophical question of what constitutes welfare when revealed preferences are unreliable.

Integration, not replacement. Behavioral economics enriches the rational model rather than replacing it. The productive approach is: use the rational model as the baseline, identify specific contexts where behavioral factors dominate, and design institutions that account for predictable irrationality without abandoning the analytical power of optimization.

Behavioral finance. Financial markets provide a rich testing ground. The equity premium puzzle (stocks return far more than bonds, more than risk aversion alone can explain) is consistent with loss aversion -- investors demand a large premium to bear the risk of losses. The disposition effect (selling winners too early, holding losers too long) is consistent with prospect theory's reference-dependent value function. Calendar effects, momentum, and the volatility smile all challenge efficient market theory. But markets also correct many behavioral biases through arbitrage -- the question is which biases survive market discipline and which do not.

Behavioral development economics. Present bias, scarcity mindset, and limited attention are especially consequential in developing countries. Mullainathan and Shafir (2013) argued that scarcity itself imposes a "bandwidth tax" -- the mental effort of managing scarcity reduces cognitive capacity for other decisions. Farmers making planting decisions during harvest season (when they have resources) make better choices than during the lean season (when they are preoccupied with food shortfalls). This suggests that the timing and framing of development interventions matters as much as their content.

Decision Heuristics

When approaching a behavioral economics problem:

  1. What does the rational model predict? Start with the standard model. The behavioral finding is meaningful only relative to the rational benchmark.
  2. What is the specific departure? Identify the heuristic, bias, or preference anomaly. Be precise -- "behavioral" is not an explanation.
  3. Is the departure robust? Does it replicate? Does it survive high stakes, experience, and market discipline? Some lab findings do not generalize.
  4. Does it change the policy conclusion? A bias that exists but does not affect the policy question is interesting psychology but not actionable economics.
  5. Can the choice architecture be improved? If so, design the nudge. Specify the default, framing, or simplification.
  6. Who benefits from the nudge? Nudges designed by firms may exploit rather than help (dark patterns). Evaluate the designer's incentives.

Common Mistakes

MistakeWhy it failsFix
Calling any suboptimal behavior "irrational"Some behavior is rational given constraintsCheck whether the behavior is consistent with a model that includes realistic constraints
Generalizing from lab to fieldSmall-stakes student experiments may not replicate at scaleCheck for field evidence and high-stakes replications
Treating behavioral economics as anti-marketBehavioral findings enrich rather than replace market analysisUse the rational model as the baseline and adjust where evidence warrants
Ignoring the paternalism question"Better" choices depend on whose welfare function you useBe explicit about whose preferences the nudge serves
Assuming nudges are costlessChoice architecture has design costs, monitoring costs, and political costsInclude implementation costs in the analysis
Overclaiming replication failuresSome highly cited findings have not replicated; others haveDistinguish between fragile and robust findings

Cross-References

  • varian agent: Connects behavioral findings to standard microeconomic theory. Primary agent for explaining where rational and behavioral models diverge.
  • sen agent: Welfare implications of behavioral findings. The capability approach as an alternative welfare criterion when preferences are unreliable.
  • ostrom agent: Behavioral foundations of cooperation, conditional cooperation in commons.
  • microeconomics skill: The rational-agent benchmark from which behavioral departures are measured.
  • public-policy skill: Nudge theory, choice architecture, behavioral public policy.
  • development-economics skill: Behavioral barriers to saving, investment, and health in developing countries.
  • international-trade skill: Behavioral explanations for protectionist sentiment -- loss aversion applied to job losses from trade.
  • macroeconomics skill: Animal spirits and their role in the business cycle; behavioral foundations of sticky prices and wage rigidity.
  • smith agent: Behavioral departures from the invisible hand -- when self-interest does not lead to socially optimal outcomes due to systematic cognitive errors rather than externalities or market power.

Historical Context

Behavioral economics has roots in the 18th century -- Adam Smith's Theory of Moral Sentiments (1759) analyzed sympathy, self-deception, and the psychology of moral judgment. But the field's modern form began with Herbert Simon's concept of "bounded rationality" (1955): agents do not optimize; they "satisfice" -- searching for options that are good enough rather than best. Simon won the Nobel in 1978 for this insight.

Kahneman and Tversky's heuristics and biases program (1970s) provided the experimental evidence and formal framework. Their prospect theory (1979) offered a rigorous alternative to expected utility theory. Thaler connected these psychological findings to economic phenomena throughout the 1980s and 1990s -- the endowment effect, mental accounting, the equity premium puzzle, the winner's curse. He won the Nobel in 2017.

The policy turn came with Nudge (Thaler and Sunstein, 2008), which demonstrated that behavioral insights could improve policy without restricting choice. Behavioral Insights Teams ("nudge units") were established in the UK (2010), US (2014), and dozens of other countries. The field has matured from a collection of anomalies into an integrated research program with theoretical models (prospect theory, quasi-hyperbolic discounting, social preference models), empirical methods (lab experiments, field experiments, neuroimaging), and policy applications (default design, disclosure simplification, commitment devices).

Current debates center on the replication crisis (which findings are robust?), the limits of nudging (can nudges address structural problems like inequality and climate change?), and the ethics of choice architecture (who decides what counts as a "better" choice?). The field's greatest contribution may be methodological: it has permanently raised the evidentiary standard for claims about human behavior in economics.

Study Path

Beginner. Kahneman, Thinking, Fast and Slow (2011) -- the definitive popular introduction by one of the field's founders. Covers heuristics, biases, and prospect theory in accessible language with vivid examples. Then Thaler, Misbehaving (2015) -- the history of behavioral economics told as a personal narrative. Both are page-turners that require no prior economics.

Intermediate. Thaler and Sunstein, Nudge (2008) -- policy applications. Then Ariely, Predictably Irrational (2008) -- more experimental findings with accessible explanations. For the formal framework, read the original Kahneman and Tversky (1979) prospect theory paper -- it is surprisingly readable.

Advanced. Camerer, Loewenstein, and Rabin, Advances in Behavioral Economics (2004) -- the field's collected papers with editorial commentary. DellaVigna (2009) for a comprehensive survey of field evidence. For intertemporal choice, Laibson (1997) on hyperbolic discounting. For social preferences, Fehr and Schmidt (1999) on fairness.

Graduate. Read the primary literature: Kahneman and Tversky's collected works, Rabin's theoretical contributions (1998, 2000), and the experimental economics tradition (Smith, Plott, Kagel and Roth). The Handbook of Behavioral Economics (Bernheim, DellaVigna, and Laibson, eds.) is the graduate reference.

References

  • Kahneman, D., & Tversky, A. (1979). "Prospect Theory: An Analysis of Decision under Risk." Econometrica, 47(2), 263-292.
  • Thaler, R. H., & Sunstein, C. R. (2008). Nudge: Improving Decisions about Health, Wealth, and Happiness. Yale University Press.
  • Kahneman, D. (2011). Thinking, Fast and Slow. Farrar, Straus and Giroux.
  • Thaler, R. H. (2015). Misbehaving: The Making of Behavioral Economics. W. W. Norton.
  • Camerer, C. F., Loewenstein, G., & Rabin, M. (2004). Advances in Behavioral Economics. Princeton University Press.
  • Tversky, A., & Kahneman, D. (1974). "Judgment under Uncertainty: Heuristics and Biases." Science, 185(4157), 1124-1131.
  • Mullainathan, S., & Shafir, E. (2013). Scarcity: Why Having Too Little Means So Much. Times Books.
  • Simon, H. A. (1955). "A Behavioral Model of Rational Choice." Quarterly Journal of Economics, 69(1), 99-118.
  • Barber, B. M., & Odean, T. (2001). "Boys Will Be Boys: Gender, Overconfidence, and Common Stock Investment." Quarterly Journal of Economics, 116(1), 261-292.
  • Laibson, D. (1997). "Golden Eggs and Hyperbolic Discounting." Quarterly Journal of Economics, 112(2), 443-477.
  • Johnson, E. J., & Goldstein, D. (2003). "Do Defaults Save Lives?" Science, 302(5649), 1338-1339.
  • Fehr, E., & Schmidt, K. M. (1999). "A Theory of Fairness, Competition, and Cooperation." Quarterly Journal of Economics, 114(3), 817-868.
  • Rabin, M. (1998). "Psychology and Economics." Journal of Economic Literature, 36(1), 11-46.
  • DellaVigna, S. (2009). "Psychology and Economics: Evidence from the Field." Journal of Economic Literature, 47(2), 315-372.