Case Study 2: Welfare Cliffs and Bug Bounties -- The Cobra Effect in Systems That Help
"The system isn't broken. It's fixed." -- Common saying about perverse institutional incentives
Two Domains, One Trap
This case study examines the cobra effect in two domains where the incentive was explicitly designed to help: social welfare programs that aim to reduce poverty, and bug bounty programs that aim to improve software security. Both domains are populated by well-intentioned designers who genuinely want to solve the problem. Both domains produce cobra effects that actively undermine their own goals. And both domains illustrate a variant of the cobra effect that is particularly difficult to address: the cobra that emerges not from malice or greed but from the rational behavior of the very people the system is trying to help.
Part I: The Welfare Cliff -- A Machine for Manufacturing Poverty Traps
The Architecture of the Cliff
The welfare cliff is not an accident. It is the mathematical consequence of a design choice that seems entirely reasonable: allocating scarce public resources to the people who need them most. Means-testing -- conditioning benefits on income -- is the mechanism that implements this principle. If you earn below the threshold, you receive support. If you earn above it, the support is reduced or eliminated. The logic is unimpeachable: public money should go to the poor, not the rich.
The cobra emerges from the transition zone.
Consider the specific case of a single mother with two children in a typical U.S. state. At an earned income of $20,000 per year, she may qualify for the following programs:
- SNAP (food assistance): approximately $6,000 per year
- Housing voucher: approximately $8,000 per year in rental subsidy
- Medicaid: healthcare coverage valued at approximately $8,000 per year
- CCDF (childcare subsidy): approximately $7,000 per year
- EITC (earned income tax credit): approximately $5,000 per year
- LIHEAP (heating assistance): approximately $1,000 per year
Her total resources: $20,000 in wages plus approximately $35,000 in benefits, for an effective income of approximately $55,000.
Now consider what happens as her earned income rises:
At **$25,000**, she begins to lose SNAP benefits and some EITC value. Total resources: approximately $54,000.
At **$30,000**, she hits the Medicaid cliff. She loses healthcare coverage. She now needs to purchase health insurance on the exchange, costing approximately $4,000 per year after subsidies. She also loses her childcare subsidy. Total resources: approximately $43,000 -- she is earning $10,000 more in wages and has $12,000 less in total resources.
At **$35,000**, she loses her housing voucher. Her rent increases by $8,000 per year. Total resources: approximately $37,000.
At **$40,000**, she has lost nearly all benefits. Total resources: approximately $42,000.
At **$55,000**, her earned income finally equals what her total resources were at $20,000 in wages. She has had to nearly triple her earnings to achieve the same standard of living she had at the bottom.
The mathematics define a poverty trap: a range of income levels where rational economic behavior is to stay put. The cliff is not a single precipice but a series of them, each removing a different support, each punishing the transition toward self-sufficiency.
The Human Cost
The poverty trap is not merely a mathematical abstraction. It shapes real decisions in real lives.
A home health aide earning $24,000 per year is offered a supervisory position paying $32,000. She calculates the effect on her benefits. The raise would cost her family $11,000 in lost SNAP, Medicaid, and childcare subsidies. The net change: she would work harder, in a more stressful role, for $3,000 less in total resources. She declines the promotion.
A single father working part-time earns $18,000 per year and receives full benefits. He could work full-time and earn $36,000, but the additional $18,000 in wages would cost him $22,000 in benefits. He continues working part-time.
A couple with a disabled child receives $42,000 in combined benefits and services. One parent is offered a job paying $50,000. Accepting the job would push the family above the income threshold for the child's disability services, which are worth $25,000 per year. The job would cost the family $17,000 in net resources. They decline.
In every case, the individual is making a rational decision given the incentive structure. The cobra is the system, not the person. The system designed to lift people out of poverty is financially punishing them for trying to leave it.
Why the Cliff Persists
If the welfare cliff is so clearly counterproductive, why does it persist? Several structural factors conspire to maintain it:
Administrative simplicity. Sharp eligibility thresholds are easy to administer. A family either qualifies or it does not. Gradual phase-outs -- where benefits decrease smoothly as income increases -- are more complex to implement, require more frequent income verification, and create more administrative burden. The legibility of the threshold (you qualify / you don't) is preferred over the messiness of the gradient.
Political economy. Programs with clear eligibility rules are easier to defend politically. A sharp threshold makes it easy to say "this program serves families earning below X." A gradual phase-out creates a gray zone where some benefits flow to families who are "not really poor," which is politically vulnerable to accusations of waste.
Program fragmentation. The cliff is not the product of any single program. It is the emergent result of many programs, each designed independently, each with its own eligibility rules, each administered by a different agency. No single designer created the cliff. It assembled itself from the accumulation of individually reasonable decisions. The system-level cobra effect is invisible to any program-level designer because each program sees only its own threshold, not the compound effect of all thresholds combined.
Moral framing. Some policymakers and voters believe that welfare cliffs are features, not bugs -- that benefits should disappear when people earn more, because public assistance should be reserved for the truly needy. This moral framing treats the cliff as deserved rather than perverse, and resists reforms that would extend benefits to higher-earning families, even if such reforms would ultimately reduce welfare dependency by eliminating the poverty trap.
Connection to Chapter 16 (Legibility and Control): The welfare cliff is a legibility project. The threshold makes poverty legible to the state: families are either eligible or not, and their eligibility is determined by a single, measurable variable (income). This legibility comes at the cost of destroying the complexity of the actual situation. Real poverty is not a binary state determined by income alone. It is a multidimensional condition involving health, housing stability, childcare access, education, transportation, social support, and dozens of other factors that a simple income threshold cannot capture. The threshold makes poverty manageable for the bureaucracy. It makes it unmanageable for the family at the cliff's edge.
The Phase-Out Solution and Its Limits
The most commonly proposed reform is to replace cliffs with gradual phase-outs: instead of benefits disappearing at a threshold, they would decrease smoothly as income increases. For every additional dollar earned, the family would lose only a fraction of a dollar in benefits. The effective marginal tax rate would be high but not catastrophic -- perhaps 50 to 70 percent rather than over 100 percent.
Gradual phase-outs would eliminate the cliff. But they would not eliminate the cobra effect entirely. They would transform it.
A gradual phase-out that reduces benefits by 50 cents for every additional dollar earned creates an effective marginal tax rate of 50 percent across the entire phase-out range. For a family whose benefits phase out between $20,000 and $60,000, this means that every dollar earned in that range is worth only 50 cents in additional resources. The incentive to earn more exists, but it is half as strong as it would be without the phase-out.
Moreover, extending the phase-out range means extending benefits to higher-income families, which increases program costs. If benefits phase out gradually from $20,000 to $60,000 rather than disappearing at $25,000, more families receive benefits, and total spending increases. This creates a political problem: the program now serves families that critics will characterize as "not truly needy," making it vulnerable to political attack.
The welfare cobra cannot be fully killed. It can only be tamed -- its cliffs can be smoothed into slopes, its effective tax rates can be reduced, its poverty traps can be weakened. But as long as benefits decrease as income increases (which they must, unless benefits are universal), some perverse incentive will exist at the margin. The question is not whether the cobra exists but how dangerous it is.
Part II: Bug Bounties -- Security's Perverse Incentive Landscape
The Bug Bounty Ecosystem
Software bug bounty programs emerged in the 1990s and have become a standard practice in the technology industry. Companies from Google to the U.S. Department of Defense offer financial rewards to security researchers who discover and responsibly disclose vulnerabilities in their software. The largest bug bounty platforms -- HackerOne, Bugcrowd, Synack -- have facilitated hundreds of thousands of vulnerability reports and paid hundreds of millions of dollars in bounties.
The rationale is sound. No software development team can find every vulnerability in its own code. External researchers bring fresh perspectives, diverse skills, and adversarial mindsets that complement internal security efforts. Paying for vulnerability discovery is far cheaper than suffering a breach. A bug bounty program that pays $10,000 for a critical vulnerability is a bargain compared to the millions of dollars a successful exploit might cost.
And yet the cobra stirs.
The Strategic Landscape of Vulnerability Discovery
When a researcher discovers a vulnerability, they face a decision with multiple options:
Option 1: Report it through the bounty program. The researcher receives a bounty payment (typically $500 to $250,000 depending on severity) and public recognition. The vulnerability is fixed. The system is more secure.
Option 2: Report it to a vulnerability broker. Companies like Zerodium purchase vulnerabilities for resale to government agencies and other clients. Zerodium's public price list offers up to $2.5 million for certain iOS exploits. The vulnerability is not fixed; it is exploited.
Option 3: Sell it on the black market. Criminal organizations purchase vulnerabilities for use in ransomware, espionage, and theft. Prices vary widely but can exceed bounty payments by orders of magnitude for high-value targets.
Option 4: Stockpile it. The researcher holds the vulnerability, waiting for the bounty value to increase, for a higher-paying buyer to emerge, or for the vulnerability to become more valuable as the affected software is more widely deployed.
Option 5: Introduce a vulnerability, then "discover" it. A developer with access to the codebase could intentionally introduce a subtle vulnerability, wait an appropriate period, then "discover" and report it through the bounty program. The developer collects both a salary and a bounty. The practice is nearly undetectable if the introduced vulnerability is subtle enough to pass code review.
The bounty program assumes Option 1 is the dominant strategy. But for a rational, self-interested agent, Option 1 is dominant only when the bounty payment exceeds the expected value of all alternatives -- including the risk-adjusted value of selling to brokers, the black market, or the value of stockpiling. When it does not, the bounty program fails to capture the most dangerous vulnerabilities.
The Cobras in the Code
Several specific cobra effects have emerged in the bug bounty ecosystem:
The severity incentive. Bounty payments are typically scaled by severity: low-severity bugs pay little, critical bugs pay a lot. This creates an incentive to find -- or create -- the most severe vulnerabilities possible. A researcher who can choose between reporting a minor bug (bounty: $500) and spending additional time to weaponize it into a critical exploit (bounty: $50,000) is incentivized to make the vulnerability worse before reporting it.
The bounty treadmill. As bounty payments increase, the talent pool shifts. Skilled researchers gravitate toward bounty hunting, leaving fewer skilled developers working on the code itself. In the extreme case, a company might find itself paying to discover vulnerabilities in code that was less thoroughly reviewed because the best security minds were outside the company, hunting bounties rather than writing secure code.
The false positive economy. Some bounty programs pay for all valid reports, including minor issues that pose no real security risk. Researchers learn to generate high volumes of low-quality reports -- automated scanning results, theoretical vulnerabilities with no practical exploit path, edge cases that would never occur in production. The bounty program's inbox fills with noise, burying the signal of genuinely critical vulnerabilities.
The perverse economics of disclosure. A vulnerability that is disclosed and patched has zero future value. A vulnerability that is not disclosed retains its value indefinitely. The bounty system asks researchers to destroy the economic value of their discoveries. This works when the bounty exceeds the vulnerability's value on other markets. When it does not -- when a zero-day exploit is worth $2 million to a broker and $50,000 through the bounty program -- the rational economic choice is clear, and it is not disclosure.
The Insider Threat Amplification
The most insidious cobra effect in bug bounty programs is the insider threat. When a company pays for vulnerability discovery, every employee with code access faces a new incentive: the code they write is worth their salary, but the bugs they introduce could be worth a bounty.
The mathematics of the insider cobra are stark. A senior developer earning $200,000 per year could supplement their income by introducing a single critical vulnerability annually and collecting a $100,000 bounty. The introduced vulnerability would need to be subtle enough to survive code review and plausible enough to appear accidental. For a skilled developer, this is not technically difficult.
The detection problem is formidable. How do you distinguish between a vulnerability that was accidentally introduced and one that was intentionally planted? Code review catches obvious bugs but is ineffective against subtle, deliberate ones. Statistical analysis might detect a pattern over time -- a developer whose code has an unusually high vulnerability rate -- but the statistical power is weak when the base rate of intentional introduction is low.
Bug bounty programs did not create the insider threat. But they amplified it. Before bounty programs, a malicious insider could sell vulnerabilities to external buyers, but the transaction was cumbersome and risky. With a bounty program, the insider has a legitimate, risk-free channel for monetizing introduced vulnerabilities. The bounty program has, at the margin, made it easier and safer to profit from weakening the very system the program was designed to protect.
Connection to Chapter 15 (Goodhart's Law): The bug bounty metric -- vulnerabilities reported -- is a Goodhart target that decouples from the underlying goal of system security. More vulnerabilities reported does not necessarily mean more security. It could mean more noise (low-quality reports), more severity inflation (researchers weaponizing bugs for higher bounties), or more intentional introduction (insiders planting bugs to collect bounties). The metric measures activity at the bounty interface. It does not measure the security of the system. A system could have rising bounty payments and rising vulnerability counts while simultaneously becoming less secure -- if the bounties are incentivizing the creation of vulnerabilities rather than their elimination.
Synthesis: The Helper's Cobra
The welfare cliff and the bug bounty illustrate a particularly painful variant of the cobra effect: the cobra that emerges from systems designed to help.
The welfare system is designed to help the poor. Its incentive structure traps them in poverty. The bug bounty program is designed to make software secure. Its incentive structure can make software less secure. In both cases, the designers' intentions are genuinely good, the logic of the incentive is genuinely sound within its model, and the cobra effect is genuinely perverse.
What makes these examples different from the colonial bounties and emissions trading is the moral dimension. The cobra farmers of Delhi were motivated by profit. The HFC-23 manufacturers were motivated by profit. These are simple cobra effects: rational self-interest exploiting a poorly designed incentive.
The welfare cliff victim is not gaming the system. She is trapped by it. The home health aide who declines a promotion is not a strategic actor exploiting a loophole -- she is a mother calculating how to feed her children. The cobra effect of the welfare cliff is not the product of greed or gaming. It is the product of a system that punishes the behavior it was designed to encourage.
The bug bounty researcher who sells a zero-day to a broker rather than reporting it through the bounty program is making a rational economic decision, but the decision exists only because the bounty program created the market context in which the decision is meaningful. Before bounty programs formalized vulnerability pricing, the market for zero-days was smaller, less organized, and less legitimate. The bounty program, by establishing that vulnerabilities have monetary value, contributed to the growth of the very market that competes with it for researcher loyalty.
The General Principle
From these two case studies, a general principle emerges: the harder you try to help through incentives, the more important it is to map the full incentive ecology before deploying them.
The welfare cliff exists because program designers focused on the first-order question -- "who deserves help?" -- and did not fully reckon with the second-order question -- "what will the help structure incentivize?" The bug bounty cobra exists because program designers focused on the first-order question -- "how do we find bugs?" -- and did not fully reckon with the second-order question -- "what else will paying for bugs incentivize?"
The cobra effect does not mean that welfare programs or bug bounties should not exist. Both have produced genuine benefits. The welfare system keeps millions of families from destitution. Bug bounty programs have fixed thousands of critical vulnerabilities. The benefits are real.
But the cobra effects are also real. And they are structural -- built into the incentive architecture, not the result of anyone's malice or incompetence. Addressing them requires not better intentions but better design: the patient, rigorous mapping of the incentive ecology, the identification of the strategic responses the system will produce, and the courage to redesign systems that are producing the opposite of their intended effects.
Spaced Review -- Iatrogenesis (Ch. 19): Both the welfare cliff and the bug bounty are iatrogenic: the interventions designed to help are causing harm. The welfare system's help creates the poverty trap. The bug bounty's help creates the vulnerability market. In both cases, the iatrogenic harm is not a side effect -- it is a structural consequence of the incentive architecture. The Intervention Calculus from Chapter 19 applies directly: before deploying an incentive to help, map not just the first-order effects (benefits delivered, bugs reported) but the second-order effects (poverty traps created, vulnerability markets stimulated) and the third-order effects (dependency on the system, erosion of intrinsic motivation, normalization of the practice of trading in the very harms the system was designed to prevent).
The cobras in helping systems are the cruelest cobras of all -- not because they are the most destructive, but because they are the most heartbreaking. They are the proof that good intentions, unaccompanied by careful incentive analysis, can produce outcomes that are indistinguishable from malice.