Alright, so I watched A Beautiful Mind last week — the Ron Howard film, Russell Crowe as John Nash — and somewhere around the chalkboard scenes it hit me: Daniel is becoming a whiteboard guy. Like, fully committed. Hannah's probably finding equations on the bathroom mirror at this point. And that got me thinking about game theory, about Nash, and then Herman reminded me that we actually know someone in this space. Not just "someone who read about game theory" — an actual game theorist. Gideon Blocq. Technion PhD, Google Doctoral Fellow, applied Nash bargaining to network routing problems, and then pivoted to found an AI startup in Tel Aviv. Daniel knows him personally. So today we're going wide: the movie, the math, the man, and where Nash's ideas are showing up in two thousand twenty-six. There's a lot here.
Herman Poppleberry, by the way, for anyone finding us for the first time. And yes — Gideon is a real connection, which makes this episode feel different. It's one thing to read about game theory in a textbook. It's another when you can trace the ideas from Nash's 1950 paper through a specific person's PhD thesis and out the other side into a funded AI company. That through-line is what I want to spend time on today.
But we have to start with the movie, because the movie is where most people's mental model of game theory comes from. And here's the thing — that mental model is wrong in a very specific and instructive way.
The bar scene.
The bar scene. For anyone who hasn't seen it: Nash and his grad school friends are at a bar, a group of women walk in including one blonde, all the men prefer the blonde, and Nash has this eureka moment where he proposes — ignore the blonde, all go for the brunettes, because if everyone chases the blonde they block each other and the brunettes won't talk to them either.
And the film presents this as the discovery of Nash equilibrium. The moment of genius. Except it's not a Nash equilibrium. It's actually the opposite.
Walk me through why.
So Nash equilibrium is a state where no individual player can improve their outcome by unilaterally changing their strategy, assuming everyone else holds steady. That's the definition. Now, in the bar scene: if all the men are going for the brunettes, what's the optimal move for any single man? Go for the blonde. She's uncontested. He has every incentive to deviate. So the "everyone ignore the blonde" outcome is inherently unstable — it fails the basic test.
The actual Nash equilibria in that game are messier.
Much messier. You've got two pure-strategy equilibria: one man gets the blonde while the others take brunettes, or a different man gets the blonde while the others take brunettes. And then there's a mixed-strategy equilibrium where each man goes for the blonde with some probability and randomizes. The film's solution is closer to what you'd call an envy-proof or egalitarian allocation — a different concept entirely, from a different branch of game theory.
The film also has Nash claiming he disproved Adam Smith. "The best result comes not from everyone doing what's best for themselves." Which is...
Not what Nash showed. Nash's work complicated Smith's invisible hand — it showed that self-interested equilibria can be inefficient, which is a real and important result. But "disproved" is doing a lot of lifting there. The Prisoner's Dilemma had already established that point before Nash. What Nash contributed was a general framework for finding equilibria in any finite game, not a refutation of market economics.
And yet the film won Best Picture. Go figure.
To be fair to Ron Howard — the equations on the chalkboards in the film are actually real Nash equations. They brought in mathematicians to make sure the math on screen was authentic. It's just the dramatic depiction of the eureka moment that's wrong. Which is a strange combination: accurate equations, inaccurate insight.
Before we get into the actual math, I want to spend a minute on Nash the person, because the real story is more interesting than the film in almost every dimension. The film focuses heavily on the illness. The recovery is where it gets genuinely extraordinary.
Nash was diagnosed with paranoid schizophrenia around 1958, when he was about thirty. He suffered for roughly twenty-five years — hospitalizations, primitive antipsychotics, periods where he was essentially unable to function mathematically. And then around 1970, he stopped taking medication. And gradually, over years, he recovered. Not partially — he recovered to the point where he received the Nobel Prize in Economics in 1994 and was considered intellectually sharp.
Stopping antipsychotic medication and recovering from schizophrenia is almost unheard of.
Essentially unprecedented for a case that severe. Nash described it in his Nobel interview as something like a rational choice — a decision to reject delusional thinking rather than just experience it passively. Whether that framing is accurate neurologically is an open question, but it's how he experienced it. He also said something striking: "Madness can be an escape." The delusions offered something. Rejecting them required choosing reality even when reality was less comfortable.
And Princeton supported him through this. He was wandering the campus for years during his worst periods, writing strange equations on blackboards, and the institution essentially let him exist there.
There's something worth noting about that. Princeton in the 1950s and 60s was also home to John von Neumann and had recently hosted the publication of Theory of Games and Economic Behavior by von Neumann and Morgenstern in 1944 — the foundational text Nash was building on. Nash arrived in that environment at twenty-one, with a PhD thesis that extended their framework dramatically, and then spent decades unable to access his own mind. The contrast is stark.
He died in 2015, coming back from Oslo where he'd just received the Abel Prize — which is essentially the Nobel Prize of mathematics. Taxi accident in New Jersey. He was eighty-six. The timing has this brutal irony to it: he'd just received recognition for work that his mathematical peers considered his deepest contribution, which wasn't even game theory.
The Nash Embedding Theorem. His work on Riemannian manifold embeddings. Mathematicians who knew him well often said that was his most technically impressive achievement — that the game theory, as influential as it became, was almost the accessible part of his genius.
Right. The game theory made him a household name. The differential geometry made him legendary among mathematicians. Okay — let's get into the actual substance of what he discovered, because this is where I think most coverage goes wrong by conflating two different things.
This is important. Nash Equilibrium and Nash Bargaining Solution are not the same concept. They're from different branches of game theory, they address different questions, and Nash actually published the bargaining paper first — in Econometrica in April 1950, when he was twenty-one — and the equilibrium paper came in 1951 as his PhD thesis. The more famous concept came second.
And they're solving different problems. The equilibrium is about what happens when players can't cooperate. The bargaining solution is about what happens when they can.
Non-cooperative versus cooperative game theory. In Nash equilibrium, players act independently, there are no binding agreements, and you're asking: what's the stable state? In Nash bargaining, players can communicate, make binding commitments, form coalitions — and you're asking: what's the fair outcome?
And Nash's bargaining paper is built on four axioms. Which is a beautiful approach — instead of proposing a solution and arguing for it, he said: here are four properties any reasonable fair solution should have, and then proved those four properties uniquely determine a single outcome.
The four axioms are: symmetry — if the players are identical, they get equal shares. Pareto efficiency — the solution leaves no value on the table, you can't make one player better off without making the other worse off. Independence of irrelevant alternatives — removing options that weren't chosen shouldn't change what gets chosen. And scale invariance — the solution shouldn't depend on the units you use to measure utility.
Each of those sounds almost obvious when you state it.
That's the genius of it. Individually they're almost inarguable. Collectively they force a unique answer: maximize the product of both players' utility gains above their disagreement points. If you call the disagreement points d-one and d-two, and the utilities u-one and u-two, the Nash solution maximizes the product of u-one minus d-one times u-two minus d-two.
And the disagreement point is the crux of everything practically useful about this framework.
The disagreement point is what you get if no deal is reached. Your outside option. Nash's formula makes explicit something negotiators have known intuitively forever: the better your alternative, the more you extract from any deal. But he proved it. He showed mathematically that your disagreement point shifts the Nash solution in your favor — not just a little, proportionally.
Which is why every piece of salary negotiation advice in the world says "get a competing offer before you negotiate." That's Nash bargaining intuition, even when the person giving the advice has never heard of John Nash.
And there's a corollary that I find genuinely uncomfortable: risk aversion costs you. If you're more risk-averse than your counterpart — if you value certainty more than they do — the Nash solution gives you less. Not because you contributed less, not because your position is weaker, but because your preferences make you easier to extract concessions from. The math works out to something like: if Alice values money as the square root of x and Bob values it linearly, Alice gets one-third of the pie and Bob gets two-thirds.
That has real implications for groups that are systematically more risk-averse in negotiation contexts.
It does. And Nash bargaining as a normative framework actually offers a path to correcting that — if an arbitrator can observe both parties' situations and apply the formula, they can produce an outcome that doesn't penalize risk aversion. Salary-capped sports leagues are actually the closest real-world context to this working: both sides' situations are relatively transparent, the numbers are known, and an arbitrator could in principle calculate the Nash solution.
The practical limitation is that the formula requires both parties' utility functions to be common knowledge. And people lie about their outside options constantly.
Which is why it's more useful as a normative benchmark — what should the outcome be — than as a descriptive model of what actually happens in messy real-world negotiations. Nash knew this. His 1950 paper is explicit that it's an idealized framework.
I want to pick up on something you said earlier about Gideon Blocq, because I think his work is a great bridge between the abstract framework and what happens when you try to apply it to a real system. He spent his entire PhD asking a specific version of the question: in a network, when selfish agents are allowed to bargain with each other and form coalitions, how much efficiency do you gain compared to pure selfishness?
His key paper — "How Good is Bargained Routing?" published in IEEE/ACM Transactions on Networking in 2016 with his advisor Ariel Orda at the Technion — introduced what they called the "Price of Selfishness." Which is a deliberate echo of the Price of Anarchy.
Explain the distinction, because they sound similar.
Price of Anarchy is a well-established concept: it compares the social optimum to the worst possible Nash equilibrium outcome when everyone acts selfishly. It measures how bad things can get when there's no cooperation. Price of Selfishness asks a subtler question: how does the Nash bargaining solution — where agents can cooperate — compare to the social optimum? Even cooperative bargaining isn't perfectly efficient. Bargainers are still maximizing their own gains, just doing it through negotiation rather than unilateral action. Blocq and Orda were quantifying that residual gap.
And in network routing, this matters because you have autonomous systems — different organizations' networks — that interact and route traffic. They're not perfectly aligned. They're not perfectly selfish either. They're somewhere in the middle, which is exactly the space Nash bargaining describes.
The findings were network-topology dependent, which is itself an interesting result. The gains from allowing coalition formation and bargaining over pure selfishness varied substantially depending on how the network was structured. There's no universal answer — the value of cooperation is a function of the specific game being played.
I dabbled in game theory during my own formative years, actually. In the canopy. You'd think sloths don't have much to negotiate over, but territorial rights to prime canopy positions are genuinely contested. There's a classic evolutionary stable strategy situation where being slow enough that predators lose interest is self-reinforcing — if everyone is slow, the fastest sloth is actually the most conspicuous, which is bad. So there's this equilibrium where "be extremely slow" is stable not because it's optimal in isolation but because deviating from it is penalized.
I love that you're bringing this up because it's a real example of an evolutionarily stable strategy, which is Nash equilibrium applied to population dynamics. Maynard Smith formalized that in the 1970s. The key insight is the same: the stability comes from the fact that no individual can improve their outcome by unilaterally changing strategy given what everyone else is doing.
And the canopy spot negotiation is essentially Nash bargaining — there's a disagreement point, which is whatever spot you'd end up with if you just fought over it, and the bargained outcome is you both recognizing that the cost of conflict exceeds the value of the disputed territory, so you find an allocation that leaves you both better off than fighting.
Which is also, not coincidentally, the logic behind most arms control agreements and trade treaties. Nash's framework isn't just for bar scenes and sports contracts.
Okay — so Blocq finishes his PhD in June 2017 and immediately co-founds VineSight. That pivot is worth examining, because it's not an obvious move.
On the surface it looks like a hard left turn. Cooperative game theory and network routing to AI-powered disinformation detection. But I think the thread is actually quite clear when you look at what both problems have in common: strategic behavior in networks, information asymmetry, and adversarial actors trying to game a system.
VineSight's core claim is that there's a four-to-twelve-hour window between a narrative forming and a crisis calcifying. And if you can detect the formation of a toxic narrative before it reaches mainstream platforms, you can intervene before the options narrow.
The technical approach is interesting — and it's where Blocq's network routing background arguably prepared him directly. Rather than analyzing content, which is computationally expensive and prone to false positives, VineSight focuses on share patterns. How content propagates across platforms. They monitor fringe channels and closed communities where narratives form before they reach Twitter or TikTok.
Which is essentially what Blocq spent his PhD doing: modeling how flows propagate through networks under different routing strategies. The analogy from routing to information diffusion is not a stretch.
And they claim a ninety-nine percent positive identification rate on this basis. They raised a four million dollar seed round in September 2022, led by AnD Ventures, with notable participation from the Technion's own investment arm — meaning his alma mater literally bet on his startup. They'd already reached profitability before the seed round, which is unusual, and achieved five-times revenue growth in eighteen months.
Blocq's quote from the company is good: "A company can have the greatest product, a candidate can be a fantastic leader, and a cause might be just, but in the age of online bad actors, one viral, toxic post has the power to destroy it all before they have a chance to defend themselves." That's a very game-theoretic framing — the first mover in narrative space has an enormous structural advantage.
And detecting first-mover activity in information networks is, functionally, a routing and propagation problem. You're trying to identify where traffic is originating, what path it's taking, and how fast it's moving. That's Blocq's PhD research applied to a different kind of network.
By the way — today's episode is being written by Claude Sonnet four point six, which I find appropriately meta for an episode about strategic behavior and information systems.
Very on-brand. Alright — I want to spend some time on where Nash bargaining is showing up in AI agent design right now, because there's a March 2026 paper on arXiv that's directly relevant and genuinely impressive.
What's the problem it's solving?
AI agents are increasingly being deployed in high-stakes negotiation contexts — supply chain contracting, spectrum allocation, legal dispute resolution. And when you train an AI on human negotiation data, you get a model that inherits human biases. Anchoring on initial offers. Loss aversion. Accepting suboptimal outcomes early because the human data contains those patterns. The result is AI negotiators that are systematically far from Pareto efficient.
So you're building AI that negotiates like a nervous person at a car dealership.
Essentially. The paper proposes a "guided graph diffusion framework" that enforces Nash bargaining axioms during the AI's generation process — steering the output toward individually rational, Pareto efficient outcomes rather than letting it reproduce whatever biased patterns existed in the training data. The results are striking: ninety-nine point four five percent Nash efficiency on synthetic data. On the CaSiNo human negotiation corpus — which is real messy human data — they achieved fifty-four point two four percent Nash efficiency compared to near-zero for unconstrained AI. On the Deal or No Deal corpus, eighty-eight point six seven percent.
And one hundred percent individual rationality compliance across all datasets.
Which is actually the more important result in some ways. Individual rationality means neither agent accepts an outcome worse than their disagreement point. Unconstrained AI trained on human data will sometimes accept those outcomes — because humans sometimes accept bad deals. The Nash framework as a normative constraint prevents that.
There's a fairness dimension here too, which connects back to the risk aversion point.
The paper explicitly notes that AI trained on historical bargaining records reproduces documented demographic disparities in final outcomes. If women historically accepted lower salary offers, an AI trained on that data will produce lower offers to women. Nash bargaining as a normative framework — if you can actually enforce it — corrects that structural bias by grounding outcomes in the formal axioms rather than historical patterns.
Which is a genuinely interesting use of formal game theory as an alignment tool. You're not just trying to make the AI smart, you're trying to make it fair by definition — by building the fairness axioms into the generation process.
And this connects to something broader about what Nash's 1950 paper actually achieved. He wasn't just solving a bargaining problem. He was asking: what would a rational, fair outcome look like if we could design one from scratch? The axioms are a statement of values — symmetry, efficiency, independence, scale invariance — and the theorem proves those values have a unique implication. That's a template for normative AI design.
Let's bring it back to the real-world applications for a moment, because I think some of this can sound abstract until you see where it's actually being used.
Patent damages is an interesting one. Some patent licensing experts argue for a fifty-fifty split of profits using Nash bargaining logic — the argument being that under idealized conditions, when two parties cooperate to produce value neither could produce alone, they should split the surplus equally. This has become genuinely controversial in patent litigation. The "fifty percent rule" derived from Nash bargaining gets invoked in court cases.
International trade negotiations are another context. Every nation negotiating a trade treaty has a threat point — their unilateral tariff policy if no deal is reached. Nash bargaining formalizes the intuition that your leverage in a trade negotiation is a function of how well you can do without the deal. A country that's self-sufficient or has many trading partners has a higher threat point and should, in principle, extract more favorable terms.
Which is also why sanctions work the way they do — they're an attempt to lower the target country's threat point by degrading their outside options. Nash bargaining makes that mechanism explicit.
Labor negotiations. The NBA's collective bargaining structure is essentially a Nash bargaining problem where the threat points are clear: players' threat point is their ability to hold out or go to alternative leagues, owners' threat point is the value of not having the player. The salary cap changes the structure of the game by constraining the feasible set of outcomes.
And this is where the framework is most practically applicable — when both sides' situations are relatively transparent. The opacity problem that makes Nash bargaining hard to apply in general negotiation is reduced in contexts like sports labor where a lot of the relevant information is public.
What about diplomacy specifically? Because I think that's where the stakes are highest and the framework is most underutilized.
Nash bargaining formalizes something diplomats have always known intuitively: your leverage is your outside option. But it goes further — it tells you that the outcome should be the point that maximizes the product of both parties' gains above their threat points. Not the midpoint. Not the point that makes one party happy. The product. Which means a party with a very low threat point has enormous leverage — if your threat point is zero, you gain everything above zero, and the Nash solution weights that heavily.
It also implies that improving your own threat point and improving your counterpart's threat point are both strategically significant moves, in opposite directions. Sanctions lower their threat point. Alliance-building raises yours.
And the four axioms give you a principled way to evaluate whether a proposed agreement is fair — not just "is everyone somewhat happy" but "does this outcome satisfy symmetry, efficiency, independence, and scale invariance?" Those are testable properties.
The challenge in real diplomacy is that utility functions are not common knowledge. Countries don't reveal their true preferences.
Which is why Nash bargaining works better as a normative benchmark than a predictive model in diplomacy. It tells you what the outcome should be if everyone were rational and transparent. The gap between that and what actually happens is a measure of how much strategic misrepresentation, domestic political constraints, and irrationality are distorting the process.
Let me ask you something that I've been sitting with through this whole conversation: Nash published the bargaining solution in 1950 and the equilibrium in 1951. The equilibrium became the famous concept — the one that won the Nobel, the one in the film. The bargaining solution is arguably more directly useful for real-world applications. Why did the equilibrium dominate?
A few reasons. The equilibrium is easier to explain — "no one can do better by changing strategy alone" is a clean sentence. The bargaining solution requires explaining axioms and utility functions and disagreement points. But I think the deeper reason is that non-cooperative game theory matched the Cold War intellectual moment. The Prisoner's Dilemma, arms race dynamics, nuclear deterrence — these are all non-cooperative equilibrium problems. The 1950s and 60s were saturated with questions about what happens when adversaries can't cooperate. Nash equilibrium was the tool for that moment.
Nash himself noted in his Nobel interview that the bargaining work "is also important, but it wasn't cited for the Nobel Prize." There's something almost melancholy about that — the paper he published first, when he was twenty-one, gets overshadowed by the thesis he published a year later.
And his mathematicians peers would say both of those are overshadowed by the Nash Embedding Theorem. The man's career is a hierarchy of underappreciated contributions.
Alright — practical takeaways, because I think this episode has a lot of directly actionable content buried in it.
The most important one is the disagreement point insight. Before any significant negotiation — salary, contract, partnership — the most valuable thing you can do is improve your outside option. Not your position, not your arguments. Your alternative. Because the Nash solution shifts in your favor proportionally to the quality of your outside option. This isn't negotiation advice, it's a theorem.
The risk aversion point is uncomfortable but important. If you're more risk-averse than your counterpart, you will tend to accept worse outcomes even in a "fair" framework. Knowing this is the first step to correcting for it — either by managing your own risk preferences consciously or by structuring negotiations in ways that reduce the salience of risk.
For anyone thinking about AI systems that involve negotiation or resource allocation — the March 2026 research suggests Nash bargaining axioms can be enforced as normative constraints during generation, and the efficiency gains are substantial. If you're building an AI agent that negotiates anything, this is worth knowing about.
And the VineSight angle — the idea that narrative formation happens in a narrow window before it reaches mainstream platforms — is something any organization with a public reputation should be thinking about. Blocq's insight is that the network dynamics of information propagation create a detectable signature before the crisis is visible. That's a monitoring problem, not just a response problem.
The broader lesson from Blocq's career is about the transferability of formal frameworks. He spent years asking how much efficiency you gain when network agents can bargain rather than just being selfish. Then he applied that question to information networks and built a company around it. The framework traveled; the domain changed.
I think that's the real legacy of Nash's work — not any specific application, but the demonstration that strategic interaction can be formalized, that fairness has mathematical content, and that the structure of a game determines its outcomes in ways that intuition often misses. The bar scene in the film gets the specific answer wrong, but the impulse — to find the hidden structure of social situations — is exactly right.
And Nash found that structure twice, in two different forms, in consecutive years, before he was twenty-three. Whatever the film gets wrong about the math, it gets something right about the quality of that mind.
Alright. That's a good place to land. Thanks as always to our producer Hilbert Flumingtop for keeping everything running, and big thanks to Modal for providing the GPU credits that power this show. Find us at myweirdprompts dot com if you want the RSS feed and all the ways to subscribe — and if you're already listening there, you know where to find us. This has been My Weird Prompts.
See you next time.