#2308: When AI Forecasts Collide: Geopol Model Divergence

Five AI models forecast the Iran-Israel-US crisis — and their disagreements reveal surprising insights about geopolitical reasoning.

0:000:00
Episode Details
Episode ID
MWP-2466
Published
Duration
24:51
Audio
Direct link
Pipeline
V5
TTS Engine
chatterbox-regular
Script Writing Agent
Claude Sonnet 4.6

AI-Generated Content: This podcast is created using AI personas. Please verify any important information independently.

In a groundbreaking experiment, five AI models from diverse training lineages were tasked with forecasting the Iran-Israel-US conflict across three time horizons: 24 hours, one week, and one month. Known as the Geopol Forecast Council, this project aimed to explore how independently reasoning models converge and diverge when analyzing the same fast-moving geopolitical scenario.

The models — Claude Sonnet 4.6, Google’s Gemini 3 Flash Preview, Zhipu’s GLM 5.1, DeepSeek V3.2, and Moonshot’s Kimi K2.5 — were chosen deliberately to represent different training backgrounds and architectural approaches. The blind-parallel structure ensured that each model worked independently, preventing groupthink and allowing for meaningful divergence.

The experiment revealed both convergence and divergence on key predictions. All five models agreed on near-term outcomes, such as the nominal status of the Lebanon ceasefire and the likelihood of IRGC vessel harassment continuing. They also concurred on longer-term predictions, like the improbability of a comprehensive nuclear deal within a month and the deepening strain in US-Israel relations. Notably, all five independently predicted partial maritime de-escalation driven by mutual economic pressures — a second-order insight that underscores the value of independent reasoning.

However, sharp disagreements emerged, particularly on the maritime track. For example, Claude Sonnet predicted sustained disruption in the Strait of Hormuz with a confidence score of 0.61, while GLM forecasted mutual de-escalation and strait reopening with a score of 0.35. These divergences were not arbitrary; each model provided coherent reasoning and countervailing precedents to support its position, highlighting how different historical analogies and data diets influence forecasts.

The methodology itself was a key focus of the experiment. The three-stage pipeline began with a grounding stack of RSS feeds and news briefings, followed by a Situation Report (SITREP) produced by the Council Head (Claude Sonnet in this run). Each council member then generated three predictions per time horizon, supported by reasoning, historical precedents, and confidence scores. This structured approach ensured transparency and epistemic rigor, forcing models to argue against their own predictions before filing them.

The experiment raises intriguing questions about the role of AI in geopolitical forecasting. While the blind-parallel structure successfully surfaced meaningful divergence, the synthesis layer — handled by the same model that produced the SITREP — introduced potential editorial bias. This tension highlights the need for further refinement in methodology.

Ultimately, the Geopol Forecast Council demonstrates both the promise and the challenges of AI-driven geopolitical analysis. By leveraging diverse models and structured reasoning, it offers a framework for generating insights that are both independent and epistemically robust. As AI continues to evolve, experiments like this will play a crucial role in shaping its application to complex, real-world scenarios.

Downloads

Episode Audio

Download the full episode as an MP3 file

Download MP3
Transcript (TXT)

Plain text transcript file

Transcript (PDF)

Formatted PDF with styling

#2308: When AI Forecasts Collide: Geopol Model Divergence

Corn
Daniel ran an experiment last night — literally at eleven fifty-four UTC — and the thing that jumped out immediately wasn't the geopolitics. It was that five AI models, reasoning independently about the same crisis, didn't agree with each other. Not in minor ways. In ways that matter. He's calling it the Geopol Forecast Council, a lean spin-off of his Geopol-Forecaster project, and the question it's really asking is: when you line up a blind-parallel panel of models from genuinely different training lineages and point them all at the same fast-moving situation, what do you learn from where they converge — and what do you learn from where they don't?
Herman
By the way, today's episode is powered by Claude Sonnet four point six — which is, not incidentally, one of the five models that sat on this council. So there's a certain recursive quality to today's show that I find delightful.
Corn
The friendly AI down the road wrote the script and also filed a forecast. The scenario Daniel chose was the Iran-Israel-US conflict, three time horizons: twenty-four hours, one week, one month. That's the test bed. What we're actually here to talk about is the methodology, what the council produced, and what the divergences reveal about how different models see the world differently — sometimes very differently.
Herman
Which is the part that surprised me most when I went through the outputs. You'd expect some variance on confidence scores, maybe some stylistic differences in how models frame uncertainty. What you don't necessarily expect is one model calling the Lebanon ceasefire effectively collapsed while another reads it as holding in name — and both of them producing internally coherent reasoning to support their position. That's not noise. That's signal.
Corn
That's exactly the design question worth unpacking. Because if you want divergence to be signal rather than groupthink, you have to build the system so that groupthink can't happen in the first place.
Herman
Right, and that's where the blind-parallel structure becomes really interesting from a methodology standpoint. So, Corn, how was this thing actually built?
Corn
The Geopol Forecast Council is essentially the leaner sibling of Daniel's Geopol-Forecaster, which simulates around forty actors across multiple timesteps and gets expensive quickly. This is the cheaper cousin — three pipeline stages, five council members, three time horizons. Designed to be fast and repeatable on a fast-moving situation.
Herman
The name "council" is doing real work there. It's not one model asked to forecast — it's a panel, deliberately constituted from different training lineages. Zhipu's GLM five point one, DeepSeek V three point two, Google's Gemini three Flash Preview, Claude Sonnet four point six, and Moonshot's Kimi K two point five. You're spanning Chinese labs, American labs, different data diets, different architectural choices. That's not accidental.
Corn
The scenario — Iran, Israel, the US, across twenty-four hours, one week, one month — was chosen precisely because it's the kind of situation where the ground shifts fast enough that you can actually stress-test a forecasting system. If everything is stable, convergence tells you nothing. You want a scenario where there are genuine open questions.
Herman
Which this absolutely was. The April twenty-second deadline alone creates a real binary inflection point cascading across three separate tracks simultaneously — Lebanon, the diplomatic channel, maritime. It's not a neat problem. It's exactly the kind of multi-actor, multi-domain tangle where you'd expect different models to weight things differently.
Corn
The scenario isn't really the point. It's the test bed. What Daniel is actually probing is whether a blind-parallel structure with genuine lineage diversity produces forecasts that are more epistemically honest than a single model asked the same question — and whether the disagreements, when they show up, are meaningful rather than arbitrary.
Herman
That framing matters. Because the alternative design — sequential deliberation, where models see each other's outputs — would just produce laundered consensus. The whole value of the blind structure is that you can't accidentally agree. So, to make this concrete, let’s break down the three-stage pipeline.
Corn
Yeah, let’s dig into that pipeline. Because "three stages" is doing a lot of heavy lifting — walk me through how it actually works.
Herman
Stage one is a grounding stack — RSS feeds from Times of Israel, Al Jazeera, BBC World, a Perplexity Sonar briefing, and Tavily news search, all running in parallel. That's your raw material. Stage two is where it gets interesting: the Council Head — which in this run was Claude Sonnet four point six — takes all of that and produces a single reconciled, timestamped SITREP. Key actors, stated positions, the load-bearing uncertainties. Every council member reasons from that identical frame.
Corn
That's a meaningful design choice. You're not letting each model do its own grounding independently.
Herman
Which could go either way. On one hand, you eliminate the possibility that one model grabbed a different headline and is now reasoning from a slightly different factual baseline. On the other hand, the SITREP itself reflects the Council Head's interpretive choices — what it decided was load-bearing, what it left out. There's a thumb on the scale before the council even convenes.
Corn
A very well-intentioned thumb, but still a thumb.
Herman
And then stage three — the council members each get that SITREP and produce exactly three concrete predictions per time horizon. Twenty-four hours, one week, one month. Nine predictions per model. Forty-five total across the panel. And here's the structural detail I find elegant: each prediction has six required fields. The prediction itself, supporting reasoning, a supporting historical precedent, a countervailing historical precedent, a single observable change factor, and a confidence score.
Corn
The countervailing precedent is the one that catches my attention. You're forcing each model to argue against itself.
Herman
Before it can file the prediction. You can't just assert "Iran will do X because historically Iran does X." You also have to surface the case where Iran didn't, or where a similar dynamic resolved differently. That's a real epistemic constraint. It's the kind of thing a good analyst does naturally and a bad one skips entirely.
Corn
GLM five point one — how did that play out in practice on the twenty-four hour horizon?
Herman
GLM filed a relatively low confidence score on IRGC vessel harassment — around zero point four zero. The reasoning was grounded, the historical precedent was solid, but the confidence was modest. And that's actually where the divergence with Claude becomes stark, because Claude came in at zero point seven eight on the same prediction. Same SITREP, same structured schema, same question. A thirty-eight point gap on how hot maritime escalation gets in the next day.
Corn
That's not a rounding difference. That's a genuine disagreement about what the situation implies.
Herman
The schema forced both of them to show their working. GLM's countervailing precedent emphasized prior instances where IRGC harassment plateaued rather than escalated. Claude's supporting precedent leaned on more recent harassment patterns. Neither is wrong to invoke the precedent they chose. They're weighting different historical analogies. That's lineage showing up as a real variable.
Corn
Which is exactly what the blind structure is designed to surface. If GLM had seen Claude's output first, does it anchor toward seventy-eight? You'd never know the real distribution.
Herman
The Report Author — also Claude Sonnet four point six in this run — then takes all forty-five predictions, clusters them, assigns consensus strength, and flags the sharp disagreements. That's the synthesis layer. Not averaging the confidence scores, but actually characterizing the epistemic landscape. Where did the panel agree? Where did it fracture? Those are different things and they warrant different treatment in the final report.
Corn
The Report Author being the same model as the Council Head is a detail worth sitting with. There's a question of whether that introduces a kind of editorial gravity — whether Claude's own forecast subtly shapes how it characterizes the disagreements it's summarizing.
Herman
I don't think Daniel has a clean answer to that yet. It's a real tension in the design. You want the synthesis layer to be as model-agnostic as possible, but someone has to do the clustering, and whoever does it brings priors. It's not unique to AI — any human analyst writing up a panel's findings has the same problem.
Corn
Just usually slower about it.
Herman
And a human analyst also doesn't have to worry about whether their clustering algorithm is subtly biased toward their own prior outputs.
Corn
Let's get into what the council actually produced. Where did five independently reasoning models, from different lineages, look at the same SITREP and say the same thing?
Herman
There's more convergence than you might expect, actually. All five agreed that the Lebanon ceasefire is nominal — kinetic incidents near-certain within twenty-four hours. All five agreed Iran would not reopen the Strait of Hormuz in that same window, but that IRGC harassment of vessels continues. All five agreed there's no comprehensive US-Iran framework deal coming in twenty-four hours, at best procedural progress. And all five flagged April twenty-second as a genuine binary inflection point cascading across Lebanon, diplomacy, and maritime simultaneously.
Corn
On the high-confidence, near-term stuff, the panel converges. Which makes a kind of intuitive sense — the closer the horizon, the more constrained the possibilities.
Herman
The convergence holds at one month too, on certain things. All five agreed a comprehensive nuclear deal within a month is structurally unlikely — the gap on enrichment is too wide. All five saw US-Israel alignment as visibly strained and deepening as Washington pursues a deal Tel Aviv opposes. And all five predicted some form of partial maritime de-escalation within a month driven by dual-blockade economic pressure — US blocking Iranian exports, Iran blocking Hormuz, both sides eventually needing an off-ramp.
Corn
That last one is interesting because it's not obvious. It's a second-order prediction. The first-order read is "blockade equals escalation." The second-order read is "mutual economic pain drives partial de-escalation." The fact that all five landed there independently is actually more meaningful than if one model had been told to look for it.
Herman
That's the value of the blind structure showing up in the convergence, not just the divergence. When five models reason independently and arrive at the same non-obvious conclusion, that's a stronger signal than any single model producing the same prediction. You haven't just gotten one analyst's read. You've gotten five independent derivations of the same result.
Corn
Now tell me where it fell apart.
Herman
The maritime track is where the sharpest single disagreement lives. On the one-month Hormuz outlook, Claude sits at zero point six one for sustained disruption. GLM sits at zero point three five — predicting mutual de-escalation and the strait reopening. That's a direct contradiction. Not a confidence gap. A directional disagreement on how the maritime situation resolves.
Corn
Both of them went through the same countervailing-precedent exercise to get there.
Herman
Which is what makes it informative rather than just noisy. GLM's historical anchor was the pattern of IRGC brinkmanship that stops short of full closure — the 2019 tanker incidents, the repeated harassment that never quite became a sustained blockade. Claude's anchor was the current economic pressure being qualitatively different from prior episodes. Same schema, different historical weighting, opposite directional predictions.
Corn
DeepSeek was the one I found most interesting from a framing standpoint. Walk me through that.
Herman
DeepSeek's divergence is more subtle but in some ways more revealing. Where Claude and Kimi both read the Lebanon ceasefire as effectively collapsed, DeepSeek read it as "holds in name" — a framing disagreement rather than a confidence disagreement. And on the April twenty-second inflection point, DeepSeek filed zero point five five on full ceasefire collapse. Claude was at zero point four eight, Kimi at zero point five zero. Nobody is above zero point five five. That's not consensus — that's genuine uncertainty spread across a narrow band, which is actually a more honest characterization of the situation than a high-confidence call would have been.
Corn
The calibration note on this is worth flagging. The confidence scores across the whole panel were deliberately modest — mostly between zero point three and zero point seven. And the report explicitly says that reflects the fluidity of the situation, not model uncertainty per se.
Herman
Which is a meaningful distinction. A well-calibrated forecaster on a uncertain multi-actor crisis should not be filing zero point nine. If you see a model producing very high confidence scores on a situation this fluid, that's a red flag about calibration, not a sign of superior insight.
Corn
Gemini was the outlier on nuclear escalation.
Herman
Only member to surface Iran moving to sixty percent enrichment — confidence zero point five five. Nobody else raised it. Which is interesting on its own, but what it really illustrates is the lineage diversity point in its sharpest form. Gemini noticed something the other four didn't flag. Whether that's because Gemini's training data weighted certain Iranian nuclear escalation precedents more heavily, or because its grounding of the SITREP led it to a different inference chain — you can't fully separate those variables from outside the model. But the fact that it's the only one to surface that prediction tells you something about what different lineages literally see in the same document.
Corn
Then there are the single-source predictions. One model raised it, nobody else did.
Herman
Kimi flagged a possible French UNIFIL drawdown. DeepSeek flagged a French-led emergency UN Security Council meeting. Gemini flagged a Pakistan-Turkey maritime corridor. GLM flagged an Iranian domestic security crackdown. Four predictions, four different models, zero overlap. Each one saw something the others missed entirely.
Corn
That's the lineage diversity argument in its most concrete form. It's not that the models disagreed on the same prediction. It's that they didn't even look at the same parts of the problem.
Herman
That's actually the strongest case for this kind of panel design. A single model, however capable, has a particular attentional profile — what it finds salient, what historical analogies it reaches for first, what knock-on effect it's primed to notice. A lineage-diverse panel doesn't just produce more variance. It produces broader coverage of the possibility space.
Corn
The question is whether the Report Author's clustering can actually capture that. If you've got four single-source predictions from four different models and the synthesis layer treats them as low-confidence outliers to be noted and set aside, you've potentially discarded the most interesting signal in the whole dataset.
Herman
That's the design tension Daniel hasn't fully resolved yet, I think. The Report Author flags sharp disagreements — that's the explicit design goal. But there's a difference between flagging a disagreement where four models went one way and one went another, versus flagging a prediction that only one model raised at all. The latter might be the more important category. It's not disagreement. It's a blind spot in the other four.
Corn
The Report Author being Claude means it's summarizing its own blind spots alongside everyone else's. Which is a strange epistemic position to be in.
Herman
It really is. Though to be fair, any synthesis layer has that problem. You're always asking someone to characterize their own uncertainty relative to others. The honest response is probably to treat single-source predictions as a distinct category in the output — not higher or lower confidence, but flagged as "one lineage saw this, others didn't, worth watching.
Corn
Which is actually a practical design recommendation rather than just a critique.
Herman
And I think that's where the lineage diversity analysis gets useful for anyone building one of these panels. It's not just "use diverse models." It's "build your synthesis layer so that single-source predictions get their own treatment, because they represent what a particular training lineage uniquely surfaces, and that's different from disagreement." But the real challenge is translating that into practice—what does it actually look like to implement?
Corn
So if you're building one of these panels, what does that mean concretely? You don't just throw five models at a question and call it a diverse panel. How do you ensure that synthesis layer actually captures the unique contributions of each model?
Herman
No, and I think that's the most important practical takeaway from this whole experiment. Lineage diversity is not the same as model diversity. You could pick five models that all trace back to similar training corpora, similar RLHF pipelines, similar architectural choices — and you'd get the appearance of a panel with the substance of one very confident analyst. The selection criterion has to be about training lineage, not just capability tier.
Corn
Which in practice means being deliberate about geography of origin as much as anything else. GLM and Kimi are both Chinese-lineage models, but they still surfaced different single-source predictions. DeepSeek is a different lineage again. Gemini is Google's stack. Claude is Anthropic's. The attentional profiles differ.
Herman
The six-field prediction schema is doing real work here too. If you just ask a model for its top predictions, you get whatever it finds most salient first. The schema forces it to surface a countervailing historical precedent — to argue against itself before filing the prediction. That's a meaningful constraint. It's not foolproof, but it at least requires the model to have encountered the counterargument before committing to a confidence score.
Corn
I'd also argue the confidence range is a practical output worth tracking over time. If a particular model in your panel is consistently filing scores above zero point eight on fast-moving geopolitical situations, that's a calibration problem you can identify and correct for — either by weighting that model's output differently or by building a recalibration step into the pipeline.
Herman
That's actually something the existing forecasting literature has a lot to say about. Philip Tetlock's superforecaster work found that calibration — knowing how confident to be — was as predictive of accuracy as raw analytical skill. The models that consistently overestimate their own certainty are going to drag the synthesis layer in the wrong direction even when their directional calls are good.
Corn
For anyone who wants to run something like this: pick models from different training lineages, use a structured prediction schema that includes forced self-contradiction, treat single-source predictions as their own output category rather than noise, and watch the confidence distributions over time as a calibration signal.
Herman
Keep the horizons short if you want the council to be useful. The design note on this is explicit — the council is cheaper than the full actor-simulation Geopol-Forecaster, but it's also clearly weaker on long-horizon or counterfactual reasoning. Twenty-four hours, one week, one month — that's the sweet spot. Push it to six months and you're asking the panel to reason about a world that may have structurally changed in ways none of the models can anticipate from the current SITREP.
Corn
Which is an honest limitation. Short-horizon forecasting on fast-moving situations is a real and useful thing. It doesn't have to be everything.
Herman
The other thing I'd flag for anyone building on this — the Council Head SITREP is a genuine leverage point. Getting the world-state anchoring right before the panel reasons is probably worth more effort than most pipeline designers will give it. Garbage in, garbage out applies here, but it's more subtle. It's not garbage — it's a slightly tilted frame. And a slightly tilted frame, applied to five independently reasoning models simultaneously, produces a systematically tilted output that looks like consensus.
Corn
The grounding stack — the RSS feeds, the Perplexity briefing, the Tavily search — that's not just housekeeping. That's load-bearing.
Herman
It really is. And running those in parallel rather than sequentially matters too. If you run them sequentially and let each source update your world model before the next one comes in, you've already introduced an ordering effect before the Council Head even starts writing. That raises another question: how much of the divergence we see is stable versus just noise?
Corn
That's the broader question I don't think the experiment answers, and I'm not sure it's supposed to. If you run this council on the same scenario at different timestamps, how much of the divergence is stable and how much is noise? Do GLM and Claude reliably disagree on maritime escalation, or did that zero point four zero versus zero point seven eight gap reflect something specific about what was in the news cycle at twenty-three fifty-four UTC on April eighteenth?
Herman
That's probably the most important open question for anyone trying to use this operationally. A forecasting panel you can only trust once is a research instrument. A panel whose divergence patterns are stable across runs — that's a tool. And right now we don't know which one this is.
Corn
Which is a interesting research agenda. Run the same scenario at six-hour intervals. See whether Gemini reliably surfaces nuclear escalation pathways that the others don't. See whether GLM reliably anchors toward de-escalation. If the attentional profiles are stable properties of the training lineage rather than artifacts of the specific grounding, that's a very different kind of signal.
Herman
It would let you build a meta-layer on top of the council — not just what the panel predicts, but what each lineage tends to see and not see, so you can weight the synthesis accordingly. That's where I think this methodology goes if it matures.
Corn
The geopolitical analysis use case is obvious, but I keep thinking about how this transfers. Any domain where you have a fast-moving situation, multiple actors with opaque intentions, and high stakes on getting the short-horizon call right. This is not a narrow tool.
Herman
It really isn't. And the fact that Daniel built it as a leaner version of the full actor-simulation pipeline matters — it suggests there's a cost-accessible version of this kind of structured AI forecasting that doesn't require spinning up forty simulated actors per run. That's a meaningful design contribution on its own.
Corn
Big thanks to Hilbert Flumingtop for producing this one. And to Modal for keeping our GPU pipeline running — if you're building anything that needs serverless compute, they're worth a look.
Herman
This has been My Weird Prompts. If you've got a minute to leave us a review, we read every one — find us on Spotify or at myweirdprompts.
Corn
We'll see you on the next one.

This episode was generated with AI assistance. Hosts Herman and Corn are AI personalities.