#608: The RAMpocalypse: Why AI is Starving Your PC

Why is a 32GB RAM kit now $400? Herman and Corn dive into how OpenAI is gobbling up 40% of the world's memory supply for its "Stargate" project.

0:000:00
Episode Details
Published
Duration
31:52
Audio
Direct link
Pipeline
V4
TTS Engine
LLM

AI-Generated Content: This podcast is created using AI personas. Please verify any important information independently.

On a rainy February afternoon in Jerusalem, hosts Herman and Corn Poppleberry sat down to discuss a crisis currently rattling the tech world: the "RAMpocalypse." What began as a personal anecdote about their housemate Daniel’s struggle to find affordable replacement memory for a home server quickly spiraled into a deep dive into the structural realignment of the global semiconductor industry. According to the hosts, the days of RAM being the "cheap part" of a computer build are officially over, replaced by a market that Herman describes as a "post-apocalyptic wasteland."

The 40 Percent Statistic

The central hook of the discussion is a staggering figure from recent industry reports: OpenAI is estimated to be consuming 40% of the global Dynamic Random Access Memory (DRAM) supply. While this sounds like a hyperbole, Herman explains that the math aligns with the sheer scale of current AI infrastructure projects. Specifically, he points to the "Stargate" project—a joint venture between Microsoft and OpenAI—which aims to house over a million GPUs in a single supercomputing cluster.

Training and running large language models (LLMs) with trillions of parameters isn't just a matter of processing power; it is a matter of memory. Herman introduces the concept of the "memory wall," explaining that even the fastest processors are useless if they have to wait for data to travel from slow storage. To solve this, AI giants are pivoting toward High Bandwidth Memory (HBM), a sophisticated and expensive alternative to standard desktop RAM.

The Zero-Sum Game of Manufacturing

A critical takeaway from the episode is that the production of AI-grade memory and consumer-grade RAM is a zero-sum game. The "Big Three" manufacturers—Samsung, SK Hynix, and Micron—utilize the same fabrication lines for both products. However, the manufacturing process for HBM3E and HBM4 is significantly more complex than standard DDR5.

Herman notes that HBM involves stacking memory dies vertically and connecting them with "Through-Silicon Vias" (TSVs). This complexity leads to much lower yields; if one layer in a stack is faulty, the entire unit is often scrapped. Because these manufacturers can sell HBM to enterprise AI customers at a massive premium, they have little incentive to prioritize the lower-margin consumer sticks found on retail shelves. In many cases, AI companies are even "pre-renting" entire factory lines years in advance, effectively pricing the average consumer out of the market.

A Structural Shift, Not a Bubble

Corn draws a comparison to the GPU shortages during the cryptocurrency boom of 2021, but Herman is quick to point out a fundamental difference. While crypto was driven by speculative mining, the current demand for RAM is fueled by the foundational infrastructure of the next era of computing. Companies like Google, Meta, and OpenAI are in a "winner-take-all" race to build the most capable models. To these entities, the price of RAM is secondary to the goal of achieving computational dominance.

This has led to a centralization of resources that threatens the "democratization of technology." For decades, it was assumed that high-end computing power would eventually become affordable for the average person. The RAMpocalypse suggests the opposite: a future where hardware is so expensive that most users may be forced to rely on thin clients and cloud subscriptions, while the physical silicon sits locked away in massive corporate silos.

The Bottleneck Beyond the Fabs

While new semiconductor factories are being built in places like Ohio and Arizona, relief is not expected anytime soon. Herman highlights a second-order bottleneck: the machines required to build the chips. ASML, the Dutch company that produces extreme ultraviolet (EUV) lithography machines, has a multi-year backlog. Even with the capital to build new facilities, manufacturers are stuck waiting for the precision tools needed to equip them.

The episode concludes with a somber look at the future of the hobbyist and enthusiast market. As long as the "Big Five" AI players continue to funnel more than 75% of high-end memory into training silos, users like Daniel will continue to see $400 price tags on components that used to cost a fraction of that. The "God Models" of the 21st century are being built, but the cost of their creation is being felt in the pockets of every PC user on the planet.

Downloads

Episode Audio

Download the full episode as an MP3 file

Download MP3
Transcript (TXT)

Plain text transcript file

Transcript (PDF)

Formatted PDF with styling

Read Full Transcript

Episode #608: The RAMpocalypse: Why AI is Starving Your PC

Daniel Daniel's Prompt
Daniel
I've been dealing with a home server saga and discovered that RAM prices have skyrocketed globally. Prices for 32GB kits have jumped from around $130 to over $400. This crisis, dubbed the "RAMpocalypse" or "RAMgeddon," is reportedly being driven by AI training data centers, with OpenAI alone consuming an estimated 40% of the global DRAM supply. How is it possible for the supply chain to reach a point where one company consumes nearly half of the global supply? What is going on with the RAMpocalypse, and is there any hope for prices to return to normal, or should consumers try to wait it out?
Corn
Hey everyone, welcome back to My Weird Prompts. I am Corn, and I am sitting here in our living room in Jerusalem with my brother, the man who probably has more browser tabs open than there are stars in the observable universe. It is a rainy February afternoon here, and the hum of the space heater is almost being drowned out by the sound of Herman’s mechanical keyboard.
Herman
That is a fair assessment. Herman Poppleberry here, and honestly, those tabs are all necessary research. I have been tracking the spot prices for semiconductors since six in the morning, and let me tell you, the charts look like a mountain range that only goes up. It is good to be back, even if the news I am bringing is a bit grim for the wallet.
Corn
So, we have a bit of a situation in the house. Our housemate Daniel has been going through what I can only describe as a home server odyssey. He thought he was in the clear after a motherboard failure last month—he actually managed to find a replacement board for his older Epyc system—but then he hit a wall that I think a lot of people are hitting right now. He was looking to replace some memory that got fried in the power surge, and he found that the market has essentially turned into a post-apocalyptic wasteland.
Herman
It is being called the RAMpocalypse, or RAMgeddon, depending on which corner of the internet you frequent. And Daniel is not exaggerating. The price jumps we are seeing in early twenty twenty-six are absolutely unprecedented in the history of consumer electronics. We are talking about thirty-two gigabyte kits of D-D-R-five that used to be one hundred thirty dollars now pushing four hundred dollars or more. If you are looking for high-density E-C-C modules for a server like Daniel’s, you might as well be trading in precious metals.
Corn
It is wild. I remember when RAM was the cheap part of the build. You would just throw in an extra stick like it was nothing. It was the "free space" on the bingo card of PC building. But Daniel sent us this prompt because he found a statistic that is almost hard to believe. Apparently, OpenAI alone is consuming an estimated forty percent of the global dynamic random access memory supply.
Herman
That is the figure currently circulating in the Q-one industry reports. It sounds like a typo, right? How can one single entity, even one as massive as OpenAI, gobble up nearly half of the entire planet's production of memory? But when you look at the architecture of the clusters they are building right now—specifically the phase five expansion of the Stargate project—the math starts to make a terrifying kind of sense.
Corn
Well, that is what I want to dig into today. I want to understand how the supply chain gets this brittle, what is actually happening in those data centers that requires that much silicon, and whether there is any light at the end of the tunnel for people like Daniel who just want to run a home server without taking out a second mortgage.
Herman
It is a deep rabbit hole, Corn. We are looking at a fundamental shift in how silicon is allocated globally. It is not just a temporary shortage like we saw during the pandemic; it is a structural realignment of the entire industry. The factories haven't stopped working; they've just stopped working for us.
Corn
Okay, so let's start with that forty percent number. That is the one that really sticks in my throat. We are talking about a global market that supplies every laptop, every smartphone, every gaming console, and every server on earth. How does one company, even one running the world's most famous large language models, take up forty percent of that?
Herman
You have to look at the scale of the clusters they are building for training and inference. We aren't talking about a few racks of servers in a rented cage anymore. We are talking about projects like the Stargate supercomputer, which is a joint venture between Microsoft and OpenAI. By some estimates, that single project aims to house over a million G-P-Us. When you train a model with trillions of parameters—like the rumored G-P-T-five or whatever they are calling the latest iteration—you aren't just limited by the processing power of the graphics processing units. You are limited by the memory wall.
Corn
The memory wall. Explain that for the non-engineers.
Herman
Think of the G-P-U as a world-class chef who can chop vegetables at lightning speed. But the chef only has a tiny cutting board. If the ingredients—the data—aren't sitting right there on the board, the chef has to stop, walk to the pantry, and come back. That walk to the pantry is the latency of going to a hard drive or even standard system RAM. To keep the chef working at full speed, you need a massive cutting board. In the AI world, that cutting board is H-B-M, or High Bandwidth Memory.
Corn
Right, because if the data isn't in the memory, the processor is just sitting there idling, waiting for the network to catch up. And at the scale of a hundred-billion-dollar supercomputer, an idling processor is losing you thousands of dollars a second.
Herman
Exactly. And in these high-end AI servers, they aren't using the kind of RAM you and I buy for our desktops. They are using something called H-B-M-three-E right now, and they are already moving toward H-B-M-four. But here is the kicker: the production lines that make H-B-M are the same ones that make the standard D-D-R-four and D-D-R-five memory that goes into Daniel's home server.
Corn
Ah, so it is a zero-sum game. If Samsung or S-K Hynix decides to dedicate a factory line to H-B-M for OpenAI's next big cluster, that is a line that is no longer making the sticks of RAM we see on the shelves at the local computer store.
Herman
Precisely. And it is actually worse than a one-to-one trade. The yield for H-B-M is much lower. It is a more complex manufacturing process where you are stacking memory dies on top of each other and connecting them with thousands of tiny vertical wires called Through-Silicon Vias. Because the stack is so tall and complex, if one layer is bad, the whole stack is often trash. It takes about twice as many wafer starts to get the same amount of usable memory in H-B-M format compared to standard D-D-R-five. So, when OpenAI says they need enough memory to power a cluster of a million G-P-Us, they are effectively pre-ordering the entire capacity of these fabrication plants for months or years in advance.
Corn
So, it isn't just that they are buying the finished product. They are essentially renting the factories and telling the manufacturers to ignore everyone else.
Herman
In some cases, they are literally providing the capital to build the factories. We have seen these massive multi-billion dollar deals where the big AI players basically say, "Here is five billion dollars upfront. Build the capacity, and we get first dibs on everything that comes off the line for the next three years." For the manufacturers like Micron or S-K Hynix, it is a no-brainer. Why sell a thirty-two gigabyte kit to a consumer for a small margin when you can sell the equivalent amount of silicon to an enterprise customer for ten times the price as part of an H-B-M stack?
Corn
It feels like the consumer is being priced out of the silicon age. I mean, if this trend continues, does the concept of a home computer even stay viable? Or do we all just become thin-client users because the hardware is too expensive to own?
Herman
That is the dark side of this, Corn. We have talked about the democratization of technology for decades—the idea that a kid in their bedroom could have the same computing power as a university lab from ten years ago. But the RAMpocalypse is a massive step in the opposite direction. It is a centralization of resources. If forty percent of the supply is going to one company, that means everyone else, from Dell and H-P to the small-time enthusiasts like Daniel, is fighting over the remaining sixty percent. And that sixty percent has to cover every smartphone, every car, every smart fridge, and every laptop on the planet.
Corn
And I imagine the other big players like Google, Meta, and Microsoft aren't exactly sitting on their hands. If OpenAI is taking forty percent, what are the others taking?
Herman
That is the thing. If you add up the top five AI players—OpenAI, Google, Meta, Amazon, and Microsoft—you are looking at a situation where more than seventy-five percent of the global high-end memory supply is being funneled into these massive training silos. This is why Daniel is seeing those four hundred dollar price tags. The supply for the consumer market has been squeezed to a fraction of what it was just two years ago. We are living on the leftovers of the AI giants.
Corn
It reminds me of the graphics card shortage during the crypto mining boom around twenty twenty-one, but this feels different. Crypto was a speculative bubble. This AI demand feels like it's built into the core of how the global economy is trying to function now. It doesn't feel like it's just going to pop and go away.
Herman
You are right. Crypto was about hashing for rewards—it was an external incentive. AI is about the foundational infrastructure of the next era of computing. The companies buying this RAM aren't just trying to make a quick buck; they are trying to build the "God Models" of the twenty-first century. They see this as a winner-take-all race. If you have the most memory and the most compute, you have the best model. If you have the best model, you own the market for everything from medical diagnosis to automated coding. So, they are willing to pay almost any price to secure that supply. If the price of RAM triples, they don't care. They just add another zero to the debt facility.
Corn
Let's talk about the manufacturers for a second. You mentioned Samsung, S-K Hynix, and Micron. These are the "Big Three" that control over ninety percent of the world's D-RAM production. Are they just laughing all the way to the bank? Or are they struggling to keep up with the technical demands too?
Herman
It is a bit of both. Their profits have reached record highs—S-K Hynix recently reported its most profitable quarter in history. But they are also under immense pressure. The technical requirements for H-B-M-three-E are insane. We are talking about stacking twelve or sixteen layers of memory with micron-level precision. The heat management alone is a nightmare. I was reading an analyst report recently where the chief operating officer of Dell, Jeff Clarke, mentioned that they have never witnessed cost escalation at this pace. When even a giant like Dell, who buys millions of units, is complaining about the price of components, you know the manufacturers are in a very powerful, but very stressed, position.
Corn
But surely they are trying to build more factories? If the demand is this high, the logical economic response is to increase supply.
Herman
They are trying, but you can't just snap your fingers and build a semiconductor fab. It takes three to five years and tens of billions of dollars. We have the new fabs coming online in Ohio and Arizona, but those won't reach full capacity until late twenty twenty-six or twenty twenty-seven. And there is another bottleneck: the machines that make the chips. Companies like A-S-M-L, which make the extreme ultraviolet lithography machines, have their own multi-year backlogs. So even if Samsung wants to build a new line today, they might have to wait until twenty twenty-eight just to get the equipment to put in it.
Corn
So we are in this perfect storm where demand has jumped by an order of magnitude, but supply is physically constrained by the laws of construction and high-precision engineering. It’s like trying to widen a highway while the number of cars is doubling every week.
Herman
Exactly. And there is a second-order effect that people don't often think about. When manufacturers shift their focus to the high-end, high-margin AI memory, they stop innovating on the lower-end consumer stuff. Why spend research and development money on making D-D-R-four cheaper or more efficient when you can spend it on making H-B-M-four faster for a customer who will buy every single chip you make? We are seeing a stagnation in the technology that regular people use, even as the price of that technology goes through the roof.
Corn
That is a really depressing thought. It is like the luxury car market taking over so much of the steel supply that even a basic bicycle becomes unaffordable and stops getting better.
Herman
It is a very apt analogy. And it is not just RAM. Daniel mentioned N-A-N-D flash memory too. The same thing is happening there. Large language models need massive amounts of fast storage—petabytes of it—to feed the data into the memory. So, solid state drive prices are also climbing. The entire storage and memory ecosystem is being sucked into the AI gravity well. We are seeing enterprise S-S-Ds being sold for five times their twenty twenty-four prices.
Corn
Okay, so let's look at the numbers Daniel mentioned. He saw a two hundred to four hundred percent price increase. Is that consistent across the board? Or is it specific to certain types of memory?
Herman
It is most severe in the high-capacity kits. If you are looking for sixteen gigabytes of standard D-D-R-five, you might only see a sixty or seventy percent increase. But once you move into the thirty-two, sixty-four, or one hundred twenty-eight gigabyte range—the kind of stuff Daniel needs for his server—the prices go exponential. This is because the high-density chips required for those kits are exactly what the data centers want. They are competing for the same physical silicon dies.
Corn
So Daniel's home server, which probably needs a lot of RAM for virtualization or running local containers, is the exact use case that is getting hit the hardest. He’s essentially competing with Sam Altman for the same piece of silicon.
Herman
Precisely. If you are a gamer and you can get by with sixteen gigabytes, you are annoyed. But if you are a professional or a home lab enthusiast who needs sixty-four gigabytes to run your environment, you are looking at a thousand-dollar bill just for memory. That is more than the cost of the rest of the server combined in many cases. It has turned the economics of home computing upside down.
Corn
I want to go back to the OpenAI thing for a second. Forty percent. If they are consuming forty percent of the world's supply, what does that actually look like in terms of hardware? Are we talking about millions of sticks of RAM arriving at their data centers every week?
Herman
It is more about the integrated modules. Think about the Nvidia Blackwell B-two-hundred chips. Each one of those units has one hundred ninety-two gigabytes of H-B-M-three-E physically bonded to the processor. When OpenAI orders hundreds of thousands of these units for a single cluster, they are effectively taking millions of gigabytes of the highest-quality memory off the market in one fell swoop. And it is not just the training. It is the inference. Every time someone asks an AI a question, that model has to be loaded into memory somewhere. As the user base grows into the hundreds of millions, the amount of RAM needed just to keep the service running becomes astronomical.
Corn
So it is a double whammy. You need the RAM to build the model, which takes months, and then you need even more RAM to let people use it in real-time.
Herman
And as the models get bigger, the RAM requirements grow. We went from models that could fit on a single high-end card to models that require hundreds of cards working in parallel just to generate one sentence. The appetite is insatiable. There is no "enough" in the current AI arms race.
Corn
It makes me wonder about the efficiency of these models. Are we just throwing silicon at a problem that could be solved with better software? Are we being lazy because we assume the hardware will always be there?
Herman
That is the trillion-dollar question. There is a lot of research into model compression, quantization, and more efficient architectures like Mamba or Liquid Neural Networks. But right now, the brute force approach—scaling laws—is what is winning the race. And as long as brute force is the path to the most capable AI, companies will keep buying every scrap of memory they can find. They would rather spend ten billion dollars on RAM than risk falling six months behind a competitor.
Corn
You know, we have done a lot of episodes over the years. I think back to some of our earlier discussions about the cloud and how everything was moving to these centralized data centers. It felt like a convenience then. Now it feels like a resource war.
Herman
It really does. It is a struggle for the fundamental building blocks of computation. And the consumer is currently losing that war because we don't have the collective bargaining power of a Microsoft or a Google. We are just individuals looking at a Newegg listing and crying.
Corn
So what is the play for someone like Daniel? He is sitting there with a server that won't boot because of a memory mismatch, and he is looking at a four hundred dollar replacement cost. Does he wait? Or is this the new normal?
Herman
This is the toughest part. Usually, in the semiconductor industry, things are cyclical. We have a shortage, prices go up, manufacturers build more capacity, we have an oversupply, and prices crash. We have seen this cycle dozens of times since the nineteen seventies. But the AI demand might be what economists call a structural break.
Corn
Meaning the old rules don't apply anymore?
Herman
Meaning the demand curve has shifted so far to the right that the traditional cycle can't keep up. If the demand for AI compute continues to grow at its current pace, we might not see a return to those one hundred thirty dollar kits for a very long time, if ever. We might be entering an era where high-performance local hardware is a luxury item, like a fine watch or a sports car.
Corn
That is a grim outlook, Herman. You are saying there is no light at the end of the tunnel?
Herman
Well, there is some hope. First, as newer types of memory like H-B-M-four come online in twenty twenty-seven, some of the pressure might ease off the older D-D-R-five lines. If the AI giants move to the latest and greatest, they might leave some crumbs for the rest of us.
Corn
"Crumbs." That is a great way to describe the consumer market right now. We’re the birds waiting under the table at the silicon banquet.
Herman
It is. But also, there is the possibility of a plateau in model size. If we hit a point of diminishing returns where making a model ten times bigger doesn't make it ten times better, the insane scramble for more RAM might slow down. But looking at the roadmap for the next eighteen months, I don't see that happening yet. The "Stargate" project is just the beginning.
Corn
What about the secondary market? If new RAM is too expensive, should people be looking at used parts? Or is that market also picked clean?
Herman
The used market is also being affected. When new prices go up, used prices follow. Plus, you have to be careful. A lot of the used enterprise RAM coming onto the market has been run at high temperatures in data centers twenty-four-seven for years. It is a bit of a gamble. But for a home server like Daniel’s, looking for "pulled" modules from decommissioned servers might be the only viable path right now. You just have to be prepared for a higher failure rate.
Corn
It is funny, we always think of technology as getting cheaper and better over time. This is one of the few instances where the opposite is happening. My computer from three years ago might actually be worth more today than when I bought it, just because of the RAM and storage inside.
Herman
It is a weirdly appreciating asset right now. If you have a sixty-four gigabyte kit of D-D-R-four or D-D-R-five sitting in a drawer, you are basically sitting on a small gold bar. I’ve seen people on Reddit trading RAM for high-end graphics cards or even car parts.
Corn
I should go check our spare parts box in the hallway. We might be sitting on a gold mine.
Herman
Don't get too excited, Corn. Most of our stuff is probably D-D-R-three, which is ancient history at this point. It’s like having a collection of very nice V-H-S tapes.
Corn
Fair point. But let's look at the broader implications. If RAM stays this expensive, how does it change the software we use? Do developers start getting more disciplined about memory usage again?
Herman
You would hope so. For the last decade, we have been very lazy with memory. We have these massive web-based applications—looking at you, Electron apps—that eat up gigabytes of RAM just to show a chat window. If RAM stays at four hundred dollars a kit, companies might actually have to hire engineers who know how to optimize code. We might see a return to the efficiency of the late nineties era, out of pure necessity. We might see the "bloatware" era finally come to an end because the hardware can no longer subsidize the bad code.
Corn
That would be a silver lining. I am tired of my browser using eight gigabytes of RAM just to show me a few news articles.
Herman
Exactly. But it also means that new, innovative software that requires a lot of memory might never get off the ground because the hardware requirements are too high for the average user. It could stifle innovation in areas like local AI—where you run the model on your own machine for privacy—or high-end gaming and complex simulations. If only the top one percent can afford the hardware, the market for that software shrinks.
Corn
That is the part that worries me the most. We are building these massive central brains in the cloud, but we are making the individual's computer weaker in the process. It is a shift from personal computing back to a sort of mainframe-and-terminal model. We’re losing our digital autonomy.
Herman
It is very much like the nineteen sixties and seventies. The "Big Iron" is in the data center, and we just get to peek at it through a browser window. The dream of the "Personal Supercomputer" is taking a serious hit right now.
Corn
So, for Daniel and our listeners who are facing this RAMpocalypse, what is the practical advice? If you absolutely need the RAM now to get a system running, do you just bite the bullet?
Herman
If it is for work or a mission-critical project, you might have to. But if it is a hobby project, my advice would be to look for alternatives. Can you optimize your current setup? Can you use "swap space" on a fast N-V-Me drive to compensate, even if it is slower? Or can you look at refurbished enterprise gear that uses older, slightly more available memory standards like D-D-R-four? It won't be as fast, but it will be a lot cheaper.
Corn
And what about the timing? Is there any window in the next year where things might dip?
Herman
Some analysts are pointing to the end of twenty twenty-six as a potential easing point, as several new fabrication plants in the United States and Europe are scheduled to come online. But that is a long time to wait if your server is down today.
Corn
Yeah, that doesn't help Daniel get his Home Assistant and media server running this weekend.
Herman
No, it doesn't. One thing I will say is to keep an eye on the smaller manufacturers. Sometimes they have inventory that hasn't been snatched up by the big players yet. And avoid the high-performance "gaming" brands if you don't need the flashy lights and heat sinks. Plain old "green P-C-B" sticks from companies like Crucial or Samsung are sometimes significantly cheaper because they aren't marketed to the same demographic.
Corn
That is a good tip. Function over form, especially when form costs an extra two hundred dollars.
Herman
Exactly. And hey, if you are listening to this and you have managed to find a reliable source for affordable memory in this climate, let us know through the contact form on our website. We are all in this together, trying to survive the RAMgeddon.
Corn
Yeah, myweirdprompts.com is the place to go. We would love to hear your survival stories or even your "I can't believe I paid this much" stories.
Herman
It is a wild time to be a nerd, Corn. We are seeing history happen in the most literal, silicon-based way possible. We are watching the physical limits of our digital ambition.
Corn
It really is. I mean, forty percent of the global supply for one company. It is a staggering statistic. It speaks to the incredible ambition of these AI projects, but also to the fragility of our global supply chains. We have built a world where a breakthrough in software in San Francisco can make it impossible for a guy in Jerusalem to fix his home server.
Herman
The butterfly effect of the twenty-first century. A transformer model flaps its wings in a data center, and your bank account for computer parts disappears.
Corn
Well put. I think we have covered the what and the why. The "when" is still a bit up in the air, but at least we understand the forces at play. It is not just corporate greed; it is a fundamental shift in what our civilization values in terms of computation. We’ve decided that "knowing everything" is more important than "everyone having a computer."
Herman
And as long as the value of AI continues to skyrocket, the price of the silicon that powers it will follow. It’s the new oil, Corn. And we’re just trying to fill up our tanks.
Corn
It is a sobering thought. But I think it is important for people to understand that they aren't crazy. When you see those prices, it is not your imagination. The world really has changed.
Herman
It has. And we are just trying to navigate it one stick of RAM at a time.
Corn
Well, Herman, thanks for diving into the data on this one. I feel a lot more informed, even if my wallet feels a bit lighter just thinking about it.
Herman
Always a pleasure, Corn. I will keep an eye on the spot prices and let you know if I see any movement. I’ve got a script running that pings the markets every five minutes.
Corn
Of course you do. And to everyone listening, thanks for joining us on this deep dive into the RAMpocalypse. If you have been enjoying the show, we would really appreciate it if you could leave us a review on your podcast app or on Spotify. It genuinely helps other people find the show and helps us keep doing this.
Herman
It really does. We love seeing your feedback and hearing your thoughts on these prompts. It makes the research worth it.
Corn
Absolutely. This has been My Weird Prompts. You can find all our past episodes, including our deep dives into the semiconductor industry from last year, at myweirdprompts.com. We are also on Spotify and wherever you get your podcasts.
Herman
Thanks to Daniel for sending in this prompt and sparking such a great discussion. We will see you all in the next episode.
Corn
Take care, everyone. And good luck with your hardware builds. You are going to need it.
Herman
Bye for now.
Corn
Goodbye.

This episode was generated with AI assistance. Hosts Herman and Corn are AI personalities.