.webp)
San Mateo, California. Sometime in 2012.
Two men stand before a whiteboard covered in diagrams, arrows, and crossed-out ideas. They've been at this for months. Coffee cups stack up. Takeout containers accumulate. The diagrams keep getting more complex, then simpler, then complex again.
Benoît Dageville and Thierry Cruanes are arguing. Again.
They've left Oracle—the database empire where they spent over a decade as data architects. They've left behind stable salaries, stock options, the comfort of being cogs in a very profitable machine. Now they're in a cramped office, drawing boxes and arrows, trying to solve a problem that the entire data industry has declared unsolvable.
The prevailing wisdom is clear: Hadoop is the future. Big data means distributed files, MapReduce jobs, armies of engineers wrestling with clusters. Everyone knows this.
Dageville and Cruanes think everyone is wrong.
"We were convinced," Dageville would later say, "there were many reasons why Hadoop would fail."
Their bet: the cloud can do something that on-premise hardware never could. It can provide infinite elasticity—computing resources that appear when you need them and disappear when you don't. If you could build a data warehouse that truly embraced this elasticity, you wouldn't need Hadoop. You wouldn't need anything the industry was building.
You'd need something entirely new.
Sixteen Years in the Empire
Benoît Dageville joined Oracle in January 1996, arriving at the legendary Redwood Shores campus in California. He was a database researcher, a PhD from France, passionate about the arcane mathematics of query optimization and execution engines.
Oracle in the 1990s was the center of the database universe. Every major company ran Oracle. Every serious database engineer wanted to work there. Dageville spent sixteen years in the trenches—building, optimizing, learning the limits of relational databases at scale.
It was at Oracle that he met Thierry Cruanes.
The Bromance
"Thierry and I were very good friends working as architects in the core Oracle database team," Dageville later explained.
They weren't just colleagues. They were intellectual partners—"bros," as Dageville would affectionately call them. They spent years debating database architecture, arguing about performance bottlenecks, dreaming about what databases could be if you weren't constrained by the decisions made decades earlier.
Two Revolutions
By 2012, Dageville and Cruanes saw something that Oracle's leadership seemed to miss: two revolutions were happening simultaneously.
The first was the cloud. Amazon Web Services had launched years earlier, but by 2012 it was clear that cloud computing wasn't a fad. It was a fundamental shift in how infrastructure worked. "For us, the cloud was a miracle for analytics, resources, and elasticity," Dageville said.
The second was big data. Companies were generating data at rates that traditional warehouses couldn't handle. The answer the industry had settled on was Hadoop—an open-source framework for distributed processing that required armies of engineers to maintain.
Dageville and Cruanes saw the opportunity: What if you could handle big data without Hadoop? What if the cloud's elasticity could solve the problem more elegantly?
Oracle wasn't pursuing this path. The empire was focused on protecting its existing business.
So the two architects decided to leave.
July 2012: The Founding
In July 2012, Dageville and Cruanes officially founded Snowflake in San Mateo, California. They brought in a third co-founder: Marcin Żukowski, a database researcher from Poland who had co-founded Vectorwise (later acquired by Actian). Żukowski brought expertise in columnar storage and vectorized execution—the low-level optimizations that would make Snowflake fast.
The name "Snowflake" came from the idea that every data workload is unique—no two snowflakes are alike.
The Architecture Debates
For months, the founders did nothing but argue about architecture. They scrawled designs on whiteboards, tore them up, started over.
The key insight emerged: separate everything.
Traditional data warehouses coupled storage and compute on the same machines. If you needed more computing power, you had to buy more storage. If you needed more storage, you had to buy more compute. Everything was tangled together.
Snowflake would be different. Three layers, completely independent:
Storage Layer: All data lives in cheap cloud object storage (like Amazon S3), compressed in a columnar format optimized for analytics.
Compute Layer: "Virtual warehouses"—clusters of compute that spin up when needed and disappear when not. You could run ten warehouses simultaneously, each handling different workloads, none interfering with the others.
Services Layer: The brain—handling metadata, query optimization, security, transaction management.
Each layer could scale independently. Pay for storage separately from compute. Spin up massive clusters for five minutes, then shut them down. The economics of data analysis would fundamentally change.
Stealth Mode
For two full years—2012 to 2014—Snowflake operated in stealth mode. No public announcements. No marketing. Just building.
"For us, it was about building a product, not a company," Dageville said. "We created a company as collateral damage."
They needed time. The architecture was ambitious. They were betting against Hadoop, against the conventional wisdom, against what every VC and analyst was saying about the future of big data.
They couldn't afford to tip off Oracle or Amazon or any of the incumbents who might see what they were building and crush them.
The Funding
Mike Speiser at Sutter Hill Ventures saw the vision early. In 2012, Sutter Hill led a $5 million Series A. Speiser was so committed he became Snowflake's first CEO—a placeholder until they could find a permanent executive.
In 2014, they found their leader: Bob Muglia, a former Microsoft executive who had run the company's server and tools division. With Muglia in place, Snowflake emerged from stealth.
The Launch
In 2014, Snowflake made its public debut. The pitch was audacious: forget Hadoop. Forget the complexity. Just run SQL queries, like you always have, but now they'll work at any scale.
The early customers were skeptical. Everyone had been told Hadoop was the future. Everyone had investments in Hadoop ecosystems. Snowflake was saying: none of that matters.
But when they ran queries—queries that took hours on traditional systems, queries that required Hadoop clusters and teams of engineers—and saw results in seconds... the skepticism evaporated.
The Cloud-Native Advantage
Snowflake's secret weapon was being born in the cloud.
Competitors like Amazon Redshift had taken existing data warehouse architectures and ported them to AWS. The underlying design was still rooted in the on-premise era. Snowflake was designed from scratch for cloud economics. Every architectural decision assumed infinite elasticity.
This meant:
- No hardware provisioning
- No capacity planning
- No downtime for maintenance
- No fighting for resources when multiple teams needed analytics
You just... ran queries. The infrastructure handled itself.
The Workload Isolation Miracle
One feature became legendary: workload isolation.
In traditional data warehouses, if the CFO ran a massive report, the data scientists' models would slow down. Everyone competed for the same resources. With Snowflake, you could spin up separate virtual warehouses for different teams. The CFO's queries ran in complete isolation from data science. No interference. No contention.
For large companies with multiple analytics teams, this was transformative.
September 16, 2020
On September 16, 2020, Snowflake went public on the New York Stock Exchange. The IPO was, at the time, the largest software IPO in history.
The stock opened at $245—more than double the IPO price of $120. The company's market cap reached $70 billion on its first day of trading. Warren Buffett's Berkshire Hathaway, legendary for avoiding tech stocks, had invested $735 million in the IPO.
The Oracle architects who bet against Hadoop—who spent months arguing over whiteboards in San Mateo—had built one of the most valuable data companies in history.
The Dageville Legacy
Benoît Dageville remained with Snowflake as President of Products, overseeing the continued evolution of the platform. He became a billionaire multiple times over.
But when asked about the journey, he kept returning to the same theme: they built a product, and the company was "collateral damage."
The database nerds from Oracle had changed how the world stores and analyzes data. Not by following the consensus. By proving it wrong.
Where Snowflake Is Now
Today, Snowflake powers analytics for thousands of companies—from Fortune 500 enterprises to startups. The platform handles petabytes of data, serves millions of queries, and has become the de facto standard for cloud data warehousing.
The separation of storage and compute—the insight that emerged from heated arguments on a San Mateo whiteboard—became the blueprint for an entire industry.
Every major cloud provider now offers some version of this architecture. The idea that seemed contrarian in 2012 became obvious in retrospect.
That's how the best ideas work. They look crazy until they don't.