It’s all about Data Pipelines. Join Pure Storage Field Solution Architect Chad Hendron and Solutions Director Andrew Silifant for a deep dive into the evolution of data management, focusing on the Data Lakehouse architecture and its role in the age of AI and ML. Our discussion looks at the Data Lakehouse as a powerful combination of a data lake and a data warehouse, solving problems like "data swamps” and proprietary formats of older systems. Viewers will learn about technological advancements, such as object storage and open table formats, that have made this new architecture possible, allowing for greater standardization and multiple tooling functions to access the same data.
Our guests also explore current industry trends, including a look at Dremio's 2025 report showing the rapid adoption of Data Lakehouses, particularly as a replacement for older, inefficient systems like cloud data warehouses and traditional data lakes. Gain insight into the drivers behind this migration, including the exponential growth of unstructured data and the need to control cloud expenditure by being more prescriptive about what data is stored in the cloud versus on-premises. Andrew provides a detailed breakdown of processing architectures and the critical importance of meeting SLAs to avoid costly and frustrating pipeline breaks in regulated industries like banking.
Finally, we provide practical takeaways and a real-world case study. Chad shares a customer success story about replacing a large, complex Hadoop cluster with a streamlined Dremio and Pure Storage solution, highlighting the massive reduction in physical space, power consumption, and management complexity. Both guests emphasize the need for better governance practices to manage cloud spend and risk. Andrew underscores the essential, full-circle role of databases—from the "alpha" of data creation to the "omega" of feature stores and vector databases for modern AI use cases like Retrieval-Augmented Generation (RAG). Tune in to understand how a holistic data strategy, including Pure’s Enterprise Data Cloud, can simplify infrastructure and future-proof your organization for the next wave of data-intensive workloads.
To learn more, visit https://www.purestorage.com/solutions/ai/data-warehouse-streaming-analytics.html
Check out the new Pure Storage digital customer community to join the conversation with peers and Pure experts:
https://purecommunity.purestorage.com/
00:00 Intro and Welcome
03:15 Data Lakehouse Primer
08:31 Stat of the Episode on Lakehouse Usage
10:50 Challenges with Data Pipeline access
13:58 Assessing Organization Success with Data Cleaning
16:07 Use Cases for the Data Lakehouse
20:41 Case Study on Data Lakehouse Use Case
24:11 Hot Takes Segment
All content for The Pure Report is the property of Pure Storage and is served directly from their servers
with no modification, redirects, or rehosting. The podcast is not affiliated with or endorsed by Podjoint in any way.
It’s all about Data Pipelines. Join Pure Storage Field Solution Architect Chad Hendron and Solutions Director Andrew Silifant for a deep dive into the evolution of data management, focusing on the Data Lakehouse architecture and its role in the age of AI and ML. Our discussion looks at the Data Lakehouse as a powerful combination of a data lake and a data warehouse, solving problems like "data swamps” and proprietary formats of older systems. Viewers will learn about technological advancements, such as object storage and open table formats, that have made this new architecture possible, allowing for greater standardization and multiple tooling functions to access the same data.
Our guests also explore current industry trends, including a look at Dremio's 2025 report showing the rapid adoption of Data Lakehouses, particularly as a replacement for older, inefficient systems like cloud data warehouses and traditional data lakes. Gain insight into the drivers behind this migration, including the exponential growth of unstructured data and the need to control cloud expenditure by being more prescriptive about what data is stored in the cloud versus on-premises. Andrew provides a detailed breakdown of processing architectures and the critical importance of meeting SLAs to avoid costly and frustrating pipeline breaks in regulated industries like banking.
Finally, we provide practical takeaways and a real-world case study. Chad shares a customer success story about replacing a large, complex Hadoop cluster with a streamlined Dremio and Pure Storage solution, highlighting the massive reduction in physical space, power consumption, and management complexity. Both guests emphasize the need for better governance practices to manage cloud spend and risk. Andrew underscores the essential, full-circle role of databases—from the "alpha" of data creation to the "omega" of feature stores and vector databases for modern AI use cases like Retrieval-Augmented Generation (RAG). Tune in to understand how a holistic data strategy, including Pure’s Enterprise Data Cloud, can simplify infrastructure and future-proof your organization for the next wave of data-intensive workloads.
To learn more, visit https://www.purestorage.com/solutions/ai/data-warehouse-streaming-analytics.html
Check out the new Pure Storage digital customer community to join the conversation with peers and Pure experts:
https://purecommunity.purestorage.com/
00:00 Intro and Welcome
03:15 Data Lakehouse Primer
08:31 Stat of the Episode on Lakehouse Usage
10:50 Challenges with Data Pipeline access
13:58 Assessing Organization Success with Data Cleaning
16:07 Use Cases for the Data Lakehouse
20:41 Case Study on Data Lakehouse Use Case
24:11 Hot Takes Segment
In this episode, we welcome Lead Principal Technologist Hari Kannan to cut through the noise and tackle some of the biggest myths surrounding AI data management and the revolutionary FlashBlade//EXA platform. With GPU shipments now outstripping CPUs, the foundation of modern AI is shifting, and legacy storage architectures are struggling to keep up. Hari dives into the implications of this massive GPU consumption, setting the stage for why a new approach is desperately needed for companies driving serious AI initiatives.
Hari dismantles three critical myths that hold IT leaders back. First, he discusses how traditional storage is ill-equipped for modern AI's millions of small, concurrent files, where metadata performance is the true bottleneck—a problem FlashBlade//EXA solves with its metadata-data separation and single namespace. Second, he addresses the outdated notion that high-performance AI is file-only, highlighting FlashBlade//EXA's unified, uncompromising delivery of both file and object storage at exabyte scale and peak efficiency. Finally, Hari explains that GPUs are only as good as the data they consume, countering the belief that only raw horsepower matters. FlashBlade//EXA addresses this by delivering reliable, scalable throughput, efficient DirectFlash Modules up to 300 TB, and the metadata performance required to keep expensive GPUs fully utilized and models training faster.
Join us as we explore the blind spots in current AI data strategies during our "Hot Takes" segment and recount a favorite FlashBlade success story. Hari closes with a compelling summary of how Pure Storage's complete portfolio is perfectly suited to provide the complementary data management essential for scaling AI. Tune in to discover why FlashBlade//EXA is the non-compromise, exabyte-scale solution built to keep your AI infrastructure running at its full potential.
For more information, visit: https://www.pure.ai/flashblade-exa.html
Check out the new Pure Storage digital customer community to join the conversation with peers and Pure experts:
https://purecommunity.purestorage.com/
00:00 Intro and Welcome
04:30 Primer on FlashBlade
11:32 Stat of the Episode on GPU Shipments
13:25 What is FlashBlade//EXA
18:58 Myth #1: Traditional Storage Challenges for AI Data
22:01 Myth #2: AI Workloads are not just File-based
26:42: Myth #3: AI Needs more than just GPUs
31:35 Hot Takes Segment
The Pure Report
It’s all about Data Pipelines. Join Pure Storage Field Solution Architect Chad Hendron and Solutions Director Andrew Silifant for a deep dive into the evolution of data management, focusing on the Data Lakehouse architecture and its role in the age of AI and ML. Our discussion looks at the Data Lakehouse as a powerful combination of a data lake and a data warehouse, solving problems like "data swamps” and proprietary formats of older systems. Viewers will learn about technological advancements, such as object storage and open table formats, that have made this new architecture possible, allowing for greater standardization and multiple tooling functions to access the same data.
Our guests also explore current industry trends, including a look at Dremio's 2025 report showing the rapid adoption of Data Lakehouses, particularly as a replacement for older, inefficient systems like cloud data warehouses and traditional data lakes. Gain insight into the drivers behind this migration, including the exponential growth of unstructured data and the need to control cloud expenditure by being more prescriptive about what data is stored in the cloud versus on-premises. Andrew provides a detailed breakdown of processing architectures and the critical importance of meeting SLAs to avoid costly and frustrating pipeline breaks in regulated industries like banking.
Finally, we provide practical takeaways and a real-world case study. Chad shares a customer success story about replacing a large, complex Hadoop cluster with a streamlined Dremio and Pure Storage solution, highlighting the massive reduction in physical space, power consumption, and management complexity. Both guests emphasize the need for better governance practices to manage cloud spend and risk. Andrew underscores the essential, full-circle role of databases—from the "alpha" of data creation to the "omega" of feature stores and vector databases for modern AI use cases like Retrieval-Augmented Generation (RAG). Tune in to understand how a holistic data strategy, including Pure’s Enterprise Data Cloud, can simplify infrastructure and future-proof your organization for the next wave of data-intensive workloads.
To learn more, visit https://www.purestorage.com/solutions/ai/data-warehouse-streaming-analytics.html
Check out the new Pure Storage digital customer community to join the conversation with peers and Pure experts:
https://purecommunity.purestorage.com/
00:00 Intro and Welcome
03:15 Data Lakehouse Primer
08:31 Stat of the Episode on Lakehouse Usage
10:50 Challenges with Data Pipeline access
13:58 Assessing Organization Success with Data Cleaning
16:07 Use Cases for the Data Lakehouse
20:41 Case Study on Data Lakehouse Use Case
24:11 Hot Takes Segment