It’s all about Data Pipelines. Join Pure Storage Field Solution Architect Chad Hendron and Solutions Director Andrew Silifant for a deep dive into the evolution of data management, focusing on the Data Lakehouse architecture and its role in the age of AI and ML. Our discussion looks at the Data Lakehouse as a powerful combination of a data lake and a data warehouse, solving problems like "data swamps” and proprietary formats of older systems. Viewers will learn about technological advancements, such as object storage and open table formats, that have made this new architecture possible, allowing for greater standardization and multiple tooling functions to access the same data.
Our guests also explore current industry trends, including a look at Dremio's 2025 report showing the rapid adoption of Data Lakehouses, particularly as a replacement for older, inefficient systems like cloud data warehouses and traditional data lakes. Gain insight into the drivers behind this migration, including the exponential growth of unstructured data and the need to control cloud expenditure by being more prescriptive about what data is stored in the cloud versus on-premises. Andrew provides a detailed breakdown of processing architectures and the critical importance of meeting SLAs to avoid costly and frustrating pipeline breaks in regulated industries like banking.
Finally, we provide practical takeaways and a real-world case study. Chad shares a customer success story about replacing a large, complex Hadoop cluster with a streamlined Dremio and Pure Storage solution, highlighting the massive reduction in physical space, power consumption, and management complexity. Both guests emphasize the need for better governance practices to manage cloud spend and risk. Andrew underscores the essential, full-circle role of databases—from the "alpha" of data creation to the "omega" of feature stores and vector databases for modern AI use cases like Retrieval-Augmented Generation (RAG). Tune in to understand how a holistic data strategy, including Pure’s Enterprise Data Cloud, can simplify infrastructure and future-proof your organization for the next wave of data-intensive workloads.
To learn more, visit https://www.purestorage.com/solutions/ai/data-warehouse-streaming-analytics.html
Check out the new Pure Storage digital customer community to join the conversation with peers and Pure experts:
https://purecommunity.purestorage.com/
00:00 Intro and Welcome
03:15 Data Lakehouse Primer
08:31 Stat of the Episode on Lakehouse Usage
10:50 Challenges with Data Pipeline access
13:58 Assessing Organization Success with Data Cleaning
16:07 Use Cases for the Data Lakehouse
20:41 Case Study on Data Lakehouse Use Case
24:11 Hot Takes Segment
All content for The Pure Report is the property of Pure Storage and is served directly from their servers
with no modification, redirects, or rehosting. The podcast is not affiliated with or endorsed by Podjoint in any way.
It’s all about Data Pipelines. Join Pure Storage Field Solution Architect Chad Hendron and Solutions Director Andrew Silifant for a deep dive into the evolution of data management, focusing on the Data Lakehouse architecture and its role in the age of AI and ML. Our discussion looks at the Data Lakehouse as a powerful combination of a data lake and a data warehouse, solving problems like "data swamps” and proprietary formats of older systems. Viewers will learn about technological advancements, such as object storage and open table formats, that have made this new architecture possible, allowing for greater standardization and multiple tooling functions to access the same data.
Our guests also explore current industry trends, including a look at Dremio's 2025 report showing the rapid adoption of Data Lakehouses, particularly as a replacement for older, inefficient systems like cloud data warehouses and traditional data lakes. Gain insight into the drivers behind this migration, including the exponential growth of unstructured data and the need to control cloud expenditure by being more prescriptive about what data is stored in the cloud versus on-premises. Andrew provides a detailed breakdown of processing architectures and the critical importance of meeting SLAs to avoid costly and frustrating pipeline breaks in regulated industries like banking.
Finally, we provide practical takeaways and a real-world case study. Chad shares a customer success story about replacing a large, complex Hadoop cluster with a streamlined Dremio and Pure Storage solution, highlighting the massive reduction in physical space, power consumption, and management complexity. Both guests emphasize the need for better governance practices to manage cloud spend and risk. Andrew underscores the essential, full-circle role of databases—from the "alpha" of data creation to the "omega" of feature stores and vector databases for modern AI use cases like Retrieval-Augmented Generation (RAG). Tune in to understand how a holistic data strategy, including Pure’s Enterprise Data Cloud, can simplify infrastructure and future-proof your organization for the next wave of data-intensive workloads.
To learn more, visit https://www.purestorage.com/solutions/ai/data-warehouse-streaming-analytics.html
Check out the new Pure Storage digital customer community to join the conversation with peers and Pure experts:
https://purecommunity.purestorage.com/
00:00 Intro and Welcome
03:15 Data Lakehouse Primer
08:31 Stat of the Episode on Lakehouse Usage
10:50 Challenges with Data Pipeline access
13:58 Assessing Organization Success with Data Cleaning
16:07 Use Cases for the Data Lakehouse
20:41 Case Study on Data Lakehouse Use Case
24:11 Hot Takes Segment
Taming the Beast: Managing HPC & AI Data at Scale with Zero Move Tiering on FlashBlade
The Pure Report
47 minutes 51 seconds
4 months ago
Taming the Beast: Managing HPC & AI Data at Scale with Zero Move Tiering on FlashBlade
Data management in high-performance computing (HPC) and AI environments presents unique challenges, especially with heavy file usage and evolving data lifecycles. This edition of the Pure Report podcast delves into these complexities, exploring how data transitions through development, simulation/training, inference/optimization, and long-term retention phases. Together with Pure technologists Bikash Choudhury and Don Poorman, we’ll tackle the critical problem statement: reducing total cost of ownership (TCO) by automatically moving inactive data to different tiers without sacrificing accessibility or performance during crucial operational stages.
Join us as we introduce Pure Storage's innovative Zero Move Tiering (ZMT) solution. Discover how ZMT delivers dynamic Service Level Agreements (SLAs) across performance and capacity tiers without data movement, leading to lower TCO and zero operational overhead. We'll also examine how ZMT fits into various data lifecycle stages, from development build workloads to AI-augmented modeling and simulation, and compare Pure's approach to the common deficiencies seen in other vendor solutions.
The Pure Report
It’s all about Data Pipelines. Join Pure Storage Field Solution Architect Chad Hendron and Solutions Director Andrew Silifant for a deep dive into the evolution of data management, focusing on the Data Lakehouse architecture and its role in the age of AI and ML. Our discussion looks at the Data Lakehouse as a powerful combination of a data lake and a data warehouse, solving problems like "data swamps” and proprietary formats of older systems. Viewers will learn about technological advancements, such as object storage and open table formats, that have made this new architecture possible, allowing for greater standardization and multiple tooling functions to access the same data.
Our guests also explore current industry trends, including a look at Dremio's 2025 report showing the rapid adoption of Data Lakehouses, particularly as a replacement for older, inefficient systems like cloud data warehouses and traditional data lakes. Gain insight into the drivers behind this migration, including the exponential growth of unstructured data and the need to control cloud expenditure by being more prescriptive about what data is stored in the cloud versus on-premises. Andrew provides a detailed breakdown of processing architectures and the critical importance of meeting SLAs to avoid costly and frustrating pipeline breaks in regulated industries like banking.
Finally, we provide practical takeaways and a real-world case study. Chad shares a customer success story about replacing a large, complex Hadoop cluster with a streamlined Dremio and Pure Storage solution, highlighting the massive reduction in physical space, power consumption, and management complexity. Both guests emphasize the need for better governance practices to manage cloud spend and risk. Andrew underscores the essential, full-circle role of databases—from the "alpha" of data creation to the "omega" of feature stores and vector databases for modern AI use cases like Retrieval-Augmented Generation (RAG). Tune in to understand how a holistic data strategy, including Pure’s Enterprise Data Cloud, can simplify infrastructure and future-proof your organization for the next wave of data-intensive workloads.
To learn more, visit https://www.purestorage.com/solutions/ai/data-warehouse-streaming-analytics.html
Check out the new Pure Storage digital customer community to join the conversation with peers and Pure experts:
https://purecommunity.purestorage.com/
00:00 Intro and Welcome
03:15 Data Lakehouse Primer
08:31 Stat of the Episode on Lakehouse Usage
10:50 Challenges with Data Pipeline access
13:58 Assessing Organization Success with Data Cleaning
16:07 Use Cases for the Data Lakehouse
20:41 Case Study on Data Lakehouse Use Case
24:11 Hot Takes Segment