BBC Radio 5 live’s award winning gaming podcast, discussing the world of video games and games culture.
…
continue reading
内容由Tobias Macey提供。所有播客内容(包括剧集、图形和播客描述)均由 Tobias Macey 或其播客平台合作伙伴直接上传和提供。如果您认为有人在未经您许可的情况下使用您的受版权保护的作品,您可以按照此处概述的流程进行操作https://zh.player.fm/legal。
Player FM -播客应用
使用Player FM应用程序离线!
使用Player FM应用程序离线!
Accelerated Computing in Modern Data Centers With Datapelago
Manage episode 470343101 series 3449056
内容由Tobias Macey提供。所有播客内容(包括剧集、图形和播客描述)均由 Tobias Macey 或其播客平台合作伙伴直接上传和提供。如果您认为有人在未经您许可的情况下使用您的受版权保护的作品,您可以按照此处概述的流程进行操作https://zh.player.fm/legal。
Summary
In this episode of the Data Engineering Podcast Rajan Goyal, CEO and co-founder of Datapelago, talks about improving efficiencies in data processing by reimagining system architecture. Rajan explains the shift from hyperconverged to disaggregated and composable infrastructure, highlighting the importance of accelerated computing in modern data centers. He discusses the evolution from proprietary to open, composable stacks, emphasizing the role of open table formats and the need for a universal data processing engine, and outlines Datapelago's strategy to leverage existing frameworks like Spark and Trino while providing accelerated computing benefits.
Announcements
Parting Question
…
continue reading
In this episode of the Data Engineering Podcast Rajan Goyal, CEO and co-founder of Datapelago, talks about improving efficiencies in data processing by reimagining system architecture. Rajan explains the shift from hyperconverged to disaggregated and composable infrastructure, highlighting the importance of accelerated computing in modern data centers. He discusses the evolution from proprietary to open, composable stacks, emphasizing the role of open table formats and the need for a universal data processing engine, and outlines Datapelago's strategy to leverage existing frameworks like Spark and Trino while providing accelerated computing benefits.
Announcements
- Hello and welcome to the Data Engineering Podcast, the show about modern data management
- Data migrations are brutal. They drag on for months—sometimes years—burning through resources and crushing team morale. Datafold's AI-powered Migration Agent changes all that. Their unique combination of AI code translation and automated data validation has helped companies complete migrations up to 10 times faster than manual approaches. And they're so confident in their solution, they'll actually guarantee your timeline in writing. Ready to turn your year-long migration into weeks? Visit dataengineeringpodcast.com/datafold today for the details.
- Your host is Tobias Macey and today I'm interviewing Rajan Goyal about how to drastically improve efficiencies in data processing by re-imagining the system architecture
- Introduction
- How did you get involved in the area of data management?
- Can you start by outlining the main factors that contribute to performance challenges in data lake environments?
- The different components of open data processing systems have evolved from different starting points with different objectives. In your experience, how has that un-planned and un-synchronized evolution of the ecosystem hindered the capabilities and adoption of open technologies?
- The introduction of a new cross-cutting capability (e.g. Iceberg) has typically taken a substantial amount of time to gain support across different engines and ecosystems. What do you see as the point of highest leverage to improve the capabilities of the entire stack with the least amount of co-ordination?
- What was the motivating insight that led you to invest in the technology that powers Datapelago?
- Can you describe the system design of Datapelago and how it integrates with existing data engines?
- The growth in the generation and application of unstructured data is a notable shift in the work being done by data teams. What are the areas of overlap in the fundamental nature of data (whether structured, semi-structured, or unstructured) that you are able to exploit to bridge the processing gap?
- What are the most interesting, innovative, or unexpected ways that you have seen Datapelago used?
- What are the most interesting, unexpected, or challenging lessons that you have learned while working on Datapelago?
- When is Datapelago the wrong choice?
- What do you have planned for the future of Datapelago?
Parting Question
- From your perspective, what is the biggest gap in the tooling or technology for data management today?
- Datapelago
- MIPS Architecture
- ARM Architecture
- AWS Nitro
- Mellanox
- Nvidia
- Von Neumann Architecture
- TPU == Tensor Processing Unit
- FPGA == Field-Programmable Gate Array
- Spark
- Trino
- Iceberg
- Delta Lake
- Hudi
- Apache Gluten
- Intermediate Representation
- Turing Completeness
- LLVM
- Amdahl's Law
- LSTM == Long Short-Term Memory
462集单集
Manage episode 470343101 series 3449056
内容由Tobias Macey提供。所有播客内容(包括剧集、图形和播客描述)均由 Tobias Macey 或其播客平台合作伙伴直接上传和提供。如果您认为有人在未经您许可的情况下使用您的受版权保护的作品,您可以按照此处概述的流程进行操作https://zh.player.fm/legal。
Summary
In this episode of the Data Engineering Podcast Rajan Goyal, CEO and co-founder of Datapelago, talks about improving efficiencies in data processing by reimagining system architecture. Rajan explains the shift from hyperconverged to disaggregated and composable infrastructure, highlighting the importance of accelerated computing in modern data centers. He discusses the evolution from proprietary to open, composable stacks, emphasizing the role of open table formats and the need for a universal data processing engine, and outlines Datapelago's strategy to leverage existing frameworks like Spark and Trino while providing accelerated computing benefits.
Announcements
Parting Question
…
continue reading
In this episode of the Data Engineering Podcast Rajan Goyal, CEO and co-founder of Datapelago, talks about improving efficiencies in data processing by reimagining system architecture. Rajan explains the shift from hyperconverged to disaggregated and composable infrastructure, highlighting the importance of accelerated computing in modern data centers. He discusses the evolution from proprietary to open, composable stacks, emphasizing the role of open table formats and the need for a universal data processing engine, and outlines Datapelago's strategy to leverage existing frameworks like Spark and Trino while providing accelerated computing benefits.
Announcements
- Hello and welcome to the Data Engineering Podcast, the show about modern data management
- Data migrations are brutal. They drag on for months—sometimes years—burning through resources and crushing team morale. Datafold's AI-powered Migration Agent changes all that. Their unique combination of AI code translation and automated data validation has helped companies complete migrations up to 10 times faster than manual approaches. And they're so confident in their solution, they'll actually guarantee your timeline in writing. Ready to turn your year-long migration into weeks? Visit dataengineeringpodcast.com/datafold today for the details.
- Your host is Tobias Macey and today I'm interviewing Rajan Goyal about how to drastically improve efficiencies in data processing by re-imagining the system architecture
- Introduction
- How did you get involved in the area of data management?
- Can you start by outlining the main factors that contribute to performance challenges in data lake environments?
- The different components of open data processing systems have evolved from different starting points with different objectives. In your experience, how has that un-planned and un-synchronized evolution of the ecosystem hindered the capabilities and adoption of open technologies?
- The introduction of a new cross-cutting capability (e.g. Iceberg) has typically taken a substantial amount of time to gain support across different engines and ecosystems. What do you see as the point of highest leverage to improve the capabilities of the entire stack with the least amount of co-ordination?
- What was the motivating insight that led you to invest in the technology that powers Datapelago?
- Can you describe the system design of Datapelago and how it integrates with existing data engines?
- The growth in the generation and application of unstructured data is a notable shift in the work being done by data teams. What are the areas of overlap in the fundamental nature of data (whether structured, semi-structured, or unstructured) that you are able to exploit to bridge the processing gap?
- What are the most interesting, innovative, or unexpected ways that you have seen Datapelago used?
- What are the most interesting, unexpected, or challenging lessons that you have learned while working on Datapelago?
- When is Datapelago the wrong choice?
- What do you have planned for the future of Datapelago?
Parting Question
- From your perspective, what is the biggest gap in the tooling or technology for data management today?
- Datapelago
- MIPS Architecture
- ARM Architecture
- AWS Nitro
- Mellanox
- Nvidia
- Von Neumann Architecture
- TPU == Tensor Processing Unit
- FPGA == Field-Programmable Gate Array
- Spark
- Trino
- Iceberg
- Delta Lake
- Hudi
- Apache Gluten
- Intermediate Representation
- Turing Completeness
- LLVM
- Amdahl's Law
- LSTM == Long Short-Term Memory
462集单集
所有剧集
×欢迎使用Player FM
Player FM正在网上搜索高质量的播客,以便您现在享受。它是最好的播客应用程序,适用于安卓、iPhone和网络。注册以跨设备同步订阅。