Artwork

内容由Confluent, founded by the original creators of Apache Kafka® and Founded by the original creators of Apache Kafka®提供。所有播客内容(包括剧集、图形和播客描述)均由 Confluent, founded by the original creators of Apache Kafka® and Founded by the original creators of Apache Kafka® 或其播客平台合作伙伴直接上传和提供。如果您认为有人在未经您许可的情况下使用您的受版权保护的作品,您可以按照此处概述的流程进行操作https://zh.player.fm/legal
Player FM -播客应用
使用Player FM应用程序离线!

Security for Real-Time Data Stream Processing with Confluent Cloud

48:33
 
分享
 

Manage episode 346008718 series 2355972
内容由Confluent, founded by the original creators of Apache Kafka® and Founded by the original creators of Apache Kafka®提供。所有播客内容(包括剧集、图形和播客描述)均由 Confluent, founded by the original creators of Apache Kafka® and Founded by the original creators of Apache Kafka® 或其播客平台合作伙伴直接上传和提供。如果您认为有人在未经您许可的情况下使用您的受版权保护的作品,您可以按照此处概述的流程进行操作https://zh.player.fm/legal

Streaming real-time data at scale and processing it efficiently is critical to cybersecurity organizations like SecurityScorecard. Jared Smith, Senior Director of Threat Intelligence, and Brandon Brown, Senior Staff Software Engineer, Data Platform at SecurityScorecard, discuss their journey from using RabbitMQ to open-source Apache Kafka® for stream processing. As well as why turning to fully-managed Kafka on Confluent Cloud is the right choice for building real-time data pipelines at scale.
SecurityScorecard mines data from dozens of digital sources to discover security risks and flaws with the potential to expose their client’ data. This includes scanning and ingesting data from a large number of ports to identify suspicious IP addresses, exposed servers, out-of-date endpoints, malware-infected devices, and other potential cyber threats for more than 12 million companies worldwide.
To allow real-time stream processing for the organization, the team moved away from using RabbitMQ to open-source Kafka for processing a massive amount of data in a matter of milliseconds, instead of weeks or months. This makes the detection of a website’s security posture risk happen quickly for constantly evolving security threats. The team relied on batch pipelines to push data to and from Amazon S3 as well as expensive REST API based communication carrying data between systems. They also spent significant time and resources on open-source Kafka upgrades on Amazon MSK.
Self-maintaining the Kafka infrastructure increased operational overhead with escalating costs. In order to scale faster, govern data better, and ultimately lower the total cost of ownership (TOC), Brandon, lead of the organization’s Pipeline team, pivoted towards a fully-managed, cloud-native approach for more scalable streaming data pipelines, and for the development of a new Automatic Vendor Detection (AVD) product.
Jared and Brandon continue to leverage the Cloud for use cases including using PostgreSQL and pushing data to downstream systems using CSC connectors, increasing data governance and security for streaming scalability, and more.
EPISODE LINKS

  continue reading

章节

1. Intro (00:00:00)

2. What is SecurityScorecard? (00:01:42)

3. Scaling with Kafka clusters (00:06:36)

4. Moving from RabbitMQ to Kafka and Confluent Cloud (00:16:36)

5. Upsides of using Kafka for data processing (00:24:50)

6. Downsides of ingesting large amounts of data (00:27:12)

7. Leveraging the Cloud to increase data governance and security (00:37:08)

8. Helping teams create and consume new feeds of data (00:41:25)

9. It's a wrap! (00:47:01)

265集单集

Artwork
icon分享
 
Manage episode 346008718 series 2355972
内容由Confluent, founded by the original creators of Apache Kafka® and Founded by the original creators of Apache Kafka®提供。所有播客内容(包括剧集、图形和播客描述)均由 Confluent, founded by the original creators of Apache Kafka® and Founded by the original creators of Apache Kafka® 或其播客平台合作伙伴直接上传和提供。如果您认为有人在未经您许可的情况下使用您的受版权保护的作品,您可以按照此处概述的流程进行操作https://zh.player.fm/legal

Streaming real-time data at scale and processing it efficiently is critical to cybersecurity organizations like SecurityScorecard. Jared Smith, Senior Director of Threat Intelligence, and Brandon Brown, Senior Staff Software Engineer, Data Platform at SecurityScorecard, discuss their journey from using RabbitMQ to open-source Apache Kafka® for stream processing. As well as why turning to fully-managed Kafka on Confluent Cloud is the right choice for building real-time data pipelines at scale.
SecurityScorecard mines data from dozens of digital sources to discover security risks and flaws with the potential to expose their client’ data. This includes scanning and ingesting data from a large number of ports to identify suspicious IP addresses, exposed servers, out-of-date endpoints, malware-infected devices, and other potential cyber threats for more than 12 million companies worldwide.
To allow real-time stream processing for the organization, the team moved away from using RabbitMQ to open-source Kafka for processing a massive amount of data in a matter of milliseconds, instead of weeks or months. This makes the detection of a website’s security posture risk happen quickly for constantly evolving security threats. The team relied on batch pipelines to push data to and from Amazon S3 as well as expensive REST API based communication carrying data between systems. They also spent significant time and resources on open-source Kafka upgrades on Amazon MSK.
Self-maintaining the Kafka infrastructure increased operational overhead with escalating costs. In order to scale faster, govern data better, and ultimately lower the total cost of ownership (TOC), Brandon, lead of the organization’s Pipeline team, pivoted towards a fully-managed, cloud-native approach for more scalable streaming data pipelines, and for the development of a new Automatic Vendor Detection (AVD) product.
Jared and Brandon continue to leverage the Cloud for use cases including using PostgreSQL and pushing data to downstream systems using CSC connectors, increasing data governance and security for streaming scalability, and more.
EPISODE LINKS

  continue reading

章节

1. Intro (00:00:00)

2. What is SecurityScorecard? (00:01:42)

3. Scaling with Kafka clusters (00:06:36)

4. Moving from RabbitMQ to Kafka and Confluent Cloud (00:16:36)

5. Upsides of using Kafka for data processing (00:24:50)

6. Downsides of ingesting large amounts of data (00:27:12)

7. Leveraging the Cloud to increase data governance and security (00:37:08)

8. Helping teams create and consume new feeds of data (00:41:25)

9. It's a wrap! (00:47:01)

265集单集

Alle afleveringen

×
 
Loading …

欢迎使用Player FM

Player FM正在网上搜索高质量的播客,以便您现在享受。它是最好的播客应用程序,适用于安卓、iPhone和网络。注册以跨设备同步订阅。

 

快速参考指南