InfoQ Homepage Presentations Developing Real-time Data Pipelines with Apache Kafka
Developing Real-time Data Pipelines with Apache Kafka
Summary
Joe Stein makes an introduction for developers about why and how to use Apache Kafka. Apache Kafka is a publish-subscribe messaging system rethought of as a distributed commit log. Kafka is designed to allow a single cluster to serve as the central data backbone. A single Kafka broker can handle hundreds of megabytes of reads and writes per second from thousands of clients.
Bio
Joe Stein is CEO at Elodina Inc.
About the conference
This is a one-of-a-kind conference for application developers, solution and data architects: people who develop business applications, create multi-device aware web applications, process vast quantities of enterprise data, design cloud architectures, and manage high performance infrastructure. The sessions are specifically tailored for Developers and Architects using the popular open source Spring IO Projects, Groovy & Grails, Cloud Foundry, RabbitMQ, Redis, Geode, Hadoop and Tomcat technologies. Whether you're building mission-critical web or business applications, crunching huge amounts of distributed data, or designing the next killer cloud native application, SpringOne2GX will keep you up to date with the latest enterprise technology.
This content is in the AI, ML & Data Engineering topic
Related Topics:
Sponsored Content
-
Related Editorial
-
Related Sponsors
-
Popular across InfoQ
-
AWS Introduces ECS Managed Instances for Containerized Applications
-
Producing a Better Software Architecture with Residuality Theory
-
GitHub Introduces New Embedding Model to Improve Code Search and Context
-
Google DeepMind Introduces CodeMender, an AI Agent for Automated Code Repair
-
Building Distributed Event-Driven Architectures across Multi-Cloud Boundaries
-
Elena Samuylova on Large Language Model (LLM)-Based Application Evaluation and LLM as a Judge
-