This device is not compatible.
PROJECT
Setting up a Streaming Data Pipeline With Kafka
In this project, we’ll learn about configuring and starting the Kafka server. To check the working of the Kafka server, we’ll begin by creating a topic on our server. After creating a topic, we’ll create a new producer and consumer to demonstrate data transmission using Kafka.
You will learn to:
Configure Zookeeper and Kafka
Start Zookeeper and Kafka
Create a topic using a console
Carry out console-based streaming between producers and consumers
Skills
Data Pipeline Engineering
Live Streaming
Prerequisites
Basic knowledge of Python
Basic knowledge of streaming
Basic knowledge of message broker architecture.
Basic knowledge of Kafka and Zookeeper
Technologies
Kafka
Python
Project Description
Real-time data processing requires systems that handle high-throughput message flows with low latency. Apache Kafka is the industry-standard stream-processing platform for building real-time data pipelines, enabling applications to publish, subscribe to, and process event streams at scale.
In this project, we'll set up a complete Kafka streaming pipeline using Apache Kafka and Zookeeper for distributed coordination. Zookeeper manages the Kafka cluster by tracking node status, topic metadata, and partition assignments, ensuring reliable message delivery across distributed systems. We'll configure both services on localhost, create Kafka topics for message organization, and implement console-based producers and consumers to verify end-to-end data flow through the streaming pipeline.
We'll start by configuring Zookeeper for service synchronization, then launch the Kafka broker and create topics using the Kafka CLI. We'll build a producer to publish messages and a consumer to subscribe and process them, demonstrating the publish-subscribe pattern for event-driven architecture. By the end, we'll have hands-on experience with Kafka setup, Zookeeper configuration, topic management, producer-consumer patterns, and stream processing applicable to any real-time analytics, event streaming, or data integration system.
Project Tasks
1
Installation and Configuration
Task 0: Getting Started
Task 1: Configure Kafka Broker Network Settings
2
Starting Zookeeper and Kafka
Task 2: Start Zookeeper
Task 3: Start Kafka
3
Creating a Topic, Producer, and Consumer
Task 4: Creating a New Topic
Task 5: Things To Know
Task 6: Creating a New Producer
Task 7: Create a New Consumer
Congratulations!
Subscribe to project updates
Andres G.
Senior Software Engineer @ Citi
Atabek BEKENOV
Senior Software Engineer
Pradip Pariyar
Senior Software Engineer
Relevant Courses
Use the following content to review prerequisites or explore specific concepts in detail.