Job Title : Kafka Developer
Location : Noida / Chennai / Pune
Education : B.E. / B.Tech.
Year of Experience : 8-10 Years
About us
At
R Systems
we are shaping the future of technology by designing cutting-edge software
#products , platforms, and
#digital
experiences that drive business growth for our clients.
Our product mindset and advanced engineering capabilities in
#Cloud , Data,
#AI,
and Customer Experience empower us to deliver innovative solutions to key players across the high-tech industry. This includes ISVs, SaaS, and Internet companies, as well as leading organizations in
#telecom ,
#media ,
#healthcare ,
#finance , and
#manufacturing .
We are
Great Place To Work Certified
in 10 countries where we have a full-time workforce - India, the USA, Canada, Poland, Romania, Moldova, Indonesia, Singapore, Malaysia, and Thailand.
Mandatory Requirements :
A degree in Computer Science, IT, or a related field.
Kafka Expertise – Strong knowledge of Kafka architecture, brokers, producers, consumers, and stream processing.
Programming Skills – Proficiency in Java, Scala, or Python for developing Kafka-based applications.
Big Data & Streaming Technologies – Experience with Spark, Flink, or Apache Storm is a plus.
Database Knowledge – Familiarity with SQL and NoSQL databases like Cassandra, MongoDB, or PostgreSQL.
Cloud & DevOps – Experience with cloud platforms (AWS, Azure, GCP) and Kubernetes / Docker.
Event-Driven Architecture – Understanding of event-driven and microservices architectures.
Monitoring & Debugging – Experience with Kafka monitoring tools like Confluent Control Center, Kafka Manager, or ELK stack.
Security & Scalability – Knowledge of Kafka security, access control, and scaling strategies.
Problem-Solving & Communication – Strong analytical skills and ability to work in cross-functional teams
Roles & Responsibilities :
Kafka Application Development – Design, develop, and maintain real-time data streaming applications using Apache Kafka.
Topic Management – Create and manage Kafka topics, partitions, and replication factors.
Data Pipeline Development – Build and optimize data pipelines for real-time data processing.
Producer & Consumer Implementation – Develop Kafka producers and consumers for seamless data exchange between systems.
Integration & Connectivity – Integrate Kafka with databases, cloud services, and other messaging platforms.
Performance Optimization – Monitor and fine-tune Kafka clusters for low latency and high throughput.
Security & Compliance – Implement security best practices, including SSL, SASL, and authentication mechanisms.
Cluster Administration – Manage Kafka clusters, brokers, ZooKeeper, and ensure high availability.
Monitoring & Logging – Use tools like Prometheus, Grafana, or Kafka Manager for monitoring and troubleshooting.
Documentation & Best Practices – Maintain documentation for Kafka configurations, data flow, and architectural decisions.
Mandatory Skills
Java, Apache Kafka, Kafka Streams, Kafka, Apache Zookeeper, Kafka Cluster, Kafka Administrator, Microservices
Developer Kafka • India