Talent.com
This job offer is not available in your country.
Technical Architect - Data Streaming Platforms

Technical Architect - Data Streaming Platforms

Delphie Consulting servicesIndia
30+ days ago
Job description

Job Summary :

We are seeking a highly skilled Technical Architect with deep expertise in real-time data streaming platforms, especially within the Apache Kafka ecosystem, to lead the architecture and implementation of scalable, secure, and high-throughput event-driven systems. This role involves end-to-end solution design and optimization of streaming data pipelines using technologies like Kafka, Kafka Connect, Spark Streaming, Flink, Beam, and cloud-native tools such as Google Cloud Pub / Sub.

You will work with cross-functional teams, including data engineers, DevOps, product managers, and cloud architects, to define robust and high-performance streaming architectures that power data-intensive applications and analytics.

Key Responsibilities :

Kafka Architecture & Management :

  • Design and architect scalable, fault-tolerant, and high-throughput streaming systems using Apache Kafka.
  • Build and manage Kafka clusters for high availability, scalability, disaster recovery, and performance.
  • Optimize cluster configuration parameters, partitioning strategies, consumer group behavior, and replication factors.

Streaming Pipeline Development :

  • Architect and integrate Kafka Streams, Kafka Connect, Apache Flink, Beam, and Spark Streaming to enable complex real-time data processing pipelines.
  • Design end-to-end pipelines that support ingestion, transformation, enrichment, validation, and delivery of streaming data across systems.
  • Platform Integration & Ecosystem :

  • Implement and manage integrations with Schema Registry, Kafka Connect connectors, GCP Pub / Sub, and third-party systems.
  • Drive seamless data interoperability across operational databases, data lakes, cloud services, and analytics platforms.
  • Security, Monitoring, and Compliance :

  • Enforce security best practices for Kafka, including authentication, authorization (ACLs), TLS encryption, and audit logging.
  • Implement observability solutions using monitoring tools (e.g., Prometheus, Grafana, Confluent Control Center, etc.) to track cluster health, latency, throughput, lag, and performance.
  • Ensure compliance with enterprise policies and governance standards.
  • Collaboration & Leadership :

  • Translate business and technical requirements into scalable Kafka-based architectural solutions.
  • Collaborate with development and DevOps teams to automate Kafka deployments using infrastructure-as-code (e.g., Terraform, Ansible, Helm).
  • Serve as a subject matter expert, mentor, and advisor for engineering teams on Kafka and stream processing best practices.
  • Stay current with industry trends, contribute to architectural roadmaps, and evaluate emerging technologies in the real-time data space.
  • Required Technical Skills :

  • Apache Kafka (core) architecture, broker configuration, topic management, partitioning, replication, consumer groups, offset management.
  • Kafka Connect development and management of source / sink connectors (custom and Confluent-provided).
  • Schema Registry schema design (Avro / Protobuf), evolution, compatibility settings.
  • Kafka Streams, Apache Flink, Beam, Spark Streaming hands-on experience in real-time stream processing.
  • Cloud Platforms especially GCP (Google Cloud Pub / Sub, Dataflow, BigQuery); exposure to AWS or Azure is a plus.
  • Strong understanding of distributed systems, event-driven architecture, message serialization formats (Avro, Protobuf, JSON), and data consistency models.
  • Hands-on with monitoring, logging, and alerting tools for real-time systems.
  • Proficient in Java or Scala; Python is a bonus.
  • Experience with containerization and orchestration tools like Docker, Kubernetes, and Helm.
  • (ref : hirist.tech)

    Create a job alert for this search

    Platform Architect • India