Required Skills :
Must be proficient in the major languages such as Scala, Spark, and Kafka (any variant). Write elegant and maintainable code, and be comfortable with picking up new technologies.
Proficient in working with distributed systems and have experience with different distributed processing frameworks that can handle data in batch and near real-time pipelines.
Experience in working with AWS cloud services and Databricks to build end-to-end data lakehouse solutions that bring different systems together
Capability to understand and implement metadata-based data processing framework and methodologies
Experience in working with CICD process - Jenkins, & GIT Actions
Experience in working with monitoring and control stack - Prometheus / Grafana
Key job responsibilities :
Drive the backbone of our organizational data platform by building robust pipelines that turn complex data into actionable insights using AWS and the Databricks platform
Design scalable, maintainable data systems with strong observability
Monitor, log, alert, and ensure system robustness
Build efficient pipelines for structured and unstructured data
Interface with Product engineering, regulatory, and legal teams as necessary to ensure Lakehouse's compliance adherence
Write clean, efficient code that handles massive amounts of structured and unstructured data
Interested candidates can share their updated resume to
Senior Data Engineer • chennai, India