About the job :
As part of our Innovation Team, we are seeking a Certified Senior Databricks Engineer / Tech Lead with 78 years of hands-on experience in building scalable data platforms.
This role will focus on designing, building, and operationalizing data solutions on the Databricks platform to accelerate advanced analytics and AI use cases.
Key Responsibilities :
- Architect, develop, productionize and maintain end to end solutions in Databricks
- Implement and optimize ETL / ELT processes for structured and semi-structured data
- Leverage Delta Lake for ACID transactions, data versioning, and time-travel features
- Drive adoption of the Lakehouse architecture to unify data warehousing and AI / ML workloads
- Implement CI / CD pipelines using Databricks Repos, Asset Bundles, and integration with DevOps tools
- Configure and enforce Unity Catalog for secure, governed access to data assets
- Design and implement data quality and validation frameworks to ensure trusted data
- Lead performance tuning and optimization efforts for Spark jobs and queries
- Integrate with external systems such as Kafka, Event Hub, and REST APIs for real-time and batch processing
- Collaborate with data scientists and business stakeholders to build feature-rich datasets and reusable assets
- Troubleshoot and debug complex data workflows in development and production environments
- Guide junior engineers and contribute to best practices in data engineering and platform usage
- Ensure platform security, access controls, and compliance with enterprise data governance standards.
Required Skills :
Expertise in Apache Spark and Databricks platformExperience with Databricks Lakehouse architectureDelta Lake conceptsProficient in PySpark, SQL, and Delta LakeStrong knowledge of Data Engineering conceptsExperience with data ingestion, ETL / ELT pipelinesFamiliarity with Unity Catalog and data governanceHands-on with Databricks Notebooks and JobsCI / CD automation with Databricks Repos and DevOps, Asset BundlesDatabricks Asset Bundle implementation knowledgeStrong understanding of performance tuning in SparkData quality and validation framework implementationExperience in handling structured, semi-structured dataProficient in debugging and troubleshootingCollaboration with data scientists and analystsGood understanding of security and access controlExperience with Mosaic AI or Databricks ML capabilitiesExposure to streaming pipelines using Structured StreamingFamiliarity with data observability and lineage tools(ref : hirist.tech)