Role Overview
The Technical Architect - Databricks designs and implements scalable data architectures and solutions. The jobholder has expertise in Databricks Lakehouse, data modeling, and cloud integration, ensuring high performance, security, and reliability.
Responsibilities
- Design and implement Databricks-based data architectures to meet business requirements.
- Develop and optimize data pipelines using PySpark, Scala, or SQL.
- Establish the Databricks Lakehouse architecture for batch and streaming data.
- Collaborate with cross-functional teams to integrate Databricks with cloud platforms (e.g., AWS, Azure, GCP).
- Ensure data security and compliance with best practices.
- Monitor and troubleshoot Databricks environments for performance and reliability.
- Stay updated on Databricks advancements and industry trends.
Key Technical Skills & Responsibilities
12+ years of experience in data engineering using Databricks or Apache Spark-based platforms.Proven track record of building and optimizing ETL / ELT pipelines for batch and streaming data ingestion.Hands-on experience with Azure services such as Azure Data Factory, Azure Data Lake Storage, Azure Synapse Analytics, or Azure SQL Data Warehouse.Proficiency in programming languages such as Python, Scala, or SQL for data processing and transformation.Expertise in Spark (PySpark, Spark SQL, or Scala) and Databricks notebooks for large-scale data processing.Familiarity with Delta Lake, Delta Live Tables, and medallion architecture for data lakehouse implementations.Build and query deltalake storage solutionsProcess streaming data with Azure Databricks structured streamingDesign Azure Databricks security and data protection solutionsFlatten nested structures and explode arrays with sparkTransfer data outside using sparkpools using pyspark connectorOptimizing spark jobsImplementing best practices in spark / databricksExperience with orchestration tools like Azure Data Factory or Databricks Jobs for scheduling and automation.Knowledge of Git for source control and CI / CD integration for Databricks workflows, cost optimization, performance tuning.Familiarity with Unity Catalog, RBAC, or enterprise-level Databricks setups.Ability to create reusable components, templates, and documentation to standardize data engineering workflows.Solutioning and presales - Architecting frameworks, defining roadmaps, and engaging with stakeholders.Experience in defining data strategy, evaluating new tools / technologies, and driving adoption across the organization.Must have experience of working with streaming data sources and Kafka (preferred).Eligibility Criteria
Bachelor's degree in computer science, Information Technology, or related fieldProven experience as a Databricks Architect or similar roleComplete knowledge in Azure Databricks platform architectureDatabricks certification (e.g., Certified Data Engineer, Associate Developer)Expertise in Python / Scala / SQL / RExperience with cloud platforms like AWS, Azure, or GCPStrong understanding of data modeling and cloud integrationExperience with cluster sizing and security implementationExcellent problem-solving and communication skillsSkills Required
Spark SQL, Scala, Pyspark, Apache Spark, Kafka, Sql, Git, Azure Data Factory, Azure Synapse Analytics, Databricks