Description : About the Role :
We are seeking a highly skilled and motivated Databricks Consultant to support our data engineering and analytics initiatives. The ideal candidate will have a strong background in data pipeline development, Delta Lake optimization, and data governance using Unity Catalog on the Databricks Lakehouse Platform. Youll work with large datasets, support complex data transformations, and contribute to building scalable and secure data architecture solutions using the Azure ecosystem.
This is a remote, contract-based position with long-term potential.
Key Responsibilities :
- Design and develop highly scalable, reliable, and secure data pipelines using Azure Databricks and Azure Data Factory (ADF).
- Implement and optimize Delta Lake architecture for large-scale data ingestion, transformation, and real-time analytics.
- Work with Databricks Unity Catalog for data discovery, access control, governance, and cross-workspace data management.
- Perform advanced data wrangling and Spark transformations on large structured and semi-structured datasets.
- Develop and orchestrate complex ETL / ELT workflows with ADF and Apache Airflow, including dynamic and parameterized pipelines.
- Build and optimize SQL queries, stored procedures, and indexing strategies to support data integration with SQL Server and other relational databases.
- Participate in data modeling activities involving Star Schema, Snowflake Schema, Facts, and Dimensions.
- Apply data validation and unit testing methodologies to ensure pipeline and transformation integrity.
- Collaborate with Data Architects, Business Analysts, and other stakeholders to gather requirements and deliver data solutions aligned with business goals.
- Follow SDLC best practices : version control, code reviews, CI / CD (using Azure DevOps), and automated deployment of data pipelines.
Required Skills & Experience :
3 - 6 years of professional experience in Data Engineering, Data Integration, or Big Data Development.Strong experience with :
Azure Databricks (Spark, Delta Lake, MLflow)Azure Data Factory (ADF)Unity Catalog for data governance and access managementProficiency in PySpark, SQL, and experience with large-scale data processing and tuning.In-depth knowledge of RDBMS (SQL Server) - strong SQL development skills, query tuning, and schema management.Solid understanding of ETL / ELT concepts, data modeling techniques (transactional and dimensional), and data warehousing principles.Experience with at least two end-to-end data integration or migration projects using the Azure data stack.Familiarity with CI / CD practices, source control (Git), and deployment pipelines using Azure DevOps.Good understanding of data testing frameworks, writing test cases for transformations, and validating data output.(ref : hirist.tech)