Job Title : Associate Architect - Data Engineering
Experience : 6 to 9 Years
Location : Bangalore / Mumbai (any) (Hybrid WFO)
About the Role :
We are looking for a skilled Data Engineers with hands-on experience in the Microsoft Azure
data ecosystem. The ideal candidate will play a critical role in designing, developing, and
managing data pipelines and transformation frameworks that enable analytics, reporting, and
business intelligence capabilities across the enterprise.
You will be working closely with data architects, platform engineers, and DevOps teams to build
scalable, resilient, and efficient data engineering solutions using modern Azure tools and best
practices.
Key Responsibilities :
- Design and build scalable ETL / ELT pipelines using Azure Data Factory (ADF), Azure Databricks (Spark), and Azure Synapse Analytics.
- Develop and optimize PySpark / Spark SQL jobs for large-scale batch and streaming data transformations.
- Ingest data from various sources including Apache Kafka, REST APIs, and RDBMS, ensuring real-time or near-real-time availability.
- Implement data modeling strategies (star schema, snowflake schema) for analytics consumption layers in Synapse or ADLS.
- Collaborate with DevOps teams to automate deployment using CI / CD pipelines (Azure DevOps, GitHub Actions, etc.).
- Monitor, troubleshoot, and optimize data workflows for performance, cost-efficiency, and reliability.
- Follow coding standards, participate in peer reviews, and maintain version-controlled code in Git repositories.
- Support data quality checks, logging, alerting, and observability mechanisms for production workloads.
- Participate in sprint ceremonies and contribute to task estimation and delivery planning.
Must-Have Skills :
5+ years of experience in data engineering roles.Hands-on experience with :○ Azure Data Factory (ADF) – building pipelines, triggers, linked services.
○ Azure Databricks – building and managing Spark jobs in PySpark.
○ Azure Synapse Analytics – data warehousing, SQL queries, workspace
orchestration.
○ Apache Kafka – consuming and processing real-time data streams.
Strong in SQL, Python, and Spark for data manipulation and transformation.Exposure to CI / CD practices (Azure DevOps, Git workflows, build / release pipelines).Understanding of data lake architecture and modern data warehousing principles.Nice-to-Have Skills :
Experience with Delta Lake, Event Hub, or Azure Stream Analytics.Familiarity with infrastructure-as-code tools like Terraform or ARM templates.Exposure to Data Governance, Data Quality frameworks, or metadata management.Understanding of networking, security, and access control in Azure data servicesPreferred Certifications :
Microsoft Certified : Azure Data Engineer AssociateDatabricks Certified Associate Developer for Apache SparkAzure Fundamentals (AZ-900) – optional but helpful