About ADM :
We are one of the world’s largest nutrition companies and a global leader in human and animal nutrition. We unlock the power of nature to provide nourishing quality of life by transforming crops into ingredients and solutions for foods, beverages, supplements, livestock, aquaculture, and pets.
About ADM India Hub :
At ADM, we have long recognized the strength and potential of India’s talent pool, which is why we have maintained a presence in the country for more than 25 years. Building on this foundation, we have now established ADM India Hub, our first GCC in India.
At ADM India Hub, we are hiring for IT and finance roles across diverse technology and business functions. We stand at the intersection of global expertise and local excellence, enabling us to drive innovation and support our larger purpose of unlocking the power of nature to enrich quality of life.
Job Title : Senior Data Engineer
Overview :
We are seeking a skilled and motivated Azure Databricks Data Engineer to join our dynamic team. The ideal candidate will have strong experience with Python, Spark programming, and expertise in building and optimizing data pipelines in Azure Databricks. You will play a pivotal role in leveraging Databricks workflows, Databricks Asset Bundles, and CI / CD pipelines using GitHub to deliver high-performance data solutions. A solid understanding of Data Warehousing and Data Mart architecture in Databricks is critical for success in this role. If you’re passionate about data engineering, cloud technologies, and scalable data architecture, we’d love to hear from you!
Key Responsibilities :
Required Qualifications :
Python and Spark Programming :
- Minimum of 7 years of experience in Python programming, especially in data engineering, ETL processes, and distributed computing.
- Solid experience using Apache Spark (PySpark) for large-scale data processing and transformation within Databricks.
- Proficiency in writing and optimizing Spark-based jobs for high performance on large datasets.
Databricks Workflows :
Strong hands-on experience with Databricks “Workflows” for orchestrating data pipelines and batch processes.Ability to design and optimize multi-step workflows with task dependencies, retries, and monitoring.Databricks Asset Bundles :
Experience in creating and managing Databricks Asset Bundles to promote reusability and modularization of notebooks, libraries, and models.CI / CD for Databricks Artifacts using GitHub :
Experience with implementing CI / CD pipelines for Databricks using GitHub and GitHub Actions for automating deployment of notebooks, jobs, and libraries.Expertise in version control practices and integrating Databricks with external Git repositories for collaborative development.Data Warehousing & Data Mart Experience :
Strong experience in designing and implementing Data Warehouses and Data Marts using Databricks and Spark.Understanding of dimensional modeling (star and snowflake schemas) and the ability to create optimized data structures for reporting and analytics.Hands-on experience integrating data from multiple sources and managing the ETL process within a Data Warehouse or Data Mart environment.Cloud Experience :
Solid experience working with the Azure ecosystem, including Azure Data Lake, Azure Blob Storage, and Azure SQL Database.Experience working in cloud environments and leveraging cloud-based tools for building and managing data pipelines.Data Engineering Best Practices :
Knowledge of best practices for designing and managing scalable, efficient, and cost-effective data pipelines.Experience in performance tuning and query optimization within Databricks and Spark.Collaboration and Communication :
Excellent teamwork and communication skills, with the ability to collaborate effectively across cross-functional teams.Ability to document technical processes and communicate progress and results to stakeholders.Preferred Qualifications :
Cloud Certifications :
Azure certifications, particularly in Databricks, Data Engineering, or Cloud Solutions, are a plus.Big Data Technologies :Familiarity with other big data tools such as Kafka, Hadoop, or Flink for streaming and real-time data processing is a plus.Data Science / ML Experience :
Exposure to machine learning workflows and model management within Databricks (e.g., using MLflow) is beneficial.