About the Role :
We are looking for a skilled Data Engineer to design and build scalable data solutions that power analytics, reporting, and machine learning. You will work with a modern data stack including Python , SQL , Apache Airflow , Apache Spark , and AWS services to manage and transform large datasets.
Responsibilities :
- Develop robust data workflows and pipelines using Python and SQL .
- Orchestrate data processing jobs with Apache Airflow .
- Perform large-scale data processing using Apache Spark .
- Design and implement data solutions on AWS , including S3, Lambda, Glue , and Redshift .
- Ensure data reliability, availability, and performance for downstream applications.
- Collaborate with engineering, analytics, and data science teams to provide clean, accessible data.
- Maintain and optimize data infrastructure for scalability and cost-efficiency.
Required Skills :
3+ years of experience in data engineering or a similar role.Proficient in Python and advanced SQL .Hands-on experience with Airflow for scheduling and orchestration.Strong understanding of distributed data processing with Apache Spark .Solid experience with AWS services : S3, Lambda, Glue , etc.Ability to troubleshoot data issues and optimize data workflows.ref : hirist.tech)