Job Description :
We are seeking a highly skilled AWS Data Engineer with strong hands-on experience in Python, PySpark, and a wide range of AWS services. The ideal candidate should have excellent analytical skills and the ability to develop scalable data processing solutions while meeting complex business requirements.
Key Responsibilities :
- Develop and maintain data pipelines using AWS services such as S3, Lambda, Glue, EMR, Athena, SNS, IAM, etc.
- Write efficient and scalable code using Python and PySpark for big data processing.
- Perform in-depth SQL queries to analyze data and fulfill complex business needs.
- Work with tools such as EMR Notebooks, MapReduce, and Visual Studio Code or equivalent.
- Collaborate with cross-functional teams to gather requirements and deliver robust data solutions.
- Troubleshoot and optimize existing data workflows for performance and scalability.
Required Skills :
Strong hands-on experience with AWS services including : S3, Lambda, EMR, SNS, IAM, Glue, Athena, etc.Proven development experience with Python and PySpark.Proficiency in SQL for data analysis and transformation.Familiarity with EMR Notebooks, MapReduce processing, and modern development tools like VS Code.Strong analytical and problem-solving abilities.Experience :
5+ years of experience in data engineering or relevant field.(ref : hirist.tech)