Job Role : -AWS Data Engineer
Job Location : -Pune / Hyderabad / Chennai / Mysuru / Bhubaneswar / Mangalore / Trivandrum / Chandigarh / Jaipur / Nagpur / Indore / Gurgaon
Experience : - 7+ Years
Job Roles & Responsibilities : -
- Design, develop, and maintain data pipelines and assets on AWS.
- Optimize and refactor legacy PySpark / Spark SQL code for performance and maintainability.
- Implement unit testing / TDD to ensure robust, bug-free code.
- Debug and resolve complex performance, concurrency, and logic issues .
- Manage code versioning and repositories (Git, JFrog Artifactory).
- Leverage AWS services (S3, EC2, Lambda, Redshift, CloudFormation) for scalable data solutions.
Job Skills & Requirements : -
7+ years hands-on experience in Python, PySpark, Boto3 , and related frameworks / libraries.Proven expertise in Spark SQL & PySpark optimization .Strong knowledge of AWS architecture (S3, EC2, Lambda, Redshift, CloudFormation).Experience in code refactorization for clean, maintainable solutions.Familiarity with Git, JFrog Artifactory , and modern CI / CD practices.Strong debugging and problem-solving skills.Solid understanding of unit testing and TDD methodologies .