Role Overview
We are seeking a skilled and motivated Data Engineer with 5–8 years of experience in building scalable data pipelines using Python, PySpark, and AWS services. The ideal candidate will have hands-on expertise in big data processing, orchestration using AWS Step Functions, and serverless computing with AWS Lambda. Familiarity with DynamoDB and deployment of ETL programs in AWS is essential.
Key Responsibilities
Design, develop, and maintain robust data pipelines using Python and PySpark
Handle large-scale data processing and transformation using AWS services
Implement orchestration workflows using AWS Step Functions
Develop and manage serverless components using AWS Lambda
Deploy and monitor ETL programs in AWS environments
Configure and optimize DynamoDB for data storage and retrieval
Collaborate with cross-functional teams to understand data requirements and deliver scalable solutions
Ensure data quality, integrity, and security across all stages of the pipeline
Required Skills & Qualifications
5–8 years of experience in data engineering or related field
Strong proficiency in Python and PySpark
Solid understanding of AWS services including S3, Lambda, Step Functions, Glue, and DynamoDB Experience deploying and managing ETL workflows in AWS
Familiarity with NoSQL databases, especially DynamoDB
Knowledge of CI / CD practices and infrastructure-as-code tools (e.g., CloudFormation, Terraform) is a plus
Excellent problem-solving and communication skills
Ability to work independently in a remote setup
What We Offer
Fully remote work environment
Opportunity to work on cutting-edge data engineering projects
Collaborative and inclusive team culture
Competitive compensation and benefits
Data Engineer • Secunderabad, Telangana, India