Experience : 5 to 8 years
Job description : Python API + Data Engineer
We are looking for a self-motivated Python Data Engineer with strong skills in backend / API development and large-scale data processing , who can work independently and thrive in a fast-paced, agile environment.
Key Responsibilities
- Design, develop, and optimize data pipelines using PySpark, AWS Glue, and EMR .
- Build and maintain Python backend / API services (FastAPI, async preferred).
- Support ML / MLOps workflows leveraging AWS SageMaker and other cloud-native tools.
- Write clean, testable, and scalable code with unit testing (pytest) and CI / CD practices.
- Work with AWS services (S3, Athena, Glue, EMR, SageMaker) for data engineering and model deployment.
- Collaborate cross-functionally to provide production support, troubleshooting, and performance optimization .
- Use modern DevOps tools (Git, Docker, Kubernetes, Terraform) to automate deployments and ensure reliability.
- Communicate effectively with global teams and adapt quickly to changing requirements.
Required Skills & Experience
Strong expertise in Python programming .Hands-on experience with PySpark and large-scale data processing.Proven experience in Python API / backend development (FastAPI, REST / async).Proficiency in AWS cloud services : Glue, EMR, SageMaker, S3, Athena.Exposure to ML / MLOps workflows and production-level deployment.Solid knowledge of SQL / PostgreSQL and database query optimization.Familiarity with unit testing (pytest), Git, Docker, Kubernetes, Terraform .Strong problem-solving, production support, and communication skills.Work Environment
Remote-friendly, with expectation of independent ownership and accountability.Timing : 2 PM - 11 PM IST ( 4 hours overlap with EST)