Experience : 5 to 8 years
Job description :
Python API + Data Engineer
We are looking for a
self-motivated Python Data Engineer
with strong skills in
backend / API development and large-scale data processing , who can work independently and thrive in a fast-paced, agile environment.
Key Responsibilities
Design, develop, and optimize
data pipelines
using
PySpark, AWS Glue, and EMR .
Build and maintain
Python backend / API services
(FastAPI, async preferred).
Support
ML / MLOps workflows
leveraging
AWS SageMaker
and other cloud-native tools.
Write clean, testable, and scalable code with
unit testing (pytest)
and CI / CD practices.
Work with
AWS services
(S3, Athena, Glue, EMR, SageMaker) for data engineering and model deployment.
Collaborate cross-functionally to provide
production support, troubleshooting, and performance optimization .
Use
modern DevOps tools
(Git, Docker, Kubernetes, Terraform) to automate deployments and ensure reliability.
Communicate effectively with global teams and adapt quickly to changing requirements.
Required Skills & Experience
Strong expertise in
Python programming .
Hands-on experience with
PySpark
and large-scale data processing.
Proven experience in
Python API / backend development
(FastAPI, REST / async).
Proficiency in
AWS cloud services : Glue, EMR, SageMaker, S3, Athena.
Exposure to
ML / MLOps workflows
and production-level deployment.
Solid knowledge of
SQL / PostgreSQL
and database query optimization.
Familiarity with
unit testing (pytest), Git, Docker, Kubernetes, Terraform .
Strong problem-solving, production support, and communication skills.
Work Environment
Remote-friendly, with expectation of independent ownership and accountability.
Timing : 2 PM - 11 PM IST ( 4 hours overlap with EST)
Data Engineer Python • Dehra Dun, Uttarakhand, India