Job Title : ETL Developer – DataStage, AWS, Snowflake
Experience : 5–7 Years
Location : Remote
Job Type : Full-time
About the Role
We are looking for a talented and motivated ETL Developer / Senior Developer to join our data engineering team. You will work on building scalable and efficient data pipelines using IBM DataStage (on Cloud Pak for Data) , AWS Glue , and Snowflake . You will collaborate with architects, business analysts, and data modelers to ensure timely and accurate delivery of critical data assets supporting analytics and AI / ML use cases.
Key Responsibilities
- Design, develop, and maintain ETL pipelines using IBM DataStage (CP4D) and AWS Glue / Lambda for ingestion from varied sources like flat files, APIs, Oracle, DB2, etc.
- Build and optimize data flows for loading curated datasets into Snowflake , leveraging best practices for schema design, partitioning, and transformation logic.
- Participate in code reviews , performance tuning, and defect triage sessions.
- Work closely with data governance teams to ensure lineage, privacy tagging, and quality controls are embedded within pipelines.
- Contribute to CI / CD integration of ETL components using Git, Jenkins, and parameterized job configurations.
- Troubleshoot and resolve issues in QA / UAT / Production environments as needed.
- Adhere to agile delivery practices, sprint planning, and documentation requirements.
Required Skills and Experience
4+ years of experience in ETL development with at least 1–2 years in IBM DataStage (preferably CP4D version) .Hands-on experience with AWS Glue (PySpark or Spark) and AWS Lambda for event-based processing.Experience working with Snowflake : loading strategies, stream-task, zero-copy cloning, and performance tuning.Proficiency in SQL , Unix scripting , and basic Python for data handling or automation.Familiarity with S3 , version control systems (Git), and job orchestration tools.Experience with data profiling, cleansing, and quality validation routines.Understanding of data lake / data warehouse architectures and DevOps practices.Good to Have
Experience with Collibra, BigID , or other metadata / governance toolsExposure to Data Mesh / Data Domain modelsExperience with agile / Scrum delivery and Jira / Confluence toolsAWS or Snowflake certification is a plus