We are looking for a highly skilled Data Engineer (AWS) to join our growing data engineering team. This role involves designing, developing, and optimizing scalable data pipelines on AWS to support analytics, reporting, and business intelligence needs. You will be working with modern cloud-native services, migrating legacy data systems, and ensuring reliable, secure, and high-performance data flows.
If youre passionate about solving complex data problems, automating workflows, and building next-generation data platforms in the cloudthis opportunity is for you!
Key Responsibilities :
- Design, build, and optimize ETL / ELT pipelines using AWS Glue, Redshift, S3, and Lambda.
- Migrate legacy data systems (SQL Data Warehouse & SSIS ETL jobs) into modern AWS data services.
- Automate data workflows including ingestion, transformation, partitioning, and lifecycle management.
- Ensure data solutions are reliable, scalable, and cost-efficient with performance tuning and monitoring.
- Work closely with data architects, analysts, and stakeholders to deliver clean, well-modeled, and high-quality datasets.
- Debug, troubleshoot, and document data pipelines to maintain high availability.
- Contribute to the best practices, standards, and architecture guidelines for cloud-based data :
- 3+ years of hands-on experience as a Data Engineer, preferably with recent AWS cloud projects.
- Strong expertise in :
1. AWS Glue (ETL jobs, crawlers, catalog)
2. Amazon Redshift (data loading, schema design, performance optimization)
3. Amazon S3 (data partitioning, lifecycle management, storage optimization)
4. AWS Lambda (serverless data processing)
Proficiency in SQL and Python, with a solid understanding of ETL best practices.Experience with data migration from SQL DW / SSIS ETL to AWS cloud data platforms.Familiarity with cloud-native architectures and modern data engineering principles.Strong problem-solving skills, with an ability to work independently and Excellent communication and documentation Skills :Experience with CI / CD for data pipelines (GitHub Actions, CodePipeline, etc.).Exposure to data orchestration tools like Airflow, Step Functions, or Glue Workflows.Knowledge of data lake / lakehouse architectures (Delta Lake, Apache Iceberg, etc.).Hands-on with data quality frameworks and monitoring tools.(ref : hirist.tech)