Position : Senior Data Engineer
Location : Bengaluru, India (Work From Anywhere)
Job Type : Full-time
Job Summary :
We are seeking an experienced Senior Data Engineer to design, build, and maintain robust data infrastructure and scalable pipelines.
The ideal candidate will have deep expertise in AWS-based data platforms, big data technologies, and data architecture to effectively support our analytics and machine learning initiatives.
This role requires advanced proficiency in Python and SQL to automate workflows and ensure high data quality and security across the organization.
Key Responsibilities :
- Data Pipeline Development : Design, develop, and maintain highly scalable and efficient data pipelines utilizing core AWS services such as S3, Glue, Lambda, and RDS.
- ETL / ELT Processes : Build robust ETL (Extract, Transform, Load) or ELT processes to efficiently ingest, transform, and store large volumes of data for consumption.
- Database Management : Manage and optimize relational databases (e.g., PostgreSQL, MySQL) to ensure performance, scalability, and integrity, supporting evolving business requirements.
- Automation & Workflow : Automate complex data workflows and processing tasks using advanced scripting in Python, SQL, and Bash.
- Data Quality & Governance : Implement thorough data validation, monitoring, and governance best practices to ensure high data quality and integrity across all systems.
- DevOps & Infrastructure : Implement DevOps and CI / CD practices for data pipelines using tools like Terraform (for Infrastructure as Code) and GitHub Actions.
- Data Enablement : Enable data accessibility and facilitate insights generation by preparing data structures for BI tools such as Tableau or Metabase.
- Cross-Functional Collaboration : Collaborate effectively with analysts, data scientists, and other cross-functional teams to translate complex business needs into effective and scalable data solutions.
Qualifications :
Experience : Proven experience as a Senior Data Engineer (46+ years suggested by skill depth).Cloud Expertise : Strong background and hands-on experience with AWS-based data platforms (S3, Glue, Lambda, RDS).Programming Proficiency : Advanced proficiency in Python for scripting, automation, and data transformation logic.Database Skills : Strong SQL skills and proven experience with database management (PostgreSQL, MySQL, etc.).Data Architecture : Solid grasp of ETL / ELT development principles and building efficient data pipelines.IaC & DevOps : Experience using Infrastructure as Code tools like Terraform or CloudFormation, and familiarity with CI / CD workflows and Git version control.Data Quality : Strong commitment to data quality, security, and governance best practices.Preferred Skills :
Machine Learning : Exposure to ML technologies such as SageMaker.BI Tools : Familiarity with BI tools like Tableau or Metabase for reporting and visualization.Agile Experience : Experience working in Agile / Scrum environments with tools like Jira and Confluence.Testing : Familiarity with automated testing methodologies for data pipelines.(ref : hirist.tech)