About the Role :
We are seeking a talented and motivated Data Engineer to join our engineering team. In this role, you will be responsible for designing, developing, and maintaining robust data pipelines that enable the seamless flow of data across our systems. You will work with diverse data sources, ensuring the data is ingested, cleaned, transformed, and stored in a way that supports analytics, reporting, and decision-making. This is a unique opportunity to shape our data architecture and directly impact business outcomes through high-quality, reliable data delivery.
Key Responsibilities :
- Design, build, and maintain scalable ETL / ELT pipelines using Python and orchestration tools like Airflow, Prefect, or similar.
- Ingest, clean, transform, and structure data from multiple sources for downstream applications.
- Develop efficient, reusable, and well-documented Python scripts for data manipulation, automation, and integration tasks.
- Design and optimize data models to support analytics, reporting, and operational needs.
- Write complex and optimized SQL queries for data extraction, transformation, and analysis.
- Work with relational databases such as PostgreSQL, MySQL, or SQL Server to ensure efficient storage and retrieval of data.
- Collaborate closely with business stakeholders, analysts, and engineers to understand data requirements and deliver fit-for-purpose solutions.
- Monitor, debug, and optimize data pipelines for performance, scalability, and reliability.
- Ensure data quality, integrity, and consistency across systems.
- Participate in code reviews, sprint planning, and agile development processes to maintain engineering best practices.
- Continuously research and adopt new technologies and frameworks to enhance data engineering capabilities.
Required Skills & Qualifications :
Bachelors or Masters degree in Computer Science, Engineering, Information Systems, or related field.3+ years of hands-on experience in data engineering, with a strong focus on Python.Proficiency in Python programming for data processing, scripting, and automation.Solid understanding of data structures, algorithms, and data modeling concepts.Hands-on experience with ETL / ELT development and workflow orchestration tools (Airflow, Prefect, Luigi, etc.).Strong SQL skills with experience in relational databases (PostgreSQL, MySQL, SQL Server).Knowledge of data warehouse concepts and familiarity with cloud platforms (AWS, Azure, or GCP) is a plus.Experience with version control (Git) and collaborative software development practices.Strong problem-solving, debugging, and analytical skills.Excellent communication and teamwork abilities, with the ability to work in a fast-paced, agile environment.Preferred Qualifications :
Experience with cloud-native data services (e.g., AWS Glue, Redshift, BigQuery, Snowflake, or Azure Data Factory).Exposure to streaming data platforms such as Kafka or Kinesis.Familiarity with containerization (Docker, Kubernetes) and CI / CD pipelines.Knowledge of data governance, data security, and compliance best practices.(ref : hirist.tech)