We are looking for a highly skilled Data Engineer with strong expertise in PostgreSQL and hands-on experience in building data warehouses and data lakes.
In this role, you will design and implement a scalable PostgreSQL-based data warehouse and manage a data lake on AWS infrastructure, leveraging primarily open-source technologies.
The ideal candidate has a strong foundation in SQL engineering, cloud data architectures, and modern data pipelines.
Experience with hosted data warehouse platforms such as DBT, Snowflake, or Databricks is a plus.
Key Responsibilities :
Data Engineering & Architecture :
- Design, build, and maintain a PostgreSQL-based data warehouse for scalable analytics and reporting.
- Develop and manage data pipelines (batch and streaming) for ingestion, transformation, and storage.
- Architect and maintain a data lake on AWS infrastructure (e.g., S3, Glue, Athena, Redshift Spectrum).
- Optimize queries, indexing, and schema design for performance and scalability in PostgreSQL.
- Ensure solutions align with modern open-source data engineering best practices.
Collaboration & Delivery :
Work closely with product, analytics, and engineering teams to deliver high-quality, reliable data solutions.Translate business and analytics requirements into scalable data models and pipelines.Provide technical expertise to support data-driven decision-making across the Quality & Governance :Implement data quality checks, lineage tracking, and metadata management.Ensure compliance with data security and governance standards (e.g., GDPR, :Stay current with modern data engineering tools and frameworks.Provide guidance on evaluating and integrating DBT, Snowflake, or Databricks as needed.Required Qualifications :
3+ years of hands-on experience in data engineering or data platform development.Expertise in PostgreSQL : advanced SQL, query optimization, schema design, and data modeling.Strong experience building ETL / ELT pipelines using Python or open-source frameworks (e.g., Airflow, dbt).Proficiency with AWS services (S3, Glue, Athena, Redshift Spectrum, Lambda, etc.).Experience managing structured and unstructured data at scale.Solid programming skills in Python (or similar).Familiarity with open-source data frameworks (e.g., Apache Spark, Kafka).Preferred Qualifications :
Exposure to DBT, Snowflake, or Databricks for hosted data warehousing.Experience with infrastructure-as-code tools (Terraform, CloudFormation).Knowledge of data governance and cataloging tools.Cloud certifications (e.g., AWS Data Analytics Specialty).Experience in regulated industries (healthcare, finance, life We Offer :Opportunity to design and own enterprise-level data platforms with modern tech.A collaborative environment with mentorship from senior data and engineering leaders.Flexibility to innovate and work with open-source-first approaches.Remote-first culture with growth opportunities in data architecture and leadership.(ref : hirist.tech)