Job Title : Data Integration Engineer
Location : [Add Location or mention “Remote / Hybrid”]
Company : LumenData
Role Overview
We are seeking a Data Integration Engineer with expertise in building and orchestrating data pipelines using Apache Airflow to integrate data from diverse sources into Snowflake.
The ideal candidate will have strong experience with JDBC and API-based integrations (REST / JSON), hands-on proficiency with Postman, and solid skills in SQL encryption / decryption, Python development, and pipeline monitoring.
Key Responsibilities
- Design, develop, and maintain Airflow DAGs to orchestrate end-to-end data workflows.
- Integrate structured and unstructured data from multiple systems into Snowflake using JDBC connectors, APIs, and flat-file ingestion.
- Use Postman and other tools to test, validate, and automate API integrations.
- Implement SQL encryption / decryption techniques to protect sensitive datasets.
- Perform data quality checks, including row-level validation, hash-based reconciliation, and exception handling.
- Develop transformation logic using Python and SQL, ensuring performance, scalability, and maintainability.
- Implement detailed logging, monitoring, and alerting to ensure pipeline reliability and compliance.
- Collaborate with stakeholders to understand requirements and deliver scalable, production-ready solutions.
Required Skills & Experience
Strong proficiency in Apache Airflow for workflow orchestration.Hands-on experience with Snowflake as a data warehouse.Proven ability to integrate data via JDBC drivers, REST APIs, and Postman-tested endpoints.Advanced knowledge of SQL, including encryption / decryption techniques.Strong programming skills in Python for ETL / ELT development.Experience with logging, monitoring, and data observability practices.