Job Title :
Data Integration Engineer
Location :
[Add Location or mention “Remote / Hybrid”]
Company : LumenData
Role Overview
We are seeking a
Data Integration Engineer
with expertise in building and orchestrating data pipelines using
Apache Airflow
to integrate data from diverse sources into
Snowflake .
The ideal candidate will have strong experience with
JDBC
and
API-based integrations (REST / JSON) , hands-on proficiency with
Postman , and solid skills in
SQL encryption / decryption ,
Python development , and
pipeline monitoring .
Key Responsibilities
Design, develop, and maintain
Airflow DAGs
to orchestrate end-to-end data workflows.
Integrate structured and unstructured data from multiple systems into
Snowflake
using
JDBC connectors ,
APIs , and
flat-file ingestion .
Use
Postman
and other tools to test, validate, and automate API integrations.
Implement
SQL encryption / decryption
techniques to protect sensitive datasets.
Perform
data quality checks , including row-level validation, hash-based reconciliation, and exception handling.
Develop transformation logic using
Python
and
SQL , ensuring performance, scalability, and maintainability.
Implement detailed
logging ,
monitoring , and
alerting
to ensure pipeline reliability and compliance.
Collaborate with stakeholders to understand requirements and deliver
scalable, production-ready solutions .
Required Skills & Experience
Strong proficiency in
Apache Airflow
for workflow orchestration.
Hands-on experience with
Snowflake
as a data warehouse.
Proven ability to integrate data via
JDBC drivers ,
REST APIs , and
Postman-tested endpoints .
Advanced knowledge of
SQL , including
encryption / decryption techniques .
Strong programming skills in
Python
for
ETL / ELT development .
Experience with
logging ,
monitoring , and
data observability
practices.
Data Integration Engineer • Delhi, India