Role : Data Engineering Remote
Looking for 5-10 years of Exp
Skill Set :
AWS, Snowflake, Kafka, Airflow, GitHub, PySpark, Python
Key Responsibilities :
Design, develop, and maintain scalable ETL / ELT pipelines
Ingest data from various sources (APIs, databases, files, etc.)
Implement both real-time and batch processing solutions based on use case requirements
Ensure data quality through validation and cleansing processes
Collaborate with Product Managers and Business Stakeholders to gather and understand data requirements
Translate business needs into technical specifications
Ensure data security, access control, and compliance with relevant policies
Maintain documentation and follow best practices for data engineering
Data Integration Specialist • Delhi, Delhi, India