Job Title : Data Engineer
Experience : 5–10 Years
Location : Bangalore (ONLY)
Mandatory Skills :
SQL, ETL, PySpark, Python
Key Responsibilities
Design, develop, and maintain
scalable ETL / ELT pipelines
to ingest and transform large datasets from diverse sources.
Build and optimize
PySpark-based data processing workflows
for batch and real-time data use cases.
Write high-performance
SQL
queries and manage complex data models for analytics and reporting.
Collaborate with data architects, analysts, and cross-functional teams to understand business requirements and translate them into technical solutions.
Ensure
data quality, accuracy, and reliability
through robust validation and monitoring frameworks.
Optimize data storage and processing performance on cloud or distributed systems.
Implement automation for recurring tasks using
Python
and related tools.
Work with CI / CD pipelines and version control to maintain clean, scalable, and maintainable code.
Troubleshoot and resolve issues in data pipelines, ensuring minimal disruption to business operations.
Required Skills & Qualifications
5–10 years
of hands-on experience in
Data Engineering .
Strong proficiency in
SQL , including complex queries, performance tuning, and data modeling.
Expertise in
ETL / ELT development , data integration, and transformation frameworks.
Solid experience in
PySpark
for distributed data processing.
Strong programming experience in
Python , including writing modular, production-grade code.
Experience working with big data ecosystems (e.g., Hadoop, Hive, Spark).
Familiarity with data warehousing concepts and cloud platforms (AWS / Azure / GCP).
Experience with version control (Git), CI / CD, and workflow orchestration tools (Airflow / Luigi) is a plus.
Strong analytical, problem-solving, and communication skills.
Data Engineer • Delhi, Delhi, India