Job Title : Data Engineer
Experience : 5–10 Years
Location : Bangalore (ONLY)
Mandatory Skills : SQL, ETL, PySpark, Python
Key Responsibilities
Design, develop, and maintain scalable ETL / ELT pipelines to ingest and transform large datasets from diverse sources.
Build and optimize PySpark-based data processing workflows for batch and real-time data use cases.
Write high-performance SQL queries and manage complex data models for analytics and reporting.
Collaborate with data architects, analysts, and cross-functional teams to understand business requirements and translate them into technical solutions.
Ensure data quality, accuracy, and reliability through robust validation and monitoring frameworks.
Optimize data storage and processing performance on cloud or distributed systems.
Implement automation for recurring tasks using Python and related tools.
Work with CI / CD pipelines and version control to maintain clean, scalable, and maintainable code.
Troubleshoot and resolve issues in data pipelines, ensuring minimal disruption to business operations.
Required Skills & Qualifications
5–10 years of hands-on experience in Data Engineering .
Strong proficiency in SQL , including complex queries, performance tuning, and data modeling.
Expertise in ETL / ELT development , data integration, and transformation frameworks.
Solid experience in PySpark for distributed data processing.
Strong programming experience in Python , including writing modular, production-grade code.
Experience working with big data ecosystems (e.g., Hadoop, Hive, Spark).
Familiarity with data warehousing concepts and cloud platforms (AWS / Azure / GCP).
Experience with version control (Git), CI / CD, and workflow orchestration tools (Airflow / Luigi) is a plus.
Strong analytical, problem-solving, and communication skills.
Data Engineer • Srikakulam, Andhra Pradesh, India