Role-Senior Data Engineer
CTC- 20-50 LPA(Based on experience)
Location-Bangalore
We are seeking a skilled Data Engineer with strong expertise in Python , SQL , and modern big data technologies to design and build scalable, efficient data flows and pipelines in a cloud environment . The ideal candidate will have hands-on experience with real-time data ingestion using Kafka , proficiency in Spark for data processing, and familiarity with AWS cloud services such as Glue and Athena . This role demands strong programming skills, experience with ETL frameworks , version control proficiency, and a proactive, collaborative mindset for seamless delivery in Agile teams .
Key Responsibilities
- Develop, design, and optimize data processing pipelines leveraging Python , SQL , and Spark .
- Architect scalable and repeatable data flows and management solutions in a Cloud environment .
- Utilize Kafka platform for real-time, high-volume data ingestion and processing.
- Implement infrastructure-as-code practices for deployment consistency and reliability.
- Maintain and use version control systems such as GIT ; develop and update technical documentation .
- Collaborate effectively with cross-functional teams, ensuring strong communication and agile delivery.
- Apply Agile , Scrum , or Kanban methodologies to manage product delivery efficiently.
Qualifications
Bachelor’s or Master’s degree in Computer Science, Information Systems, or related field.Minimum 4 years of professional experience building data pipelines and managing big data technologies.Strong hands-on experience with ETL frameworks (e.g., Airflow, Jenkins) for production deployment.Deep understanding of data structures and willingness to adopt new technologies.Experience with AWS cloud tools , including Glue and Athena .Proficient in English , with excellent communication skills.Demonstrated ability to deliver data engineering solutions with a product-oriented mindset .