Job Description : Data Engineer
3–5 Years Experience
Location : Remote
RedPluto Analytics is seeking an experienced Data Engineer who is passionate about designing, developing, and optimizing scalable data pipelines. The ideal candidate will have strong hands-on expertise in modern data engineering tools, frameworks, and cloud platforms.
Key Responsibilities
- Design, develop, and maintain scalable data pipelines using PySpark and Big Data technologies.
- Develop, schedule, and manage workflows using Apache Airflow.
- Work on data ingestion, transformation, and storage using Data Lakes and Data Warehouses.
- Optimize PySpark jobs for improved performance and cost efficiency.
- Implement real-time and batch data processing solutions.
- Collaborate with cross-functional teams to understand data requirements and deliver high-quality solutions.
- Ensure data quality, reliability, security, and governance across the ecosystem.
Must-Have Skills
3–5 years of hands-on experience in Data Engineering.Strong proficiency in SQL.Solid working experience with PySpark and its optimization techniques.Strong understanding of Big Data concepts and distributed data processing.Hands-on real-time experience with Apache Airflow.Knowledge of Data Lakes and Data Warehouses architectures.Good-to-Have Skills
Experience working with any of the following :
Cloud Platforms : AWS / Azure / GCPDatabricksKafka (for streaming / real-time processing)Relevant certifications in data engineering or cloud technologies.