We are seeking a highly skilled and passionate Scala Data Engineer to join our growing data team.
In this role, you will be instrumental in designing, developing, and maintaining high-performance data pipelines and applications using Scala.
You will work within a dynamic software development environment, contributing to all phases of the software development lifecycle, from architectural design to testing and deployment, with a strong focus on cloud-based data engineering solutions.
Key Responsibilities :
Scala Application Development :
- Design, develop, and maintain robust, scalable, and efficient data processing applications and services primarily using Scala.
- Implement complex data transformations, aggregations, and orchestrations to support various analytical and operational use cases.
Architectural Contributions :
Actively participate in all phases of architectural development tasks related to the application and data platform.Contribute to the design of data models, schemas, and data flow architectures.Provide input on technology choices and best practices for building resilient and performant data systems.Code Quality & Standards :
Write clean, well-documented, and highly optimized code in accordance with established application requirements, coding standards, and design patterns.Perform thorough software analysis, including requirements analysis, code reviews, and performance profiling.Collaborative Development :
Work closely as an integral member of a fast-paced software development team, ensuring that the developed programs and data solutions meet high-quality standards and business objectives.Participate in agile ceremonies, including sprint planning, daily stand-ups, and Collaboration :Collaborate effectively with cross-functional teams, including Data Scientists, Analysts, Product Managers, and other engineering teams, to understand data needs, deliver solutions, and ensure seamless data integration.Required Skills & Qualifications (Must-Haves) :
Programming Language : Strong proficiency and hands-on experience with Scala for building data-intensive applications and backend services.Extensive practical experience with Cloud Data Engineering principles and services.Demonstrated experience with Amazon EMR (Elastic MapReduce) for big data processing (Spark on EMR).Proven experience designing, developing, and managing data workflows and pipelines using Apache Airflow.Desired Skills (Good-to-Haves) :
Data Warehousing : Experience working with Snowflake for cloud data warehousing and analytics.Data Ingestion / Integration : Familiarity with StreamSets or similar data ingestion and integration platforms.Big Data Ecosystem : Experience with other components of the Apache Hadoop ecosystem (HDFS, Hive, etc.) and Apache Spark.Database Knowledge : Strong understanding of relational databases (SQL) and NoSQL databases.DevOps Practices : Experience with CI / CD pipelines, Git, and automation tools for data deployments.Education :
Bachelor's or Master's degree in Computer Science, Software Engineering, Data Science, or a related technical field
ref : hirist.tech)