Roles and Responsibility
- Design, develop, and implement data processing pipelines using Pyspark and AWS Glue.
- Collaborate with cross-functional teams to identify and prioritize project requirements.
- Develop and maintain large-scale data warehouses using Snowflake and Snowpark.
- Optimize data processing workflows for performance and scalability.
- Troubleshoot and resolve technical issues related to data processing and storage.
- Ensure data quality and integrity by implementing data validation and testing procedures.
Job Requirements
Strong proficiency in Pyspark, SQL, and AWS Glue.Experience with Snowflake and Snowpark is desirable.Excellent problem-solving skills and attention to detail.Ability to work collaboratively in a team environment.Strong communication and interpersonal skills.Ability to adapt to changing priorities and deadlines.Skills Required
Pyspark, Sql, AWS Glue, Data Modeling, Big Data, Cloud Computing