Key Result Areas and Activities :
- Study existing technology landscape and understand current data integration framework and do impact assessment for the requirements.
- Develop spark jobs using Scala for new project requirements.
- Enhance existing spark jobs for any ongoing product enhancement.
- Performance tuning of spark jobs, stress testing etc.
- Create new data pipelines for developed / enhanced spark jobs using AWS Lambda or Apache Airflow
- Responsible for database design process, logical design, physical design, star schema, snowflake schema etc.
- Analyze data processing, integration, modelling and reporting requirements & Define data loading strategies considering volume, data types, frequency and analytics specifications.
- Ensure optimal balance between cost and performance.
- Project documentation, Adheres to Quality guidelines & Schedules.
- Works hand in hand with PM for successful delivery of project and provide Estimation, scoping, scheduling assistance.
- Manage build phase and quality assure code to ensure fulfilling requirements and adhering to Cloud architecture. Resolve difficult design and develop issues.
Work and Technical Experience :
Must-Have :
Overall, 7-9 years of IT Exp5+ years of AWS related projectGood to have Associate Level and Professional Level AWS CertificationIn depth knowledge of following AWS Services required : S3, EC2, EMR, Severless, Athena, AWS Glue, Lambda, Step FunctionsCloud Databases (Must have) – AWS Aurora, Singlestore, RedShift, SnowflakeBig Data (Must have) - Hadoop, Hive, Spark, YARNProgramming Language (Must have) – Scala, Python, Shell Scripts, PySparkOperating System (Must have) - Any flavor of Linux, WindowsMust have very strong SQL SkillsOrchestration Tools (Must have) : Apache AirflowExpertise in developing ETL workflows comprising complex transformations like SCD, deduplications, aggregations etc.Should have thorough conceptual understanding of AWS VPC, Subnets, Security Groups, & Route TablesShould be a quick and self-learner and be ready to adapt to new AWS Services or new Big Data Technologies as and when requiredQualifications :
Bachelor’s degree in computer science, engineering, or related field (Master’s degree is a plus)Demonstrated continued learning through one or more technical certifications or related methodsMinimum 5 years of experience on Cloud related projectsQualities :
Hold strong technical knowledge and experienceShould have the capability to deep dive and research in various technical related fieldsSelf-motivated and focused on delivering outcomes for a fast-growing team and firmAble to communicate persuasively through speaking, writing, and client presentationsAble to consult, write, and present persuasivelyAble to work in a self-organized and cross-functional teamAble to iterate based on new information, peer reviews, and feedbackPrior experience of working in a large media company would be added advantage