Profile : AWS Data Engineer
Mandate Skills : AWS + Databricks + Pyspark + SQL role
Location : Bangalore / Pune / Hyderabad / Chennai / Gurgaon
Notice Period : Immediate
Key Responsibilities :
- Design, build, and maintain scalable data pipelines to collect, process, and store data from multiple datasets
- Optimize data storage solutions for performance, scalability, and cost-efficiency
- Develop and manage ETL / ELT processes with schema transformations and data slicing / dicing
- Collaborate with cross-functional teams to understand requirements and accelerate feature development
- Create curated datasets for downstream consumption and end-user reporting
- Automate deployment and CI / CD processes using GitHub workflows
- Ensure compliance with data governance, privacy regulations, and security protocols
- Work with AWS cloud services and Databricks for data processing with S3 storage
- Utilize big data technologies : Spark, SQL, Delta Lake
- Integrate SFTP for secure data transfer from Databricks to remote locations
- Analyze Spark query execution plans for performance optimization
- Troubleshoot issues in large-scale distributed systems
Required Skills :
3+ years AWS experience (S3, EMR, Glue, Lambda, Redshift)Proficiency in Spark, SQL, Delta Lake, and DatabricksStrong Python / Scala programming skillsExperience with ETL / ELT processes and data pipelinesGit / GitHub workflows and CI / CD automationKnowledge of data governance and security protocolsProblem-solving skills in distributed systemsPreferred :
AWS certificationsReal-time data processing experienceBachelor's in Computer Science or related field(ref : hirist.tech)