We are seeking skilled and motivated Spark & Databricks Developers to join our dynamic team for a long-term project. The ideal candidate will have strong hands-on experience in Apache Spark, Databricks, and GitHub-based development workflows.
Key Responsibilities :
- Design, develop, and optimize big data pipelines using Apache Spark.
- Build and maintain scalable data solutions on Databricks.
- Collaborate with cross-functional teams for data integration and transformation.
- Manage version control and code collaboration using GitHub.
- Ensure data quality, performance tuning, and job optimization.
- Participate in code reviews, testing, and documentation activities.
Must-Have Skills :
5–8 years of experience in Data Engineering or related rolesStrong hands-on expertise in Apache Spark (Batch & Streaming)Proficiency in Databricks for developing and managing data workflowsExperience with GitHub (version control, pull requests, branching strategies)Good understanding of Data Lake and Data Warehouse architecturesStrong SQL and Python scripting skillsIn-depth knowledge of Python programmingGood-to-Have Skills :
Experience with Azure Data Lake, AWS S3, or GCP BigQueryFamiliarity with Delta Lake and Databricks SQLExposure to CI / CD pipelines and DevOps practicesExperience with ETL tools or data modelingUnderstanding of data governance, security, and performance tuning best practices