We are seeking skilled and motivated Spark & Databricks Developers to join our dynamic team for a long-term project. The ideal candidate will have strong hands-on experience in Apache Spark , Databricks , and GitHub-based development workflows .
Key Responsibilities :
- Design, develop, and optimize big data pipelines using Apache Spark .
- Build and maintain scalable data solutions on Databricks .
- Collaborate with cross-functional teams for data integration and transformation.
- Manage version control and code collaboration using GitHub .
- Ensure data quality , performance tuning, and job optimization.
- Participate in code reviews , testing, and documentation activities.
Must-Have Skills :
5–8 years of experience in Data Engineering or related rolesStrong hands-on expertise in Apache Spark (Batch & Streaming)Proficiency in Databricks for developing and managing data workflowsExperience with GitHub (version control, pull requests, branching strategies)Good understanding of Data Lake and Data Warehouse architecturesStrong SQL and Python scripting skillsIn-depth knowledge of Python programmingGood-to-Have Skills :
Experience with Azure Data Lake , AWS S3 , or GCP BigQueryFamiliarity with Delta Lake and Databricks SQLExposure to CI / CD pipelines and DevOps practicesExperience with ETL tools or data modelingUnderstanding of data governance , security , and performance tuning best practices