Description :
Location Pune & Nagpur
Experience 5-7 Years
Employment Type Full-Time
Job Summary :
We are seeking an experienced Data Engineer with strong hands-on expertise in Databricks on any cloud platform (AWS, Azure, or GCP), along with proven skills in SQL, Python, and PySpark. The ideal candidate will have a solid background in database design, data processing, and building scalable data pipelines in a cloud-based environment.
Key Responsibilities :
- Design, develop, and maintain scalable data pipelines and ETL processes using Databricks and PySpark
- Work with structured and semi-structured data across various cloud platforms (AWS, Azure, or GCP)
- Write optimized and complex SQL queries for data extraction and transformation
- Implement data ingestion from various sources into the data lake / warehouse
- Collaborate with data analysts, business stakeholders, and other engineering teams to understand requirements and deliver solutions
- Monitor and improve the performance of existing data pipelines
- Ensure data quality, security, and governance across the data lifecycle
Required Skills & Qualifications :
57 years of professional experience in Data Engineering or related rolesStrong experience with Databricks on any cloud platform (AWS, Azure, or GCP)Proficiency in SQL, Python, and PySparkSolid understanding of database structures, data modeling, and data warehousing conceptsExperience with cloud data services and storage solutionsFamiliarity with CI / CD practices and version control tools like GitExcellent problem-solving and communication skillsGood to Have :
Exposure to Delta Lake, Lakehouse architecture, or MLflowKnowledge of Apache Airflow or other workflow orchestration toolsExperience in performance tuning and job optimization in Spark environments(ref : hirist.tech)