GCP Data Engineer | Gurugram / Bengaluru | 6 - 10 years
Job Description :
We are seeking experienced GCP Data Engineers to design, build, and optimize scalable data pipelines and analytics solutions on Google Cloud Platform (GCP). The ideal candidate will have strong expertise in Python, PySpark, SQL, and ETL concepts, with hands-on experience across GCP’s data engineering services such as BigQuery, Dataflow, and Dataproc .
Key Responsibilities :
- Design, build, and maintain scalable data pipelines for processing high-volume datasets.
- Develop, optimize, and manage ETL workflows using Python, PySpark, and SQL .
- Implement solutions on BigQuery, Dataflow, Dataproc, and Cloud Storage to enable analytics and reporting.
- Write and optimize SQL queries (basic to advanced) for data analysis and warehousing.
- Collaborate with cross-functional teams to understand data needs and deliver reliable solutions.
- Ensure data quality, security, and governance across all pipelines and workflows.
- Automate deployment and monitoring of data pipelines using CI / CD tools (Jenkins, GitHub Actions, etc.).
- Leverage orchestration tools such as Apache Airflow / Cloud Composer for workflow management.
Must-Have Skills
Strong programming experience in Python with object-oriented design .Hands-on experience with PySpark APIs for large-scale data processing.Deep understanding of ETL concepts, data warehousing principles, and SQL .Expertise in BigQuery (data modeling, query optimization, performance tuning).