Only immediate joiner- Within 7-15 days
5+ Years
Chennai, Bangalore, Hyderabad, Pune (HYRBID)
Shift Timing : 2 PM to 11 PM
Job Summary :
We are looking for an experienced GCP Data Engineer to design, build, and optimize scalable data pipelines and analytical solutions on Google Cloud Platform. The ideal candidate will have strong hands-on experience with BigQuery , DataProc (PySpark) , and a good understanding of Informatica for data integration.
Key Responsibilities :
- Design, develop, and maintain data pipelines and ETL processes using GCP services such as BigQuery , DataProc , and Cloud Storage .
- Develop and optimize PySpark scripts for data transformation and processing.
- Work closely with data architects, analysts, and business stakeholders to understand data requirements and deliver robust data solutions.
- Ensure data quality, governance, and security across all pipelines and datasets.
- Automate and orchestrate workflows using tools such as Cloud Composer (Airflow) or similar.
- Troubleshoot performance bottlenecks and optimize queries and pipelines for efficiency.
- Collaborate with cross-functional teams to integrate data from various sources, including Informatica-based pipelines.
- Support migration of legacy data processes to GCP.
Required Skills & Qualifications :
5+ years of experience in data engineering.Strong experience with Google BigQuery (query optimization, partitioning, clustering).Expertise in DataProc and PySpark for distributed data processing.Proficient in SQL and Python .Good understanding of data modeling, ETL design, and data warehouse concepts.Experience with version control (Git) and CI / CD pipelines for data projects.