Location : Trivandrum, Bangalore, Chennai, Kochi.
Job Description :
Responsibilities include :
- Design and implement scalable, secure, and cost-effective data architectures using GCP.
- Lead the design and development of data pipelines with BigQuery, Dataflow, and Cloud Storage.
- Architect and implement data lakes, data warehouses, and real-time data processing solutions on GCP.
- Ensure data architecture aligns with business goals, governance, and compliance requirements.
- Collaborate with stakeholders to define data strategy and roadmap.
- Design and deploy BigQuery solutions for optimized performance and cost efficiency.
- Build and maintain ETL / ELT pipelines for large-scale data processing.
- Leverage Cloud Pub / Sub, Dataflow, and Cloud Functions for real-time data integration.
Requirement :
10 years of experience in data engineering, with at least 6 years in GCP.Expertise in BigQuery, Cloud Dataflow, Cloud Pub / Sub, Cloud Storage, and related GCP services.Strong experience in data warehousing, data lakes, and real-time data pipelines.Proficiency in SQL, Python, or other data processing languages.Experience with cloud security, data governance, and compliance frameworks.Strong problem-solving skills and ability to architect solutions for complex data environments.Google Cloud Certification (Professional Data Engineer, Professional Cloud Architect) preferred.Leadership experience and ability to mentor technical teams.Excellent communication and collaboration skills.Immediate joiner (15 days NP max).ref : hirist.tech)