Role Overview :
We are seeking a highly skilled Data Architect - GCP with 6–8 years of experience in designing, developing, and managing enterprise data solutions on Google Cloud Platform (GCP) along with Data Governance. The ideal candidate will have a strong background in cloud data architecture, data warehousing, big data processing, and data integration, with proven expertise in data governance, delivering scalable, secure, and efficient data platforms.
Key Responsibilities :
- Design and architect end-to-end data solutions on GCP, aligning with business and technical requirements.
- Hands on experience with Data Governance
- Define data models, storage strategies, data ingestion, processing, and consumption frameworks.
- Implement data lakes, data warehouses, and data marts using services like BigQuery, Cloud Storage, Dataflow, Dataproc, Pub / Sub, and Composer.
- Collaborate with business stakeholders, data scientists, and engineering teams to understand data needs and translate them into scalable architectures.
- Design and implement data governance, security, and compliance frameworks for cloud-based data platforms.
- Optimize data workflows, query performance, and storage costs in the GCP environment.
- Lead data migration and modernization initiatives from on-premise or other cloud platforms to GCP.
- Stay updated with GCP services, features, and industry best practices to recommend improvements and innovation.
- Provide technical leadership and mentoring to data engineering teams.
Required Skills and Qualification :
Educational Background :Bachelor’s degree in Computer Science, Software Engineering, or a related field.Experience :6–8 years of experience in data architecture and engineering roles, with at least 3 years hands-on on GCP.Hands-on Data Governance experience.Strong expertise in GCP data services : BigQuery, Dataflow, Pub / Sub, Dataproc, Cloud Storage, Cloud Composer, Data Catalog.Proficient in data modeling, data warehousing concepts, ETL / ELT pipelines, and big data processing frameworks.Experience with SQL, Python, and Terraform (preferred) for infrastructure as code.Hands-on experience in data security, encryption, access control, and governance on GCP.Experience in integrating with real-time data pipelines and event-driven architectures.Strong understanding of DevOps, CI / CD pipelines for data workflows, and cloud cost optimization.GCP Professional Data Engineer / Cloud Architect certification is a plus.Soft Skills :Strong problem-solving and analytical skills.Excellent communication and collaboration abilities.Ability to work independently and within a team in an Agile / Scrum environment.Traits we are looking for :
Curiosity-driven : Passionate about exploring new tools, technologies, and methods to solve problems creatively.Problem-Solver : Thrives on identifying and fixing complex issues, with a natural inclination for troubleshooting.Self-starter : Takes initiative, requires minimal supervision, and demonstrates ownership of tasks from start to finish.Innovative Mindset : Enjoys experimenting with different approaches and is not afraid to break things to learn.Continuous Learner : Actively seeks to expand their knowledge through reading, online courses, and experimenting with side projects.Adaptability : Quick to adjust to new technologies, tools, or changes in requirements.Detail-Oriented : Pays attention to both the big picture and the small details, ensuring quality in execution.Good To Have :
Exposure to AI / ML workflows, data preparation for ML models.Experience with third-party tools like Apache Airflow, Looker, or Dataplex.Knowledge of other cloud platforms (AWS, Azure) for hybrid / multi-cloud strategies.What We Offer :
Competitive salary and benefits package.Opportunity to work on cutting-edge security challenges.A collaborative and growth-oriented work environment with opportunities for career development.