Experience : 5-10years
Location : Hyderabad
Google Cloud + Ingestion :
Experience on BigQuery, Cloud Storage or equivalent cloud platforms
Knowledge of BigQuery ingress and egress patterns
Experience in writing Airflow DAGs
Knowledge of pubsub,dataflow or any declarative data pipeline tools using batch and streaming ingestion
Other GCP Services : Vertex AI, Model Registry, Secret Manager, KMS, Composer, KubeFlow,
Container Registry, Artefact Registry, Cloud Build, Cloud Run, OAuth2.0, Scheduler, GKE, Model Registry, MIG, Cloud Function, Pub / Sub
Extensive experience in Google Cloud Platform (GCP) and related services (e.g. IAM, BigQuery, cloud storage, functions, compute).
Creating data models and standard patterns for big data (ingestion, storage, analytics, etc.)
Data Warehousing (BigQuery)
Strong SQL and data analysis skills
Experience on BigQuery, using partitions, clustering, arrays (atleast one level) and struct
Understanding of Bigquery quotas, service accounts and query types / patterns
(append, replace, partition types,unnest, aggregate to create arrays, Analytic functions)
Experience in design and implementation of ELT or dataflow pipelines (using Google Dataflow or similar tool)
Understanding of nested data sources (json)
Exposure and understanding of bigdata processing
Data Modelling experience to define target data warehouse model from source till business / reporting layer
Proficient ETL / ELT experience in workflows, data modelling and data pipelines
Programming
Strong python programming experience in using Google Cloud python client libraries
Unix Shell scripting experience to automate operational activities
Visualisation
Cloud Google Cloud • Hyderabad, Telangana, India