Key Responsibilities :
- Design and develop data pipelines and ETL processes to ingest, process, and store large volumes of data.
- Implement and manage big data technologies such as Kafka, Dataflow, BigQuery, CloudSQL, PubSub
- Collaborate with stakeholders to understand data requirements and deliver high-quality data solutions.
- Monitor and troubleshoot data pipeline issues and implement solutions to prevent future occurrences.
Required Skills and Experience :
Generally, we use Google Cloud Platform (GCP) for all software deployed at Wayfair.Data Storage and ProcessingBigQueryCloudSQLPostgreSQLDataProcPub / SubData modeling :
Breaking the business requirements(KPIs) to data points.Building the scalable data modelETL Tools :
DBTSQLData Orchestration and ETLDataflowCloud ComposerInfrastructure and DeploymentKubernetesHelmData Access and ManagementLookerTerraformIdeal Business Domain Experience :
Supply chain or warehousing experienceThe project is focused on building a normalized data layer which ingests information from multiple WarehouseManagement Systems (WMS) and projects it for backoffice analysis.(ref : hirist.tech)