About Costco Wholesale
Costco Wholesale is a multi-billion-dollar global retailer with warehouse club operations in eleven countries. They provide a wide selection of quality merchandise, plus the convenience of specialty departments and exclusive member services, all designed to make shopping a pleasurable experience for their members.
About Costco Wholesale India
At Costco Wholesale India, we foster a collaborative space, working to support Costco Wholesale in developing innovative solutions that improve members’ experiences and make employees’ jobs easier. Our employees play a key role in driving and delivering innovation to establish IT as a core competitive advantage for Costco Wholesale.
Position Title : Staff, Data Engineer
Job Description :
Roles & Responsibilities :
- Shape and drive enterprise-wide data architecture strategy : Define and evolve the long-term technical vision for scalable, resilient data infrastructure across multiple business units and domains.
- Lead large-scale, cross-functional initiatives : Architect and guide the implementation of data platforms and pipelines that enable analytics, AI / ML, and BI at an organizational scale.
- Pioneer advanced and forward-looking solutions : Introduce novel approaches in real-time processing, hybrid / multi-cloud, and AI / ML integration to transform how data is processed and leveraged across the enterprise.
- Mentor and develop senior technical leaders : Influence Principal Engineers, Engineering Managers, and other Staff Engineers; create a culture of deep technical excellence and innovation.
- Establish cross-org technical standards : Define and enforce best practices for data modeling, pipeline architecture, governance, and compliance at scale.
- Solve the most complex, ambiguous challenges : Tackle systemic issues in data scalability, interoperability, and performance that impact multiple teams or the enterprise as a whole.
- Serve as a strategic advisor to executive leadership : Provide technical insights to senior executives on data strategy, emerging technologies, and long-term investments.
- Represent the organization as a thought leader : Speak at industry events / conferences, publish thought leadership, contribute to open source and standards bodies, and lead partnerships with external research or academic institutions.
Technical Skills :
15+ years of experienceMastery of data architecture and distributed systems at enterprise scale : Deep experience in GCP .Advanced programming and infrastructure capabilities : Expertise in writing database queries, Python, or Java, along with infrastructure-as-code tools like Terraform or Cloud Deployment Manager.Leadership in streaming and big data systems : Authority in tools such as BigQuery, Dataflow, Dataproc, Pub / sub for both batch and streaming workloads.Enterprise-grade governance and compliance expertise : Design and implement standards for data quality, lineage, security, privacy (e.g., GDPR, HIPAA), and auditability across the organization.Strategic integration with AI / ML ecosystems : Architect platforms that serve advanced analytics and AI workloads (Vertex AI, TFX, MLflow).Exceptional ability to influence across all levels : Communicate technical vision to engineers, influence strategic direction with executives, and drive alignment across diverse stakeholders.Recognized industry leader : Demonstrated track record through conference presentations, publications, open-source contributions, or standards development.Must Have Skills :
Deep expertise in data architecture, distributed systems, and GCP.Python or Java, infrastructure-as-code (e.g. Terraform)Big data tools : BigQuery(Expert level. Having experience on performance tuning and UDFs), Dataflow, Dataproc, Pub / Sub (batch + streaming)Data governance, privacy, and compliance (e.g. GDPR, HIPAA)Data modeling and architecture - level expert, have experience on hybrid architecturesSQL Skills level - ExpertDeep understanding of BigQuery, have experience on partitioning, clustering and performance optimizationsExperience on Cloud function, Composer and Cloud run, dataflow flex templates - should be able to writeUnderstanding of full concepts of cloud architecture.