Talent.com
This job offer is not available in your country.
▷ Urgent : Data Engineer L3 - Data Architecture [T500-20144]

▷ Urgent : Data Engineer L3 - Data Architecture [T500-20144]

Costco ITHyderabad, Telangana, India
12 hours ago
Job description

About Costco Wholesale

Costco Wholesale is a multi-billion-dollar global retailer with warehouse club operations in eleven countries. They provide a wide selection of quality merchandise, plus the convenience of specialty departments and exclusive member services, all designed to make shopping a pleasurable experience for their members.

About Costco Wholesale India

At Costco Wholesale India, we foster a collaborative space, working to support Costco Wholesale in developing innovative solutions that improve members’ experiences and make employees’ jobs easier. Our employees play a key role in driving and delivering innovation to establish IT as a core competitive advantage for Costco Wholesale.

Position Title : Data Engineer L3

Job Description :

Roles & Responsibilities :

  • Lead the design and implementation of enterprise data platforms : Architect and oversee the deployment of scalable, reliable, and secure data infrastructure for large organizations.
  • Drive innovation and adoption of new technologies : Research and integrate cutting-edge tools and frameworks for data ingestion, processing, and governance.
  • Mentor and guide junior and mid-level data engineers : Provide technical leadership, code reviews, and career development support.
  • Collaborate with stakeholders across teams : Align data engineering initiatives with business objectives and ensure cross-functional alignment.
  • Establish and enforce data engineering best practices : Define standards for pipeline architecture, data quality, security, and compliance across the organization.
  • Present findings and recommendations to senior leadership : Communicate technical concepts and business impacts to executives and decision-makers.

Technical Skills :

  • 8 – 12 years of experience
  • Expert-level proficiency in programming, automation, and orchestration : Mastery of Python, and workflow orchestration tools.
  • Deep understanding of data storage, processing, and governance : Advanced knowledge of data warehousing, Lakehouse architectures, and real-time streaming.
  • Proven ability to build and deploy scalable data systems : Design and implement robust, production-grade data platforms on GCP.
  • Experience with big data technologies : Use Dataflow, Dataproc, Pub / sub, or similar for large-scale data processing.
  • Strong security and compliance expertise : Implement and enforce security controls, encryption, audit logging for data systems, and compliance standards or data systems.
  • Excellent communication and presentation skills : Articulate technical concepts and business value to diverse audiences.
  • Must Have Skills :

  • Python, orchestration tools (e.g. Airflow, Cloud Composer)
  • Data architecture : data lakes, warehouses, streaming (e.g. Pub / Sub, Dataflow, Dataproc)
  • Experience with GCP and production-grade data platform deployment
  • Data security, compliance, and governance standards
  • Data modeling skills - Experience with different data modeling techniques (dimensional modeling, relational modeling, should be able to design scalable data models)
  • SQL Skills level - Expert
  • Deep understanding of bigquery, have experience on partitioning, clustering and performance optimizations
  • Experience on Cloud function, Composer and Cloud run, dataflow flex templates - should be able to write.
  • Create a job alert for this search

    Urgent Data Engineer • Hyderabad, Telangana, India