Key Qualifications 6+ years of experience in building and managing big data platforms and programming experience in Java, Scala or Python In-depth knowledge of Spark, Ray or other distributed computing frameworks Understanding of SQL query engines like Trino, Hive etc. Experience with Docker, Kubernetes or EKS Experience with debugging issues on distributed systems. Knowledge of software engineering practices and standard methodologies for the full software development lifecycle Experience with public cloud (AWS / GCP) Have created frameworks to deploy platforms in AWS / Azure / GCP. Good understanding of AI / ML stack - GPUs, MLFlow, LLM models is plus. Experience in building, tuning, scaling, and monitoring applications to process real-time, near-real-time and batch data Understanding of data modeling, data warehousing, and ETL concepts is a plus Description The GBI Big Data Engineering team is responsible for building and managing data platforms at scale on Cloud, that help Apple process, store and access petabytes of data. We’re looking for a Big Data Engineer with strong understanding of big data and ML platforms, passionate about building and managing cloud infrastructure and can demonstrate to Apple’s high standards. As a part of our team, you’ll be responsible for building and integrating various compute and storage infrastructure using open source or other solutions to aid various critical applications, such as analytic, reporting, AI / ML apps. Additionally, you will be deploying cloud stacks supporting big data and AI / ML platforms, locating production errors and issues. PySpark