About the Job
We are building a Customer Data Platform (CDP) designed to unlock the full potential of customer experience (CX) across our products and services. This role offers the opportunity to design and scale a platform that unifies customer data from multiple sources, ensures data quality and governance, and provides a single source of truth for analytics, personalization, and engagement.
As a key member of the data engineering team, you will architect and implement the pipelines, storage layers, and integrations that power our CDP. You’ll work with product, data science, and engineering stakeholders to deliver a robust platform that supports real-time decision-making, personalization at scale, and actionable customer insights.
This is a high-impact, hands-on engineering role where you’ll shape the data foundation that directly influences how we understand and serve our customers.
What you will be doing
- Design and implement Five9’s Customer Data Platform solutions and enable production use cases for customers
- Design and develop highly scalable and resilient services for ingesting large scale datasets
- Demonstrate strong ownership by ensuring operational excellence with a sharp focus on monitoring, observability, and system reliability
- Develop and orchestrate ETL / ELT pipelines using Apache Airflow
- Collaborate with cross functional partners and lead technical initiatives independently end to end.
- Design, build, and optimize distributed query engines such as Apache Spark, or Snowflake to support complex data workloads.
- Write, review, or provide feedback on a technical design proposal from others.
Skills
5-7 years of software / data engineering and data platform experience,Extensive experience with data transformation and modeling, including advanced features and best practicesDeep knowledge of Snowflake for data warehousing, including optimization, security, and cost managementGood understanding of data streaming technologies such as Kafka / Kafka ConnectStrong knowledge of highly scalable distributed systems, microservices, Rest APIsStrong proficiency in object oriented and / or functional programming language such as Java,. NET, Python and SQL for data processing, transformation, and pipeline developmentKnowledge of Apache Airflow for workflow orchestration is nice to haveUnderstanding of containerization and Kubernetes conceptsExperience with the AWS cloud platform and infrastructure as code practicesMaintain high standards of code quality with keen eye for test automation and operational excellence.Track record of delivering scalable data engineering solutions that support analytics, machine learning, and operational use casesExcellent written and verbal communication and interpersonal skillsBachelor's degree in Computer Science, Engineering or related field, or equivalent training, fellowship, or work experience