Sonatype is the software supply chain security company. We provide the worlds best end-to-end software supply chain security solution, combining the only proactive protection against malicious open source, the only enterprise grade SBOM management and the leading open source dependency management platform. This empowers enterprises to create and maintain secure, quality, and innovative software at scale.
As founders of Nexus Repository and stewards of Maven Central, the worlds largest repository of Java open-source software, we are software pioneers and our open source expertise is unmatched. We empower innovation with an unparalleled commitment to build faster, safer software and harness AI and data intelligence to mitigate risk, maximize efficiencies, and drive powerful software development.
More than 2,000 organizations, including 70% of the Fortune 100 and 15 million software developers, rely on Sonatype to optimize their software supply chains.
The Opportunity :
Were looking for a Senior Data Engineer to join our growing Data Platform team. This role is a hybrid of data engineering and business intelligence, ideal for someone who enjoys solving complex data challenges while also building intuitive and actionable reporting solutions.
Youll play a key role in designing and scaling the infrastructure and pipelines that power analytics, dashboards, machine learning, and decision-making across Sonatype. Youll also be responsible for delivering clear, compelling, and insightful business intelligence through tools like Looker Studio and advanced SQL queries.
What Youll Do :
- Design, build, and maintain scalable data pipelines and ETL / ELT processes.
- Architect and optimize data models and storage solutions for analytics and operational use.
- Create and manage business intelligence reports and dashboards using tools like Looker Studio, Power BI, or similar.
- Collaborate with data scientists, analysts, and stakeholders to ensure datasets are reliable, meaningful, and actionable.
- Own and evolve parts of our data platform (e.g., Airflow, dbt, Spark, Redshift, or Snowflake).
- Write complex, high-performance SQL queries to support reporting and analytics needs.
- Implement observability, alerting, and data quality monitoring for critical pipelines.
- Drive best practices in data engineering and business intelligence, including documentation, testing, and CI / CD.
- Contribute to the evolution of our next-generation data lakehouse and BI architecture.
What Were Looking For :
5+ years of experience as a Data Engineer or in a hybrid data / reporting role.Strong programming skills in Python, Java, or Scala.Proficiency with data tools such as Databricks, data modeling techniques (e.g., star schema, dimensional modeling), and data warehousing solutions like Snowflake or Redshift.Hands-on experience with modern data platforms and orchestration tools (e.g., Spark, Kafka, Airflow).Proficient in SQL with experience in writing and optimizing complex queries for BI and analytics.Experience with BI tools such as Looker Studio, Power BI, or Tableau.Experience in building and maintaining robust ETL / ELT pipelines in production.Understanding of data quality, observability, and governance best practices.Why Youll Love Working Here :
Data with purpose : Work on problems that directly impact how the world builds secure software.Full-spectrum impact : Use both engineering and analytical skills to shape product, strategy, and operations.Modern tooling : Leverage the best of open-source and cloud-native technologies.Collaborative culture : Join a passionate team that values learning, autonomy, and real-world impact.(ref : hirist.tech)