We are seeking a results-driven Senior Data Engineer for our German client to lead the development and implementation of advanced data infrastructure with a strong focus on graph databases (Neo4j) and cloud-based architecture (AWS, GCP, Azure). In this role, you will be responsible for transforming complex data systems into scalable, secure, and high-performing platforms to power next-generation analytics. This is an opportunity to drive meaningful outcomes while building expertise in a fast-moving, DevOps-oriented environment. Career growth opportunities include technical leadership and potential progression into data architecture or platform engineering leadership roles.
Job Responsibilities
- Lead the design and implementation of a scalable graph data architecture leveraging Neo4j, ensuring optimal query performance and alignment with business analytics needs within the first 90 days.
- Own the full lifecycle of data ingestion and ETL pipelines , including development, orchestration, monitoring, and optimization of data flows across multiple cloud environments, with measurable reduction in latency and improved data quality within 6 months.
- Deploy and maintain infrastructure using Terraform , supporting infrastructure-as-code best practices across AWS and GCP, ensuring automated, reproducible deployments and >
99.9% system uptime.
Automate CI / CD workflows using GitHub Actions , reducing manual operations by 50% within the first 6 months through effective automation of test, deploy, and monitoring pipelines.Collaborate with cross-functional engineering, data science, and security teams to integrate and support graph-based solutions in production with clear documentation, support structures, and knowledge sharing.Establish and enforce data security and access protocols , working with DevOps and InfoSec to ensure compliance with organizational standards by month 9.Desired Qualifications
Demonstrated success designing and maintaining production-grade graph database systems (preferably Neo4j).Strong command of Cypher, SQL, and Python for scalable data engineering solutions.Hands-on experience with Terraform , including managing multi-cloud infrastructure.Proven capability in building robust, reliable ETL pipelines and orchestrating workflows.Experience with CI / CD tools like GitHub Actions in a production DevOps environment.Strong communication skills, able to explain complex data concepts clearly across teams.Ability to independently identify and implement performance optimizations in data systems.