About the Role
We are looking for a highly skilled Backend Engineer with a strong background in Python, system design, and infrastructure to join our team. You will be responsible for designing, building, and maintaining scalable backend systems, while collaborating with cross-functional teams to deliver robust and efficient solutions.
This role requires someone who can think end-to-end , from designing high-level architecture, implementing core services, to ensuring production-grade reliability and performance.
Key Responsibilities
- Develop and maintain backend services and APIs using Python & NodeJs
- Design scalable, resilient, and maintainable systems, focusing on system architecture and distributed systems .
- Integrate AI and large language models (LLMs) into applications, ensuring performance, scalability, and cost-efficiency.
- Collaborate with AI / ML teams to deploy models into production pipelines.
- Optimize infrastructure for AI workloads (GPU usage, caching, batch processing)
- Build and maintain monitoring, logging, and observability for AI-powered systems.
- Troubleshoot and resolve issues in production systems while maintaining high reliability.
- Participate in design and code reviews, and drive engineering best practices across the team.
- Automate deployment pipelines for backend and AI services (CI / CD, IaC).
Required Skills & Qualifications
Strong experience in Python (FastAPI (most-preferred), Flask, Django, or similar) or NodeJS (Express (most-preferred), Fastify or similar)Solid understanding of system design principles : scalability, fault tolerance, distributed systems.Experience with infrastructure and DevOps : Docker, Kubernetes, Terraform, CI / CD pipelines.Hands-on experience with cloud platforms (AWS, Azure, GCP), especially for AI workloads.Knowledge of databases (SQL & NoSQL) and caching systems (Redis, Memcached).Experience integrating LLMs or AI APIs into production systems (OpenAI, HuggingFace, LangChain, etc.).Familiarity with messaging / streaming systems (Kafka, RabbitMQ).Monitoring and observability experience (Prometheus, Grafana, ELK).Strong problem-solving, debugging, and analytical skills.Excellent communication and collaboration skills.Nice to Have
Experience with generative AI pipelines , vector databases, and embeddings.Familiarity with ML Ops tools (MLflow, BentoML, Ray Serve, etc.).Knowledge of event-driven architectures and microservices.Prior experience in AI / LLM-focused startups or high-scale AI systems .What We Offer
Opportunity to work on challenging, large-scale systems with real-world impact.Collaborative team culture with focus on learning and innovation .Competitive compensation and growth opportunities.