Location : Remote (India-based candidates preferred)
Employment : Full-time (Hired through VS Technology, working on ServeFirst product)
ServeFirst is a scrappy, well-funded AI startup (recently closed a £4.5m seed round) that’s redefining how businesses understand their customers. We transform raw customer feedback into real-time insights for companies that care deeply about CX.
Our 2025 roadmap is ambitious :
- Break apart our Node.js monolith into scalable microservices
- Double our AI-driven workflows
- Harden our infra to handle 100× traffic growth
- Maintain a culture where everyone ships, everyone is on-call, bureaucracy is nil, and velocity is high
This is a chance to join early, own core architecture decisions, and scale AI-first systems in production
What You’ll Do
Break up the monolithDefine service boundaries, transition to microservices, and implement REST + SQS communication containerised on ECS Fargate.Own AI workflow integrationOrchestrate AI workflows with OpenAI / Claude APIs and frameworks like LangChain. Design multi-model pipelines (prompt orchestration, vector DBs, caching, telemetry) and lead our shift toward Bedrock / RAG-native infra.Build and scale AI infraStand up inference / workflow infra on AWS (Bedrock, SageMaker, containers). Ensure observability, security, and cost efficiency.Architect the AI platformGo beyond “just wrapping GPT.” Define orchestration vs. inference boundaries, expose tracing & prompt history, and build systems for experimentation & scale.Champion testing and correctnessEnforce unit, integration, and load testing. Design clear contracts, mocks, and fast CI.Estimate, scope, deliverBreak complex features into milestones, highlight risks, and communicate trade-offs clearly.Make it observableImplement structured logs, metrics, traces, alerts. Make LLM behavior traceable (token usage, prompt mutation).Think security firstHandle sensitive customer data (PII, IAM, secrets management, GDPR readiness).Ship backend codeWork in Node.js / TypeScript, with MongoDB (Mongoose), Redis, and schedulers (cron / EventBridge).Keep infra flexibleAWS-first today, modular Terraform for GCP portability tomorrow.Mentor & raise the barLead reviews, mentor engineers, and reinforce best practices while maintaining velocity.✅ Must-Haves
8+ years backend engineering; deep AWS + Node.js / TypeScript experienceStrong system design skills (event-driven, autoscaling systems)Production LLM workflow experience (OpenAI, Claude, Bedrock, etc.)Hands-on with LangChain / agentic frameworks / workflow automationProven track in deploying / scaling AI infra (Bedrock, SageMaker, containers)MongoDB & Redis tuning / debugging expertiseFluent with Terraform & DockerTesting mindset with CI / CD pipelinesStrong async communication (docs, writing, code clarity)Security awareness (IAM, GDPR, SOC2 basics)Demonstrated ability to scope, build, and deliver complex featuresNice-to-Haves
CX / survey / analytics SaaS backgroundExperience with Bedrock, LangChain, or RAG-native infraLLMOps exposure (prompt versioning, feedback loops, telemetry)GCP infra knowledgeReact familiarity / empathy for frontend engineersIncident response & blameless postmortems