About Kapiva
Kapiva (Series-C funded) is on a journey of transformation — from being one of India’s leading modern Ayurvedic nutrition brands to becoming a health-tech company that leverages technology to drive better health outcomes for millions of people across India and internationally.We believe the next wave of innovation in health will be driven by AI-first solutions — from personalized recommendations and intelligent coaching, to automated platforms that scale care and wellness delivery. At Kapiva, you’ll be part of building this future.
Role Summary
We are looking for a DevOps Engineer (Level 3) to lead the design, automation, and optimization of Kapiva’s infrastructure supporting our Ecommerce and Health Tech platforms . This is a senior, hands-on role that blends deep technical expertise in cloud infrastructure, CI / CD, observability, and DevSecOps with strategic ownership of cost, performance, and reliability .You will also lead the adoption of AI-driven automation and monitoring , helping improve developer productivity and system efficiency across multiple environments.
Key Responsibilities
- Design, implement, and manage scalable, secure, and reliable infrastructure across AWS and GCP.
- Architect and maintain CI / CD pipelines for multi-service, multi-environment deployments.
- Implement and maintain infrastructure as code (Terraform, Helm, Ansible) for consistent and automated provisioning.
- Lead observability, monitoring, and incident response initiatives using Prometheus, Grafana, ELK, or Datadog.
- Own infrastructure cost optimization , identifying waste, optimizing compute / storage usage, and implementing right-sizing strategies to ensure efficient resource utilization.
- Collaborate with engineering teams to design cost-aware and scalable architectures .
- Champion DevSecOps practices , embedding security, compliance (like PCI-DSS, HIPAA), and governance into pipelines.
- Drive performance optimization across build, deploy, and runtime workflows.
- Automate operational workflows and log analysis using Python / Bash scripts or AI-based tooling .
- Evaluate and integrate AI-powered DevOps solutions (e.g., AI log analysis, anomaly detection, predictive scaling).
- Define and maintain SRE best practices — SLIs, SLOs, and error budgets.
- Mentor junior DevOps engineers and conduct design, code, and post-incident reviews.
Required Skills and Qualifications
4–6 years of experience in DevOps, SRE, or Infrastructure Engineering roles.Hands-on expertise with GCP and / or AWS cloud infrastructure .Deep understanding of Kubernetes, Docker, and container orchestration .Proficiency in CI / CD tools (GitHub Actions, Jenkins, GitLab CI, ArgoCD).Strong command over infrastructure as code (Terraform, Helm, Ansible).Experience with monitoring and logging frameworks (Prometheus, Grafana, ELK, Datadog).Practical experience in infrastructure performance tuning and cost optimization .Understanding of networking concepts (VPC, load balancing, DNS, firewalls).Scripting proficiency in Python, Bash, or Go .Strong troubleshooting and problem-solving skills.Bonus Points
Experience in Ecommerce or Health Tech platforms with high reliability requirements.Exposure to security practices (IAM, secrets management, vulnerability scanning).Knowledge of cost optimization strategies on AWS / GCP.Familiarity with incident management and on-call practices .What We Offer
Ownership of infrastructure scalability, reliability, and cost efficiency across all product lines.Opportunity to lead AI-driven automation and monitoring initiatives .Collaborative culture focused on engineering excellence, optimization, and continuous learning .Competitive compensation and clear growth path to DevOps Architect or Platform Engineering Lead .Work on distributed systems powering real-world Ecommerce and Health Tech applications.