Talent.com
Principal Devops / Edge Ai Engineer

Principal Devops / Edge Ai Engineer

Mulya TechnologiesBengaluru, Republic Of India, IN
10 hours ago
Job description

Principal DevOps / Edge AI Engineer

Bangalore

Founded in 2023,by Industry veterans HQ in California,US

  • We are revolutionizing sustainable AI compute through intuitive software with composable silicon

Principal DevOps / Edge AI Engineer

Overview :

You will be responsible for building, deploying, and maintaining the local infrastructure that powers high-performance multimodal AI models (text, image, audio, video) on a compact AI appliance. You’ll bridge the gap between hardware, ML inference, and user-facing applications - ensuring reliability, scalability, and efficiency of on-device AI workloads.

Key Responsibilities :

  • System Deployment & Orchestration
  • Containerize AI inference services and web applications using Docker or Podman.
  • Design lightweight orchestration layers for local systems (Kubernetes, Nomad, or custom orchestration).
  • Automate build, test, and deployment pipelines (CI / CD) for local-first AI workloads.
  • Performance Optimization & Resource Management
  • Optimize compute utilization for concurrent multimodal workloads.
  • Develop monitoring tools for system health, thermal management, and memory / bandwidth usage.
  • Tune OS, drivers, and I / O subsystems for maximum throughput and low latency.
  • Edge Infrastructure & Networking
  • Configure low-latency local networking for browser-based access to the AI appliance.
  • Set up secure local APIs and data isolation layers — ensuring zero external data leakage.
  • Integrate hardware accelerators and manage firmware updates across different SKUs.
  • Reliability, Testing, and Scaling
  • Build test harnesses to validate multimodal model performance (e.G., LLM + diffusion + ASR pipelines).
  • Implement over-the-air (OTA) update mechanisms for edge devices without exposing user data.
  • Develop monitoring dashboards and alerting for real-time performance metrics.
  • Required Qualifications :

  • Strong background in Linux systems engineering and containerization (Docker, Podman, LXC).
  • Experience deploying AI inference services locally or at the edge (llama.Cpp, ollama, vLLM, ONNX).
  • Proficiency in CI / CD tools (GitHub Actions, Jenkins, ArgoCD) and infrastructure-as-code (Terraform, Ansible).
  • Expertise in GPU / accelerator optimization, CUDA stack management, or similar.
  • Solid understanding of networking, security, and firewall configurations for local appliances.
  • Scripting and automation skills (Python, Bash, Go, or Rust).
  • Preferred Qualifications :

  • Experience with embedded systems or edge AI devices (e.G., Jetson, Coral, FPGA-based accelerators).
  • experience minimum 10 years
  • Familiarity with low-bit quantization, model partitioning, or distributed inference.
  • Background in hardware / software co-design or systems integration.
  • Knowledge of browser-based local apps (WebSocket, WebRTC, RESTful APIs) and AI service backends.
  • Prior work in privacy-preserving AI systems or local-first architectures.
  • Contact : Uday

    Mulya Technologies

    muday_bhaskar@yahoo.com

    "Mining The Knowledge Community"

    Create a job alert for this search

    Principal Ai Engineer • Bengaluru, Republic Of India, IN