Education : Bachelors or Masters degree in Computer Science, AIML / Data Science, Statistics or a related field only.
Preferred Institution : NIT (Any)
Responsibilities :
- Architect and deploy scalable Large language models (LLM) and transformer-based architectures.
- Implement MLOps / LLMOps best practices for CI / CD, automated model retraining, and lifecycle management.
- Develop advanced AI RAG applications, fine-tune models with PEFT / LoRA, RLHF.
- Productionize RAG pipelines : embedding strategy, vector-DB design (Weaviate / Pinecone / Chroma). Hybrid search, evaluations and guardrails.
- Optimize AI / ML pipelines for distributed computing, leveraging cloud platforms and accelerators (GPUs / TPUs).
- Optimize inference : quantization (INT4 / 8), speculative decoding, TensorRT-LLM / vLLM, or Ray Serve to reduce token costs.
- Work closely with data engineers, software developers, and product managers to integrate AI into production systems.
- Translate complex AI concepts into actionable insights for business stakeholders.
- Establish governance frameworks for ethical AI, ensuring compliance with industry standards.
Required Skills :
5+ years of experience as a AIML Engg.Strong background in statistics, linear algebra, and probability theoryStrong proficiency in C++ / Python and SQLExperience with deep learning frameworks such as TensorFlow, PyTorch, or JAXBuild the Agentic LLM applications prototype and launch LLM agents (planner-executor, multi-agent, tool-calling).Experience with LangChain, LlamaIndex, CrewAI, and vector DBs like Weaviate, Pinecone, Qdrant.Experience in Quantization, distillation, GPU kernel tuning.Strong expertise in fine-tuning transformer models, hyperparameter optimization, and reinforcement learning.Proficiency in Azure cloud platform AIML lifecycle CICD - MLOps / LLMOps. Experience in Kubernetes, KubeRay. Familiar with Terraform / Pulumi and observability tools.Lead from the front, responsible for coding, designing, and ensuring best practices & frameworks are adhered by the team.Knowledge of big data processing frameworks (Apache Spark, Hadoop, Data bricks).Create end to end AI systems with responsible AI principlesExcellent communication skills to facilitate interactions with stakeholdersref : hirist.tech)