JOB SUMMARY :
We are seeking a highly skilled Generative AI Engineer to lead and support the development and deployment of cutting-edge AI solutions. You will work at the intersection of LLMs, prompt engineering, RAG (Retrieval-Augmented Generation) frameworks, and scalable cloud infrastructure. This role demands a strong foundation in AWS services, deep understanding of LLMs like Claude, and hands-on experience with tools like LangChain and LangGraph.
Key Responsibilities :
- Design and implement Generative AI applications leveraging LLMs (e.g., Claude, OpenAI, Mistral, etc.)
- Develop and optimize RAG pipelines using LangChain, LangGraph, and vector databases
- Integrate with LLM APIs (Anthropic Claude, OpenAI, AWS Bedrock, etc.)
- Architect scalable AI solutions on AWS infrastructure (EC2, Lambda, S3, Bedrock, SageMaker)
- Collaborate with ML Engineers and Product teams to define LLM use cases and prototypes
- Deploy and maintain AI pipelines in production environments
- Implement secure, robust data pipelines for knowledge ingestion and retrieval
- Explore and experiment with fine-tuning and prompt optimization techniques
- Stay current with the latest trends in Generative AI, Foundation Models, and ML Ops
Required Skills :
3+ years of experience working with LLMs and Generative AI frameworksProficiency with LangChain and LangGraph for building RAG-based applicationsHands-on experience with AWS AI services, including Bedrock, SageMaker, and LambdaExperience integrating with Claude, OpenAI GPT, or similar APIsStrong understanding of RAG (Retrieval-Augmented Generation) architecturesProficiency in Python, with knowledge of relevant AI / ML and data librariesExperience deploying AI solutions using Docker, CI / CD pipelines, and serverless architectureFamiliarity with Vector Databases (e.g., FAISS, Pinecone, Weaviate, Amazon Kendra)Understanding of data security, privacy, and AI governance