About Codewalla
Codewalla is a New York–based product studio with engineering teams in India. Since 2005, we’ve built innovative products that scale. We work at the intersection of design, engineering, and AI developing systems shaped by real business needs and tested in the real world. Our team moves fast, thinks deeply, and cares about pushing what software can do to empower people and businesses
About the Role
We’re hiring an AI Applications Developer with 5 to 8 years of engineering experience, including at least 1 year building and shipping LLM-powered features in production.
Your mission : translate raw model capability into lean, reliable, and user-ready features. You’ll work on Model Context Protocol (MCP) servers, build agentic clients, architect RAG pipelines, and automate LLM evaluations to ensure every release delivers measurable value.
If you thrive on rapid iteration, prompt experimentation, and seeing your code make it into users’ hands we’d love to hear from you.
What You’ll Work On
Build MCP servers and agentic clients that handle user intent parsing, tool orchestration, and structured response generation
Architect efficient RAG pipelines with chunk decay, latency budgeting, and cost-aware vector search
Automate evaluation pipelines that test LLM outputs for relevance, accuracy, and coherence
Work closely with DevOps to codify and deploy infrastructure using CDK or Terraform
Set up observability dashboards for prompt performance, latency, and failure traceability
Continuously refine prompts, embeddings, and model behavior based on user feedback and regression tests
What Makes You a Great Fit
5–8 years of full-stack or backend development experience, with at least 1 year building AI-powered or LLM-based applications
AI-native mindset : test fast, trace deeply, pause to reframe when needed
Strong Python and TypeScript skills
Experience with either AWS or GCP stacks, such as :
AWS : Lambda, Bedrock, DynamoDB, OpenSearch Vector Search
GCP : Cloud Functions, Vertex AI, Firestore, BigQuery, Vector Search
Familiarity with LangChain, Bedrock SDK, and vector database schema design
Understanding of prompt design, embeddings, and agentic workflows
CI / CD fluencyGitHub Actions, containerized deployment, test-first habits
Experience with LLM evaluation tools like Promptfoo, LangSmith, or Guardrails
Bonus : Experience with MLflow, LaunchDarkly, Inferentia / GPU tuning
Tools & Tech We Work With
Languages : Python, TypeScript
Frameworks : LangChain, FastAPI, Next.js
Cloud : AWS (Bedrock, Lambda, DynamoDB, OpenSearch Vector Search) or GCP (Vertex AI, Cloud Functions, Firestore, BigQuery, Vector Search)
Dev Tools : GitHub Copilot, Cursor
Evaluation & Safety : Promptfoo, LangSmith, Guardrails
DevOps : GitHub Actions, CDK or Terraform, Docker, Prometheus, Grafana
Why Join Codewalla?
Build for scale : Work on products that get investor love and scale to global audiences
Be part of AI-native product development : Integrate AI tools into how we design, build, and ship
Global exposure : Directly collaborate with US-based startups, VCs, and product teams
Learn fast : Small, elite teams with zero bureaucracy and high ownership
See your impact : Every decision you make will shape user experience and business outcomes
Inclusion Matters
We’re an equal opportunity employer. We celebrate diversity and are committed to creating an inclusive environment for all team members.
Ready to build the future with LLMs, without waiting for the future to catch up?
Apply now and let’s build together.
Generative Ai Engineer • Mumbai, Maharashtra, India