Mission
Transform Cerebry Research designs into production-grade GenAI features —retrieval-grounded, safe, observable, and ready for seamless product rollout. Architect, code, evaluate, and package GenAI services that power Cerebry end-to-end.
Why this is exciting (Ownership-Forward)
Founder-mindset equity. We emphasize meaningful ownership from day one.
Upside compounds with impact. Initial grants are designed for real participation in value creation, with refresh opportunities tied to scope and milestones.
Transparent offers. We share the full comp picture (salary, equity targets, vesting cadence, strike / valuation context) during the process.
Long-term alignment. Packages are crafted for builders who want to grow the platform and their stake as it scales.
What you’ll build
Retrieval & data grounding : connectors for warehouses / blobs / APIs; schema validation and PII-aware pipelines; chunking / embeddings; hybrid search with rerankers; multi-tenant index management.
Orchestration & reasoning : function / tool calling with structured outputs; controller logic for agent workflows; context / prompt management with citations and provenance.
Evaluation & observability : gold sets + LLM-as-judge; regression suites in CI; dataset / version tracking; traces with token / latency / cost attribution.
Safety & governance : input / output filtering, policy tests, prompt hardening, auditable decisions.
Performance & efficiency : streaming, caching, prompt compression, batching; adaptive routing across models / providers; fallback and circuit strategies.
Product-ready packaging : versioned APIs / SDKs / CLIs, Helm / Terraform, config schemas, feature flags, progressive delivery playbooks.
Outcomes you’ll drive
Quality : higher factuality, task success, and user trust across domains.
Speed : rapid time-to-value via templates, IaC, and repeatable rollout paths.
Unit economics : measurable gains in latency and token efficiency at scale.
Reliability : clear SLOs, rich telemetry, and smooth, regression-free releases.
Reusability : template repos, connectors, and platform components adopted across product teams.
How you’ll work
Collaborate asynchronously with Research, Product, and Infra / SRE.
Share designs via concise docs and PRs; ship behind flags; measure, iterate, and document.
Enable product teams through well-factored packages, SDKs, and runbooks.
Tech you’ll use
LLMs & providers : OpenAI, Anthropic, Google, Azure OpenAI, AWS Bedrock; targeted OSS where it fits.
Orchestration / evals : LangChain / LlamaIndex or lightweight custom layers; test / eval harnesses.
Retrieval : pgvector / FAISS / Pinecone / Weaviate; hybrid search + rerankers.
Services & data : Python (primary), TypeScript; FastAPI / Flask / Express; Postgres / BigQuery; Redis; queues.
Ops : Docker, CI / CD, Terraform / CDK, metrics / logs / traces; deep experience in at least one of AWS / Azure / GCP.
What you bring
A track record of shipping and operating GenAI / ML-backed applications in production.
Strong Python , solid SQL , and systems design skills (concurrency, caching, queues, backpressure).
Hands-on RAG experience (indexing quality, retrieval / reranking) and function / tool use patterns.
Experience designing eval pipelines and using telemetry to guide improvements.
Clear, concise technical writing (design docs, runbooks, PRs).
Success metrics
Evaluation scores (task success, factuality) trending upward
Latency and token-cost improvements per feature
SLO attainment and incident trends
Adoption of templates / connectors / IaC across product teams
Clarity and usage of documentation and recorded walkthroughs
Hiring process
Focused coding exercise (2–3h) : ingestion → retrieval → tool-calling endpoint with tests, traces, and evals
Systems design (60m) : multi-tenant GenAI service, reliability, and rollout strategy
GenAI deep dive (45m) : RAG, guardrails, eval design, and cost / latency tradeoffs
Docs review (30m) : discuss a short design doc or runbook you’ve written (or from the exercise)
Founder conversation (30m)
Apply
Share links to code (GitHub / PRs / gists) or architecture docs you authored, plus a brief note on a GenAI system you built—problem, approach, metrics, and improvements over time.
Email : info@cerebry.co
Implementation Engineer • Tirupur, Tamil Nadu, India