Talent.com
This job offer is not available in your country.
Cerebry — GenAI Implementation Engineer (AI Growth Lead)

Cerebry — GenAI Implementation Engineer (AI Growth Lead)

CerebryPalakkad, Kerala, India
7 days ago
Job description

Mission

Transform Cerebry Research designs into production-grade GenAI features —retrieval-grounded, safe, observable, and ready for seamless product rollout. Architect, code, evaluate, and package GenAI services that power Cerebry end-to-end.

Why this is exciting (Ownership-Forward)

Founder-mindset equity. We emphasize meaningful ownership from day one.

Upside compounds with impact. Initial grants are designed for real participation in value creation, with refresh opportunities tied to scope and milestones.

Transparent offers. We share the full comp picture (salary, equity targets, vesting cadence, strike / valuation context) during the process.

Long-term alignment. Packages are crafted for builders who want to grow the platform and their stake as it scales.

What you’ll build

Retrieval & data grounding : connectors for warehouses / blobs / APIs; schema validation and PII-aware pipelines; chunking / embeddings; hybrid search with rerankers; multi-tenant index management.

Orchestration & reasoning : function / tool calling with structured outputs; controller logic for agent workflows; context / prompt management with citations and provenance.

Evaluation & observability : gold sets + LLM-as-judge; regression suites in CI; dataset / version tracking; traces with token / latency / cost attribution.

Safety & governance : input / output filtering, policy tests, prompt hardening, auditable decisions.

Performance & efficiency : streaming, caching, prompt compression, batching; adaptive routing across models / providers; fallback and circuit strategies.

Product-ready packaging : versioned APIs / SDKs / CLIs, Helm / Terraform, config schemas, feature flags, progressive delivery playbooks.

Outcomes you’ll drive

Quality : higher factuality, task success, and user trust across domains.

Speed : rapid time-to-value via templates, IaC, and repeatable rollout paths.

Unit economics : measurable gains in latency and token efficiency at scale.

Reliability : clear SLOs, rich telemetry, and smooth, regression-free releases.

Reusability : template repos, connectors, and platform components adopted across product teams.

How you’ll work

Collaborate asynchronously with Research, Product, and Infra / SRE.

Share designs via concise docs and PRs; ship behind flags; measure, iterate, and document.

Enable product teams through well-factored packages, SDKs, and runbooks.

Tech you’ll use

LLMs & providers : OpenAI, Anthropic, Google, Azure OpenAI, AWS Bedrock; targeted OSS where it fits.

Orchestration / evals : LangChain / LlamaIndex or lightweight custom layers; test / eval harnesses.

Retrieval : pgvector / FAISS / Pinecone / Weaviate; hybrid search + rerankers.

Services & data : Python (primary), TypeScript; FastAPI / Flask / Express; Postgres / BigQuery; Redis; queues.

Ops : Docker, CI / CD, Terraform / CDK, metrics / logs / traces; deep experience in at least one of AWS / Azure / GCP.

What you bring

A track record of shipping and operating GenAI / ML-backed applications in production.

Strong Python , solid SQL , and systems design skills (concurrency, caching, queues, backpressure).

Hands-on RAG experience (indexing quality, retrieval / reranking) and function / tool use patterns.

Experience designing eval pipelines and using telemetry to guide improvements.

Clear, concise technical writing (design docs, runbooks, PRs).

Success metrics

Evaluation scores (task success, factuality) trending upward

Latency and token-cost improvements per feature

SLO attainment and incident trends

Adoption of templates / connectors / IaC across product teams

Clarity and usage of documentation and recorded walkthroughs

Hiring process

Focused coding exercise (2–3h) : ingestion → retrieval → tool-calling endpoint with tests, traces, and evals

Systems design (60m) : multi-tenant GenAI service, reliability, and rollout strategy

GenAI deep dive (45m) : RAG, guardrails, eval design, and cost / latency tradeoffs

Docs review (30m) : discuss a short design doc or runbook you’ve written (or from the exercise)

Founder conversation (30m)

Apply

Share links to code (GitHub / PRs / gists) or architecture docs you authored, plus a brief note on a GenAI system you built—problem, approach, metrics, and improvements over time.

Email : info@cerebry.co

Create a job alert for this search

Implementation Engineer • Palakkad, Kerala, India