Responsibilities : Design, build, and maintain our AI-driven knowledge graph creation pipeline
Integrate third-party APIs for web scraping (e.g., Browserless), search (e.g., Tavily), and multimedia processing
Develop and refine sophisticated prompts for Large Language Models (Amazon Bedrock, Claude) to perform high-accuracy data extraction from unstructured text and video transcripts
Build and manage the backend data workflow using Python and Django, populating a PostgreSQL database
Implement an agentic AI system that translates natural language user queries into structured database lookups via the Django ORM
Develop data validation, cleaning, and entity deduplication logic to ensure knowledge graph integrity
Optimize the end-to-end pipeline for performance, scalability, and cost-effectiveness
Required Technical Skills : 4+ years of software development experience with strong Python proficiency
Proven experience working with major LLM APIs (Amazon Bedrock, OpenAI, Google Gemini)
Expertise in prompt engineering for complex data extraction and agentic tasks
Hands-on experience with web scraping tools or services (e.g., Beautiful Soup, Scrapy, Browserless, Apify)
Strong experience with backend frameworks, preferably Django
Proficiency with relational databases (PostgreSQL) and ORMs
Experience integrating various REST APIs
Proficiency with version control (Git)
Preferred Skills : Experience with agentic frameworks like LangChain or LangGraph
Familiarity with cloud services, particularly AWS (S3, Lambda, RDS, Transcribe)
Knowledge of workflow automation tools (e.g., n8n, Airflow)
Understanding of graph databases (Amazon Neptune, Neo4j) and query languages (Gremlin, Cypher)
Experience with multimedia processing libraries (e.g., ffmpeg)
CI / CD for AI / ML systems
Familiarity with vector databases and semantic search concepts
Technical Requirements : Deep understanding of Retrieval-Augmented Generation (RAG) architecture
Experience building secure Text-to-SQL or Text-to-ORM systems
Proficiency in processing both unstructured (text, transcripts) and structured (database) data
Strong knowledge of data structures, algorithms, and system design principles
Understanding of the trade-offs between different data storage solutions (SQL vs. Graph vs. Vector)
Qualifications : Bachelor's in Computer Science, AI, Data Science, or a related field
2+ years of relevant software engineering experience, with a significant focus on AI solutions
Strong analytical and problem-solving abilities, especially with messy, real-world data
Excellent communication skills to articulate complex technical concepts
Experience working in Agile / Scrum environments
Benefits : Competitive salary and benefits package
Opportunity to work on challenging projects using cutting-edge technologies
Professional development opportunities
Flexible work arrangements
Collaborative and innovative work environment
Health insurance and other benefits
Regular team events and activities
Contact us : Send us your resume with your preferred position to below address
careers@gumo.co.in
Ai Engineer • Malappuram, Kerala, India