Search by job, company or skills

A

Senior GenAI & MLOps Engineer

Save
new job description bg glownew job description bg glow
  • Posted 12 hours ago
  • Be among the first 10 applicants
Early Applicant

Job Description

Company Description

Amunra Advisors LLP is a quantitative investment firm focused on the Indian markets. Following a multi-strategy, multi-manager approach, the firm drives consistent risk-adjusted returns. Amunra employs rigorous, institutional-grade manager research to identify top talent and strategically allocate capital across established and emerging managers. By leveraging innovative strategies and expertise, Amunra provides a dynamic and results-driven investment framework.

Role Description

Amunra is building OpenBrain, a firm-wide intelligence platform that integrates Generative AI, Second Brain knowledge architecture, and a dynamic Knowledge Graph to enhance investment decision-making.

We are looking for a Senior GenAI / MLOps / Data Engineer to own the complete data-to-intelligence lifecycle. This includes procuring and cleaning high-quality financial data, building robust data pipelines, constructing and maintaining a Knowledge Graph, and productionising advanced Generative AI systems.

This is a high-impact, hands-on role requiring strong technical depth across Data Procurement, Data Cleaning, Data Engineering, Knowledge Graph Engineering, Generative AI, and MLOps.

Key Responsibilities

Data Procurement & Ingestion

  • Identify, evaluate, and integrate diverse financial and alternative data sources, including APIs, Bloomberg, regulatory filings, earnings transcripts, research reports, web sources, and other structured or unstructured datasets.
  • Build automated, reliable, and compliant data procurement pipelines that can support research, investment workflows, and AI-native applications.

Data Cleaning & Ingestion

  • Design and implement large-scale data cleaning, deduplication, normalisation, validation, and enrichment processes.
  • Work extensively with unstructured and semi-structured financial documents such as PDFs, filings, transcripts, reports, and research notes.
  • Use both traditional data processing techniques and LLM-powered methods to ensure data is accurate, well-structured, and AI-ready.

Data Engineering

  • Build scalable batch and real-time ETL/ELT pipelines using modern data engineering tools and frameworks such as Spark, Kafka, Airflow, Dagster, dbt, or equivalent technologies.
  • Design and maintain data lakes, lakehouses, and storage architectures optimised for Generative AI, Knowledge Graph, and retrieval workloads.

Knowledge Graph & Second Brain Architecture

  • Architect, build, and evolve a large-scale Knowledge Graph for financial and investment intelligence.
  • Develop LLM-augmented systems for entity extraction, relationship mining, ontology management, temporal reasoning, and knowledge enrichment.
  • Implement advanced retrieval systems, including vector search, graph retrieval, hybrid retrieval, and agentic RAG.

Generative AI & Agent Development

  • Build, fine-tune, and productionise Generative AI systems, including multi-agent workflows, financial copilots, reasoning engines, and domain-specific LLM applications.
  • Develop robust evaluation frameworks to assess accuracy, reliability, hallucination risk, retrieval quality, and system performance.

MLOps & Production AI Systems

  • Own the end-to-end ML and AI lifecycle, including experimentation, training, deployment, monitoring, drift detection, observability, and governance.
  • Design scalable, cost-efficient inference platforms using Kubernetes and tools such as vLLM, Ray, MLflow, LangSmith, Docker, LangChain, and LangGraph.

Requirements

Must Have

  • 3-6 years of experience across Data Engineering, Generative AI, MLOps, or related infrastructure-heavy AI/ML roles.
  • Hands-on expertise in data procurement, data cleaning, and processing unstructured financial documents such as PDFs, transcripts, regulatory filings, and research reports.
  • Experience building Knowledge Graphs and advanced RAG systems using technologies such as Neo4j or equivalent graph databases, along with vector databases.
  • Strong proficiency in LLM workflows, including prompt engineering, fine-tuning, evaluation, retrieval augmentation, and agentic systems.
  • Production-grade experience with data pipelines and MLOps tools such as Python, Spark, Kafka, Airflow, Kubernetes, Docker, MLflow, LangChain, LangGraph, or similar platforms.
  • Strong software engineering fundamentals, clean coding practices, system design ability, and a production-first mindset.
  • Ability to operate in a hands-on, high-ownership environment where reliability, scalability, and quality matter.

Strong Plus

  • Experience working in quantitative hedge funds, asset management, capital markets, financial research, or trading environments.
  • Experience with multimodal document intelligence models and tools such as LayoutLM, Donut, LLaVA, or similar systems.
  • Open-source contributions, technical publications, or demonstrated thought leadership in RAG, Knowledge Graphs, AI infrastructure, or Data Engineering for AI.

Ideal Candidate

  • The ideal candidate is a senior hands-on builder who can move fluently across data infrastructure, AI engineering, and production systems.
  • You are comfortable working with messy financial data, designing reliable pipelines, building intelligent retrieval systems, and deploying GenAI applications that must be accurate, observable, and useful in real-world investment workflows.
  • You combine strong engineering discipline with curiosity about markets, research, and how institutional knowledge can be transformed into durable investment intelligence.

More Info

Job Type:
Industry:
Function:
Employment Type:

About Company

Job ID: 147461421