Search by job, company or skills

J

LLM Engineer

Save
new job description bg glownew job description bg glownew job description bg svg
  • Posted 4 days ago
  • Be among the first 10 applicants
Early Applicant

Job Description

Company Description

Janooma is India's first agent-driven AI marketplace, enabling seamless buyer-seller interactions by leveraging an innovative conversational layer. With a database of over 4 million businesses across India, Janooma delivers live quotes from verified vendors to buyers and connects service providers with high-intent leads through WhatsApp. This platform facilitates everything from electronics to home services in one conversation, providing convenience and efficiency for all users. By solving the cold-start problem before its launch, Janooma is poised to revolutionize the marketplace with upcoming expansions to 10 Indian cities by 2026 and Southeast Asia by 2027.

LLM Systems Engineer

Location: Bengaluru (Hybrid)

Experience: 0-1 year (Freshers from 2024-2026 batches are strongly encouraged)

We are building next-generation large language model infrastructure focused on advanced model composition, optimization, and continuous capability enhancement across multiple domains and categories.

About the Role

You will play a foundational role in designing and implementing modular, scalable architectures for working with large language models. This includes building automated systems for model discovery, intelligent model merging/composition, architecture parsing, performance evaluation, gap analysis, and iterative improvement pipelines that power multiple specialized LLMs.

You will work on end-to-end LLM engineering workflows — from low-level model introspection to high-level automated training and deployment loops — enabling the creation and continuous evolution of high-performance models across various categories and use cases.

Key Responsibilities

  • Design and implement automated pipelines for discovering, evaluating, and selecting high-performing open-source LLMs for different categories
  • Develop advanced model composition techniques (union-style capability aggregation and intersection-style conservative merging) using state-of-the-art merging frameworks
  • Parse and introspect LLM architectures in detail (similar to DOM tree parsing) — working with layers, attention mechanisms, state dictionaries, and parameter structures
  • Build and maintain iterative improvement loops: task processing, multi-model comparison, automated evaluation, knowledge gap detection, synthetic data generation, and targeted fine-tuning/merging
  • Implement efficient fine-tuning workflows using parameter-efficient methods (LoRA/QLoRA, PEFT) and delta adapter techniques
  • Create robust evaluation frameworks using multiple metrics (ROUGE, BERTScore, faithfulness, hallucination detection, etc.) and LLM-as-Judge systems
  • Optimize inference pipelines for high throughput using tools like vLLM or TGI
  • Set up MLOps practices for model versioning, experiment tracking, documentation, and deployment to Hugging Face Hub
  • Contribute to building reusable, modular components that can support multiple LLM categories and future domain expansions

Requirements (Freshers Welcome!)

  • B.Tech / M.Tech in Computer Science, Artificial Intelligence, Machine Learning, or equivalent (2024–2026 batch)
  • Strong proficiency in Python and PyTorch
  • Hands-on experience with Hugging Face Transformers library
  • Practical experience fine-tuning LLMs (Llama-3, Qwen2, Phi-3, or similar) using PEFT/LoRA/QLoRA
  • Solid understanding of Transformer architecture (attention mechanisms, layers, embeddings, LayerNorm, MLP blocks, etc.)
  • Comfort with model introspection tools (named_modules, state_dict, named_parameters)
  • Good knowledge of Git, Linux, and basic MLOps workflows
  • Strong problem-solving skills and excitement about working at the systems level of large language models

Big Advantages (Not Mandatory)

  • Exposure to model merging frameworks (mergekit or similar) and techniques like DARE, TIES, or SLERP
  • Experience with synthetic data generation or distillation methods
  • Familiarity with preference optimization (DPO, ORPO, etc.)
  • Experience running large models with vLLM or Text Generation Inference (TGI)
  • Active Hugging Face profile with public fine-tunes or experiments
  • Understanding of evaluation benchmarks and LLM-as-Judge patterns

What We Offer

  • Direct ownership of critical components in a cutting-edge LLM engineering stack
  • Access to high-end GPU cloud resources (A100/H100)
  • Fast learning curve and rapid growth — strong performers can move into senior LLM Architect roles quickly
  • Opportunity to work on multiple LLM categories and advanced model evolution systems
  • Collaborative environment with clear architecture guidance and mentorship

This is a high-impact role where you will gain deep, production-grade expertise in the most advanced areas of LLM engineering in 2026.

How to Apply

Please send your resume + GitHub profile link + a brief note or link showcasing one relevant LLM project (fine-tuning, merging, evaluation pipeline, or inference work) to [Confidential Information]

Subject Line: LLM Systems Engineer – 2026

We review applications on a rolling basis.

More Info

Job Type:
Industry:
Function:
Employment Type:

Job ID: 145770885

Similar Jobs