Search by job, company or skills

  • Posted 8 days ago
  • Over 200 applicants
Quick Apply

Job Description

What You'll Do

Define technical vision and architecture: Lead the design and architecture of scalable, production-grade GenAI systems using LangChain, LangGraph, FastAPI, and modern AI frameworks. Make critical technical decisions that balance innovation, scalability, and maintainability

Drive GenAI innovation: Research, prototype, and implement cutting-edge GenAI techniques including advanced prompt engineering, RAG optimization, agentic AI systems, fine-tuning, and multi-agent orchestration

Lead AI experimentation and evaluation: Design and implement comprehensive evaluation frameworks for LLM performance, establish best practices for prompt engineering, and drive data-driven decision making through rigorous experimentation

Own GenAI at utility and program scale: Architect and optimize LLM workloads that support millions of calls across multiple utilities and channels, under strict cost and latency targets.

Architect event-driven AI systems: Design and implement sophisticated event-driven architectures using SQS, asynchronous workflows, distributed processing, and microservices patterns

Establish GenAI best practices: Define and enforce coding standards, design patterns, testing strategies, and operational excellence for GenAI applications. Champion observability, monitoring, and reliability

Technical leadership and mentorship: Mentor and guide engineers across all levels, conduct architecture reviews, provide technical direction, and foster a culture of continuous learning and innovation

Cross-functional collaboration: Partner with product management, data science, engineering leadership, and business stakeholders to translate business requirements into technical solutions. Communicate complex technical concepts to diverse audiences

Drive strategic initiatives: Identify opportunities for AI-driven innovation, evaluate emerging technologies, and lead proof-of-concept projects that align with business objectives Ensure production excellence: Establish SLAs, implement monitoring and alerting, optimize performance and cost, and ensure high availability of AI services

Thought leadership: Represent Bidgely in the GenAI community through technical blogs, conference talks, open-source contributions, and industry engagement

What Makes You Successful

Education: BS/MS+ in Computer Science, AI/ML, or equivalent from premier institutes

Experience: 5+ years of software engineering experience with at least 1-2 years focused on Generative AI, Agentic AI, and production LLM-based applications

Expert-level knowledge of LangChain, LangGraph, and LLM orchestration frameworks; deep understanding of prompt engineering, RAG architectures, agentic AI, and multi-agent systems

Production AI Systems: Proven track record of architecting and deploying production-grade AI systems at scale, handling millions of requests and complex workflows

API & Microservices: Extensive experience building scalable RESTful APIs, microservices architectures, and event-driven systems using FastAPI, Flask, or similar frameworks LLM Integration: Deep expertise integrating and optimizing LLM providers; understanding of model selection, cost optimization, and latency management

Technical Leadership: Demonstrated ability to lead technical teams, drive architectural decisions, and mentor engineers; experience with code reviews, technical design documents, and cross-team collaboration

Problem Solving: Exceptional problem-solving skills with ability to tackle ambiguous, complex challenges; strong in algorithms, data structures, and system design

Communication: Outstanding communication skills; ability to influence technical direction, present to leadership, and collaborate effectively across global distributed teams

Preferred Qualifications

Python & Architecture: Expert-level Python programming with deep knowledge of software architecture, design patterns, distributed systems, and performance optimization

Experience with vector databases (ChromaDB, Pinecone, Weaviate, FAISS) and advanced retrieval techniques

Deep knowledge of LLM observability tools (LangSmith, Langfuse, Weights & Biases) and evaluation frameworks

Expertise in AWS services (Bedrock, SQS, S3, Lambda, ECS, Parameter Store) and cloud-native architectures

Experience with fine-tuning, RLHF, prompt optimization techniques, and model evaluation methodologies

Knowledge of PostgreSQL optimization, Redis caching strategies, and database performance tuning

Experience with Docker, Kubernetes, Terraform, and infrastructure-as-code

Background in building conversational AI, chatbots, or virtual assistants

Experience with MLOps, model deployment pipelines, and A/B testing frameworks

Familiarity with emerging AI trends: multi-modal models, reasoning models, agent frameworks

Interested candidates can share their resume with : [Confidential Information]

More Info

Function:
Employment Type:
Open to candidates from:
Indian

Job ID: 142073811

User Avatar
0 Active Jobs