
Search by job, company or skills
What You'll Do
Define technical vision and architecture: Lead the design and architecture of scalable, production-grade GenAI systems using LangChain, LangGraph, FastAPI, and modern AI frameworks. Make critical technical decisions that balance innovation, scalability, and maintainability
Drive GenAI innovation: Research, prototype, and implement cutting-edge GenAI techniques including advanced prompt engineering, RAG optimization, agentic AI systems, fine-tuning, and multi-agent orchestration
Lead AI experimentation and evaluation: Design and implement comprehensive evaluation frameworks for LLM performance, establish best practices for prompt engineering, and drive data-driven decision making through rigorous experimentation
Own GenAI at utility and program scale: Architect and optimize LLM workloads that support millions of calls across multiple utilities and channels, under strict cost and latency targets.
Architect event-driven AI systems: Design and implement sophisticated event-driven architectures using SQS, asynchronous workflows, distributed processing, and microservices patterns
Establish GenAI best practices: Define and enforce coding standards, design patterns, testing strategies, and operational excellence for GenAI applications. Champion observability, monitoring, and reliability
Technical leadership and mentorship: Mentor and guide engineers across all levels, conduct architecture reviews, provide technical direction, and foster a culture of continuous learning and innovation
Cross-functional collaboration: Partner with product management, data science, engineering leadership, and business stakeholders to translate business requirements into technical solutions. Communicate complex technical concepts to diverse audiences
Drive strategic initiatives: Identify opportunities for AI-driven innovation, evaluate emerging technologies, and lead proof-of-concept projects that align with business objectives Ensure production excellence: Establish SLAs, implement monitoring and alerting, optimize performance and cost, and ensure high availability of AI services
Thought leadership: Represent Bidgely in the GenAI community through technical blogs, conference talks, open-source contributions, and industry engagement
What Makes You Successful
Education: BS/MS+ in Computer Science, AI/ML, or equivalent from premier institutes
Experience: 5+ years of software engineering experience with at least 1-2 years focused on Generative AI, Agentic AI, and production LLM-based applications
Expert-level knowledge of LangChain, LangGraph, and LLM orchestration frameworks; deep understanding of prompt engineering, RAG architectures, agentic AI, and multi-agent systems
Production AI Systems: Proven track record of architecting and deploying production-grade AI systems at scale, handling millions of requests and complex workflows
API & Microservices: Extensive experience building scalable RESTful APIs, microservices architectures, and event-driven systems using FastAPI, Flask, or similar frameworks LLM Integration: Deep expertise integrating and optimizing LLM providers; understanding of model selection, cost optimization, and latency management
Technical Leadership: Demonstrated ability to lead technical teams, drive architectural decisions, and mentor engineers; experience with code reviews, technical design documents, and cross-team collaboration
Problem Solving: Exceptional problem-solving skills with ability to tackle ambiguous, complex challenges; strong in algorithms, data structures, and system design
Communication: Outstanding communication skills; ability to influence technical direction, present to leadership, and collaborate effectively across global distributed teams
Preferred Qualifications
Python & Architecture: Expert-level Python programming with deep knowledge of software architecture, design patterns, distributed systems, and performance optimization
Experience with vector databases (ChromaDB, Pinecone, Weaviate, FAISS) and advanced retrieval techniques
Deep knowledge of LLM observability tools (LangSmith, Langfuse, Weights & Biases) and evaluation frameworks
Expertise in AWS services (Bedrock, SQS, S3, Lambda, ECS, Parameter Store) and cloud-native architectures
Experience with fine-tuning, RLHF, prompt optimization techniques, and model evaluation methodologies
Knowledge of PostgreSQL optimization, Redis caching strategies, and database performance tuning
Experience with Docker, Kubernetes, Terraform, and infrastructure-as-code
Background in building conversational AI, chatbots, or virtual assistants
Experience with MLOps, model deployment pipelines, and A/B testing frameworks
Familiarity with emerging AI trends: multi-modal models, reasoning models, agent frameworks
Interested candidates can share their resume with : [Confidential Information]
Job ID: 142073811