About Techjays
At Techjays, we build production-grade AI platforms for global clients. We operate at the intersection of backend engineering, distributed systems, and applied AI — delivering secure, scalable, and enterprise-ready intelligent systems. Our team has built and scaled products at Google, Akamai, NetApp, ADP, Cognizant, and Capgemini.
About The Role
This is not a feature-delivery role. We are looking for an
AI Lead who can architect, own, and scale intelligent backend systems end-to-end. You will drive both technical direction and execution — working across LLM integrations, RAG pipelines, agentic AI workflows, and cloud-native backend systems for global clients.
What You'll Do
- Architect and scale backend systems powering AI-driven applications
- Design and implement RAG pipelines, AI agents, and LLM integrations
- Own systems end-to-end — from architecture to deployment and scaling
- Integrate and optimize LLMs (Claude, GPT, Gemini) for real-world production use cases
- Build high-performance distributed systems with observability and cost efficiency
- Lead backend and AI initiatives with strong technical ownership
- Mentor engineers and raise the technical bar across teams
- Collaborate with product and AI teams to deliver AI-native solutions
What We're Looking For
- 6–10 years of strong backend engineering experience
- Hands-on expertise in Python (FastAPI / Django / Flask)
- Deep understanding of Generative AI and LLM-based systems
- Strong experience with RAG pipelines and Vector Databases (Pinecone, FAISS, ChromaDB, Weaviate)
- Solid knowledge of Agentic AI — building autonomous agents and multi-agent workflows
- Proficiency in AWS or GCP in production environments
- Experience with distributed systems, microservices, and system design
- Strong grasp of Data Structures, Algorithms, and Design Patterns
- Familiarity with WebSockets, Git, Linux/Unix, and CI/CD
Nice to Have
- Experience with Anthropic Claude API and Claude Code
- Familiarity with real-time data systems or streaming (Kafka, etc.)
- MLOps and AI system lifecycle experience
- Optimizing AI systems for latency, cost, and scalability
Who You Are
- You think in systems, not just features
- You take full ownership of what you build
- You are comfortable navigating fast-moving, ambiguous environments
- You stay updated with the latest in Generative AI and backend technologies
- Strong communicator who can collaborate across teams and global clients
What We Offer
- Competitive compensation (Best in Industry)
- Work on production-grade AI systems used by global clients
- Exposure to cutting-edge AI tools and frameworks
- A culture that values clarity, integrity, and continuous growth
Skills:- Retrieval Augmented Generation (RAG), Python, LangGraph, Natural Language Processing (NLP), Data Structures, Robot Framework, Google Cloud Platform (GCP), Amazon Web Services (AWS), Algorithms, WebSocket, Git, Linux/Unix and Anthropic Claude