About Techjays
At Techjays, we build production-grade AI platforms for global clients. We operate at the intersection of backend engineering, distributed systems, and applied AI — delivering secure, scalable, and enterprise-ready intelligent systems. Our team has built and scaled products at Google, Akamai, NetApp, ADP, Cognizant, and Capgemini.
About The Role
This is not a feature-delivery role. We are looking for an
AI Lead who can architect, own, and scale intelligent backend systems end-to-end. You will drive both technical direction and execution — working across LLM integrations, RAG pipelines, agentic AI workflows, and cloud-native backend systems for global clients.
What You'll Do
- Architect and scale backend systems powering AI-driven applications
- Design and implement RAG pipelines, AI agents, and LLM integrations
- Own systems end-to-end — from architecture to deployment and scaling
- Integrate and optimize LLMs (Claude, GPT, Gemini) for real-world production use cases
- Build high-performance distributed systems with observability and cost efficiency
- Lead backend and AI initiatives with strong technical ownership
- Mentor engineers and raise the technical bar across teams
- Collaborate with product and AI teams to deliver AI-native solutions
What We're Looking For
- 6–10 years of strong backend engineering experience
- Hands-on expertise in Python (FastAPI / Django / Flask)
- Deep understanding of Generative AI and LLM-based systems
- Strong experience with RAG pipelines and Vector Databases (Pinecone, FAISS, ChromaDB, Weaviate)
- Solid knowledge of Agentic AI — building autonomous agents and multi-agent workflows
- Proficiency in AWS or GCP in production environments
- Experience with distributed systems, microservices, and system design
- Strong grasp of Data Structures, Algorithms, and Design Patterns
- Familiarity with WebSockets, Git, Linux/Unix, and CI/CD
Nice to Have
- Experience with Anthropic Claude API and Claude Code
- Familiarity with real-time data systems or streaming (Kafka, etc.)
- MLOps and AI system lifecycle experience
- Optimizing AI systems for latency, cost, and scalability
Who You Are
- You think in systems, not just features
- You take full ownership of what you build
- You are comfortable navigating fast-moving, ambiguous environments
- You stay updated with the latest in Generative AI and backend technologies
- Strong communicator who can collaborate across teams and global clients
What We Offer
- Competitive compensation (Best in Industry)
- Work on production-grade AI systems used by global clients
- Exposure to cutting-edge AI tools and frameworks
- A culture that values clarity, integrity, and continuous growth
Skills:- Retrieval Augmented Generation (RAG), Python, Generative AI, Agentic AI, Data Structures, Vector database, Large Language Models (LLM), Design patterns, Google Cloud Platform (GCP), Amazon Web Services (AWS), Algorithms, WebSocket, Git, Linux/Unix and Anthropic Claude