
Search by job, company or skills
Experience required: 24 years of professional software engineering experience.
Budget for the role - Max Up to Rs. 15 LPA
100% remote Opportunity
Role Overview:
You will design, build, and operate software for data collection and processing at scale. The role is handson, with emphasis on clean design, reliability, and performance.
Key Responsibilities:
Develop and maintain Python applications for crawling, parsing, enrichment, and processing of large datasets.
Build and operate data workflows (ETL/ELT), including validation, monitoring, and errorhandling.
Work with SQL and NoSQL (plus vector databases/data lakes) for modeling, storage, and retrieval.
Contribute to system design using cloudnative components on AWS (e.g., S3, Lambda, ECS/EKS, SQS/SNS, RDS/DynamoDB, CloudWatch).
Implement and consume APIs/microservices; write clear contracts and documentation.
Write unit/integration tests, perform debugging and profiling; contribute to code reviews and maintain high code quality.
Implement observability (logging/metrics/tracing) and basic security practices (secrets, IAM, least privilege).
Collaborate with Dev/QA/Ops; ship incrementally using PRs and design docs.
Required Qualifications
24 years of professional software engineering experience.
Strong proficiency in Python; good knowledge of data structures/algorithms and software design principles.
Handson with SQL and at least one NoSQL store; familiarity with vector databases is a plus.
Experience with web scraping frameworks (e.g., Scrapy, Selenium/Playwright, BeautifulSoup) and resilient crawling patterns (respect robots/rotations/retries).
Practical understanding of system design and distributed systems basics.
Exposure to AWS services and cloudnative design; comfortable on Linux and with Git.
Preferred / Good to Have (Prioritized)
GenAI & LLMs: experience with LangChain, CrewAI, LlamaIndex, prompt design, RAG patterns, and vector stores. (Candidates with this experience will be prioritized.)
CI/CD & Containers: exposure to pipelines (GitHub Actions/Jenkins), Docker, and Kubernetes.
Data Pipelines/Big Data: ETL/ELT, Airflow, Spark, Kafka, or similar.
Infra as Code: Terraform/CloudFormation; basic cost and performanceoptimization on cloud.
Frontend/JS: not required; basic JS or frontend skills are a nicetohave only.
Exposure to GCP/Azure.
How We Work
Ownership of modules endtoend (design build deploy operate).
Clear communication, collaborative problemsolving, and documentation.
Pragmatic engineering: small PRs, incremental delivery, and measurable reliability.
WorkfromHome Requirements
Highspeed internet for calls and collaboration.
A capable, reliable computer (modern CPU, 8GB+ RAM).
Headphones with clear audio quality.
Stable power and backup arrangements.
ForageAI is an equalopportunity employer. We value curiosity, craftsmanship, and collaboration.
Job ID: 130413987