Position: Platform Engineer
Shift Timings: General Shift (Flexibility to take calls with US Team)
Location: Remote
Overview
We are seeking a Platform Support Engineer to support the operations of our on‑premise big‑data environment. The ideal candidate brings strong foundational skills in Linux/Windows systems, working knowledge of Hadoop ecosystem and components, and the ability to monitor, troubleshoot, and maintain data‑ingestion pipelines in a distributed environment. This role is hands‑on and requires a detail‑oriented individual who is comfortable working in a cross functional team.
Responsibilities:
- Monitor the health and performance of distributed systems, including Hadoop components and data‑ingestion pipelines.
- Review logs and basic metrics to support troubleshooting and operational investigations.
- Support virtualized environments, including understanding how services run across VMs in the cluster.
- Work closely with a highly cross-functional team of development engineers to implement solutions to customer issues.
- Perform user acceptance testing before deploying solutions to the field.
- Develop and implement procedures for configuration and testing of systems.
- Create documentation highlighting root cause corrective actions and ensure the developed solutions are actionable by our field service engineers.
- Mentor field service engineers .
- Travel internationally to customer sites if the issue cannot be resolved remotely.
Required Skills & Experience
- BS/MS degree in a computer science field or related discipline
- 3+ years experience in software support and troubleshooting
- Extensive programming background in one or more language (shell scripting, Java, SQL, Python, R)
- Linux fundamentals: command‑line navigation, file operations, and basic troubleshooting.
- Windows OS familiarity, incl. interacting with system tools and environments.
- Comfort working in virtualized environments (Hyper‑V or similar) and understanding the role of VMs in system operation.
- Hadoop ecosystem awareness, including major components such as HDFS, YARN, and related services (HBase, Kafka, Solr, Flume). Candidate must understand the hdfs environment.
- Experience using a cluster management UI (e.g., Ambari or similar platform) to view service status and metrics.
- Understanding of data ingestion pipelines and the general flow of data from source to processing to indexing.
- Experience using standard system/operations tools such as SSH clients (e.g., MobaXterm) and basic monitoring utilities.
- Ability to read logs and identify common failure indicators.
- Good understanding of file formats including JSON and others
- Excellent communication skills and the ability to clearly, document, articulate, and share findings with the team
- Passionate about solving problems involving complex interrelated software systems
- Able to work independently and escalate appropriately.
- Curious, adaptable, and eager to deepen their understanding of large‑scale data platforms.
Preferred Qualifications
- 6+ years experience in Hadoop, HBase, Spark, Kafka, Solr
- Experience with Jira/Confluence
- Experience with databases like Oracle, Postgres, HBase
- Experience (or at least familiarity) with relational database schemas
- Experience with semiconductor capital equipment systems, SECS/GEM, factory automation, and data storage
- Experience with Kubernetes and Docker.
- Hands-On experience with IP networking and networking equipment (Switches, Routers, Firewalls, etc.
- Exposure to Spark job monitoring or log review.
- General hardware familiarity (servers, storage, replacements). Not required but beneficial.