Search by job, company or skills

BayOne Solutions

Lead Hadoop Administrator

Save
new job description bg glownew job description bg glownew job description bg svg
  • Posted a day ago
  • Be among the first 10 applicants
Early Applicant

Job Description

Position: Platform Engineer

Shift Timings: General Shift (Flexibility to take calls with US Team)

Location: Remote

Overview

We are seeking a Platform Support Engineer to support the operations of our on‑premise big‑data environment. The ideal candidate brings strong foundational skills in Linux/Windows systems, working knowledge of Hadoop ecosystem and components, and the ability to monitor, troubleshoot, and maintain data‑ingestion pipelines in a distributed environment. This role is hands‑on and requires a detail‑oriented individual who is comfortable working in a cross functional team.

Responsibilities:

  • Monitor the health and performance of distributed systems, including Hadoop components and data‑ingestion pipelines.
  • Review logs and basic metrics to support troubleshooting and operational investigations.
  • Support virtualized environments, including understanding how services run across VMs in the cluster.
  • Work closely with a highly cross-functional team of development engineers to implement solutions to customer issues.
  • Perform user acceptance testing before deploying solutions to the field.
  • Develop and implement procedures for configuration and testing of systems.
  • Create documentation highlighting root cause corrective actions and ensure the developed solutions are actionable by our field service engineers.
  • Mentor field service engineers .
  • Travel internationally to customer sites if the issue cannot be resolved remotely.

Required Skills & Experience

  • BS/MS degree in a computer science field or related discipline
  • 3+ years experience in software support and troubleshooting
  • Extensive programming background in one or more language (shell scripting, Java, SQL, Python, R)
  • Linux fundamentals: command‑line navigation, file operations, and basic troubleshooting.
  • Windows OS familiarity, incl. interacting with system tools and environments.
  • Comfort working in virtualized environments (Hyper‑V or similar) and understanding the role of VMs in system operation.
  • Hadoop ecosystem awareness, including major components such as HDFS, YARN, and related services (HBase, Kafka, Solr, Flume). Candidate must understand the hdfs environment.
  • Experience using a cluster management UI (e.g., Ambari or similar platform) to view service status and metrics.
  • Understanding of data ingestion pipelines and the general flow of data from source to processing to indexing.
  • Experience using standard system/operations tools such as SSH clients (e.g., MobaXterm) and basic monitoring utilities.
  • Ability to read logs and identify common failure indicators.
  • Good understanding of file formats including JSON and others
  • Excellent communication skills and the ability to clearly, document, articulate, and share findings with the team
  • Passionate about solving problems involving complex interrelated software systems
  • Able to work independently and escalate appropriately.
  • Curious, adaptable, and eager to deepen their understanding of large‑scale data platforms.

Preferred Qualifications

  • 6+ years experience in Hadoop, HBase, Spark, Kafka, Solr
  • Experience with Jira/Confluence
  • Experience with databases like Oracle, Postgres, HBase
  • Experience (or at least familiarity) with relational database schemas
  • Experience with semiconductor capital equipment systems, SECS/GEM, factory automation, and data storage
  • Experience with Kubernetes and Docker.
  • Hands-On experience with IP networking and networking equipment (Switches, Routers, Firewalls, etc.
  • Exposure to Spark job monitoring or log review.
  • General hardware familiarity (servers, storage, replacements). Not required but beneficial.

More Info

Job Type:
Industry:
Employment Type:

About Company

Job ID: 147322513