- Experience in developing REST API services using one of the Scala frameworks.
- Ability to troubleshoot and optimize complex queries on the Spark platform.
- Expert in building and optimizing big data, data, and ML pipelines, architectures, and data sets.
- Knowledge in modeling unstructured to structured data design.
- Experience in Big Data access and storage techniques.
- Experience in doing cost estimation based on the design and development.
- Excellent debugging skills for the technical stack mentioned above, which even includes analyzing server logs and application logs.
- Highly organized, self-motivated, proactive, and ability to propose best design solutions.
- Good time management and multitasking skills to work to deadlines by working independently and as part of a team.
Experience - Must have:
- Scala: Minimum 2 years of experience.
- Spark: Minimum 2 years of experience.
- Hadoop: Minimum 2 years of experience (Security, Spark on YARN, Architectural knowledge).
- HBase: Minimum 2 years of experience.
- Hive: Minimum 2 years of experience.
- RDBMS (MySQL, Postgres, Maria): Minimum 2 years of experience.
- CI/CD: Minimum 1 year of experience.
Experience (Good to have):
- Kafka.
- Spark Streaming.
- Apache Phoenix.
- Caching layer (Memcache, Redis).
- Spark ML.
- FP (Scala cats, scalaz).
Qualifications:
- Bachelor's degree in IT, Computer Science, Software Engineering, Business Analytics or equivalent with at least 2 years of experience in big data systems such as Hadoop as well as cloud-based solutions.