JD for Big Data Positions:
Key Responsibilities:
- Develop and maintain big data pipelines using Hadoop and Spark (Scala)
- Process, transform, and analyze large datasets efficiently
- Collaborate with data engineers, analysts, and cross-functional teams
- Optimize data workflows for performance and scalability
- Troubleshoot and debug data-related issues
- Ensure data quality, reliability, and consistency
Required Skills (Must-have):
- Strong experience with Hadoop ecosystem
- Hands-on experience in Spark with Scala
- Good understanding of distributed data processing
- Solid programming and problem-solving skills
Preferred Skills (Nice-to-have):
- Knowledge of Git (version control)
- Understanding of JSON structure and data formats