Java+Big Data
Work exp - 5 to 7 yrs
Key Responsibilities
- Design, develop, and maintain Java-based data processing applications.
- Build and optimize Big Data pipelines using technologies such as Hadoop, Spark, Hive, HBase, Kafka etc.
- Work on ETL processes, data ingestion, data transformation, and data quality improvements.
- Implement and optimize distributed data processing solutions for large datasets.
- Integrate Java applications with Big Data ecosystems (HDFS, Hive, Kafka, etc.).
- Develop REST APIs and microservices using Java/Spring Boot for data access and processing.
- Troubleshoot performance issues across complex data pipelines.
- Collaborate with data engineers, architects, and analysts to understand business requirements.
- Ensure best practices in coding, version control, CI/CD, and cloud-based deployments.
- Write clean, maintainable, and scalable code following industry standards.
Required Skills
Java Skills
- Strong proficiency in Core Java, OOPS, Collections, Multithreading
- Experience with Spring / Spring Boot
- Hands-on experience in building scalable backend systems
- Good understanding of REST APIs and microservices
Big Data Skills
- Experience with Hadoop ecosystem HDFS, YARN, MapReduce
- Strong knowledge of Apache Spark (RDD, DataFrame, Spark SQL)
- Experience with Hive / Impala for querying large datasets
- Knowledge of Kafka or other messaging/streaming platforms
- Experience with NoSQL databases (HBase, Cassandra, MongoDB)
Additional Good-to-Have Skills
- Experience with Airflow, NiFi, or similar orchestration tools
- Exposure to AWS / Azure / GCP Big Data services
- Familiarity with Docker / Kubernetes
- Knowledge of data modeling, data warehousing concepts
- Performance tuning and optimization for distributed systems