We are looking for a highly skilled
Data Engineer with strong expertise in
real-time streaming and event-driven architectures. The ideal candidate will be responsible for building scalable data pipelines, optimizing streaming systems, and ensuring high performance across distributed environments.
Key Responsibilities
- Design and implement real-time data streaming pipelines using technologies like Apache Flink, Kafka, and Java
- Build and maintain event-driven architectures for large-scale distributed systems
- Perform JVM tuning and performance optimization for streaming applications
- Develop and deploy applications using containerization tools (Docker, Kubernetes)
- Work with Cloudera platform for data engineering and pipeline orchestration
- Implement robust design patterns and ensure high-quality coding standards
- Troubleshoot and resolve issues in distributed systems ecosystem
- Collaborate with DevOps teams to maintain CI/CD pipelines (GitHub, Jenkins)
- Work on Linux-based systems, including configuration and shell scripting
- Optimize data processing with caching mechanisms (e.g., Redis - nice to have)
Required Skills & Experience
- Strong hands-on experience in Real-Time Streaming (Flink / Kafka / Java)
- Deep understanding of event-driven architecture
- Experience with JVM performance tuning
- Proficiency in Docker and Kubernetes
- Solid experience in Linux OS and shell scripting
- Knowledge of design patterns and scalable system design
- Experience with CI/CD tools like GitHub and Jenkins
- Hands-on troubleshooting experience in distributed systems
Nice to Have
- Experience with Redis or other caching systems
- Exposure to Cloudera Data Platform engineering
- Prior experience in banking or financial domain is a plus