Roles and Responsibilities:
- Develop scalable infrastructure and platform to collect, analyze and process large amounts of structured and unstructured data with real-time data interpretation.
- Work closely across an array of various teams and organizations in the company (including partners, customers)
Basic Qualifications:
- Bachelor's degree in Engineering in Computer Science/Electronics & Telecommunication.Minimum of 2 year of Big Data experience
Preferred Qualifications:
- 5+ years of software development experience using multiple computer languages.
- Experience building large scale distributed data processing systems/applications or large-scale internet systems (cloud computing)
- Strong foundational knowledge and experience with distributed systems and computing systems in general.
- Hands-on engineering skills.
- Should be able to develop big data solution and how that big data solution can be delivered using big data technology such as Hadoop/HDFS, MapReduce, Hive, AWS EMR, MongoDB, Airflow, Oozie,, Yarn, Ambari, ZooKeeper; Sqoop, BIRT or any other big data frameworks(with full life cycle of Hadoop Solution)
- Hands-on experience with Hadoop applications (e.g. administration, configuration management, monitoring, debugging, and performance tuning).
- Firm understanding of major programming/scripting languages like Java, Linux, Ruby, Kafka, Camunda,Phyton and/or R, Shell script.
- Should be able to build a big data environment on premises or in thecloudBroad understanding and experience of real-time analytics, NoSQL data stores, data modeling and data management, analytical tools, languages, or libraries (e.g. SAS, SPSS, R, Mahout).
- Experience with deploying container based applications using tools such as DockerExperience developing API centric solutions (REST API)