Job Description :
·         Experience in Hadoop distributions (Apache / Cloudera / Hortonworks)

·         Ability to deploy and maintain multi-node Hadoop cluster

·         Experience working on SPARK RDD, SPARK SQL & SPARK STREAMING.

·         Efficient on coding spark code by using scala/python

·         Experience working with Big Data eco-system including tools such as Hadoop, Map Reduce, Yarn, Hive, Pig, Impala, Spark , Kafka, Hive, Impala and Storm to name a few  Knowledgeable in techniques for designing Hadoop-based file layout optimized to meet business needs

·         Understands the tradeoffs between different approaches to Hadoop file design

·         Experience with techniques of performance optimization for both data loading and data retrieval

·         Experience with NoSQL Databases – HBase, Apache Cassandra,  or MongoDB

·         Able to translate business requirements into logical and physical file structure design

·         Ability to build and test rapidly Map Reduce code in a rapid, iterative manner

·         Ability to articulate reasons behind the design choices being made