Job Description :
Hands on experience and proficiency in working with the Hadoop ecosystem, including: HDFS, Hive, Hbase, YARN, Sqoop, Oozie, Spark, Ambari, and Ranger.
At least 4+ years of experience with the Hadoop ecosystem and Big Data technologies.
Working knowledge of NoSQL, RDBMS, SQL, JSON, XML and ETL skills are a must.
Experience with performance/scalability and process computational complexity in Hadoop applications (Hive and Spark) for high performance and throughput.
Excellent Data Analysis skills. Must be comfortable with querying and analyzing large amount of data on Hadoop HDFS using Hive and Spark.
Experience using Spark Context, Spark-SQL, Data Frame, Pair RDD’s concepts running applications using Spark Framework
Intermediate experience in Hadoop administration - Knowledgeable in Ambari, performance tuning and configuration