Job Description :
  • Programming in Big Data technology area, specifically using Cloudera distribution
  • Highly skilled in Spark and Scala
  • Must have expert level knowledge of Hive, Impala and Kudu
  • Strong knowledge in SQL and Unix/Linux scripts
  • Exposure to other Hadoop Ecosystem technologies like YARN, Zookeeper, HDFS, Avro, Parquet etc.
  • Experience with cleansing, preparing large, complex data sets for reporting and analytics
  • Knowledge on storage design concepts including partitioning
  • Maintain, modify and improve large set of structured and unstructured data
  • Monitoring and troubleshooting Spark jobs



Similar Jobs you may be interested in ..