Job Description :
Good knowledge of database structures, theories, principles, and practices

· Experience in Java and Knowledge in Hadoop (HDFS/HBASE/KUDU/SPARKSQL and SPARK/Scala with or prior experience in MapReduce) concepts and ability to write Spark/Scala RDD jobs.

· Proven understanding with Hadoop, HBase, Hive, and HBase

· Ability to write Pig Latin scripts. Familiarity with data loading tools like Flume, Kafka and Sqoop.

· Knowledge of workflow/schedulers like Oozie.

· Good aptitude in multi-threading and concurrency concept.

· Loading data from disparate data source sets.

· Certifications like Cloudera Developer/Administrator Certifications added advantage.

· Hands on experience with at least two NO SQL databases.

· Ability to analyze, identify issues with existing cluster, suggest architectural design changes.

· Ability/Knowledge to implement Data Governance in Hadoop clusters