Job Description :
Hi,

Hope you are doing great today.

Job Role :: Hadoop Developer/Lead
Job Location :: Raleigh, NC
Job Duration :: Full Time
Mode of Interview :: Face to Face


Job Description::

Description:
Position responsible for operational support of the Hadoop cluster used at Prime Therapeutics. Technologies include Cloudera Navigator, Cloudera Manager, HDFS, Spark, Map Reduce, Hive, PIG, Hbase, YARN, Flume, Scoop, and HiveQL, Scala, Java and REST API technologies. The candidate will support these technologies under the direction of the infrastructure operation team.
Responsibilities:
Install, upgrade, configure, and apply patches for Cloudera Manager
Setup of Hadoop cluster and maintenance support of the cluster
Keeping track of Hadoop cluster connectivity and security
Capacity planning and monitoring of Hadoop cluster job performances
HDFS maintenance and support
Resource manager configuration and trouble shooting
Setting up Hadoop users
Testing HDFS, Hive, Pig, and MapReduce access for the new users
Backup, recovery and maintenance
Consult with business users to manage tasks, incidents and focus on
Create/Update and manage reports for metrics and performance
Work with Shell Scripts, Python Scripts, and Ansible
Prepare file system / mount points
Install required services LDAP, DNS, and etc.
Collaborate with engineering, development, and operation teams to troubleshoot and resolve their issues
Manage Hadoop jobs using scheduler
Cluster coordination Services
Point of contact for vendor escalation
Executes and provides feedback for operational policies, procedure, processes, and standards
Automate manual tasks
Develop infrastructure documents
Troubleshoot production problems within assigned software applications.
Other duties as assigned
Technical Qualifications:
HDFS file system experience
Experience with disaster recovery and business continuity practice with Hadoop clusters
Minimum 3 years broad-based experience with Cloudera and Hadoop technologies
Scripting skills in Linux environment
Hands on experience in Cloudera, Hadoop cluster, HCatalog, and Hive
Hadoop authentication protocol (Kerbero) knowledge
Development skills in Java, REST API, Python, Scala
Experience with Design of Data Integration using ETL, ELT or patterns.
Map Reduce, PIG, HIVE, SPARK, HADOOP