Job Description :
Hadoop Developer

Minneapolis, MN

6+ Months

$55/hr c2c

Please send candidates who can provide Passport number and Visa copy.
NO OPTs

Description:

Position responsible for operational support of the Hadoop cluster used at client.

Technologies include

Cloudera Navigator, Cloudera Manager, HDFS, Spark, Map Reduce, Hive, PIG, Hbase, YARN, Flume, Scoop, and HiveQL, Scala, Java and REST API technologies.

The candidate will support these technologies under the direction of the infrastructure operation team.

The candidate will work closely with the infrastructure engineering build team.

Responsibilities:

Install, upgrade, configure, and apply patches for Cloudera Manager

Setup of Hadoop cluster and maintenance support of the cluster

Keeping track of Hadoop cluster connectivity and security

Capacity planning and monitoring of Hadoop cluster job performances

HDFS maintenance and support

Resource manager configuration and trouble shooting

Setting up Hadoop users

Testing HDFS, Hive, Pig, and MapReduce access for the new users

Backup, recovery and maintenance

Consult with business users to manage tasks, incidents and focus on

Create/Update and manage reports for metrics and performance

Work with Shell Scripts, Python Scripts, and Ansible

Prepare file system / mount points

Install required services LDAP, DNS, and etc.

Collaborate with engineering, development, and operation teams to troubleshoot and resolve their issues

Manage Hadoop jobs using scheduler

Cluster coordination Services

Point of contact for vendor escalation

Executes and provides feedback for operational policies, procedure, processes, and standards

Automate manual tasks

Develop infrastructure documents

Troubleshoot production problems within assigned software applications.

Other duties as assigned

Technical Qualifications:

Administrator background – Hadoop cluster administration, Install, and upgrades

Experience with Cloudera software support and Cloudera Manager

HDFS file system experience

Experience with disaster recovery and business continuity practice with Hadoop clusters

Minimum 3 years broad-based experience with Cloudera and Hadoop technologies

Scripting skills in Linux environment

Hands on experience in Cloudera, Hadoop cluster, HCatalog, and Hive

Hadoop authentication protocol (Kerbero) knowledge

Development skills in Java, REST API, Python, Scala

Experience with Design of Data Integration using ETL, ELT or patterns.

Map Reduce, PIG, HIVE, SPARK, HADOOP