Job Description :
Hadoop Admin
Montana
Long term

Must have:
6+ years of relevant experience - Hadoop, System administration with sound knowledge in Unix based Operating System internals
Experience in setting up and supporting Hadoop environment (Cloud and On-premise
Ability to design the structure, plan the capacity, performance fine-tune, monitor the components of the cluster
Ability to closely work with the infrastructure, networking, development teams to ensure the timely deliverables
Ability to setup, configure and implement security for Hadoop clusters using Kerberos
Experience in setting up services like YARN, HDFS, Zookeeper, Hive, Impala, HBase etc
Linux tuning - understands kernel parameters tuning and tweaking them?• Strong scripting background (shell/bash, python/ruby, etc- Automation, configuration management
Develop and document best practices
Disk RAID - filesystem setup etc.
Ability to perform Backup And Recovery as applicable?
Experience in large scale enterprise apps that involves provisioning systems using virtualization on private clouds and using physical servers
The interest, ability, curiosity to tinker and automate everything in a repeatable fashion with least amount of manual work involved
Strong analytical skills and "Get it done" attitude

Must have at least one of these:
Chef/Puppet/Ansible awareness?
Any mongodb, Sassandra, hadoop, elastic search cluster administration experience
AWS experience (VPC, OpsWorks, IAM )


Preferred to have understanding of the following tools:
Log aggregation, ELK, Splunk, logstash, Ganglia, Nags?
BlueGreenDeployments - understanding and experience?
Measuring availability, SLAs, 99th percentile, 99.9th percentile
Open source Technology background

Job Description:
Responsible for implementation and ongoing administration of Hadoop infrastructure as part of our global Managed Services group.
Primary responsibilities would include ensuring high availability of the Services in the cluster and ensuring that the Cluster resources are utilized optimally reducing the costs.
Work closely with India and US team with some overlap hours to ensure a proper handover resulting in optimal productivity. Install and configure monitoring tools.
Diligently teaming with the infrastructure, network, database, application and business teams to guarantee high data quality and availability.
Managing the Hadoop File system and monitoring the services and performing backup and disaster recovery (when applicable) for Hadoop data.
Coordinate root cause analysis (RCA) efforts to minimize future system issues. Provide off hours support on a rotational basis. Assist with develop and maintain the system runbooks.
Create and publish various production metrics including system performance and reliability information to systems owners and management.
Work with big data developers and developers designing scalable supportable infrastructure. Perform duties as needed within the engineering team.