Job Description :
Job Title: Hadoop Architect
Location: Minneapolis, MN
Duration: Long Term

Job Description
Our Big Data team is responsible for delivering a large scale data platform . This includes the implementation and maintenance of our Hadoop platform. We are looking for someone who can develop a strategy by creating architecture blueprints, validating designs and providing recommendations on the enterprise platform strategic roadmap. Experience with design and develop in high-volume real-time big data platforms is preferred.

Responsibilities:
Design, Administration, management, and tuning of Hadoop environment.
Collaborate with the infrastructure team to coordinate OS-level patching and to identify and solve hardware related issues.
Cluster & Node Maintenance, Health Checks, Automation of Job Monitoring by creating alerts.
Assist development team in identifying the root cause of slow performing jobs / queries (HDFS
Capacity Planning / Forecasting.
Develop strategy to automate management and deployment processes (DevOps
Deploying and managing all Hadoop platform components.
Planning and conducting platform upgrades.
Work with development staff to ensure all components are ready for release / deployment.
Collaborate with Project Managers, Developers and business staff to develop products & services.
Participate in managing and maintaining the product on an on-going basis

Skills/Experience
Strong UNIX/Linux knowledge including the ability to understand the interaction between applications and the operating system.
Ability to provide recommendations and suggestions related to troubleshooting and performance improvement.
Experience designing and administering a reasonably-sized Hadoop cluster (100+ nodes
Experience in running, using and troubleshooting the Apache Big Data stack i.e. Hadoop FS, Hive, HBase, Kafka, Pig, Oozie, Yarn, Sqoop, Flume etc.
Ability to create infrastructure capacity plans based on quantitative and qualitative data points.
Experience with implementing and managing security for a multitenancy environment,
Familiarity with networking stack from TCP/IP and up
Good work ethics with extremely high standard of code quality, system reliability, and performance.
Experience processing large amounts of structured and unstructured data with MapReduce.
Experience with data movement and transformation technologies.
Experience with tuning and troubleshooting JVM environment.

Nice to Have

Experience with data virtualization using Presto-DB or JBOSS Tied or any other similar technologies.
Experience in SQL and Relation Database developing data extraction applications.
             

Similar Jobs you may be interested in ..