Job Description :
Hello, We have an urgent requirement for Hadoop Cloudera Admin. Below is the job description for it. If you are interested please reply me with updated resume. Title: Hadoop Cloudera Admin Duration: 6-12 Months contract to hire Location: Remote Overview: As a Cloudera administrator you will be responsible to set up, shape, administer, and test the applications as part of our project delivery team specific to Hadoop platforms You will be part of a team of DevOps engineers focusing on the day-to-day tasks of managing and maintaining On-Prem and Cloud environments and will be hands-on involved with CI/CD process and monitoring application servers. Candidate must be comfortable working in an agile environment. Responsibilities: Installation of Cloudera-Hadoop from scratch with SDLC cycle methodology (Dev, Test, Cert, Production, Disaster Recovery Responsible for planning of Capacity Planning, Infrastructure Planning and version fix to build Hadoop Cluster Implementation of Hadoop Security like Kerberos, Cloudera Key Trustee Server and Key Trustee Management Systems. Enable Sentry for RBAC (role-based access control) to have a privilege level access to the data in HDFS. Enriched knowledge in Hadoop ecosystem including HDFS Spark PySpark Sqoop Flume Hive Impala PIG MapReduce Proven understanding and related experience with Hadoop HBase Hive Pig Sqoop Flume Hbase Kafka flume and or MapReduce as well as Unix OS Core Java programming shell scripting experience Perform upgrades to Cloudera Manager, CDH along with support for Linux Server Patching Provide infrastructure and support for software developers to rapidly iterate on their products and services and deliver high-quality results. This includes infrastructure for automated builds and testing, continuous integration, software releases, and system deployment Designing and Implementation of Backup and Disaster Recovery strategy based out of Cloudera BDR utility for Batch applications and Kafka mirror maker for real-time streaming applications. Aligning with development and architecture teams to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments. Should be able to develop Kafka into Hadoop to enable real-time streaming applications. Monitor and coordinate all data system operations, including security procedures, and liaison with infrastructure, security, DevOps, Data Platform and Application team. Broad understanding of tools and technologies: source control, continuous integration, infrastructure automation, deployment automation, container concepts, orchestration and cloud Ensure proper resource utilization between the different development teams and processes Design and implement a toolset that simplifies provisioning and support of a large cluster environment Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments Apply proper architecture guidelines to ensure highly available services Review performance stats and query execution/explain plans; recommend changes for tuning Create and maintain detailed, up-to-date technical documentation Collaborating with cross-functional teams across hardware, platform services and operations. Solve live performance and stability issues and prevent recurrence Qualifications: Bachelor's Degree in Computer Science or equivalent work experience Expertise in administration and implementation of Cloudrea Hadoop(CDH), Cloudera Manager, HDFS, Yarn, MapReduce, Hive, Impala, KUDU, Sqoop, Spark, Kafka, HBase, Teradata Studio Express, Teradata, Tableau, Kerberos, Active Directory, Sentry, TLS/SSL, Linux/RHEL, Unix Windows, SBT, Maven, Jenkins, Oracle, MS SQL Server, Shell Scripting, Eclipse IDE, Git, At least 5 years of related experience as a Hadoop administrator with an expert level knowledge of Cloudera Hadoop components such as HDFS, Sentry, HBase, Impala, Hue, Spark, Hive, Kafka, YARN, and Zoo Keeper. Must have prior Hadoop cluster deployment experience in adding and removing nodes, troubleshooting failed jobs, configure and tune the clusters, monitor critical parts of the cluster. Hands-on experience with Cloudera, working with data delivery teams to setup new Hadoop users. This includes setting up Linux users, setting up Kerberos principals and testing HDFS, Hive, Pig and MapReduce access for the new users. Competency in Red Hat Linux administration (security, configuration, tuning, troubleshooting and monitoring Experience in utilizing and implementing Zoo Keeper and Broker with Kafka Experience in setting up Kerberos principals and testing HDFS, Hive, Impala and Spark access for the new users Strong knowledge of scripting and automation tools and strategies, e.g. Shell, Python, PowerShell, Must have the ability to identify complex problems and review related information to develop and evaluate options and implement solutions. Experience with performance tuning of Cloudera clusters, YARN & Spark Healthcare IT experience a plus. Regards, Thanks Sourabh Kumar United Software Group Inc. Phone eye