Job Description :

Description

Top 3 skillsets for the position - Big Data Lake Administrator
• Minimum 5 years work experience in Hadoop ecosystems (Horton HDP or Cloudera’s CDH/CDP)
• should demonstrate strong concepts in Unix/Linux, Windows OS, cloud platforms (MS Azure, AWS)
•  must have good exposure to Cloudera manager, Cloudera Navigator or similar cluster management tool

 

Role: manage and maintain Data Lake clusters infrastructure: installation, configuration, performance tuning and monitoring of Hadoop clusters

  • BS degree, preferably in Computer science or equivalent
  • Good communication skills with right attitude to blend in with team
  • Minimum 5 years work experience in Hadoop ecosystems (Horton HDP or Cloudera’s CDH/CDP)
  • should demonstrate strong concepts in Unix/Linux, Windows OS, cloud platforms (MS Azure, AWS)
  • must have good exposure to Cloudera manager, Cloudera Navigator or similar cluster management tool
  • Collaborate and assist developers in successful implementation of their code, monitor and fine tune their process for optimum resource utilization on cluster, ability to automate run time process
  • must have good knowledge of HDFS, Ranger/Sentry, Hive, Impala, Spark, HBase, Kudu, Kafka, Ni-Fi, sqoop and other Hadoop related services
  • exposure to Data Science collaborative tools such as data science workbench, anaconda, etc.
  • should have strong Networking concepts: topology, proxy, F5, firewall
  • Strong security concepts: Active directory ,Kerberos, LDAP, SAML, SSL, data encryption @rest
  • programming language concepts: Java, Perl, python, PySpark and Unix shell scripting
  • Experience in tuning and optimizing Hadoop environment in keeping clusters healthy and available for end users and applications with maximum cluster uptime as defined in SLA
  • important role is to keep the cluster performing well and dependable
  • should possess experience in cluster management, perform cluster upgrade, migration and testing
  • Perform periodic updates to cluster and keeping the stack current
  • Familiarity with Cloudera’s BDR tool to perform and monitor backups of critical data and able to restore data when in need
  • ability to expand clusters by adding new nodes and rebalance cluster storage systems
  • manage application databases, application integration, users, roles, permissions within cluster
  • collaborate with Unix, network, database and security teams on cluster related matters
  • must monitor cluster for maximum uptime, ability to research on cluster issues via logs and collaborate with support in a proactive way
             

Similar Jobs you may be interested in ..