Job Description :

Reasonable accommodations may be made to enable individuals with disabilities to perform the essential functions. This is not intended to be an all-inclusive list of job duties and the position will perform other duties as assigned.

  • Hands-on experience on Cloudera installation, configuration, debugging, tuning and administration.
  • Collaborate and assist developers in successful implementation of their code, monitor and fine tune their process for optimum resource utilization on cluster, ability to automate run time process
  • Must have prior Cloudera Hadoop (CDP) cluster deployment experience from the scratch.
  • Hands-on experience on Cloudera, working with data delivery teams to setup new Hadoop users. This includes setting up Linux users, setting up Kerberos principals and testing HDFS, Hive.
  • Exposure to Data Science collaborative tools such as data science workbench, Jupyter hub etc
  • Competency in Red Hat Linux administration (security, configuration, tuning, troubleshooting, and monitoring).
  • Expert knowledge on Active Directory/LDAP security integration with Cloudera Big Data platform.
  • Performance tuning of Cloudera clusters and Spark (PySpark, SparkQL,) and MapReduce routines.
  • Optimize clusters for future workloads.
  • Hands-on experience on node management, monitoring and response, support processes creation, upgrades, and patches, logging configuration and managing user rights and space quota.
  • Working knowledge of Networks, Linux OS, and Unix Shell Scripting.
  • Strong hands-on experience in implementation of security like Kerberos, Sentry, OS Upgrade and TLS/SSL implementation etc.
  • Experience with setting up and configuring the YARN queues using the YARN queue manager and Authorization through Apache sentry.
  • Important role is to keep the cluster performing well and dependable
  • Perform periodic updates to cluster and keeping the stack current
  • Familiarity with Cloudera’s BDR tool to perform and monitor backups of critical data and able to restore data when in need
  • Exposure to AWS, or any Cloud Platform
  • Exposure to Snowflake (Nice to have)