Job Description :
Responsibilities
Collaborate with internal/external business partners on big data projects
Utilize technical expertise in Hadoop administration
Evaluate new big data tools, frameworks and technologies, explore Proof of Concept (POC) to identify optimum solutions for requested capabilities
Ensure holistic understanding of the BIG DATA Ecosystem
Install, maintain, and administer software on Linux servers
Automate manual processes using tools such as Python, Ruby, Unix Shell (bash, ksh) etc.
Monitor Big Data Application/Infrastructure Performance and availability
Implement ETL processes from various data sources to Hadoop cluster

Qualifications
Bachelors'' of Science in Computer Science or a related field
4+ years'' experience in the following:
Automating build/deployment, software configuration, continuous integration/continuous delivery, release engineering related tasks in a big data Environments
Automating manual processes and Developing using Python/Java, Unix Shell (bash, ash), SQL etc.
Big Data Components/Frameworks such as Hadoop (MapR), Spark, Yarn, Kafka, Flink, ELK etc.
NoSQL databases such as HBase, Cassandra, MapR DB
Big Data querying tools such as Drill, Presto, Hive etc
Infrastructure automation tools like Ansible
Monitoring tools like Grafana, Splunk etc
Monitoring Application/Infrastructure Performance and availability.
Experience or understanding of developing machine/deep learning systems in a distributed environment.
Development tools such as GIT, and familiarity with collaboration tools such as Jira and Confluence or similar tools.
Containerization (Docker) and resource scheduling (Kubernetes