Job Description :
Job Purpose: This position is responsible for developing, integrating, testing, and maintaining existing and new applications; knowing one or more programming languages; knowing one or more development methodologies / delivery models.
Required Job Qualifications:
· Must have qualifications –
o Bachelor Degree and 4 years Information Technology experience OR Technical Certification and/or College Courses and 6 year Information Technology experience OR 8 years Information Technology experience.
o Hands on experience in developing, and maintaining software solutions in Hadoop cluster.
o Experience with strong UNIX shell scripting, SQOOP, eclipse, HCatalog, Pig scripts, Hive QL, UDF
o Hands-on experience with Spark and Scala.
o Experience in Kafka
o Familiarity with Hadoop Best Practices, Troubleshooting and Performance tuning.
o Experience with change management / DevOps tools (Github / Jenkins etc
o Familiarity to SDLC Methodology (Agile / Scrum / Iterative Development
· Nice-to-have qualifications
o Working experience with developing MapReduce programs running on the Hadoop cluster using Java/Python.
o Experience with NoSQL Databases like HBASE, Mongo or Cassandra
o Experience using Talend with Hadoop technologies.
o Working experience in the data warehousing and Business Intelligence systems
o Business requirements management and Systems change / configuration management. Familiarity with JIRA.
o Experience in ZENA (or any other scheduling tool)
o Healthcare experience