Job Description :
Role: Hadoop Architect
Location: Jersey City, NJ and Pittsburgh, PA
Duration: 12+ Months
Interview mode: Phone and F2F
Client: Life and Health Insurance Industry

Job Description:
Our direct client is seeking a Hadoop Architect in their Jersey City, NJ and Pittsburgh, PA
locations.
This is a highly visible and critical role within our Client39;s evidence based analytics system
which process data at a massive scale, allowing machine learning and analytical algorithms
to run on top of the data we have, generating analytics in real-time, using streaming
technologies that we are actively building out.
The system will allow the user to ask questions and provide deep insights into your
processes, allowing for evidence-based management.
Responsibilities:
Predominately a java open source environment.
Includes RESTful web services and APIs.
Other key elements are Hadoop, Apache Storm and Spark, Vertica, Kafka, Zookeeper,
Elastic Search, Scala, NoSQL DB39;s, CockroachDB.
Always considering adding new tools such as Apache Flink and Druid.
We are not expecting you to have experience with everything in our tech stack. All relevant
experience will be considered.
Required Skills:
Development experience in a java web environment (java, REST web services
Experience with data structures and algorithms.
Experience with highly distributed systems drawing from extremely large data sets (both
structured and unstructured will be considered
Positive attitude and the ability to learn and listen overall a great team player.
Preferably a Master39;s in Computer Science, A B.S in Computer Science is required.
Hands-on experience in dealing with large data sets and distributed computing in Big Data
and systems using Hadoop.
Proficiency in Java or Scala, writing software for distributed systems.
Experience in designing and developing RESTful API and WebServices.
Experience in writing software with Hadoop or Spark.
Experience around developing code for large clusters with huge volumes of data –
streaming as well as batch code.
Strong knowledge of Linux.
At least 3 years of experience with and strong understanding of Big Data technologies in
Hadoop ecosystem – Hive, HDFS, Map/Reduce, Yarn, Pig, Oozie, HBase, Sqoop, etc.
At least 1 year of data streaming; Spark, Kafka, Storm.
At least 1 year of NoSQL Database; Druid, Cassandra, Hbase.

Client : Life and Health Insurance Industry