Job Description :
MUST HAVE:
Required Skills and Knowledge:
Advanced knowledge of data architectures, data pipelines, real time processing, streaming, networking, and security
Proficient understanding of distributed computing principles
Good knowledge of Big Data querying tools, such as Pig or Hive
Good understanding of Lambda Architecture, along with its advantages and drawbacks •Proficiency with MapReduce, HDFS
Experience with integration of data from multiple data sources
Basic Qualifications:
Bachelors Degree in Software Engineer or similar degree
12+ years experience in software engineering
Experience developing ETL processing flows using MapReduce technologies like Spark and Hadoop
Experience developing with ingestion and clustering frameworks such as Kafka, Zookeeper, YARN
Experience with building stream-processing systems, using solutions such as Storm or Spark-Streaming
Experience with various messaging systems, such as Kafka or RabbitMQ


Preferred Experience:
Experience with DataBricks and Spark
Ability to solve any ongoing issues with operating the cluster
Management of Spark or Hadoop clusters, with all included services
Experience with NoSQL databases, such as HBase, Cassandra, MongoDB
Experience with Big Data ML toolkits, such as Mahout, SparkML, or H2O
Understanding of Service Oriented Architecture
Technical writing, system documentation, design document-management skills


Client : Direct

             

Similar Jobs you may be interested in ..