Job Description :
B.S./B.A. Computer Science or related field; or equivalent level of experience
Working knowledge of Hadoop services including HDFS, HBase, Hive, Spark and other emerging technologies
5+ years working in Information Technology including experience with UNIX/Linux operating systems and basic networking concepts
4+ years working with Hadoop clusters including hands-on experience working with HDFS, YARN, Spark, Hive, Spark, Zookeeper, Dremio, HBASE and Kerberos
3+ years of experience with different real-time data streaming technologies tools such as Apache Kafka, Strom, Amazon Kinesis, Apache Samsa etc.
4 years of Data Modeling, SQL experience. 4 years of ETL experience.
Cloud (AWS/AZURE) experience with data ingestion, data catalog, data lake etc.
Minimum of 6 years of programming experience and Minimum of 3 years of Kafka platform experience in open source kafka, zookeepers, kafka connect, schema registry (avro schemas)
Experience with automation tools like provisioning using open shift, kubernetes , Docker , Chef.
Good knowledge in administration and operations of the kafka platform like provisioning, access lists (Kerberos)
Good knowledge on tuning the real time infrastructure w.r.t performance and scalability.
Should be good with scaling, sizing, configuring the cluster and configuring the services across the cluster.
Experience working in applications, systems or IT operations
Experience in agile methodology such as sprint planning and backlog grooming, Comfort with frequent, incremental code, and testing
Previous software engineering and development experience preferred
Strong passion for technology and a willingness to learn new skills
Self-motivated and self-directed, ability to translate technical direction and apply to specific work responsibilities
Ability to work effectively with cross-functional teams
Excellent communication (written and verbal) and interpersonal skills
Strong technical aptitude
Strong analytical skills