Job Description :
Minimum of 4 years experience in architecture, design and development of Big data systems using Java/Scala, working on systems that are distributed, highly available, performant and scalable.
Strong experience with MapReduce, HDFS and Hive are required.
Experience with Spark would be a plus. • Experience with designing and implementing large-scale systems to process Terabytes to Petabytes of data. • Relational database experience preferably Teradata and demonstrated abilities in SQL and data modeling are required.
Proficiency with NoSQL databases is desired as well.
Strong experience with data deep dives and product analytical skills are required.
Experience with data visualization in Tableau or other business intelligence tools is desirable.
Experience E2E automation of data pipelines is required.
Experience with working in UNIX environment and scripting in Shell/Perl/Python is required.
Having SEO domain knowledge is a plus. •
Ability to take requirements from design through to implementation both independently and with larger teams.
Strong problem solving and debugging skills are required.
Ability to communicate effectively, both written and verbal, with technical and non-technical cross-functional teams.
Results oriented and deadline driven.

Top 3 skills: Data analyst with engineer experience ; Hadoop, big data, SDLC life cycle, design, development, production Teradata, SQL, hive python