Job Description :
 "Experience in a technical lead role in a big data environment with Hadoop distributions such as cloudera or hortonworks . They should have experience in the following
1. Good programming skills with experience in HDFS, Hive, Impala, Spark, Kafka, Flume, Pig etc
2. Data analytics, wrangling and BI solutions usage in Hadoop ecosystem
3. Worked in both Batch and Real time implementation workflows on Hadoop, usage of Autosys JIL is preferable
4. NOSQL implementation - Cassandra / HBase would be preferable
5. Should have worked with some data quality metric tools like Informatica BDM, DQ
6. MS Azure Cloud experience is preferable
Specific experience in one or more following areas would be a plus
• Assemble large, complex data sets that meet functional / non-functional business requirements.
• Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
• Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL
• Advanced working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases.
• Experience building and optimizing 39;big data39; data pipelines, architectures and data sets.
• Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
• Build processes supporting data transformation, data structures, metadata, dependency and workload management.
• A successful history of manipulating, processing and extracting value from large disconnected datasets.
• Working knowledge of message queuing, stream processing, and highly scalable 39;big data39; data stores   "

Client : Accenture Flex