Job Description :
Job Responsibilities:
Developing distributed computing Big Data applications using Spark, Elastic Search on HDP Hadoop
Designing, developing, testing, tuning and building a large-scale data processing system, for Data Ingestion and Data products that allow the Client to improve quality, velocity and monetization of our data assets for both Operational Applications and Analytical needs.
Strong experience in software engineering and development of solutions within the Hadoop Ecosystem.
Troubleshoot and develop on Hadoop technologies including HDFS, Hive, Pig, Flume, HBase, Spark, Storm, Impala and Hadoop ETL development
Experience in analyzing text, streams and complex event types with emerging Hadoop-based Big Data, NoSQL.
Good understanding & experience with Performance and Performance tuning for complex S/W projects mainly around large scale and low latency
Analyzing data with Hive, Pig and HBase
Basic Qualifications:
Min. 5 years'' of experience in Hortonworks Hadoop - hdInsight/ HDFS/ Hbase/ flume/ Scoop/ MapReduce/ Yarn
Hands on experience in job scheduling tools like Oozie and Falcon.
Hands on experience in Pig and Hive Queries.
Hands on experience in Importing and exporting the data using Sqoop from HDFS to Relational Database systems/mainframe and vice-versa.
Min. 2 years''s of experience in Open Source-JS Frameworks-Angular JS
Hands on experience in UNIX and Shell Scripting.
Experience working with Spark for data manipulation, preparation, cleansing.
Nice to have Experiences (Added advantage)
Informatica MDM, Informatica Power Center-Informatica Mart
Web dev experience, mainly around Angular
Knowledge of Informatica and Master Data management
Knowledge of GIS System and Geospatial data processing in Hadoop