Job Description :
Client: TXDOT/NTTDATA
Job Title: Senior Hadoop Developer
Job ID: 2017-191123
Duration: 8+ Months Contract
Location: Austin, TX - 78731

Job Responsibilities:
- Developing distributed computing Big Data applications using Spark, Elastic Search on HDP Hadoop
- Designing, developing, testing, tuning and building a large-scale data processing system, for Data Ingestion and Data products that allow the Client to improve quality, velocity and monetization of our data assets for both Operational Applications and Analytical needs.
- Strong experience in software engineering and development of solutions within the Hadoop Ecosystem.
- Troubleshoot and develop on Hadoop technologies including HDFS, Hive, Pig, Flume, HBase, Spark, Storm, Impala and Hadoop ETL development
- Experience in analyzing text, streams and complex event types with emerging Hadoop-based Big Data, NoSQL.
- Good understanding & experience with Performance and Performance tuning for complex S/W projects mainly around large scale and low latency
- Analyzing data with Hive, Pig and HBase

Basic Qualifications:
- Min. 5 years'' of experience in Hortonworks Hadoop - hdInsight/ HDFS/ Hbase/ flume/ Scoop/ MapReduce/ Yarn
- Hands on experience in job scheduling tools like Oozie and Falcon.
- Hands on experience in Pig and Hive Queries.
- Hands on experience in Importing and exporting the data using Sqoop from HDFS to Relational Database systems/mainframe and vice-versa.
- Min. 2 years''s of experience in Open Source-JS Frameworks-Angular JS
- Hands on experience in UNIX and Shell Scripting.
- Experience working with Spark for data manipulation, preparation, cleansing.

Nice to have Experiences (Added advantage)
- Informatica MDM, Informatica Power Center-Informatica Mart
- Web dev experience, mainly around Angular
- Knowledge of Informatica and Master Data management
- Knowledge of GIS System and Geospatial data processing in Hadoop


Client : TXDOT

             

Similar Jobs you may be interested in ..