Job Description :
*Job Location*SFO, CA

*Role *Data track - lead

*Project Duration*6 months

*Number of positions (Separate indent # for each position1

*Responsibilites - In detail*

Lead the team and take responsibility of overall deliverables
Analyze, Design, Development, integration, Unit testing, support SIT, UAT,
PFIX
Research and create POC''s
Provide insights on sanitized production data (banking)
Compare different products (client suggested and Persistent IP) and provide
recommendations

Front end with vendor partners as required
*Technical Skills - In Detail*
Essential Skills
Good understanding of Data pipeline including Extraction, acquisition,
transformation and visualization
Prior experience working with RBDMS and Big data distributions
Experience with requirements gathering, systems development, systems
integration and designing/developing API
Experience with Linux and shell programming
Experience in machine learning models and Python 2.x and 3.x
Experience with frameworks like Anaconda and developing ETL using PySpark
on any major Big Data distributions
Good understanding of XML processing using Python,Spark RDD and dataframes
Performance tuning, unit testing and integration testing
Excellent communication and articulation skills
Self starter and ability to work in dynamic and agile environment
Experience of working with Hadoop ecosystem MapReduce, Hive, HBase
Experience of working with at least one NoSQL database like C*, MongoDB
Experience of working on AWS
*Experience Level*
10-15 years
             

Similar Jobs you may be interested in ..