Job Description :
Job Title: Big Data Developer
Location: Los Angeles, CA
Duration: 6+Months

Job Description:

Responsible for administration of the Big Data platform including the areas of back up, configuration, performance tuning, security and promoting code
Install, Configure, and apply patch updates
Installing and configuring additional big data parcels like Informatica etc.
Partner with key roles (e.g. application analysts, data integration team, database administrators team, infrastructure team, and business constituents to maintain a robust and sustainable big data platform system
System recovery in the event of system failures
Implement and manage various environments including failover for BCP and DR
Maintain best practices for maintaining Hadoop in production
Troubleshoot, diagnose, tuning and solving Hadoop issues including HDFS, YARN, Hive, Impala and Spark
Work with vendor support teams to triage issues

Big Data Development

Provide expertise in architecting, designing and implementing Big Data solutions
Interface with business and IT stakeholders to identify issues and develop solutions
Responsible for full cycle development of Big Data solutions including architecting data staging , data movement , data streaming using technologies like Hive QL , Spark , Flume etc.
Perform code reviews on big data components
Code deployment across environments
Document and communicate technical architectures, standards best practices, toolsets and solutions
Responsible for standards across all the SDLC phases
Ensures solutions developed adhere to security and data privacy polices
Translate complex functional and technical requirements into detail design
Define best practices and standards
Maintain clear documentation to help increase overall productivity

Big Data as a service
Investigate and prove concepts as Big Data technologies evolves
Test prototypes and oversee operational processes
Support ad-hoc requests as requested by management and business partners

Skills and Knowledge
Experience in administering high performance and large Hadoop clusters – preferably Cloudera on Oracle Big Data Appliance
Experience with Kerberos Hadoop clusters
Experience managing and developing utilizing open source technologies and libraries
Experience with HiveQL and Spark
Experience with RDBMS and NoSQL database technologies
Hands on experience with Linux administration and shell scripting
Familiarity with version control, job scheduling and configuration management tools
Ability to lead and take ownership of projects
Some programing experience (Java, Python etc
Good oral and written communication and presentations skills with the ability to lead meetings effectively
Experience in SDLC Process
             

Similar Jobs you may be interested in ..