Job Description :
*Job title: Big Data (Sqoop, Spark) Developer*

*Location: Plano, TX*

*Duration: 7 months*

Need Passport No

Competency: Digital : BigData and Hadoop Ecosystems, Digital : IBM BigData
Exp:6 - 8

Essentials:

The Lead Engineer, Enterprise Analytics (Advanced Software Engineer) will
report to a Senior Manager - Enterprise Analytics. Using Big Data
technology, you will be responsible for designing, developing and
implementing data analysis using cloud architecture and applying data
science techniques for data exploration

Role:

Primary Responsibilities: Conducting data analysis, research
and data modeling using structured and unstructured data Process,
cleanse, and verify the integrity of data used for analysis
Perform ad-hoc analysis and present results in a clear manner
Design rich data visualizations to communicate complex ideas to customers
or company leaders Developing automated methods for
ingesting large datasets into an enterprise-scale analytical system using
Scoop, Spark and Kafka Identifying technical implementation
options and issues Plays a key role in defining the future IT
architecture and develops roadmaps of the systems, platforms and/or
processes for the company across disciplines Provides technical
leadership and coaching to other engineers and architects across the
organization Partners and communicates cross-functionally
across the enterprise Provides highest level of expertise for the
development and/or specification and/or implementation of Analytical
solutions (e.g. R, Python, Tableau & Datameer) Able to
explain technical issues to senior leaders in the company, including the
board, in non-technical, understandable terms Interact with business
teams to gather, interpret and understand their business needs creating
design specifications. Foster the continuous evolution of best practices
within the development team to ensure data standardization and consistency

Desired:

Core Competencies & Accomplishments: 10 + years of professional
experience 2+ years of experience with Big data technology and analytics
Experience in one or more languages (e.g., Python or Java)
Experience with data visualization tools like Tableau
Experience with applied data sciences techniques,
including but not limited to, machine learning approaches and statistical
modeling Proficiency in using query languages such as SQL, Hive
In-depth knowledge of statistical software packages (e.g.
SAS and R) Understanding of Big Data tools (e.g., NoSQL DB, Hadoop,
Hbase) Understanding of data preparation and manipulation using Datameer
tool Knowledge of SOA, IaaS, and Cloud Computing technologies,
particularly in the AWS environment Experience with Hadoop/Hive,
Spark and Scoop highly desirable Experience with continuous
software integration, test and deployment. Experience in agile
software development paradigm (e.g., Scrum, Kanban) The
ability to work within a dynamic programmatic environment with evolving
requirements and capability goals Self-motivated. Capable of working with
little or no supervision Maintenance of up-to-date knowledge in the
appropriate technical area Strong written and verbal
communication
             

Similar Jobs you may be interested in ..