Job Description :
new position 03/27 - LH_BDA_NY

Rate: $90/hr on C2C

Big Data Architect – Hadoop Technologies

Buffalo NY

12 Months +

Contract

Hadoop/Hive, Elastic Search needed

Banking, Ab initio a plus


Job Requirement: Solutions Architect

Responsible to work with the Portfolio and Platform Architects, Technical Lead & Engineers to define the detailed solution design (incl. Story Decomposition) for stories in backlog

Accountable for ensuring the products & services are supported by the right architectures and solutions meet the customer needs
Accountable for ensuring the design of the product solutions are cost effective and maintained through the agile development lifecycle, managing the flow of the backlog of design activities
Working with Ops Engineers to ensure operational issues are resolved and that any design related issues are addressed in a timely manner
Convert requirements into actionable product/service requirements that feed IT solutions development & influence service direction.
Responsible for ensuring solutions are aligned with the platform architecture and roadmap, group standards and policies and the overall enterprise architecture for their function

Skills
Advanced HIVE, HQL, Hadoop
HDFS concepts, UDFs, xml sqoop HBASE, SPARK, Scala
Data Modeling (Logical as well as physical)
Data mapping (ETL system to Conformed Data Models)
Solid Documentation Skills
Nexus, Jenkins, G3, Github, Defect Debugging, Ansible

Team will be responsible to ETL data from various source systems into an FCR Common Data Model in the Group Big Data System for use in AML. This is a base requirement for AMS3 and Event Triage.

Systems may be (but not limited to):

–Transaction Monitoring systems: CAMP, FCCM Mantas, NORKOM

–Core Banking Systems: HUB, HOGAN, OBS, RPS, etc.

–HSBC Systems; AMU, SCION

–3rd party data sources: ICX4, Google,

–Initially, the Team will have a large backlog with CAMP and MANTAS ingestion being the priority. To support this, a larger/temporary workforce is needed, comprised of HSBC staff and contractors.

After the initial ingestion of CAMP/MANTAS, BAU processing is expected to be minimal, but steady ingestion of various systems.
ETL scope may expand into areas such as NLP and data mining unstructured data to build structured data.

AMS3 - Deliver a strategic solution to build a narrative containing supplementary alert, case, transaction, focal entity information to aide in working an AML alert in UCM.

Current AMS3 solution will be enhanced with many new business features, additional microservices and leverage Spark HBASE.
AMS3 will also need to consume data (ETL) into an AMS consumption view from the FCR CDM which will be the strategic data warehouse for AML Data

–Transaction Monitoring systems data: CAMP, FCCM Mantas, NORKOM

–Core Banking Systems data: HUB, HOGAN, OBS, RPS, etc.

–HSBC Systems data; AMU, SCION

–3rd party data sources: ICX4, Google,

In the later term AMS3, will need to consume escalated alerts from the Event Triage process, and data stored in the CDM

With the demand for work we will need to add an additional Feature POD