Job Description :

Hi,    

I hope you are doing well. I'm Ashok from Nityo Infotech Corp. We are actively looking for SDET with ETL Big Data and please share your consultant resume.

 

Role: SDET with ETL Big Data

Location: San Ramon, CA

 

Essential Job Functions

•  Create & execute software test design, test strategies, test cases for Bank’s MDM, Digital, Marketing, Transactional systems data ingestion through batch and streaming framework.

•  Perform detailed Data Analysis, Data Reconciliation and Data Integrity in databases and file systems using SQL, Hive, Spark and other programming languages/tools.

•  Validate the data ingestion pipeline built using NIFI, Spark, Kafka, Ariflow and Denodo.

•  Create test data in various file formats an databases.

•  Analyze the logs created by data ingestion jobs and report any failures and performance issues in detail.

•  Build reusable standard TCOE automation frameworks for functional, regression, performance and E2E testing using DevOps tools and scripts.

•  Define and document software test plan and perform several phases of testing cycles and record test results, test metrics and track defects using Jira-Xray in order to ensure applications, products and/or releases are in compliance with Bank of the West`s QA standards.

•  Partner end-to-end with Product Managers, Architects, Tech Leads, Data Governance team to understand business requirements to recommend quality improvement best practices and processes.

Other Job Duties

•  High level of personal commitment to each task, a can do attitude and a drive to deliver.

•  Strong communication skills to be able to communicate at all levels.

•  Ability to understand banking financial requirements.

•  High level of responsibility and Ownership from inception through to implementation.

•  The candidate should have a proactive approach to problem solving.

•  Good analysis skills in order to aid in trouble shooting and problem solving.

•  Good testing principles as well as good defect management skills should have experience in handling.

•  Multiple assignments at same time and long with other team members.

Qualifications

Required Experience

•  Overall 7+ years of experience of IT experience with ETL/ Data Warehouse Testing or Development experience with Big data experience definitely a plus.

•  Experience in writing complex SQL, python/shell Scripts to test data ingestion framework based on the data mapping & requirements provided and perform extensive data analysis to identify the defects.

•  Strong Data Analytics, ETL, Data warehouse, Data Virtualization, BI Dashboard concepts.

•  Experience in working with large scale Big data/Enterprise Data Warehouse, Data Integration, Data Migration and upgrade projects.

•  Experience in testing complex data systems, data ingestion pipeline through batch, real time/streaming framework.

•  Experience in building/updating automating frameworks using programming languages such as Python/Java/Shell or previous proven programming experience in any relevant scripting languages.

•  Experience in test data setup in various file formats and databases.

•  Experience working with Database upgrades, tool upgrades and Interface testing.

•  Experience in creating Test plan, test cases and engineering best practices related to software test engineering, both manual and automated testing.

•  Experience coordinating testing activities and optimizing test cycles working with project team.

•  Experience with utilization of appropriate test methodologies and use of test management tools like JIRA.

•  Experience with conducting and running defect triage meeting with project teams.

•  Experience working with financial services applications.

•  Strong in SDLC process, Test Strategy/Plan, Test Estimation and experience working in Agile Methodology.

•  Effective Project & People Management Skills, exposure to Knowledge Management. Solid time management and prioritization skills.

•  Excellent verbal and written communication skills.

 

Preferred Skills

•  Domain knowledge & previous experience with Banking and Financial services.

•  Passion towards giving technical solutions and using different testing tools.

•  Experience or knowledge of Big data tools such as Spark, NIFI, Kafka, Denodo technologies, Hive, NOSQL Databases like HBase, Cassandra, MangoDB, BI tools like PowerBI, Zeppeline, etc.

•  Experience in working data integration testing pipeline built using heterogeneous source systems like transactional databases & files systems (JSON, delimited, COBOL, Parquet, Avro, etc.,), HDFS, API and Webservices.

•  Experience in source control tools like GitLab or GitHub and built DevOps CICT pipeline or similar.

•  Experience in using Data Governance, Metadata and data lineage tools like Schema Registry, Atlas, ABACUS, etc.

•  Experience in data masking, tokenization, detokenization process and testing the same.

•  Experience handling multiple assignments and be a team player

 

 

Looking forward to work with you

 

,

Ashok Raju

Nityo Infotech Corp.
Suite 1285, 666 Plainsboro Road
Plainsboro , NJ , 08536

E-mail: 

Cell: 

Desk:  EXT: 4029    

 

 “If you feel you received this email by mistake or wish to unsubscribe, kindly reply to this email with “UNSUBSCRIBE” in the subject line”

-----

             

Similar Jobs you may be interested in ..