Job Description :
responsibilities:
Ideal candidates would have working knowledge of Hadoop ecosystem and AWS ecosystem. This includes but not limited to relational data stores, Data Integration techniques, XML, Python, Spark, ETL techniques in Hadoop and AWS ecosystem. Projects leverage Agile methodology for enabling new business data capabilities.

qualifications:
Requirements/Critical Skills (4-6 years working experience):
Able to work efficiently under UNIX/Linux environment
Experience working with in-memory computing using R, Python, Spark, PySpark and Scala.
Experience in parsing and shredding XML and JSON, shell scripting, and SQL
Experience working with Hadoop ecosystem - HDFS, Hive, Ambari, Atlas, Ranger
Experience working with AWS ecosystem - S3, EMR, EC2, Lambda Cloud Formation, Cloud Watch, SNS/SQS
Experience working with SQL and No SQL databases
Experience with DevOps, CI/CD implementations of the following technologies: Docker or Jenkins or Test Driven Development patterns
Experience designing and developing data sourcing routines utilizing typical data quality functions involving standardization, transformation, rationalization, linking and matching
Knowledge of data, master data and metadata related standards, processes and technology
Experience working with multi-Terabyte data sets (structured, semi/unstructured data)
Knowledge of job scheduling and monitoring tools (Eg: Autosys, Airflow)

Bachelor''s Degree

skills: Preferred Skills:
Experience in the Financial, and information management needs are a plus
Experience in Financial Services/Secondary Mortgage industry is a plus
Demonstrated flexibility and adaptability to changes - Agile methodology experience with Jira
Demonstrated ability to manage multiple priorities and deadlines
Ability to work independently or as part of a team
Experience with Attunity CDC and Kafka