Job Description :
Sr Data Engineer/ Data Architect with experience in Data pipelines, ETL, Unstructured data and MySQL.

Healthcare/Medicare Required*

We are looking for a hands-on Data Architect with a start-up mindset that wants to join a dynamic team that is disrupting the healthcare industry.
You will use your extensive data engineering experience in building data pipelines from third-party data sources and internal applications and building analytical models and reporting framework for deriving actionable insights from the collected data.
The ideal candidate should have an in-depth understanding of real-time data harvesting, parsing, storing and modeling for reporting purposes.
The Data Architect will be working very closely with our engineering and business stakeholders for architecture and design of Enterprise Data Hub, Data Lake, to support our business operational objectives.
You''re an excellent communicator, happy to work with people from several different business units. You can translate business needs into technical requirements and implementation. You are a Rockstar when it comes to creating actionable insights from big data.


Responsibilities:
Design and code data pipeline features and data processing jobs for collecting data from disparate systems for the purpose of storing in a data warehouse.
Work with real-time data streams from multiple internal/external sources
Ensure smooth ongoing operations of various data platforms with high availability while making continuous improvements
Provide scalable solutions to manage large imports and implement operational procedures as necessary
Design robust enterprise data resource management policies including security, backup and recovery, performance monitoring and tuning and high availability
Partner with member care and operations teams to understand user acquisition, engagement, satisfaction, etc.
Participate in peer code reviews and be comfortable in writing high quality documentation about their work
Willingness and ability to work in a startup environment.
Proven expertise in building data pipelines using Spark, Python, Shell Scripting.
Experience in working with large amounts of unstructured data using AWS S3, Cloudwatch logs, Athena, or similar cloud services.
Strong data modeling skills (relational, dimensional and flattened) and analytical SQL skills for ingesting and retrieving data from MPP databases such as Snowflake or Redshift.
Experience integrating data from multiple data sources and file types such as JSON, XMS, CSV and non-traditional data structures
Proven expertise working with Docker, familiar with virtual machines, or other containerized environments
Experience with a visualization tool and/or Embedded Analytics
Experience in microservice architecture and developing AI-driven microservices for application consumption
High level of comfort in Linux environments on the command line
Knowledge of machine learning, information retrieval, analytics, statistics, deep learning techniques
Has strong interpersonal communication skills; effectively communicates in verbal and written form.
Experienced with star schema design, data warehousing, and business intelligence best practices

Knowledge on the below
Architected secure, scalable, cloud-based solutions that are HIPAA and/or PCI compliant
Knowledge of Java, Javascript and/or Node JS
Knowledge of Jupyter and other machine learning tools
Knowledge working with GIT or any other version control system.
Experience working in a fast paced/startup environment
Understanding of Tableau server architecture and maintenance
Understanding of Airflow or another workflow scheduler, such as Luigi
Masters in CS, Data Analytics, Data Science or other related advanced degree program