Job Description :
Roles and Responsibilities:
Develops ETL (Extract / Transform / Load) processes and tools for real-time and offline analytic processing.
Collaborates with data science team to gain a deep understanding of their needs.
Applies knowledge of Hadoop architecture and experience designing & optimizing queries to develop large scale data structures and pipelines to organize, collect and standardize data that helps generate insights and addresses reporting needs.
Uses strong programming skills in Python, Java, SQL or any of the major languages(Java/Scala) to build robust data pipelines and dynamic systems.
Builds data marts and data models to support Data Science and other internal customers.
Integrates data from a variety of sources, assuring that they adhere to data quality and accessibility standards.
Analyzes current information technology environments to identify and assess critical capabilities and recommend solutions.
Experiments with available tools and advises on new tools in order to determine optimal solution given the requirements dictated by the model/use case
             

Similar Jobs you may be interested in ..