Job Description :
Job Title: Data lake architect
Location: Chicago, IL
Job type: Fulltime

The Data Lake Architect, Senior Advisor is responsible for (but not limited to):
You will build and architect next-generation Hadoop data lake and analytics applications on a group of core Hadoop technologies.
You will evaluate new technologies and products, and research to identify opportunities that impact business strategy, business requirements and performance that can accelerate access to data and automate key data flows. e.g Advanced Machine Learning Technologies.
You will develop highly scalable and extensible Big Data platform which enables collection, storage, modeling, and analysis of massive data sets from numerous channels.
You will define and maintain data architecture, focusing on creating strategy, researching emerging technology, and applying technology to enable business solutions.
You will assess and provide recommendations on business relevance, appropriate timing and deployment.
You will analyze latest Big Data Analytic technologies and their innovative applications in both business intelligence analysis and new service offerings, adopt and implement these insights and best practices.
You will perform architecture design, data modeling, and implementation of CVS Big Data platform and analytic applications.
You will enable big data and batch/real-time analytical solutions that leverage emerging technologies.
You will help facilitate getting data from a variety of different sources, getting it in the right formats, assuring that it adhere to data quality standards, and assuring that downstream users can get that data quickly.
You will ensure proper configuration management and change controls are implemented during code migration.
Experience in evolving/managing technologies/tools in a rapidly changing environment to support business needs and capabilities.
Experience in conducting performance tuning of Hadoop clusters. Monitor and manage Hadoop cluster job, performance capacity planning, and security.
Hands-on experience with “big data” platforms and tools including data ingestion (batch & real time), transformation and delivery in Hadoop ecosystem (such as Hadoop, Pig, Hive, Flume, Ozie, Avro, YARN, Kafka, Storm and Apache Ni-Fi)
Proficiency in R, PySpark, SparkR, Scala, Hive
Experience in architecture and implementation of large and highly complex projects using Hortonworks (Hadoop Distributed File System) with Isilon commodity hardware.
Deep understanding of cloud computing infrastructure and platforms.
History of working successfully with multinational, cross-functional engineering teams.
Capability to architect highly scalable distributed systems, using different open source tools.
             

Similar Jobs you may be interested in ..