Job Description :
Job Title : Sr. Data Architect
Duration: 3 - 6 Months
Location: Remote (EST Timezone)
Must-have''s: Advanced knowledge Designing and implementing high-volume data ingestion and streaming pipelines using Open Source frameworks like Apache Spark, Flink, Nifi, and Kafka
Job Responsibilities:
* Contributing data sources, technologies, interfaces, performance issues and system configurations.
* Design and analyze data models, data flows, and integration patterns for structural deficiencies/soundness to build a robust and complete future state model.
* Create/Review logical entities and define their attributes; and define relationships between the various data objects.
* Perform data normalization and denormalization where needed.
* Design and Develop Big Data solutions on different cloud platforms.
* 12+ years of experience in Advanced Analytics and Big Data with the minimum implementation of at-least 2-3 projects at the enterprise level.
* Hands-on experience in architecting solutions at scale in Big Data.
* Proficiency in one or more modern programming languages like Python or Scala or Spark
* Solid functional understanding of the Open-Source Big Data Technologies and NoSQL databases such as MongoDB or Cassandra.
* Experience working with Big Data eco-system including tools such as HDFS, Nifi, Spark streaming, Map Reduce, Sqoop, HBase, Hive, Impala, Oozie.
* Performance tuning, table partitioning and indexing, process threading, bucketing, UDF and UDAF.
* Hands-on experience with commercializing Big Data applications (e.g. configuration management, monitoring, debugging, and security
* Support multiple Agile Scrum teams with planning, scoping and creation of technical solutions.
* Must have worked on upstream and downstream applications along with cloud deployment.
* Must have hands on experience with AWS or Azure platforms, GCP knowledge will be an added advantage.
* Detail oriented with strong analytical and problem-solving skills.
* Effective communicator (both verbal & written
* HDFS/Hadoop: 5 years+ (Required)
* Kafka: 3 years+ (Required)
* Big Data Analytics: 5 years+ (Required)
* Database (Oracle, MySQL, PostgreSQL, MS SQL Server): 10 years+ (Required)
* NoSQL Databases: 5 years+ (Required)
* ETL (Extract Transform Load): 6 years (Required)
* Python/Spark/Scala: 6 years (Required)
* Java: 5 years+ (Required)
* AWS/Azure/GCP experience a plus