Job Description :
Work Authorization: H1B/H4EAD/TN/GC/USC. We can work on c2c, w2.
Need 8+ years of experience. Pyspark is mandatory.


Required/Preferred Skills:
Need 8+ years of experience.
Experience with Data/ETL pipelines
Python experience is critical
Some Unix/Shell scripting experience preferred (we exclusively use the AWS CLI to submit jobs to our
EMR clusters for InfoSec reasons)
Experience with EMR/PySpark preferred
Airflow experience would be a plus but not mandatory
Any other AWS Experience such as with EC2, Athena, Redshift Spectrum would be a plus but not mandatory
Basic Understanding of HDFS/distributed computing
Basic SQL skills
Understanding of columnar file formats such as parquet
Experience with automating API calls would be a plus but not mandatory