Job Description :
Sr. Data Engineer
San Francisco, CA
Remote to start, eventually onsite

Long Term
Skype

JD

Palintair Foundry (PF) would be ideal candidate which will give us a better success rate, but key is going to be to identify a solid Python Pyspark SQL resource.

If they have worked on AWS Glue Redshift that would be good too.

PF Python Pyspark SQL



Position Summary

We are seeking a Senior Data Engineer to join the Data & Analytics department. This team follows best-practice agile techniques to deliver impactful data products for internal partners. You will actively participate and practice in an open, highly collaborative Agile environment. This Data Engineer role with Data & Analyticswill have a major impact in reducing wildfire risk in California, improving thequality of life for all Californians.

PositionResponsibilities

? Collaborate on a daily basis with the product team. This includespairing for all aspects of software delivery.

? Create and maintain optimal data pipeline architecture.

? Assemble large, complex data sets that meet functional andnon-functional business requirements.

? Identify, design, and implement internal process improvements:automating manual processes, optimizing data delivery, and re-designinginfrastructure for greater scalability.

? Build the infrastructure required for optimal extraction,transformation, and loading of data from a wide variety of data sources usingSQL and AWS ‘big data’ technologies.

Minimum Qualifications

? Bachelors Degree in Computer Science or job-related discipline orequivalent experience

? 7 years-experiencewith software delivery

? Experience delivering product with Agile / Scrum methodologies

? Advanced working SQL knowledge and experienceworking with relational databases, query authoring (SQL) as well as workingfamiliarity with a variety of databases.

? Experience with ETL flows using Python

? Experience building and optimizing ‘big data’ data pipelines,architectures and data sets.

? Strong analytic skills related to working with unstructureddatasets.

? Build processes supporting data transformation, data structures,metadata, dependency and workload management.

? Proficiency with the following tools that enable the candidate tocontribute autonomously: Glue, Kafka, Redshift (with a focuson infrastructure -as-code), Python.

Desired Qualifications

? Experience performingroot cause analysis on internal and external data and processes to answerspecific business questions and identify opportunities for improvement.

? Experience with Palantir Foundry platform – creatingdatasource, creating transformations using pyspark, experience with Foundry applicationsinclude Slate, Contour and Workbooks

? AWS DevOps skills

? Humble – is open tobeing coached, has high Emotional Quotient (EQ) and is self-aware

? Hungry – desires to get things done while honoring people, andseeks better ways to do the job, is highly motivated by the significant impactthis work will have

? Collaborative – has strong interpersonal skills; demonstratesempathy with teammates and stakeholders, cares about and works well withteammates

? Willingness to impact beyond defined role

? Experience with data & analytics product development