Build and maintain ETL pipelines to process and export record data to Sayari Graph application
Develop and improve entity resolution processes
Implement logic to calculate and export risk information
Work with product team and other development teams to collect and refine requirements
Run and maintain regular data releases
Requirements
Bachelor's degree, or foreign equivalent, in Data Engineering, Computer Science or closely related field
3 years of experience as Software Data Engineer or any occupation in which the required software data engineering experience is gained
Must have the following skills/knowledge: Python
a JVM programming language such as Scala
a SQL database (E.G., Postgres)
a NoSQL Database (e.g., Hbase, Cassandra, Elasticsearch, or memgraph, etc.)
designing, maintaining, and orchestrating ETL pipelines (e.g., Apache spark or Apache airflow) in cloud based environments (e.g., GCP, AWS, or Azure)
Tech Stack
Airflow
Apache
AWS
Azure
Cassandra
Cloud
ElasticSearch
ETL
Google Cloud Platform
HBase
NoSQL
Postgres
Python
Scala
Spark
SQL
Benefits
100% fully paid medical, vision, and dental for employees and their dependents
Generous time off; we observe all US federal holidays, close our office for a winter break (12/24-12/31), in addition to granting 18 PTO days and 10 sick days
Outstanding compensation package; competitive commissions for revenue roles and bonuses for non-revenue positions
A strong commitment to diversity, equity, and inclusion
Eligibility to participate in additional benefits such as 401k match up to 5%, 100% paid life insurance (up to $100,000 coverage), and parental leave