Build, optimize, and maintain data pipelines that power our business
Define and build out abstracted reusable data sets to be used for Business Intelligence, Marketing, and Data Science Research
Design, build, and evangelize a federated data validation framework to be used to monitor potential data inconsistencies
Protect our users’ privacy and security through best practices
Requirements
2+ years of software/data engineering experience (distributed data processing, data warehousing, data governance, Big Data, data variance, data privacy, and data quality).
Expertise in SQL and Python
Expertise in building scalable data pipelines, query optimization, data modeling, and defining reusable datasets
Experience working with orchestration tools (especially Airflow), databases (especially PostgreSQL), data warehouses (especially Snowflake)
Familiarity with SQL tuning, medallion/event-driven architectures, and telemetry.
Familiarity with healthcare or insurance
Familiarity with data security and HIPPA compliance
A desire to be a part of a high-performing, mission-driven team that operates with intense urgency, a strong sense of individual accountability, and a commitment to authentic feedback.