Design, implement, and continuously optimize platform services on our modern, industry-standard GCP-based data environment
Interact with our software architects/engineers, data scientists, and business analysts in an agile environment to define data needs and provide solutions
Develop a high-quality codebase for our data pipelines
Develop, provision, and monitor our data products
Maintain our data pipeline orchestration, data lake, and data warehouse
Requirements
Minimum 3 years of experience in a comparable data engineering/data science environment
Proficient in SQL programming and dbt
Proven track record of delivering data projects on Google Cloud Platform (GCP)
Experience with GCP native services such as BigQuery, Cloud Run, Dataproc, and Dataflow
Experience developing high-quality software, including unit and integration tests, in one or more languages (e.g., Python or Java)
Experience managing data pipeline workflows with Airflow
Experience with containerization (Docker) and infrastructure-as-code frameworks (Terraform)
Strong plus: experience with distributed computing technologies such as Apache Spark and familiarity with ML concepts/frameworks
Open to new technologies and challenges
Tech Stack
Airflow
Apache
BigQuery
Cloud
Docker
Google Cloud Platform
Java
Python
Spark
SQL
Terraform
Benefits
Room for your own data and engineering initiatives
Opportunity to join — and shape — a growing Data Team