Crane Worldwide Logistics is a leader in logistics, and they are seeking a Data Engineer II to build and maintain data pipelines and platforms. The role involves developing data products, collaborating with stakeholders, and ensuring data quality and governance across the organization.
Responsibilities:
- Build and maintain pipelines using dbt, Prefect, and Terraform
- Develop and manage connectors across sources and targets including Kafka, RDBMs, and Snowflake
- Implement schema evolution, validation rules, and automated testing
- Support high-availability and disaster recovery design for Snowflake and Materialize
- Author and review schemas and data contracts for consistency and governance
- Develop and optimize dbt models for Snowflake and Materialize analytics layers
- Configure clusters and role-based access for shared environments
- Document datasets to ensure discoverability and proper usage across teams
- Partner with BI developers, analysts, and business teams to deliver datasets that support reporting, dashboards, and integrations
- Investigate and resolve data issues, ensuring durable fixes
- Participate in design reviews to align technical solutions with business requirements
- Contribute to PR and design reviews for pipelines and models
- Support platform governance, observability, and best practices for data quality
- Work with adjacent teams (Ops & Reliability, Analytics, Product) to align on SLAs and data definitions
- Other duties as assigned
Requirements:
- Build and maintain pipelines using dbt, Prefect, and Terraform
- Develop and manage connectors across sources and targets including Kafka, RDBMs, and Snowflake
- Implement schema evolution, validation rules, and automated testing
- Support high-availability and disaster recovery design for Snowflake and Materialize
- Author and review schemas and data contracts for consistency and governance
- Develop and optimize dbt models for Snowflake and Materialize analytics layers
- Configure clusters and role-based access for shared environments
- Document datasets to ensure discoverability and proper usage across teams
- Partner with BI developers, analysts, and business teams to deliver datasets that support reporting, dashboards, and integrations
- Investigate and resolve data issues, ensuring durable fixes
- Participate in design reviews to align technical solutions with business requirements
- Contribute to PR and design reviews for pipelines and models
- Support platform governance, observability, and best practices for data quality
- Work with adjacent teams (Ops & Reliability, Analytics, Product) to align on SLAs and data definitions
- Proficiency in Python and SQL for building and optimizing data pipelines
- Hands-on experience with dbt for modeling and testing, and Terraform for infrastructure-as-code
- Familiarity with modern data platforms: Snowflake, Materialize, Kafka, HVR, Fivetran, or Stitch
- Understanding of data contracts, observability, and governance practices
- Experience with CI/CD tools (GitHub Actions, GitLab CI, or similar)
- Ability to translate business needs into scalable technical solutions
- Knowledge of compliance frameworks (e.g., GDPR, CCPA, SOC 2) a plus
- Prior experience in a data engineering or data-heavy backend software engineering role
- Quantitative bachelors' degree preferred but not required