GenLogs is a transportation-technology company building the next generation of truck intelligence. They are seeking a Data Engineer to build and maintain ETL/ELT pipelines, develop ingestion workflows, and ensure data quality while supporting machine learning workflows.
Responsibilities:
- Build and maintain ETL/ELT pipelines using Python and SQL
- Develop ingestion workflows with AWS Firehose, S3, and related services
- Create and optimize dbt models, tests, and incremental logic
- Tune Snowflake queries and warehouse usage for cost and performance
- Operate and improve Airflow DAGs for reliable execution and monitoring
- Maintain high data quality, data integrity, and pipeline SLA commitments
- Bring clarity to ambiguous requirements and propose practical solutions
- Build feature pipelines to support ML workflows
- Support model deployment, monitoring, and automated retraining
- Add data validation and quality checks across ML and analytics pipelines
Requirements:
- 3+ years of experience in data engineering or software engineering
- Strong Python and SQL skills
- Hands-on experience with Snowflake, AWS Firehose/S3, Airflow, and dbt
- Ability to work independently and execute in a dynamic environment
- Strong problem-solving skills and attention to detail
- Experience with geospatial data (e.g., spatial joins, geometry processing, or geospatial libraries)
- Experience with ML or MLOps pipelines
- Knowledge of Snowflake streams, tasks, and performance tuning
- Experience with large-scale or semi-structured datasets