AirflowAmazon RedshiftAWSAzureETLPostgresPySparkPythonSparkSQLTerraformMachine LearningMLELTSnowflakeRedshiftDatabricksGitHub ActionsCloudFormationGluePostgreSQLGitHubCI/CDRemote Work
About this role
Role Overview
Design and implement scalable data pipelines (ETL/ELT);
Lead the migration from Azure to AWS;
Work with Lakehouse architectures (Delta Lake / Iceberg);
Integrate tools such as ADF, Synapse, Glue, and Databricks;
Ensure data governance, quality, and orchestration;
Support solutions that underpin Machine Learning models.
Requirements
Strong experience with Python, Spark (PySpark), and Databricks
Experience with Airflow and/or orchestration (ADF, Step Functions)
Experience with AWS (Glue, EMR, Lake Formation) and Azure
Knowledge of Data Lakes / Lakehouse (Delta, Parquet, Iceberg)
Experience with systems such as Snowflake, Redshift, PostgreSQL, or Synapse
Experience with CI/CD (GitHub Actions, Terraform, CloudFormation)
Solid foundation in SQL and data modeling
Desirable: Experience with environment migrations (Azure → AWS); working with Feature Store and ML pipelines; knowledge of data governance (catalog, lineage, compliance).
Tech Stack
Airflow
Amazon Redshift
AWS
Azure
ETL
Postgres
PySpark
Python
Spark
SQL
Terraform
Benefits
Remote work
Opportunities for professional growth and development