AirflowApacheAWSCloudETLPySparkPythonSparkSQLUnityELTData EngineeringAnalyticsDatabricksApache SparkApache AirflowIAMCodePipelineGitVersion ControlCI/CDMentoringCommunicationTime ManagementCollaborationRemote Work
About this role
Role Overview
Responsible for operating and optimizing our cloud-based data processing environment
Refines data transformations using PySpark and Spark SQL within notebooks such as Databricks
Leverages orchestration tools like Apache Airflow to automate, schedule, and oversee data workflows
Participates in code reviews, testing, and documentation as part of the development lifecycle
Supports and troubleshoots Databricks jobs, Spark workloads, and AWS-based data processes
Optimizes Databricks clusters and jobs for performance and cost
Maintains and improves existing data pipelines built with AWS (CodePipeline), Delta Lake, and Databricks Notebooks
Works closely with data engineering and analytics teams to improve data quality and pipeline reliability
Maintains and enhances CI/CD workflows for Databricks deployments using AWS tools
Manages access controls with IAM and Unity Catalog
Performs regular monitoring, troubleshooting, and root-cause analysis of data and compute workloads
Requirements
Bachelor’s degree in Computer Science, Information Systems, Data Engineering, or similar
Master’s Degree will be considered as an asset
5+ years of experience in big data operations or cloud-based data engineering.
Strong hands-on experience with AWS, Databricks, Delta Lake, and Apache Spark
Proficient in Python, SQL, and PySpark
Experience with CI/CD, version control, and release processes (AWS CodePipeline, Git)
Experience with monitoring, debugging, and optimizing ETL/ELT and Spark workloads
Knowledge of data governance frameworks and exposure to enterprise security or regulated environments will be considered as an asset
Excellent problem-solving skills and attention to detail
Strong communication skills and the ability to work collaboratively in a team environment
Effective time management with ability to multi-task and prioritize work
Tech Stack
Airflow
Apache
AWS
Cloud
ETL
PySpark
Python
Spark
SQL
Unity
Benefits
Competitive Compensation : We offer an attractive salary, annual performance-based bonuses, and a monthly meal allowance through our ticket restaurant card
Health: Private medical insurance is provided for you and your family
Family Support: Monthly financial allowance for early education (nursery) and coverage of expenses for children with neurodiversity or disabilities—including therapeutic swimming, music therapy, horse riding, and parental support
Flexible Work Model: Our hybrid approach offers a level of remote work flexibility that supports work-life balance while preserving strong collaboration and team spirit
Modern Workspaces: Contemporary offices designed to support comfort, health, and productivity, with fully equipped workstations, quiet areas, on-site restaurant, and group fitness sessions
Lifelong Learning: Cepal supports continuous learning through access to e-learning platforms and structured professional development programs
Career Progression: We are committed to your growth, offering a clear development path supported by feedback, mentoring, and personalized learning plans
Make a Difference: Get involved in regular wellbeing, ESG and volunteering initiatives that reflect our values and foster a sense of purpose and community
Senior Cloud Data Engineer at Cepal Hellas Financial Services S.A. | JobVerse