ApacheBigQueryCloudDockerGoogle Cloud PlatformJavaPythonScalaSQLTerraformAIMLdbtGCPGoogle CloudPub/SubCloud StoragecomposerGitCI/CDCommunicationRemote Work
About this role
Role Overview
You design and operate scalable and reliable data pipelines.
Your modeling and provisioning of datasets is the crucial step in turning raw data into trustworthy, actionable insights.
In close partnership with our Data Scientists, you help responsibly bring cutting‑edge AI/ML models into production.
Your technical expertise will ensure our innovations are deployed quickly, reliably, and to the maximum benefit of ARD.
Requirements
University degree (Computer Science, Engineering, Natural Sciences or a comparable qualification)
Proven expertise in building and operating scalable cloud data solutions, ideally with Google Cloud Platform (GCP) (in particular BigQuery, Dataflow/Apache Beam, Pub/Sub, Cloud Composer, Cloud Storage)
Excellent programming skills in Python and SQL, and experience with a JVM language (e.g., Java/Scala)
Confident use of Git, DevOps principles (e.g., CI/CD, IaC with Terraform, containerization with Docker) and DataOps principles (dbt, monitoring, cataloging)
Familiarity with the Data Mesh concept and its implementation
Strong analytical skills for solving complex data problems, as well as excellent communication and teamwork skills in an international environment (English and German at least C1)
Tech Stack
Apache
BigQuery
Cloud
Docker
Google Cloud Platform
Java
Python
Scala
SQL
Terraform
Benefits
31 days of vacation
Competitive salary
Diversity and equal opportunities
Access to training and professional development tailored to your personal interests and goals
Flexible working hours
Remote work within Germany and certain European countries