Contribute to the design, implementation, and maintenance of batch and streaming data pipelines
Work with the product manager and other data engineers to enhance our AWS-based data platform with functional programming best practices and Apache Spark
Support data scientists and marketing stakeholders in operationalizing and scaling machine learning models
Build analytical workflows using SQL, Scala, and Python
Contribute to data quality validation and monitoring solutions
Participate in implementing POCs that bridge data and business users
Requirements
University degree in Computer Engineering, Computer Science, or a similar field
Solid programming skills in Python and a JVM-based language like Scala
Fluent in SQL, and familiarity with Apache Spark is a plus
Familiarity with concepts of modern data lake architecture and fundamentals of Data Engineering
Experience with AWS or other cloud providers (preferred)
Fluent in English
Tech Stack
Apache
AWS
Cloud
Python
Scala
Spark
SQL
Benefits
Hybrid working model with flexibility to work from home on Wednesdays and Fridays
The possibility to work up to two months per year within Germany or from an EU country
Up to 30 days paid vacation each year to unwind and recharge
City mobility with a subsidized public transportation ticket and eco-friendly bike leasing program
Company benefits including a pension scheme, 30% off the home24 assortment, and employee discounts at partner organizations