ApacheAWSDockerElasticSearchKafkaLinuxMacOSPythonScalaSparkApache KafkaLambdaGlueElasticsearchGitRemote Work
About this role
Role Overview
Design and build a green field data mesh platform on AWS.
Create a scalable data foundation for a large distribution company.
Help in building stream processing backend using AWS services, Spark and Kafka.
Ensure that the data pipelines are robust and handle low-quality raw data effectively.
Work collaboratively within a small team to achieve project goals.
Requirements
2+ years of hands-on experience with AWS services like Lambda functions, EMR, ElasticSearch, Lake formation and Glue.
Demonstrates solid technical skills in Scala and/or Python.
Prior experience with ZIO is a big plus.
Solid understanding (and preferably experience) of building pub-sub and asynchronous systems using Apache Kafka or any other messaging API.
Can analyze algorithms complexity and know data structures beyond “List and Stack”.
Have been using Linux/macOS, Docker and git in collaborative workflows for 3+ years.
5+ years of experience developing greenfield projects from scratch in an operational and live environment with strict high availability requirements is a plus.