Building distributed and highly parallelized Big Data processing pipeline which process massive amount of data (both structured and unstructured) in near real-time
Leverage Spark to enrich and transform corporate data to enable searching, data visualization, and advanced analytics
Work closely with DevOps, QA, and Product Management teams in a Continuous Delivery environment
Implement multiple data requirements coming from the Group Financial Crime Prevention unit
Develop streaming applications that connect to multiple data sources both internally and externally
Enable advanced analytics pipeline that produces clean, transformed data set that are being utilized to feed multiple products within the Financial Crime sector.
Requirements
Higher education level within IT or related fields are preferred
At least 5 years of experience working with Scala and Spark
Hands-on experience with Hadoop stack among others: yarn, EMR, Sqoop, Hive
Experience with Impala
Basic knowledge of Kafka
Fluency in English both speaking and writing
Tech Stack
Hadoop
Kafka
Scala
Spark
Yarn
Benefits
B2B agreement
Transparent working conditions
Current support during our cooperation
Possibility to work in an international environment
Collaborative environment in Swedish organizational culture