Understands, articulates and applies principles of the defined strategy to routine business problems that involve a single function.
Extracts data from identified databases.
Creates data pipelines and transforms data to a structure that is relevant to the problem by selecting appropriate techniques.
Develops knowledge of current analytics trends.
Supports the understanding of the priority order of requirements and service level agreements.
Helps identify the most suitable source for data that is fit for purpose.
Performs initial data quality checks on extracted data.
Analyses complex data elements, systems, data flows, dependencies, and relationships to contribute to conceptual, physical, and logical data models.
Defines relational tables, primary and foreign keys, and stored procedures to create a data model structure.
Evaluates existing data models and physical databases for variances and discrepancies.
Develops efficient data flows.
Writes code to develop the required solution and application features by determining the appropriate programming language and leveraging business, technical, and data requirements.
Creates test cases to review and validate the proposed solution design.
Translates business problems within one's discipline to data-related or mathematical solutions.
Requirements
6+ years of experience in Data Engineering.
Well versed with Hadoop, Hive, Spark using Scala, Kubernetes, Cloud, API and Data Lake concepts.
Proven track record coding with at least one programming language (e.g., Java, Python).
Experienced in computing platforms (e.g., GCP, Azure).
Skilled in data modelling & data migration protocols.
Experience with Kafka connect, Druid, Big Query and Looker is an added advantage.
Experience with the integration tools like Automic, Airflow.
Bachelor's degree in Computer Science and 3 years' experience in software engineering or related field, or 5 years' experience in software engineering or related field, or Master's degree in Computer Science and 1 year's experience in software engineering or related field.
2 years' experience in data engineering, database engineering, business intelligence, or business analytics.
Tech Stack
Airflow
Azure
BigQuery
Cloud
Google Cloud Platform
Hadoop
Java
Kafka
Kubernetes
Python
Scala
Spark
Benefits
Beyond our great compensation package, you can receive incentive awards for your performance.
Other great perks include a host of best-in-class benefits maternity and parental leave, pto, health benefits, and much more.