AzureCloudETLGoogle Cloud PlatformLinuxNoSQLPySparkPythonSQLShellArtificial IntelligenceELTData EngineeringDatabricksGCPGoogle CloudGitVersion Control
About this role
Role Overview
Work on automating and optimizing data workflows and processes;
Build technical solutions to ingest, process, and store data from multiple sources and in various formats (structured / unstructured / files / XML or JSON / Parquet / APIs);
Perform activities related to analyzing, building, and maintaining data processes and structures;
Design and contribute to data architecture, mapping, and modeling;
Maintain an enterprise-level perspective aligned with market best practices, solution blueprints, and solution designs;
Experience with Artificial Intelligence solutions and GCP is a plus.
Requirements
Experience in data projects;
Experience with Databricks platform;
Experience with Azure Data Factory;
Programming experience — Python, SQL, PySpark;
Experience in automating information processes (ETL/ELT);
Experience with version control using Git;
Experience with Relational Databases and NoSQL;
Experience with Azure Cloud, Data Factory, Synapse, ADLS Gen2, Delta Lake;
Experience in data engineering and integration (e.g., ETL, APIs, microservices);
Experience in Linux environments, basic commands, and shell scripting;
Knowledge of streaming processes with Event Hub;
Knowledge of data representation formats and scripts such as JSON, XML, YAML.