AzurePySparkPythonSparkSQLRData EngineeringData LakeAnalyticsBIPower BISnowflakeDatabricksAzure DevOpsAzure SQLBlob StorageGitPerformance OptimizationCI/CDCommunicationRemote Work
About this role
Role Overview
Design, develop, and maintain data pipelines using Azure Data Factory;
Build data processing solutions with Azure Databricks (PySpark and Spark SQL);
Structure and manage Data Lakes on Azure (Blob Storage/Data Lake), applying layered architecture (raw, curated, and trusted);
Model, optimize, and administer data in Azure SQL Database;
Develop analytical data models for consumption in Power BI;
Create and optimize datasets, semantic models, and DAX measures;
Implement code versioning and CI/CD practices for pipelines and analytical artifacts;
Ensure data quality, reliability, and traceability (data quality, lineage, and monitoring);
Collaborate with business, BI, and analytics teams to translate technical requirements;
Support data architecture decisions, Azure best practices, and cost optimization (FinOps);
Serve as a technical reference for junior professionals.
Requirements
Experience with Azure Data Factory for orchestrating pipelines, integrating multiple sources, parameterization, error handling, and monitoring;
Hands-on experience with Azure Databricks developing solutions using PySpark and Spark SQL, including notebook creation and versioning, job execution, and performance optimization;
Experience with Azure Data Lake and/or Blob Storage structuring layered architectures (raw, curated, and trusted), with a focus on security, governance, and access control;
Strong experience with Azure SQL Database, including relational modeling, advanced T-SQL, index creation, and query tuning;
Experience with Power BI for dimensional modeling, creating advanced DAX measures, and developing/maintaining dashboards and datasets;
Solid expertise in analytical data modeling, with emphasis on Star Schema (essential), Snowflake, and BI concepts;
Advanced SQL skills (CTEs, window functions, and execution plan analysis);
Experience using Python for data engineering (ingestion, transformation, and process automation);
Experience with code versioning using Git (branches, pull requests, and best practices);
Experience implementing CI/CD pipelines with Azure DevOps or similar tools.
Tech Stack
Azure
PySpark
Python
Spark
SQL
Benefits
Remote work – Monday to Friday (09:00 to 18:00);
Home-office allowance – Credit on Ifood card for meals/food worth R$ 300.00 per month;
Birthday – To celebrate your special day, Rox gifts you a voucher and a day off to enjoy it;
Courses – Full access via RoxSchool, Alura platform, Pluralsight, O’Reilly for books and talks;
Certifications – Reimbursement of up to R$ 300.00 for certifications (TECHNOLOGY) + a R$ 300.00 bonus per achieved certification from these brands;
Psychological support – Two psychotherapy sessions covered monthly by ROX with partner psychologists;
Feedz partnership – Gamified platform to stay connected, improve communication and track sentiment, engagement, feedback, IDP, and performance;
WellHub (Gympass) – Partnerships with gyms and health & wellness apps;