Design, build, and maintain robust, scalable data pipelines for ingestion, processing, and transformation of large volumes of data;
Develop and optimize data models for analytical systems;
Implement and manage storage solutions, including data lakes and data warehouses, in cloud environments;
Ensure data integrity, quality, and governance by applying best practices in management, security, and FinOps;
Collaborate with multidisciplinary teams to understand business requirements and deliver data-driven solutions;
Monitor and optimize the performance of pipelines, databases, and integration solutions;
Lead data architecture modernization initiatives using cutting-edge technologies and modern frameworks.
Requirements
Bachelor's degree in Computer Science, Software Engineering, Information Systems, or related fields, or a postgraduate degree in related areas;
Proven experience with data visualization platforms (Tableau, Power BI and/or Looker);
Experience with GCP and related services: BigQuery, Cloud Composer, Apache Airflow, Dataflow, Pub/Sub, Cloud Run, Cloud Functions, KMS, Secret Manager;
Databases: Knowledge of data modeling, relational databases (e.g., Oracle, SQL Server, PostgreSQL) and NoSQL databases (e.g., MongoDB, Cassandra);
ETL/ELT: Experience building ETL/ELT pipelines for data ingestion and transformation;
Desired knowledge: Cloud computing fundamentals;
Experience with Azure and services such as Azure Data Factory, Azure Synapse Analytics, and Azure Data Lake Storage;
Programming languages: Proficiency in Python and PySpark (for scripting and automation) and SQL (for data manipulation and querying);
Data Lakes: Knowledge of Data Lake architectures, storage and data optimization (e.g., Delta Lake);
DevOps: Experience with CI/CD pipelines and DataOps practices for automating deployments and monitoring data flows;
Security and Governance: Understanding of cloud security practices, data encryption, and implementation of governance policies;
Data Orchestration: Experience using tools such as Azure Data Factory and/or Synapse and/or Airflow/Cloud Composer for data integration and movement;
Performance and Optimization: Experience with techniques to improve the performance of data pipelines and queries in distributed and non-distributed environments;
APIs: Knowledge of extracting data via RESTful APIs.
Tech Stack
Airflow
Apache
Azure
BigQuery
Cassandra
Cloud
ETL
Google Cloud Platform
MongoDB
Oracle
Postgres
PySpark
Python
SQL
Tableau
Benefits
Health and Dental Plan – Bradesco – extends to dependents;
PAE – Financial assistance for dependents (children and/or stepchildren) with intellectual disabilities;
Pharmacy Agreement – Discounts up to 85%;
Complementary Pension – FlexPrev plan – contributions ranging from 1% to 11% based on salary;
Life Insurance – provided to all employees from the date of hire, with no employee contribution;
Extended Leave – Maternity leave (total of 180 days) and paternity leave (total of 20 days);
Meal and/or Food Allowance – Caju Benefícios;
Educational Assistance – For dependents up to high school;
TotalPass;
Solar Subscription – Possibility to obtain a special discount on the electricity bill through participation in a distributed generation plan;
Corporate Smiles Club – Discounts up to 50% on offered plans for planning travel. Also available for spouse and children 18+;
Length-of-Service Bonus – We offer an additional salary payment (based on time with the company) to be paid at the time of vacation;
Learning Platform: Vibra values learning path and curated content with more than 200 course options available on demand;
Internal Recruitment: Job openings advertised across Brazil, encouraging internal mobility.