Platform Development and Evolution: Design, develop, and evolve robust, scalable data platforms;
Architecture Ideation and Implementation: Lead the ideation and implementation of new data architectures;
Automation and Scalability: Build automated, scalable data pipelines;
Documentation and Knowledge Sharing: Document processes and systems in detail;
Reporting and Communication: Produce detailed technical reports and present them to stakeholders;
Cross-disciplinary Collaboration: Work with data science and data engineering teams to gather and understand requirements;
Evaluate and incorporate new technologies in collaboration with the Architecture team;
Mentor junior and mid-level data engineers;
Propose continuous improvements to processes and tools;
Help define and implement data architectures (Data Lake, Lakehouse, Data Mesh, etc.);
Engage with business and technology stakeholders to create a compelling vision for a data-driven organization and produce artifacts that promote code reuse;
Work closely with clients to design, implement, and manage solutions that meet their technical requirements and business objectives;
Provide technical guidance and best practices throughout project lifecycles and day-to-day operations;
Deep knowledge of Cloud products and services (AWS, Azure, S3);
Proficient in designing complex, scalable, and secure solutions tailored to the needs of each area and project;
Collect requirements, assess current infrastructure, and propose Cloud strategies, improvements, and governance;
User Management: Create, modify, and remove users, and manage their access credentials and roles;
Access Control: Grant and revoke privileges on Snowflake objects such as tables, schemas, and warehouses;
Security Configuration: Implement security policies, such as MFA (Multi-Factor Authentication) and access auditing;
Monitoring and Resource Management: Track account usage, including storage and compute consumption, to ensure efficiency and cost-effectiveness;
Data Management.
Requirements
AWS services such as Athena, Glue, Lambda, S3, DynamoDB; Azure; Databricks; Redshift;
Advanced software development skills in Python;
Advanced knowledge of PySpark;
Experience as a Data Architect;
Infrastructure-as-Code tools such as Terraform, Pulumi and CloudFormation;
Experience with CDC (Change Data Capture) tools;
Data storage and architecture: implementation of data lakes and Medallion/medallion-style architectures;
Machine Learning models in production: experience with deployment and monitoring of models;
ETL and orchestration tools: experience with Apache Airflow, dbt, and other tools.
Tech Stack
Airflow
Amazon Redshift
Apache
AWS
Azure
Cloud
DynamoDB
ETL
PySpark
Python
Terraform
Benefits
Wellhub to help you stay healthy
Emotional support: 2 free online therapy sessions per month via Zenklub
Special discounts: up to 35% on Alpa products
Pet leave: 2 consecutive days off when adopting a pet
Guapeco: pet health plan
Life insurance with broad coverage
Internal talks and courses led by our specialists
English and Spanish courses to boost your language skills
Flexible working hours to balance personal and professional life
No dress code – be yourself!
OnHappy for amazing trips
Day off on your birthday
Internal currency with a shop for company products