Design, construct, and deploy highly efficient and reliable data pipelines that seamlessly transfer data across various platforms, including Data Warehouses and real-time systems.
Develop deep expertise in these data pipelines and manage their Service Level Agreements (SLAs) to ensure optimal performance.
Optimize data ingestion, transformation, and storage for performance, reliability and scalability.
Collaborate with Data Analysts, Data Scientists and business stakeholders to create internal data products aimed at boosting operational efficiencies across the organization.
Implement data quality, observability and governance best practices.
Requirements
Up to 5 years of experience including internship within a technology company.
Bachelors or Masters degree in Computer Science, Information Technology or a related field.
Competent in SQL and Python. Can implement advanced data manipulation and build highly reliable data ingestion pipelines from scratch.
Proven ability & demonstrated flexibility to collaborate effectively with distributed teams and stakeholders across global time zones.
Exposure to cloud computing platforms (AWS, GCP, or Azure) and infrastructure as code.
Communicate clearly, especially to technical audience effectively, both verbally and in writing
The ability to adapt to new situations and learn quickly is valuable in a fast-paced work environment.
Able to collaborate and work well with others can contribute to a positive work environment.
Can manage their time effectively and prioritize tasks is more likely to be productive and efficient.