Confluent is a company focused on revolutionizing how data moves and its potential applications. As a Data Engineer, you will tackle big data challenges by building data pipelines that enhance accessibility for data analysts and stakeholders across the organization.
Responsibilities:
- Design, construct, and deploy highly efficient and reliable data pipelines that seamlessly transfer data across various platforms, including Data Warehouses and real-time systems
- Develop deep expertise in these data pipelines and manage their Service Level Agreements (SLAs) to ensure optimal performance
- Optimize data ingestion, transformation, and storage for performance, reliability and scalability
- Collaborate with Data Analysts, Data Scientists and business stakeholders to create internal data products aimed at boosting operational efficiencies across the organization
- Implement data quality, observability and governance best practices
Requirements:
- Up to 5 years of experience including internship within a technology company
- Bachelors or Masters degree in Computer Science, Information Technology or a related field
- Located in PST timezone
- Competent in SQL and Python. Can implement advanced data manipulation and build highly reliable data ingestion pipelines from scratch
- Proven ability & demonstrated flexibility to collaborate effectively with distributed teams and stakeholders across global time zones
- Exposure to cloud computing platforms (AWS, GCP, or Azure) and infrastructure as code
- Communicate clearly, especially to technical audience effectively, both verbally and in writing
- The ability to adapt to new situations and learn quickly is valuable in a fast-paced work environment
- Able to collaborate and work well with others can contribute to a positive work environment
- Can manage their time effectively and prioritize tasks is more likely to be productive and efficient