Yotascale is a company focused on building tools that empower engineers to understand their cloud utilization. They are seeking an experienced Data Engineer to develop scalable data pipelines, improve system performance, and collaborate with cross-functional teams to enhance their codebase and infrastructure.
Responsibilities:
- Develop, improve, and deploy highly scalable distributed data pipelines that ingest, analyze, and store billions of time-series metrics and events to our downstream storage and analytics services
- Ensure the system delivers high throughput and low latency to enable the delivery of highly available, resilient, and durable services
- Build and maintain our data processing streams and frameworks
Requirements:
- 3+ years of professional software engineering experience with consumer-facing products
- Experience building, maintaining, and monitoring data pipelines
- Experience developing against a large-scale OLTP database with billions of records
- Strong knowledge of Python and PySpark (or equivalent in another language)
- Some familiarity with Java, Scala, or Kotlin and associated tools
- Demonstrated ability to test, monitor, and integrate your code with other production systems
- Solid CS fundamentals (data structures, algorithms, design patterns)
- Strong understanding of SQL and experience with at least one DBMS
- Bachelor's degree or higher, in Computer Science or related field (or equivalent experience)
- Excellent communication and problem-solving skills