Lyft is a company focused on building a large transportation media network through Lyft Ads. They are seeking an Algorithms Data Scientist to develop machine learning models that enhance ad relevance and performance, collaborating with various teams to translate business needs into algorithmic solutions.
Responsibilities:
- Design, develop, and deploy production-grade machine learning models and algorithms that power core Lyft Ads capabilities, such as ad relevance, targeting, ranking, bid optimization, pacing, campaign delivery, and measurement
- Own the end-to-end lifecycle of modeling projects — including problem definition, data exploration, feature engineering, model development, offline evaluation, deployment, and monitoring
- Collaborate closely with Ads Engineering to integrate models into real-time ad-serving and batch decision systems, ensuring performance across latency, scalability, and reliability constraints
- Analyze large-scale mobility, behavioral, and ads performance datasets to identify patterns, surface opportunities, and guide ML and AI driven product improvements
- Implement rigorous model evaluation frameworks, including offline metrics, statistical tests, calibration, sensitivity analysis, and A/B experimentation to validate both model impact and system-level outcomes
- Build robust training pipelines, feature transformations, and scoring infrastructure, ensuring reproducibility, observability, and long-term maintainability
- Partner with Product, Engineering, and Sales to translate ambiguous advertiser goals (e.g., increased conversions, reach efficiency, brand lift) into measurable requirements and success metrics
- Investigate and resolve model behavior issues, production regressions, calibration drift, and performance anomalies in close partnership with Ads Infra teams
- Drive innovation by staying current with advances in ML for ranking, recommendation, causal inference, optimization, and ads measurement — and proactively identifying opportunities to apply them
- Contribute to Lyft Ads’ modeling and experimentation infrastructure, through model cards, documentation, reproducibility standards, and code quality improvements