Turing is the world’s leading research accelerator for frontier AI labs and a trusted partner for global enterprises deploying advanced AI systems. The Software Engineering evaluator will create datasets for training and evaluating large language models, collaborating closely with researchers to enhance AI-driven coding solutions.
Responsibilities:
- Create cutting-edge datasets for training, benchmarking, and advancing large language models
- Curate code examples, provide precise solutions, and make corrections in Python, C/C++, Rust, Go, Java, and JavaScript (including ReactJS)
- Evaluate and refine AI-generated code for efficiency, scalability, and reliability
- Work with cross-functional teams to enhance enterprise-level AI-driven coding solutions
- Build agents that can verify the quality of systems-level and infrastructure code and identify error patterns
- Hypothesize on steps in the software engineering cycle and evaluate model capabilities on them
- Design verification mechanisms that can automatically verify a solution to a software engineering task
Requirements:
- Several years of software engineering experience (3 years or more)
- Strong expertise in systems programming, infrastructure, or backend development using languages like Python, C/C++, Rust, and Go
- Experience building and deploying scalable, production-grade software using modern languages and tools
- Deep understanding of software architecture, design, development, debugging, and code quality/review assessment
- Excellent oral and written communication skills for clear, structured evaluation rationales
- Candidates must be based in the United States
- Ideal for engineers who have shipped high-impact products at fast-moving companies like Stripe, Airbnb, Cloudflare, Datadog, Coinbase, or similar high-growth engineering environments
- Graduates from top computer science programs such as Stanford, MIT, Carnegie Mellon, UC Berkeley, Georgia Tech, and comparable institutions