METR is developing evaluations for AI R&D capabilities, such that evaluators can determine if further AI development risks a "capabilities explosion," which could be extraordinarily destabilizing if realized.
METR is hiring ML research engineers/scientists to drive these AI R&D evaluations forward.
Responsibilities:Produce tasks/benchmarks that can determine if a model is dangerous.Run experiments to determine how elicitation techniques affect results on our evaluation suite.Run evaluations internally, and potentially support external partners in governments and/or frontier AI companies in performing evaluations for autonomous capabilities.Improve the tooling that researchers use for designing and running evaluations.Collaborate with evaluation science researchers to develop a robust evaluation procedure.Your work will increase the odds that METR's evaluation protocols can robustly predict whether a new frontier model poses catastrophic risks.
You'll help with:
Understanding what kinds of abilities we need to be evaluating for, and what properties we most need our evaluations to have.Finding the most promising directions to explore; designing experiments and a research roadmap.Collaborating with the threat modeling team, sharing your ML domain expertise to help us evaluate models for AI R&D skills.Research execution:Rapidly execute experiments, obtain reliable results.Design sensible pipelines and workflows (know which things are going to be reused and need to be good versus what things it's ok to do scrappily).Quickly interpret results - recognize what is signal vs noise, notice when things don't look right and there might be a bug, know where to look for bugs in ML experiments.Know how much work different approaches are likely to be and how promising they are; when you have uncertainties, get information as quickly as possible.What we're looking for:Significant ML research engineering experience, for example:A research publication related to machine learning in which you played a major role.Professional experience optimizing compute for inference or training of a large model.Participation in the training or optimization of a large model.An ideal candidate would be a machine learning researcher with substantial experience working with frontier LLMs and a track record of successful execution-heavy research projects.
We expect to hire multiple people for this position, and their work will focus more on either the engineering or research side of the role, depending on their strengths.
About METR:METR is a non-profit which does empirical research to determine whether frontier AI models pose a significant threat to humanity. It's robustly good for civilization to have a clear understanding of what types of danger AI systems pose, and know how high the risk is.
Some highlights of our work so far:
Establishing autonomous replication evals: Thanks to our work, it's now taken for granted that autonomous replication (the ability for a model to independently copy itself to different servers, obtain more GPUs, etc.) should be tested for.Pre-release evaluations: We've worked with OpenAI and Anthropic to evaluate their models pre-release, and our research has been widely cited by policymakers, AI labs, and within government.Inspiring lab evaluation efforts: Multiple leading AI companies are building their own internal evaluation teams, inspired by our work.Early commitments from labs: Anthropic credited us for their recent Responsible Scaling Policy (RSP), and OpenAI recently committed to releasing a Risk-Informed Development Policy (RDP). These fit under the category of "evals-based governance," wherein AI labs can commit to things like, "If we hit capability threshold X, we won't train a larger model until we've hit safety threshold Y."We've been mentioned by the UK government, Obama, and others. We're sufficiently connected to relevant parties (labs, governments, and academia) that any good work we do or insights we uncover can quickly be leveraged.
Logistics:Deadline to apply: None. Applications will be reviewed on a rolling basis.
We encourage you to apply even if you do not believe you meet every single qualification. Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you're interested in this work.
Apply for this job:We encourage you to apply even if your background may not seem like the perfect fit! We would rather review a larger pool of applications than risk missing out on a promising candidate for the position. If you lack US work authorization and would like to work in-person (preferred), we can likely sponsor a cap-exempt H-1B visa for this role.
We are committed to diversity and equal opportunity in all aspects of our hiring process. We do not discriminate on the basis of race, religion, national origin, gender, sexual orientation, age, marital status, veteran status, or disability status. We welcome and encourage all qualified candidates to apply for our open positions. Registering interest is quick; our main question can be answered with a few bullet points. Register interest!
#J-18808-Ljbffr