Employer Active
Job Alert
You will be updated with latest job alerts via emailJob Alert
You will be updated with latest job alerts via email$ 100000 - 175000
1 Vacancy
Since starting in July 2022 FAR.AI has grown to 19 FTE produced 28 academic papers and established the leading AI safety events for research and international cooperation. Our work is recognized globally with publications at leading venues such as NeurIPS ICML and ICLR that have been featured in the Financial Times Nature News and MIT Tech Review. We leverage our research insights to drive practical change through redteaming with frontier model developers. Additionally we help steer and grow the AI safety field through developing research roadmaps with renowned researchers such as Yoshua Bengio; running an AI safety focused coworking space FAR.Labs with 40 members; and through targeted grants to technical researchers.
Our research team likes to move fast. We explore promising research directions in AI safety and scale up only those showing a high potential for impact. Unlike other AI safety labs that take a bet on a single research direction FAR.AI aims to pursue a diverse portfolio of projects.
Our current focus areas include:
building a science of robustness (e.g. finding vulnerabilities in superhuman Go AIs
finding more effective approaches to value alignment (e.g. training from language feedback
Advancing model evaluation techniques (e.g. inverse scaling and codebook features and learned planning.
We also put our research into practice through redteaming engagements with frontier AI developers and collaborations with government institutes.
To build a flourishing field of AI safety research we host targeted workshops and events and operate a coworking space in Berkeley called FAR.Labs. Our previous events include the International Dialogue for AI Safety that brought together prominent scientists (including 2 Turing Award winners) from around the globe culminating in a public statement calling for global action on AI safety research and governance. We also host the semiannual Alignment Workshop with 150 researchers from academia industry and government to learn about the latest developments in AI safety and find collaborators. For more information on FAR.AIs activities please visit our recent post.
We are seeking applications from potential Research Scientists who can:
Take ownership of and accelerate existing AI Alignment research agendas.
Develop their own exciting AI Alignment research agendas.
Lead novel research projects where there may be unclear markers of progress and/or success.
Contribute to the development of best practices for AI safety research at FAR.AI and in the broader community.
Publish research findings and engage with the AI safety community.
We are excited by unconventional backgrounds.
You may have the following:
New and underexplored AI Alignment idea(s).
Experience leading and/or playing a senior role in research projects related to machine learning.
Ability to effectively communicate novel methods and solutions to both technical and nontechnical audiences.
PhD or several years research experience in computer science artificial intelligence machine learning or statistics.
As Research Scientist you would lead AI safety research projects or make essential contributions to existing projects. Examples of ongoing projects at FAR include:
Scaling laws for prompt injections. Will advances in capabilities from increasing model and data scale help resolve prompt injections or jailbreaks in language models or is progress in averagecase performance orthogonal to worstcase robustness
Robustness of advanced AI systems. Explore adversarial training architectural improvements and other changes to deep learning systems to improve their robustness. We are exploring this both in zerosum board games and language models.
Mechanistic interpretability for mesaoptimization. Develop techniques to identify internal planning in models to effectively audit the goals of models in addition to their external behavior.
Redteaming of frontier models. Apply our research insights to test for vulnerabilities and limitations of frontier AI models prior to deployment.
You could be an employee or an independent Contractor for FAR AI a 501(c)3 research nonprofit.
Location: Both remote and inperson (Berkeley CA) are possible. We sponsor visas for inperson employees and can also hire remotely in most countries.
Hours: Fulltime 40 hours/week).
Compensation: $100000$175000/year depending on experience and location. We will also pay for workrelated travel and equipment expenses. We offer catered lunch and dinner at our offices in Berkeley.
Application process: A 72minute programming assessment a short screening call two 1hour interviews and a 12 week paid work trial. If you are not available for a work trial we may be able to find alternative ways of testing your fit.
If you have any questions about the role please do get in touch at .
Full-Time