Liane Lovitt portrait
Researcher 2 reports

Liane Lovitt

Anthropic

Research scientist at Anthropic whose public work includes AI alignment, reinforcement learning from human feedback, and model behavior.

Samuel Marks portrait
Researcher 6 reports

Samuel Marks

Anthropic

Senior research engineer at Anthropic interested in agent foundations, model organisms of misalignment, and human-computer interaction.

Samuel R. Bowman portrait
Researcher 5 reports

Samuel R. Bowman

Anthropic

Member of technical staff at Anthropic and associate professor of computer science, data science, and linguistics at New York University on leave. His public homepage focuses on natural language processing, machine learning, and AI alignment.

Newton Cheng portrait
Researcher 1 reports

Newton Cheng

Anthropic

Anthropic researcher on the Frontier Red Team focused on cyber misuse evaluation and threat modeling; previously a physics PhD student at UC Berkeley and now also mentors in the MATS program.

Jack Clark portrait
Researcher 7 reports

Jack Clark

Anthropic / OpenAI

Co-founder and head of policy at Anthropic. He previously served as policy director at OpenAI, worked as a technology journalist, and writes the Import AI newsletter.

David Duvenaud portrait
Researcher 4 reports

David Duvenaud

Anthropic

Associate Professor at the University of Toronto whose research spans deep learning, probabilistic modeling, and machine learning methods for science and AI safety.

Shauna Kravec portrait
Researcher 3 reports

Shauna Kravec

Anthropic

Researcher focused on AI safety, reinforcement learning, and language models, with public work spanning red teaming, adversarial robustness, and model behavior.

Simon Goldstein portrait
Researcher 1 reports

Simon Goldstein

Anthropic

Assistant Professor of Philosophy at The University of Hong Kong and Research Fellow at Anthropic, working in ethics, epistemology, and social and political philosophy.

Jesse Mu portrait
Researcher 1 reports

Jesse Mu

Anthropic

Jesse Mu is a Research Scientist at Anthropic and a visiting researcher at Stanford University. His work spans machine learning, AI safety, reinforcement learning, and deep learning theory.

Linda Petrini portrait
Researcher 1 reports

Linda Petrini

Anthropic

Research scientist at Anthropic focused on safety and robustness for language models and reinforcement learning.

Roger Grosse portrait
Researcher 1 reports

Roger Grosse

Anthropic

Associate Professor of Computer Science at the University of Toronto and director of the machine learning group, with research spanning probabilistic models and optimization algorithms.

Amanda Askell portrait
Researcher 7 reports

Amanda Askell

Anthropic / OpenAI

Alignment researcher at OpenAI working on making AI understandable to and aligned with human values.

Jared D. Kaplan portrait
Researcher 6 reports

Jared D. Kaplan

Anthropic

Anthropic co-founder and Chief Science Officer. Formerly a physicist at Johns Hopkins, he helped develop scaling laws for neural language models and works on the science and safety of large AI systems.

Yuntao Bai portrait
Researcher 4 reports

Yuntao Bai

Anthropic

Anthropic researcher whose work includes reinforcement learning from human feedback and Constitutional AI; previously a Sherman Fairchild Postdoctoral Scholar in theoretical high-energy physics at Caltech.

David Bau portrait
Researcher 3 reports

David Bau

Anthropic

Research scientist at Anthropic and assistant professor of computer science at Northeastern University working on interpretability and model understanding.

Dan Hendrycks portrait
Researcher 1 reports

Dan Hendrycks

Anthropic

AI safety researcher and director of the Center for AI Safety; advisor to xAI and Scale AI, previously an advisor to OpenAI and Anthropic.

Carina Kauf portrait
Researcher 1 reports

Carina Kauf

Anthropic

Member of Anthropic's Societal Impacts team, where she studies the real-world impacts of AI systems.

Kamal Ndousse portrait
Researcher 5 reports

Kamal Ndousse

Anthropic

Researcher at Anthropic working on alignment, reasoning, and evaluation for large language models.

Sören Mindermann portrait
Researcher 3 reports

Sören Mindermann

Anthropic

Research scientist at Anthropic working on machine learning and AI safety.

Jan Leike portrait
Researcher 2 reports

Jan Leike

Anthropic

Anthropic researcher focused on AI safety, alignment, and auditing hidden objectives in language models.

Josh Batson portrait
Researcher 2 reports

Josh Batson

Anthropic

Member of technical staff at Anthropic interested in understanding deep learning and AI safety; previously a research scientist at OpenAI.

Henry Sleight portrait
Researcher 1 reports

Henry Sleight

Anthropic

PhD student at the University of Oxford working on AI safety, including scalable oversight and interpretability.

Jack Chen portrait
Researcher 1 reports

Jack Chen

Anthropic

Researcher at Anthropic with interests in machine learning, AI alignment, and economics.

Kshitij Sachan portrait
Researcher 1 reports

Kshitij Sachan

Anthropic

Kshitij Sachan is a research scientist at Anthropic whose public homepage and Google Scholar profile highlight work on language models, reasoning, code generation, and machine learning systems.

Michael Sellitto portrait
Researcher 1 reports

Michael Sellitto

Anthropic

Research scientist at Anthropic working on trustworthy AI and deceptive alignment.

Mrinank Sharma portrait
Researcher 1 reports

Mrinank Sharma

Anthropic

AI safety researcher who led Anthropic's Safeguards Research Team and worked on jailbreak robustness, automated red teaming, and monitoring for misuse and misalignment.

Zachary Witten portrait
Researcher 1 reports

Zachary Witten

Anthropic

Zachary Witten is a member of technical staff at Anthropic.

Ethan Perez portrait
Researcher 8 reports

Ethan Perez

Anthropic

Research scientist at Anthropic focused on scalable oversight, AI safety, and language model evaluation; previously worked at New York University and Google.

Nicholas Schiefer portrait
Researcher 8 reports

Nicholas Schiefer

Anthropic

Member of Technical Staff at Anthropic and cofounder of Oulipo Labs, working on language model safety, evaluations, and scientific forecasting.

Deep Ganguli portrait
Researcher 6 reports

Deep Ganguli

Anthropic

Co-founder and head of alignment science at Anthropic.

Dario Amodei portrait
Researcher 5 reports

Dario Amodei

Anthropic / OpenAI

CEO and co-founder of Anthropic. Before Anthropic, he served as vice president of research at OpenAI.

Nova DasSarma portrait
Researcher 5 reports

Nova DasSarma

Anthropic

Research scientist at Anthropic interested in understanding neural networks and applying that understanding to alignment.

Anna Chen portrait
Researcher 4 reports

Anna Chen

Anthropic

Researcher working on AI safety and adversarial evaluation, including Anthropic many-shot jailbreaking research.

Saurav Kadavath portrait
Researcher 4 reports

Saurav Kadavath

Anthropic

Research scientist at Anthropic interested in understanding and steering AI systems.

Tom Conerly portrait
Researcher 4 reports

Tom Conerly

Anthropic

Software engineer at Anthropic, previously at Google, with public writing on language models, agents, and reinforcement learning.

Alex Tamkin portrait
Researcher 3 reports

Alex Tamkin

Anthropic

Member of technical staff at Anthropic whose work focuses on language models, model understanding, and alignment.

Buck Shlegeris portrait
Researcher 3 reports

Buck Shlegeris

Anthropic

Buck Shlegeris is a Member of Technical Staff at Anthropic whose public homepage focuses on AI safety, model evaluations, and alignment.

Beth Barnes portrait
Researcher 2 reports

Beth Barnes

Anthropic

President of METR and former team member at Anthropic whose work focuses on evaluating and forecasting frontier AI capabilities.

Carson Denison portrait
Researcher 2 reports

Carson Denison

Anthropic

Member of Technical Staff at Anthropic and PhD student at Carnegie Mellon University focused on AI safety, evaluations, and oversight of large language models.

Jared Kaplan portrait
Researcher 2 reports

Jared Kaplan

Anthropic

Jared Kaplan is a researcher at Anthropic known for work on scaling laws and large language models.

Monte MacDiarmid portrait
Researcher 2 reports

Monte MacDiarmid

Anthropic

Member of technical staff at Anthropic working on alignment science and the evaluation of hidden objectives in language models.

Adam Jermyn portrait
Researcher 1 reports

Adam Jermyn

Anthropic

Research scientist at Anthropic and former professor of theoretical astrophysics at Stony Brook University.

Alexey Nazarov portrait
Researcher 1 reports

Alexey Nazarov

Anthropic

Member of technical staff at Anthropic focused on safe and reliable AI.

Daniel M. Ziegler portrait
Researcher 1 reports

Daniel M. Ziegler

Anthropic

Research scientist at Anthropic whose public work spans reinforcement learning from human feedback, AI alignment, and scalable language model training.

Esin Durmus portrait
Researcher 1 reports

Esin Durmus

Anthropic

Assistant professor of marketing at Stanford Graduate School of Business whose research uses AI systems to study human decision-making and related machine learning questions.

Holden Karnofsky portrait
Researcher 1 reports

Holden Karnofsky

Anthropic

Co-founder and president of Anthropic and writer of the Cold Takes blog.

Jan Brauner portrait
Researcher 1 reports

Jan Brauner

Anthropic

Computer scientist at Anthropic focused on making advanced AI systems safe and beneficial.

Johannes Treutlein portrait
Researcher 1 reports

Johannes Treutlein

Anthropic

Member of Technical Staff at Anthropic and researcher in neural circuits and mechanistic interpretability, building tools for understanding AI systems.

Owain Evans portrait
Researcher 1 reports

Owain Evans

Anthropic

Assistant Professor of Computer Science at the University of Oxford whose research spans generalization, reasoning, and large language model agents.

Paul Christiano portrait
Researcher 1 reports

Paul Christiano

Anthropic

Researcher focused on AI alignment, reasoning under uncertainty, and the long-term safety of advanced AI systems.

Rylan Schaeffer portrait
Researcher 1 reports

Rylan Schaeffer

Anthropic

Research scientist at Anthropic focused on AI alignment, language model behavior, and scalable oversight.

Scott Emmons portrait
Researcher 1 reports

Scott Emmons

Anthropic

Member of Technical Staff at Anthropic working on AI control, hidden objectives, alignment, and evaluations, with a background in language models, efficient training, and scientific machine learning.

Wes Gurnee portrait
Researcher 1 reports

Wes Gurnee

Anthropic

Member of technical staff at Anthropic working on deep learning, mechanistic interpretability, and AI safety.