LLMpeople
Home People Organizations Reports Fields Schools
Public Atlas People first, reports as evidence, organizations as context.

Atlas / People / Detail

Ethan Perez

Research scientist at Anthropic focused on scalable oversight, AI safety, and language model evaluation; previously worked at New York University and Google.

Researcher1 organizations8 reports

Profile status: updated

Ethan Perez portrait
Suggest a correction
Suggest a source

Trust signals

Profile completeness41%
Public sources1
Official sources1
Last reviewedMar 13, 2026
Official homepage
updated 1 public sources

Public links

website Personal homepage

Organizations

core Anthropic

Reports

Alignment and RLHF Constitutional AI: Harmlessness from AI Feedback Alignment and Safety Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training Alignment and Safety Alignment faking in large language models Alignment and Safety Auditing language models for hidden objectives Alignment and Safety Constitutional Classifiers: Defending against Universal Jailbreaks across Thousands of Hours of Red Teaming Alignment and Safety Constitutional Classifiers++: Defending against Universal Jailbreaks across Thousands of Hours of Red Teaming Interpretability On the Biology of a Large Language Model Interpretability Tracing the thoughts of a large language model

Official and primary sources

https://ethanperez.net/ Official source · homepage

LLMpeople is a public atlas for discovering frontier AI researchers with context, provenance, and respect.

Privacy · Terms