LLMpeople
Home People Organizations Reports Fields Schools
Public Atlas People first, reports as evidence, organizations as context.

Atlas / People / Detail

Jackson Kernion

Member of Anthropic's Interpretability team, where he works on understanding how large language models work.

Researcher1 organizations3 reports

Profile status: updated

Jackson Kernion portrait
Suggest a correction
Suggest a source

Contributions are treated as untrusted leads. Public changes happen only after review against public sources.

Trust signals

Profile completeness54%
Public sources1
Official sources1
CountryUnknown
Last reviewedMar 12, 2026
Review outcomeUpdated
Official homepage
updated Unknown location 1 public sources

Latest review note

Added official Anthropic profile, avatar, and an English bio based on the public team page.

Public links

website Anthropic profile

Organizations

core Anthropic

Reports

Alignment and RLHF Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback Alignment and RLHF Constitutional AI: Harmlessness from AI Feedback Alignment and RLHF Collective Constitutional AI: Aligning a Language Model with Public Input

Official and primary sources

https://www.anthropic.com/team/jackson-kernion Official source ยท homepage

LLMpeople is a public atlas for discovering frontier AI researchers with context, provenance, and respect.