Atlas / Reports / Detail
DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning
Large Language Models
Connected researchers
Daya Guo
DeepSeek / Moonshot AI
DeepSeek researcher focused on NLP, code intelligence, and LLM reasoning, with public work spanning DeepSeek-Coder, DeepSeekMath, DeepSeek-V2, DeepSeek-V3, and DeepSeek-R1.
Qihao Zhu
DeepSeek
Research scientist focused on foundation models and multimodal large language models; his homepage notes earlier work at DeepSeek AI and current research at the University of Southern California.
Dejian Yang
DeepSeek
DeepSeek team member and co-author of the DeepSeek-V3, DeepSeek-V2, and DeepSeek LLM technical reports.
Y. Wu
DeepSeek
Yu Wu is a researcher at DeepSeek AI and head of its LLM Alignment Team. His public homepage highlights work on reinforcement learning and alignment for the DeepSeek model family, including DeepSeek-V3, DeepSeek-R1, and DeepSeekMath, and notes prior work at Microsoft Research Asia.
Mao Zheng
DeepSeek
Co-founder of DeepSeek who leads pre-training and post-training research, with a focus on language models, reasoning models, and reinforcement learning.
Junxiao Song
DeepSeek
Member of Technical Staff at DeepSeek.
Haowei Zhang
DeepSeek
Research scientist at DeepSeek with public GitHub work on language models and AI systems.
Peiyi Wang
DeepSeek
Research scientist at DeepSeek with public GitHub projects on AI systems.
Wenfeng Liang
DeepSeek
Researcher affiliated with DeepSeek-AI and co-author of the Nature paper introducing DeepSeek-R1.
Ruoyu Zhang
DeepSeek
Researcher affiliated with DeepSeek-AI and co-author of the Nature paper introducing DeepSeek-R1.