Current frame
Researcher in multimodal intelligence and large-scale vision-language training
Atlas / People / Detail
Chunyuan Li is a multimodal AI researcher whose public homepage highlights work on large-scale language and vision training, including LLaVA, GroundingDINO, GLIP, GLIGEN, Florence, and Oscar.
Profile status: updated
Researcher in multimodal intelligence and large-scale vision-language training
According to his public homepage, Chunyuan Li focuses on multimodal intelligence with an emphasis on large-scale language and vision training. His homepage highlights contributions including LLaVA and earlier work such as GroundingDINO, GLIP, GLIGEN, Florence, and Oscar. The same page says he has held research roles at xAI, ByteDance, and Microsoft Research, Redmond, and that he earned a PhD in machine learning from Duke University under Lawrence Carin, where his doctoral work explored deep generative models. OpenReview also links his homepage and Google Scholar profile and lists Duke University PhD study from 2014 to 2018.