Atlas / Reports / Detail
NVLM: Open Frontier-Class Multimodal LLMs
Multimodal Language Models report from NVIDIA with 10 connected researchers in the LLMpeople atlas.
Connected researchers
Yuyin Zhou
NVIDIA
Assistant Professor of Computer Science and Engineering at UC Santa Cruz working on multimodal learning, computer vision, and medical image analysis.
Zhaowen Wang
NVIDIA
Research manager at NVIDIA working on large-scale distributed pretraining, synthetic data, multimodal LLMs, and computer vision.
Jianfeng Gao
NVIDIA
Jianfeng Gao is a researcher in natural language processing and multimodal foundation models whose public homepage and Google Scholar profile highlight work on dialogue systems, retrieval, reasoning, and vision-language models.
Thaddeus Culhane
NVIDIA
Research scientist at NVIDIA working on multimodal AI, especially language and vision models.
Andy Yao
NVIDIA
Research scientist at NVIDIA with public publications on multimodal language models and visual instruction tuning, including NVLM, VILA, and Video2Flow.
Caiming Xiong
NVIDIA
Vice President of AI Research and General Manager of AI Platforms at NVIDIA.
Ran Tian
NVIDIA
Research scientist at NVIDIA working on multimodal language models and vision-language research, with public publications including NVLM, VILA, and Visual Role Play.
Rulin Shao
NVIDIA
PhD student at UCLA and research intern at NVIDIA, working on multimodal reasoning, vision-language models, and embodied AI.
Shyamal Anadkat
NVIDIA
Research scientist at NVIDIA working on AI agents, multimodal systems, and robotics, including the NVLM project.
Weizhu Chen
NVIDIA
Distinguished scientist and managing director at Microsoft Research working on natural language processing and large language models.