-
NVIDIA Research
- Santa Clara
- https://chrisding.github.io/
Highlights
- Pro
Stars
DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scaling
DeepEP: an efficient expert-parallel communication library
A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 🍓 and reasoning techniques.
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)
Efficient Triton Kernels for LLM Training
A programming framework for agentic AI 🤖 PyPi: autogen-agentchat Discord: https://aka.ms/autogen-discord Office Hour: https://aka.ms/autogen-officehour
A bibliography and survey of the papers surrounding o1
Heterogeneous Pre-trained Transformer (HPT) as Scalable Policy Learner.
Eagle Family: Exploring Model Designs, Data Recipes and Training Strategies for Frontier-Class Multimodal LLMs
A paper list about diffusion models for natural language processing.
MiniCPM-o 2.6: A GPT-4o Level MLLM for Vision, Speech and Multimodal Live Streaming on Your Phone
MiniCPM3-4B: An edge-side LLM that surpasses GPT-3.5-Turbo.
InternLM-XComposer2.5-OmniLive: A Comprehensive Multimodal System for Long-term Streaming Video and Audio Interactions
The Go-To Choice for CV Data Visualization, Annotation, and Model Analysis.
[ECCV 2024] Embodied Understanding of Driving Scenarios
[ICLR2024] TopoMLP: A Simple yet Strong Pipeline for Driving Topology Reasoning
【CVPR 2024 Highlight】Monkey (LMM): Image Resolution and Text Label Are Important Things for Large Multi-modal Models
[AAAI2024] Far3D: Expanding the Horizon for Surround-view 3D Object Detection
[CVPR 2024] Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data. Foundation Model for Monocular Depth Estimation
Official Pytorch Implementation of Self-emerging Token Labeling
✨✨Latest Advances on Multimodal Large Language Models
EVA Series: Visual Representation Fantasies from BAAI