-
Improved Baselines with Visual Instruction Tuning
Paper • 2310.03744 • Published • 32 -
DeepSeek-VL: Towards Real-World Vision-Language Understanding
Paper • 2403.05525 • Published • 38 -
Qwen-VL: A Frontier Large Vision-Language Model with Versatile Abilities
Paper • 2308.12966 • Published • 6 -
LLaVA-Gemma: Accelerating Multimodal Foundation Models with a Compact Language Model
Paper • 2404.01331 • Published • 22
Collections
Discover the best community collections!
Collections including paper arxiv:2310.03744
-
Visual Instruction Tuning
Paper • 2304.08485 • Published • 8 -
Qwen-VL: A Frontier Large Vision-Language Model with Versatile Abilities
Paper • 2308.12966 • Published • 6 -
Improved Baselines with Visual Instruction Tuning
Paper • 2310.03744 • Published • 32 -
SILC: Improving Vision Language Pretraining with Self-Distillation
Paper • 2310.13355 • Published • 5
-
Improved Baselines with Visual Instruction Tuning
Paper • 2310.03744 • Published • 32 -
llava-hf/llava-v1.6-mistral-7b-hf
Image-Text-to-Text • Updated • 3.82M • 121 -
llava-hf/llava-v1.6-vicuna-7b-hf
Image-Text-to-Text • Updated • 13.8k • 9 -
llava-hf/llava-v1.6-vicuna-13b-hf
Image-Text-to-Text • Updated • 121k • 5
-
Woodpecker: Hallucination Correction for Multimodal Large Language Models
Paper • 2310.16045 • Published • 13 -
SILC: Improving Vision Language Pretraining with Self-Distillation
Paper • 2310.13355 • Published • 5 -
To See is to Believe: Prompting GPT-4V for Better Visual Instruction Tuning
Paper • 2311.07574 • Published • 13 -
MyVLM: Personalizing VLMs for User-Specific Queries
Paper • 2403.14599 • Published • 14
-
Textbooks Are All You Need
Paper • 2306.11644 • Published • 137 -
LLaVA-φ: Efficient Multi-Modal Assistant with Small Language Model
Paper • 2401.02330 • Published • 11 -
Textbooks Are All You Need II: phi-1.5 technical report
Paper • 2309.05463 • Published • 84 -
Visual Instruction Tuning
Paper • 2304.08485 • Published • 8
-
Visual Instruction Tuning
Paper • 2304.08485 • Published • 8 -
LLaVA-Plus: Learning to Use Tools for Creating Multimodal Agents
Paper • 2311.05437 • Published • 40 -
Improved Baselines with Visual Instruction Tuning
Paper • 2310.03744 • Published • 32 -
Aligning Large Multimodal Models with Factually Augmented RLHF
Paper • 2309.14525 • Published • 29
-
DocGraphLM: Documental Graph Language Model for Information Extraction
Paper • 2401.02823 • Published • 32 -
Understanding LLMs: A Comprehensive Overview from Training to Inference
Paper • 2401.02038 • Published • 59 -
DocLLM: A layout-aware generative language model for multimodal document understanding
Paper • 2401.00908 • Published • 173 -
Attention Where It Matters: Rethinking Visual Document Understanding with Selective Region Concentration
Paper • 2309.01131 • Published • 1
-
ImageBind: One Embedding Space To Bind Them All
Paper • 2305.05665 • Published • 3 -
ZoeDepth: Zero-shot Transfer by Combining Relative and Metric Depth
Paper • 2302.12288 • Published -
HuggingFaceM4/howto100m
Updated • 14 • 3 -
BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
Paper • 2201.12086 • Published • 2