-
EVA-CLIP-18B: Scaling CLIP to 18 Billion Parameters
Paper • 2402.04252 • Published • 28 -
Vision Superalignment: Weak-to-Strong Generalization for Vision Foundation Models
Paper • 2402.03749 • Published • 13 -
ScreenAI: A Vision-Language Model for UI and Infographics Understanding
Paper • 2402.04615 • Published • 44 -
EfficientViT-SAM: Accelerated Segment Anything Model Without Performance Loss
Paper • 2402.05008 • Published • 23
Collections
Discover the best community collections!
Collections including paper arxiv:2412.05271
-
One-Minute Video Generation with Test-Time Training
Paper • 2504.05298 • Published • 102 -
MoCha: Towards Movie-Grade Talking Character Synthesis
Paper • 2503.23307 • Published • 133 -
Towards Understanding Camera Motions in Any Video
Paper • 2504.15376 • Published • 155 -
Antidistillation Sampling
Paper • 2504.13146 • Published • 60
-
Expanding Performance Boundaries of Open-Source Multimodal Models with Model, Data, and Test-Time Scaling
Paper • 2412.05271 • Published • 157 -
naver-clova-ix/cord-v2
Viewer • Updated • 1k • 4.69k • 79 -
naver-clova-ix/synthdog-en
Viewer • Updated • 66k • 1.04k • 18 -
impira/layoutlm-invoices
Document Question Answering • Updated • 23.8k • 200
-
lang-uk/recruitment-dataset-job-descriptions-english
Viewer • Updated • 142k • 647 • 13 -
lang-uk/recruitment-dataset-candidate-profiles-english
Viewer • Updated • 210k • 253 • 7 -
cnamuangtoun/resume-job-description-fit
Viewer • Updated • 8k • 1.39k • 50 -
shashu2325/resume-job-matcher-lora
Updated • 1.08k • 7
-
VL-Rethinker: Incentivizing Self-Reflection of Vision-Language Models with Reinforcement Learning
Paper • 2504.08837 • Published • 42 -
OpenVLThinker: An Early Exploration to Complex Vision-Language Reasoning via Iterative Self-Improvement
Paper • 2503.17352 • Published • 23 -
Expanding Performance Boundaries of Open-Source Multimodal Models with Model, Data, and Test-Time Scaling
Paper • 2412.05271 • Published • 157 -
R1-Omni: Explainable Omni-Multimodal Emotion Recognition with Reinforcing Learning
Paper • 2503.05379 • Published • 37
-
BigDocs: An Open and Permissively-Licensed Dataset for Training Multimodal Models on Document and Code Tasks
Paper • 2412.04626 • Published • 14 -
GMAI-VL & GMAI-VL-5.5M: A Large Vision-Language Model and A Comprehensive Multimodal Dataset Towards General Medical AI
Paper • 2411.14522 • Published • 39 -
Both Text and Images Leaked! A Systematic Analysis of Multimodal LLM Data Contamination
Paper • 2411.03823 • Published • 49 -
Infinity-MM: Scaling Multimodal Performance with Large-Scale and High-Quality Instruction Data
Paper • 2410.18558 • Published • 20
-
Smarter, Better, Faster, Longer: A Modern Bidirectional Encoder for Fast, Memory Efficient, and Long Context Finetuning and Inference
Paper • 2412.13663 • Published • 149 -
Qwen2.5 Technical Report
Paper • 2412.15115 • Published • 367 -
Are Your LLMs Capable of Stable Reasoning?
Paper • 2412.13147 • Published • 95 -
Byte Latent Transformer: Patches Scale Better Than Tokens
Paper • 2412.09871 • Published • 102