OPUS: Towards Efficient and Principled Data Selection in Large Language Model Pre-training in Every Iteration Paper • 2602.05400 • Published 11 days ago • 314
DeepGen 1.0: A Lightweight Unified Multimodal Model for Advancing Image Generation and Editing Paper • 2602.12205 • Published 4 days ago • 75
Zooming without Zooming: Region-to-Image Distillation for Fine-Grained Multimodal Perception Paper • 2602.11858 • Published 4 days ago • 51
camenduru/dinov3-vitl16-pretrain-lvd1689m Image Feature Extraction • 0.3B • Updated Dec 17, 2025 • 4.54k • 1
IBBI-bio/dinov3-vitl16-pretrain-lvd1689m Image Feature Extraction • 21.6M • Updated Aug 29, 2025 • 584 • 3
Ming-V2 Collection Ming is the multi-modal series of any-to-any models developed by Ant Ling team. • 14 items • Updated 2 days ago • 33
Ming-Flash-Omni: A Sparse, Unified Architecture for Multimodal Perception and Generation Paper • 2510.24821 • Published Oct 28, 2025 • 40
Ming-Omni: A Unified Multimodal Model for Perception and Generation Paper • 2506.09344 • Published Jun 11, 2025 • 30