🍃 MINT-1T Collection Data for "MINT-1T: Scaling Open-Source Multimodal Data by 10x: A Multimodal Dataset with One Trillion Tokens" • 13 items • Updated Jul 24, 2024 • 58
GTE models Collection General Text Embedding Models Released by Tongyi Lab of Alibaba Group • 19 items • Updated 22 days ago • 19
PixMo Collection A set of vision-language datasets built by Ai2 and used to train the Molmo family of models. Read more at https://molmo.allenai.org/blog • 9 items • Updated 5 days ago • 53
Sparsh Collection Models and datasets for Sparsh: Self-supervised touch representations for vision-based tactile sensing • 15 items • Updated Oct 24, 2024 • 12
MIMO: Controllable Character Video Synthesis with Spatial Decomposed Modeling Paper • 2409.16160 • Published Sep 24, 2024 • 33
Qwen2.5 Collection Qwen2.5 language models, including pretrained and instruction-tuned models of 7 sizes, including 0.5B, 1.5B, 3B, 7B, 14B, 32B, and 72B. • 45 items • Updated Nov 28, 2024 • 457
Qwen2.5-Coder Collection Code-specific model series based on Qwen2.5 • 40 items • Updated Nov 28, 2024 • 260
Qwen2.5-Math Collection Math-specific model series based on Qwen2.5 • 9 items • Updated Nov 28, 2024 • 60
NuminaMath Collection Datasets and models for training SOTA math LLMs. See our GitHub for training & inference code: https://github.com/project-numina/aimo-progress-prize • 6 items • Updated Jul 21, 2024 • 70
INDUS: Effective and Efficient Language Models for Scientific Applications Paper • 2405.10725 • Published May 17, 2024 • 32
view article Article Fine-tuning Florence-2 - Microsoft's Cutting-edge Vision Language Models Jun 24, 2024 • 182