DAPO: An Open-Source LLM Reinforcement Learning System at Scale Paper • 2503.14476 • Published Mar 18 • 125
DAPO: An Open-Source LLM Reinforcement Learning System at Scale Paper • 2503.14476 • Published Mar 18 • 125
Phi-3 Collection Phi-3 family of small language and multi-modal models. Language models are available in short- and long-context lengths. • 26 items • Updated 10 days ago • 566
view article Article Cosmopedia: how to create large-scale synthetic data for pre-training Large Language Models Mar 20, 2024 • 86
Running 937 937 FineWeb: decanting the web for the finest text data at scale 🍷 Generate high-quality web text data for LLM training
view post Post EVA-CLIP 🦖 is the CLIP scaled to the moon! 🔥 The new SotA CLIP-like model 🏆 Highlights ✨ - Performs better in linear probing- Outperforms in Zero-Shot Image-Text Retrieval- Higher zero-shot accuracy in IN-1K As usual, try it with the notebook I built for you https://colab.research.google.com/drive/1K7DdCORC3x4qyhwhuB4fT4wcfJ_BQLKw?usp=sharing#scrollTo=0ZS_lJ7SK6YsI also built a Space for you to compare the output probabilities to CLIP, seems that EVACLIP is more "sure" of it's results 😊 merve/EVACLIPThe authors have shared 8B checkpoints open with Apache 2.0 license 💜 and it's built on top of transformers, super easy to use! BAAI/EVA-CLIP-8BRead the paper EVA-CLIP-18B: Scaling CLIP to 18 Billion Parameters (2402.04252) 📄 ❤️ 9 9 + Reply