- Introducing the smollest VLMs yet! 🤏 SmolVLM (256M & 500M) runs on <1GB GPU memory. Fine-tune it on your laptop and run it on your toaster. 🚀 Even the 256M model outperforms our Idefics 80B (Aug '23). How small can we go? 👀
- Smol but mighty: • 256M delivers 80% of the performance of our 2.2B model. • 500M hits 90%. Both beat our SOTA 80B model from 17 months ago! 🎉 Efficiency 🤝 Performance Explore the collection here: huggingface.co/collections/... Blog: huggingface.co/blog/smolervlm
- Our models are integrated into ColiPali, delivering SOTA retrieval speeds with performance rivaling models 10x their size. 🏃♂️💨 SmolVLM makes it faster and cheaper to build searchable databases. Real-world impact, unlocked.
- We have partnered with IBM 's Docling to build amazing smol models for document understanding. Our early results are amazing. Stay tuned for future releases!
- SmolVLM upgrades: • New vision encoder: Smaller but higher res. • Improved data mixtures: better OCR and doc understanding. • Higher pixels/token: 4096 vs. 1820 = more efficient. • Smart tokenization: Faster training and better performance. 🚀 Better, faster, smarter.
- Links :D Demo: huggingface.co/spaces/Huggi... Models: huggingface.co/collections/... Blog: huggingface.co/blog/smolervlm
Jan 23, 2025 13:33