Anton
Feeding LLMs @ Hugging Face
- LLM Reasoning labs will be eating good today🍔 We commandeered the HF cluster for a few days and generated 1.2M reasoning-filled solutions to 500k NuminaMath problems with DeepSeek-R1 🐳 Have fun!
- 🤗 Dataset: huggingface.co/datasets/ope...
- Introducing 📐FineMath: the best open math pre-training dataset with 50B+ tokens! Math remains challenging for LLMs and by training on FineMath we see considerable gains over other math datasets, especially on GSM8K and MATH. 🤗 huggingface.co/datasets/Hug... Here’s a breakdown 🧵
- First let’s break down how AI labs curate math pre-training datasets 🕵️ DeepSeekMath and QwenMath train a fastText classifier on data like OpenWebMath (OWM). They iteratively filter and recall math content from Common Crawl, focusing on the most relevant domains.
- Reposted by AntonThe Open LLM Leaderboard got a new front page for Christmas Check it out at huggingface.co/spaces/open-...
- Reposted by Anton[Not loaded yet]
- Reposted by AntonLet's go! We are releasing SmolVLM, a smol 2B VLM built for on-device inference that outperforms all models at similar GPU RAM usage and tokens throughputs. SmolVLM can be fine-tuned on a Google collab and be run on a laptop! Or process millions of documents with a consumer GPU!
- Reposted by AntonSmall yet mighty! 💫 We are releasing SmolVLM: a new 2B small vision language made for on-device use, fine-tunable on consumer GPU, immensely memory efficient 🤠 We release three checkpoints under Apache 2.0: SmolVLM-Instruct, SmolVLM-Synthetic and SmolVLM-Base huggingface.co/collections/...
- Check out how easy it is to do LLM evals with LightEval! * any dataset on the 🤗 Hub can become an eval task in a few lines of code: customize the prompt, metrics, parsing, few-shots, everything! * model- and data-parallel inference * auto batching with the new vLLM backend
- Repo: github.com/huggingface/... Here's how we use it for SmolLM 🤏 github.com/huggingface/...
- Reposted by Anton[Not loaded yet]
- Reposted by Anton[Not loaded yet]