- WOW! 🤯 Language models are becoming smaller and more capable than ever! Here's SmolLM2 running 100% locally in-browser w/ WebGPU on a 6-year-old GPU. Just look at that speed! ⚡️😍 Powered by 🤗 Transformers.js and ONNX Runtime Web! How many tokens/second do you get? Let me know! 👇
- 🤏 Learn more about SmolLM2: github.com/huggingface/... 🔗 Online WebGPU demo: huggingface.co/spaces/Huggi...
Nov 27, 2024 13:51