Next-generation reasoning model that runs locally in-browser
Explore LLM performance across hardware