klotz: university of washington* + distillation*

0 bookmark(s) - Sort by: Date ↓ / Title / - Bookmarks from other users for this tag

  1. AI researchers at Stanford and the University of Washington trained an AI 'reasoning' model named s1 for under $50 using cloud compute credits. The model, which performs similarly to OpenAI’s o1 and DeepSeek’s R1, is available on GitHub. It was developed using distillation from Google’s Gemini 2.0 Flash Thinking Experimental model and demonstrates strong performance on benchmarks.

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: Tags: university of washington + distillation

About - Propulsed by SemanticScuttle