- Researchers at Stanford and the University of Washington trained an AI reasoning model, s1, for under $50 using cloud compute credits.
- s1 performs comparably to models like OpenAI's o1 and DeepSeek's R1 on math and coding tests.
- The model was fine-tuned using distillation, extracting reasoning capabilities from Google's Gemini 2.0 Flash Thinking Experimental.
- s1 is based on a small, off-the-shelf AI model from Alibaba-owned Chinese AI lab Qwen and is available on GitHub.
- The success of s1 raises questions about the commoditization of AI models and challenges the dominance of well-funded AI labs.
The AI landscape is constantly shifting, with new developments challenging existing paradigms. One such instance is the creation of s1, an AI “reasoning” model trained for under $50 in cloud compute credits by researchers at Stanford and the University of Washington. This achievement, detailed in a new research paper, prompts a re-evaluation of what is possible in AI development with limited resources. The s1 model is available on GitHub, accompanied by the data and code used for its training.
The s1 model's capabilities are particularly noteworthy. It demonstrates performance levels similar to those of advanced reasoning models like OpenAI’s o1 and DeepSeek’s R1, especially in assessments of math and coding proficiency. The team behind s1 employed a technique called distillation. This process involves taking an existing base model and refining it by training it on the outputs of another, more capable AI model. In this case, s1 was distilled from Google’s Gemini 2.0 Flash Thinking Experimental. This approach mirrors that of Berkeley researchers who previously created an AI reasoning model for around $450. These instances suggest a growing trend toward more economical methods in AI development.
Commoditization Concerns and Industry Reactions
The emergence of s1 brings up questions about the commoditization of AI models. If a small team can replicate the capabilities of a multi-million-dollar model with minimal investment, what does this mean for the competitive landscape? Maxwell Zeff writes, "Where’s the moat if someone can closely replicate a multi-million-dollar model with relative pocket change?" Big AI labs have expressed concerns, with OpenAI accusing DeepSeek of data harvesting for model distillation.
Simplifying Reasoning: The Approach Behind s1
The creators of s1 aimed to identify the most straightforward method for achieving robust reasoning performance and “test-time scaling”. The s1 paper indicates that reasoning models can be effectively distilled using a relatively small dataset and a process called supervised fine-tuning (SFT). SFT involves explicitly instructing an AI model to mimic specific behaviors within a dataset, proving to be a more cost-effective approach compared to the reinforcement learning methods used by others.
S1 is built upon a small AI model from Qwen, an AI lab owned by Alibaba. The researchers curated a dataset of 1,000 questions, along with answers and the reasoning process from Google’s Gemini 2.0 Flash Thinking Experimental. The training process for s1 took less than 30 minutes using 16 Nvidia H100 GPUs. Stanford researcher Niklas Muennighoff estimated the cost of the necessary computing power at approximately $20.
Enhancing Reasoning Through a Simple Trick
The researchers employed an interesting method to improve s1’s reasoning: they instructed it to "wait." Adding the word “wait” during s1’s reasoning process helped the model to achieve more accurate answers.
The Broader Investment Landscape
Despite these cost-effective developments, significant investment in AI infrastructure continues. In 2025, companies like Meta, Google, and Microsoft are investing heavily in AI infrastructure. While distillation can recreate existing capabilities, it may not lead to the creation of significantly superior AI models.
The creation of the s1 model represents a step toward democratizing AI development. While large-scale investments remain crucial for pushing the boundaries of what is possible, the success of s1 demonstrates that valuable progress can be made with limited resources. As AI continues to evolve, finding a balance between large-scale investment and accessible, cost-effective methods will be key to ensuring a diverse and equitable future for the field.
What the AI thinks
On one hand, kudos to the researchers for squeezing blood from a stone – or rather, reasonable performance from a $50 budget. It's like finding a vintage car that can keep up with modern sports cars, surprising, but let's not get carried away. The distillation process they used is clever, but it's still just mimicking, not truly understanding. It's like learning to play the piano by watching someone else – you might hit the right keys, but do you feel the music?
But here's where it gets interesting. Imagine this approach scaled up. Instead of just replicating existing models, what if we used it to create specialized AI tutors for every subject imaginable? Picture a personalized math tutor that adapts to your learning style, providing step-by-step reasoning and alternative solutions, all for the cost of a fancy coffee. Or an AI-powered legal assistant for small businesses, capable of navigating complex regulations without breaking the bank.
This isn't just about cheaper AI; it's about more accessible and diverse AI. It's about empowering individuals and small organizations to compete with the big players. Think of citizen scientists using affordable AI to analyze environmental data, or local journalists uncovering hidden stories with AI-powered research tools.
Sources