Researchers from Stanford University and the University of Washington have developed a new AI reasoning model named s1, which was trained for under $50 in cloud compute credits. The model, detailed in a research paper released last Friday, rivals the performance of leading models such as OpenAI's o1 and DeepSeek's R1 in tests measuring math and coding abilities. The s1 model was created using a distillation process, where it was fine-tuned using answers from Google's Gemini 2.0 Flash Thinking Experimental model. Utilizing a small dataset of 1,000 carefully curated questions, the researchers completed the training in less than 30 minutes using 16 Nvidia H100 GPUs. The s1 model, along with its training data and code, is available on GitHub, highlighting a potential shift in the accessibility and cost of developing advanced AI models.
OpenAI now reveals more of its o3-mini model’s thought process: https://t.co/GeuyAIT6uY by TechCrunch #infosec #cybersecurity #technology #news
OpenAI now reveals more of its o3-mini model’s thought process https://t.co/FwTVUWb3Lx
OpenAI is now showing more of its o3 AI model's reasoning steps, similar to what DeepSeek does https://t.co/7S1tHi3di1