The research team modified the model and called the result s1. Preliminary training involved 1,000 question-and-answer pairs they had designed carefully to give their model a leg up on learning.
OpenAI and Microsoft recently accused DeepSeek of using their copyrighted data to train its ultra-cost-effective model. s1's training process took less than 30 minutes using 16 NVIDIA H100 GPUs.
o1 is the inference model that OpenAI first released last year ... They also mentioned that the computing performance needed to train s1 could be utilized for about $20. Researchers from Stanford ...