DeepSeek R1 is even better than OpenAI o1 and Claude 3.5 Sonnet
So i got to play around with DeepSeek R1, and based on the benchmarks I've seen and my test results, I can say it's just as good if not better (at certain things) than OpenAI o1 and Claude 3.5 Sonnet. It's a lot cheaper too (very small fraction of o1's and claude's pricing) but delivered results.
here are some of its technical specs:
Total Parameters: 671 billion
- Active Parameters per Token: 37 billion
- Context Length: Up to 128K tokens
- Training Data: Trained on 14.8 trillion tokens
- Training Compute Cost: Approximately 2.664 million H800 GPU hours
(taken from this detailed article: https://blog.getbind.co/2025/01/23/deepseek-r1-vs-gpt-o1-vs-claude-3-5-sonnet-which-is-best-for-coding/