Artificial Intelligence thread

Eventine

Junior Member
Registered Member
Please, Log in or Register to view URLs content!

Hi. Can some experts help to explain the significance of
1) the slow speed for Deepseek R1 relative to other models?
2) the higher price of 500% more for Deepseek R1 when compared to Gemini 2.0?

How will these factors matter for end-users and determine the eventual winning model?
Thank you very much.
On top of what was already said, o3-mini and Gemini 2.0 Flash are both assumed to be significantly smaller models than R1; they are, however, highly optimized for specific use cases, especially coding and logic. R1 by contrast is more of a generalist model like O1 or Claude Sonnet. This is why R1 is much better at creative writing, for example, but it comes at the cost of model size and inference speed.

Google, Open AI, and Anthropic are the top Western AI companies, and despite all the jokes we're making, the reality is that they will remain tough competition for China in global AI dominance due to their vast financial resources and access to top end GPUs. A battle was won, but the war is just getting started.
 

tphuang

Lieutenant General
Staff member
Super Moderator
VIP Professional
Registered Member
FT article about China having so much AI chips that H100 pricing dropped significantly last few months.

again, 250 public data centers being built. No chip access issues or funding issues. Biggest constraint is know how for people actually building them. Imagine that not all the SOEs building data centers has a clue what they are doing.


Steve with his chart of Nvidia revenue by region


More on re export to China via intermediate countries

 
Top