Artificial Intelligence thread

OptimusLion

New Member
Registered Member
Tsinghua University's open source project breaks through the bottleneck of large model computing power: RTX 4090 single card achieves full-blooded operation of DeepSeek-R1

It is very difficult for ordinary users to run the real full-blooded version of DeepSeek-R1 on ordinary hardware, and the cost of renting a server is stressful even for developers

001ZzMwgly1hylfxqtku9j60u0094t9502.jpg


001ZzMwgly1hylfxqyngoj60hy0k040r02.jpg

001ZzMwgly1hylfxr2k6ij60ne09maby02.jpg

001ZzMwgly1hylfxrdpv4j61400l3qrd02.jpg

001ZzMwgly1hylfxrncz1j60zk0tggyj02.jpg

Please, Log in or Register to view URLs content!
 

tphuang

Lieutenant General
Staff member
Super Moderator
VIP Professional
Registered Member
you just need to have enough VRAM to hold the model to run it. inference speed is another story.
 

Proteus

New Member
Registered Member
Some have shown that you can run the full version on a $2000 rig with a lot of system RAM, it's just pretty slow. Looks like this rig also has a lot of system RAM, so the question is how much does a single 4090 speed up the process?
1739622129967.png
Please, Log in or Register to view URLs content!
The post right above you have the answer.
This the stream video test on 14 RTX 3090s.
 
Top