Artificial Intelligence thread

9dashline

Captain
Registered Member
I know, but if you listen to his interviews it is clear that he is being pressured by shareholders to close-source Llama because of the exponential (?) increase in AI dev costs.

DeepSeek V3 just blew a huge hole on his argument that an infinitely growing multi-billion investment is necessary.

Similarly for Sam, you can see that OpenAI will get pressured.

In the West, Google seems to be the best going forward on inference and scaling up.
Yup, Google is probably almost gleeful, for now, that the likes of deepseek and qwen are helping it clear out space from its competitors....
 

diadact

New Member
Registered Member
Did Deepseek use data from western sources abt this thing or something?
They didn't post-train it well primarily due to resource constraints
The next update(1 or 2 months) will have none of these problems and will crush every LLM available rn
 

Overbom

Brigadier
Registered Member
They didn't post-train it well primarily due to resource constraints
The next update(1 or 2 months) will have none of these problems and will crush every LLM available rn
Reasoning is the big one to await for. In their paper they mentioned specifically how their R1 reasoning model helped them a lot on generating synthetic data.

If R1 was so helpful, which I don't rate it too highly tbh, then I can only imagine the leap forward with a R2 reasoning model fine tuned on a DeepSeek V3 LLM
 

luminary

Senior Member
Registered Member

Overbom

Brigadier
Registered Member
an indian OpenAI employee lol
I guess when you're outcompeted by a scale of 300x the only thing left to do is cope.

Well guess what, if you had done GPT o1 correctly, maybe you would have also gotten the o1 visa you want.


When you can't compete, just change the benchmarks:
Something is fishy with the results. Gemini 1.5 flash and Grok-beta are actually hot garbage.
Also no way Gemma 2 9b is better


Whole thing smells. Smells bad
 
Top