This is insanity. It is trained with 10% the compute it took to train LLama 405.from their paper, training took 2.788M H800 GPU hours. For 2 months training run it's roughly only 2000 H800
View attachment 141574
I had a SORA feeling about o3..... it was supposed to be part of "shipmas" but not only was o3 not shipped, it was announced with only benchmarks and not even a live demonstration.... the only reason Altman would do that is because he had something to hide and was in fake it until make it mode, otherwise he would be bragging with real live examples... instead of just a rigged chart/ fake benchmark etc... and they even told the ARC AGI dude to not disclose how much compute was used and the hardware/cost lol.... probably millions of dollars for a fake horse and pony showSome o3 commentary
Basically it seems the test creator is an openAI shill and intentionally omitted a few things that put the credibility of the test in jeopardy.
This is insanity. It is trained with 10% the compute it took to train LLama 405.
Hearing quite a few US analysts + engineers confused and coping this morning.
wait, this is that annihilation dude? lmao
CCP does not micromanage the nation at this level.Why doesn’t CCP force Huawei or Alibaba to give Deepseek more GPUs? Seems like they are doing very well
I had a SORA feeling about o3..... it was supposed to be part of "shipmas" but not only was o3 not shipped, it was announced with only benchmarks and not even a live demonstration.... the only reason Altman would do that is because he had something to hide and was in fake it until make it mode, otherwise he would be bragging with real live examples... instead of just a rigged chart/ fake benchmark etc... and they even told the ARC AGI dude to not disclose how much compute was used and the hardware/cost lol.... probably millions of dollars for a fake horse and pony show
And after I purchased the o1 pro mode for $200, I really expected, wanted and hoped that pro version of SORA would be way better than Kling, but in all my tests, its worse than Kling...
Yes, I saw that, but still quite unbelievable to do this.from their paper, training took 2.788M H800 GPU hours. For 2 months training run it's roughly only 2000 H800
View attachment 141574