Artificial Intelligence thread

diadact

New Member
Registered Member
China Telecom claims that it has trained Llama-3.1 405B parameter model using just Domestic 10k+ AI cluster (I guess when they say wan ka, I'm not sure how big it is. Could be answer where from 10k to 99k)

apparently, has MFU of 43%
They didn't train a new 405B model
They fine-tuned it on a dataset
Good for gaining experience on how to effectively utilize large domestic GPU clusters and testing the stability of cluster
They need to start building 100K H100 class GPU clusters if they want to stay competitive in 2025
Any news about ascend 910C??
Its production volume & performance??
 

tphuang

Lieutenant General
Staff member
Super Moderator
VIP Professional
Registered Member

Kling 1.5 has been released now and it looks great

They didn't train a new 405B model
They fine-tuned it on a dataset
Good for gaining experience on how to effectively utilize large domestic GPU clusters and testing the stability of cluster
They need to start building 100K H100 class GPU clusters if they want to stay competitive in 2025
Any news about ascend 910C??
Its production volume & performance??
yes, they fine tuned it. I was repeating what they said in there, but I don't see why it wasn't obvious what I meant.

And just why do they need 100k H100 GPU cluster to stay competitive? Alibaba just trained its latest Qwen 2.5 on 18 trillion tokens
Please, Log in or Register to view URLs content!

It got released just a few months after Qwen-2.0.

How much non duplicate tokens are out there globally that you can use? How much larger do their cluster really need to get?
 

antiterror13

Brigadier
I can give it twenty years and AI wouldn't replace coding or programmers - at the end of the day you need someone to actually look through, test and debug the output, which requires someone knowledgable in writing and developing software.

If an AI is capable of self developing, debugging and testing software, it must have the ability to be self aware and actually understand what it's creating. Therefore it has certainly ascended into an AGI, at which point we have bigger problems.

Certainly AI wouldn't replace top class coders (let's say top 1%) in 10 years time, however AI certainly able to replace >50% of coders now and slowly moving up and will get harder and harder
 

PCK11800

Just Hatched
Registered Member
Certainly AI wouldn't replace top class coders (let's say top 1%) in 10 years time, however AI certainly able to replace >50% of coders now and slowly moving up and will get harder and harder
Not even 5%.

Currently even the latest and greatest LLMs (looking at you ClosedAI o1) sometimes crashes and burns even with simple standalone React components. They are completely and utterly useless with writing code for my legacy backend codebase. Unless I can fit the entire codebase into the context window, no LLMs can contribute without me spoonfeeding them all the relevant context... which is like 90% of the work.

LLMs are immensely helpful and allows me to become much more productive, being able to ask my specific coding questions and receiving a specific answer insanely streamlines my development process, but that still relies on you knowing what questions to ask :)

That said with future advancements, I would say your cookie-cutter front-end only code monkeys are going to struggle in the future, but more generalist developers should still be fine.

90% within 2 years
I have to ask, had you done any professional software development? Do you have a degree in Comp Sci or write code for a living? Do you even know how to write code? Any projects, toy programs or anything?
 
Top