This is not investment advice. The author has no position in any of the stocks mentioned. Wccftech.com has a disclosure and ethics policy.
Jensen Huang had termed Elon Musk “superhuman” when he described in a recent interview how xAI was able to bring together NVIDIA’s gear and operationalize it within its own data center in just 19 days. Now, Musk appears determined to subdue his competitors by continuing to pursue a shock and awe campaign that will see xAI’s Supercluster double in size.
For the benefit of those who might not be aware, xAI’s Colossus supercomputer cluster currently consists of 100,000 units of NVIDIA’s liquid-cooled H100 GPUs. Dubbed the world’s largest AI supercomputer, the Colossus is right now training xAI’s Grok family of large language models (LLMs).
Now, NVIDIA has revealed in a dedicated press release that xAI is doubling the size of its Colossus supercluster:
“xAI is in the process of doubling the size of Colossus to a combined total of 200,000 NVIDIA Hopper GPUs.”
Bear in mind that xAI and NVIDIA were able to bring the Colossus online in just 122 days when it would ordinarily take “many months to years” to operationalize such an intricate system. What’s more, xAI was able to commence the training of its Grok LLM within 19 days of the first H100 GPU rack rolling onto the floor of the AI gigafactory.
NVIDIA goes on to note:
“Across all three tiers of the network fabric, the system has experienced zero application latency degradation or packet loss due to flow collisions. It has maintained 95% data…
Read full on Wccftech
Discover more from Technical Master - Gadgets Reviews, Guides and Gaming News
Subscribe to get the latest posts sent to your email.