Tesla has revealed its funding into an enormous compute cluster comprising 10,000 Nvidia H100 GPUs particularly designed to energy AI workloads.
The system, which went on-line this week, is designed to course of the mountains of knowledge its fleet of autos acquire with a view to accelerating the event of absolutely self-driving autosin line with its chief of AI infrastructure, Tim Zaman.
Tesla has been striving for years to achieve the purpose at which its autos will be thought of totally autonomous and has invested greater than a billion {dollars} into adopting the infrastructure to make this doable.
Tesla supercomputer
In July 2023, CEO Elon Musk revealed the agency would make investments $1 billion into constructing out its Dojo supercomputer over the following yr. Dojowhich relies on Tesla’s personal tech, started with the D1 chip, fitted with 354 customized CPU cores. Every coaching tile module contains 25 D1 chips, with the bottom Dojo V1 configuration together with 53,100 D1 cores in whole.
The agency additionally constructed a compute cluster fitted with 5,760 Nvidia A100 GPUs in June 2012. However the agency’s newest funding in 10,000 of the corporate’s H100 GPUs dwarfs the ability of this supercomputer.
This AI cluster, value greater than $300 million, will provide a peak efficiency of 340 FP64 PFLOPS for technical computing and 39.58 INT8 ExaFLOPS for AI purposes, in line with Tom’s {Hardware}.
The facility at Tesla’s disposal is definitely greater than that supplied by the Lenoardo supercomputer, the publication identified, making it one of the highly effective computer systems on the planet.
Nvidia’s chips are the parts that energy lots of the world’s main generative AI platforms. These GPUs, that are fitted into servers, have a number of different use instances from medical imaging to producing climate fashions.
Tesla is hoping to make use of the ability of those GPUs to extra effectively and successfully churn via the huge portions of knowledge it has to construct a mannequin that may efficiently rival a human.
Whereas many companies would normally lean on infrastructure hosted by the likes of Google or Microsoft, Tesla’s supercomputing infrastructure is all on-prem, meanig the agency can even have to take care of all of it.