By Max A. Cherney
(Reuters) – Google parent Alphabet on Tuesday unveiled a product called Trillium in its artificial intelligence data center chip family that it says is nearly five times as fast as its prior version.
“Industry demand for (machine learning) computer has grown by a factor of 1 million in the last six years, roughly increasing 10-fold every year,” Alphabet CEO Sundar Pichai said in a briefing call with reporters. “I think Google was built for this moment, we’ve been pioneering (AI chips) for more than a decade.”
Alphabet’s effort to build custom chips for AI data centers represents one of the few viable alternatives to Nvidia’s top-of-the-line processors that dominate the market. Together with the software that is closely tied to Google’s tensor processing units (TPUs), the chips have allowed the company to take a significant share of the market.
Nvidia commands roughly 80% of the AI data center chip market, and the vast majority of the remaining 20% is various versions of Google’s TPUs. The company doesn’t sell the chip itself, but rents access through its cloud computing platform.
The sixth-generation Trillium chip will achieve 4.7 times better computing performance compared with the TPU v5e, according to Google, a chip designed to power the tech that generates text and other media from large models. The Trillium processor is 67% more energy efficient than the v5e.
The new chip will be available to its cloud customers in “late 2024,” the company said.
Google’s engineers achieved additional performance gains by increasing the amount of high-bandwidth memory capacity and overall bandwidth. AI models require enormous amounts of advanced memory, which has been a bottleneck to further boosting performance.
The company designed the chips to be deployed in pods of 256 chips that can be scaled to hundreds of pods.
(Reporting by Max A. Cherney in San Francisco; Editing by Leslie Adler)
Comments