Customize Consent Preferences

We use cookies to help you navigate efficiently and perform certain functions. You will find detailed information about all cookies under each consent category below.

The cookies that are categorized as "Necessary" are stored on your browser as they are essential for enabling the basic functionalities of the site. ... 

Always Active

Necessary cookies are required to enable the basic features of this site, such as providing secure log-in or adjusting your consent preferences. These cookies do not store any personally identifiable data.

No cookies to display.

Functional cookies help perform certain functionalities like sharing the content of the website on social media platforms, collecting feedback, and other third-party features.

No cookies to display.

Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics such as the number of visitors, bounce rate, traffic source, etc.

No cookies to display.

Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.

No cookies to display.

Advertisement cookies are used to provide visitors with customized advertisements based on the pages you visited previously and to analyze the effectiveness of the ad campaigns.

No cookies to display.

Googles AI supercomputer faster and greener than Nvidia A100 chip

Googles AI supercomputer faster and greener than Nvidia A100 chip

Alphabet Inc‘s Google launched on Tuesday new particulars concerning the supercomputers it makes use of to coach its synthetic intelligence fashions, saying the programs are each quicker and extra power-efficient than comparable programs from Nvidia Corp.

Google has designed its personal {custom} chip known as the Tensor Processing Unit, or TPU. It makes use of these chips for greater than 90% of the corporate’s work on synthetic intelligence coaching, the method of feeding knowledge by means of fashions to make them helpful at duties reminiscent of responding to queries with human-like textual content or producing photographs.

The Google TPU is now in its fourth era. Google on Tuesday printed a scientific paper detailing the way it has strung greater than 4,000 of the chips collectively right into a supercomputer utilizing its personal custom-developed optical switches to assist join particular person machines.

Improving these connections has turn into a key level of competitors amongst firms that construct AI supercomputers as a result of so-called giant language fashions that energy applied sciences like Google’s Bard or OpenAI’s ChatGPT have exploded in dimension, which means they’re far too giant to retailer on a single chip.

The fashions should as an alternative be break up throughout 1000’s of chips, which should then work collectively for weeks or extra to coach the mannequin. Google’s PaLM mannequin – its largest publicly disclosed language mannequin up to now – was educated by splitting it throughout two of the 4,000-chip supercomputers over 50 days.

Google mentioned its supercomputers make it straightforward to reconfigure connections between chips on the fly, serving to keep away from issues and tweak for efficiency positive factors.

“Circuit switching makes it easy to route around failed components,” Google Fellow Norm Jouppi and Google Distinguished Engineer David Patterson wrote in a weblog put up concerning the system. “This flexibility even allows us to change the topology of the supercomputer interconnect to accelerate the performance of an ML (machine learning) model.”

While Google is just now releasing particulars about its supercomputer, it has been on-line inside the corporate since 2020 in a knowledge middle in Mayes County, Oklahoma. Google mentioned that startup Midjourney used the system to coach its mannequin, which generates recent photographs after being fed a couple of phrases of textual content.

In the paper, Google mentioned that for comparably sized programs, its chips are as much as 1.7 instances quicker and 1.9 instances extra power-efficient than a system based mostly on Nvidia’s A100 chip that was in the marketplace concurrently the fourth-generation TPU.

A Nvidia spokesperson declined to remark.

Google mentioned it didn’t evaluate its fourth-generation to Nvidia’s present flagship H100 chip as a result of the H100 got here to the market after Google’s chip and is made with newer know-how.

Google hinted that it is perhaps engaged on a brand new TPU that may compete with the Nvidia H100 however supplied no particulars, with Jouppi telling Reuters that Google has “a healthy pipeline of future chips.”

Source: www.anews.com.tr