image of this article category

New AI benchmarks test speed of running AI applications

03.04.2025 02:07 AM
Latest
New AI benchmarks test speed of running AI applications
dooklik website logo
As the underlying models must respond to many more queries to power AI applications such as chatbots and search engines, MLCommons developed two new versions of its MLPerf benchmarks to gauge speed Artificial intelligence group MLCommons unveiled two new benchmarks that it said can help determine how quickly top-of-the-line hardware and software can run AI applications.
share
share this article on facebook
share this article on twitter
share this article on whatsapp
share this article on facebook messenger
New AI benchmarks test speed of running AI applications
Since the launch of OpenAI's ChatGPT over two years ago, chip companies have begun to shift their focus to making hardware that can efficiently run the code that allows millions of people to use AI tools. As the underlying models must respond to many more queries to power AI applications such as chatbots and search engines, MLCommons developed two new versions of its MLPerf benchmarks to gauge speed.

One of the new benchmarks is based on Meta's so-called Llama 3.1 405-billion-parameter AI model, and the test targets general question answering, math and code generation. The new format tests a system's ability to process large queries and synthesize data from multiple sources.
Nvidia submitted several of its chips for the benchmark, and so did system builders such as Dell Technologies . There were no Advanced Micro Devices submissions for the large 405-billion-parameter benchmark, according to data provided by MLCommons.

For the new test, Nvidia's latest generation of artificial intelligence servers - called Grace Blackwell, which have 72 Nvidia graphics processing units (GPUs) inside - was 2.8 to 3.4 times faster than the previous generation, even when only using eight GPUs in the newer server to create a direct comparison to the older model, the company said at a briefing on Tuesday.

Nvidia has been working to speed up the connections of chips inside its servers, which is important in AI work where a chatbot runs on multiple chips at once.

The second benchmark is also based on an open-source AI model built by Meta and the test aims to more closely simulate the performance expectations set by consumer AI applications such as ChatGPT.

The goal is to tighten the response time for the benchmark and make it close to an instant response.

Related Articles
doolik website logo
YouTube’s testing out a handy new option for Shorts editing, which will highlight where any elements that you add to a clip, like captions and stickers, are placed relative to the actual Shorts UI, and what may interfere with your content on-screen.
doolik website logo
ROME (Reuters) - Amazon (AMZN) announced on Thursday that it had successfully completed its first delivery drone test in Italy, marking the country as the first in Europe where the e-commerce giant plans to roll out the service.

Live Video Streaming
Live video streaming lets you engage with your audience in real time with a video feed. Broadcast your daily show to your audience with no limits, no buffering and high quality videos. Reach all devices anytime anywhere with different video qualities that suits any device and any connection.
$1,120/YE*
The website uses cookies to improve your experience. We’ll assume you’re ok with this, but you can opt-out if you wish.
ACCEPT