MLPerf Client 1.0 AI benchmark released — new testing toolkit sports a GUI, covers more models and tasks, and supports more hardware acceleration paths

The user interface of the MLPerf Client 1.0 benchmark
(Image credit: Future)

The AI revolution is upon us, but unlike past shifts in computing, most of us interact with the most advanced versions of AI models in the cloud. Leading services like ChatGPT, Claude, and Gemini all remain cloud-based. For reasons including privacy, research, and control, however, locally run AI models are still of interest, and it's important to be able to reliably and neutrally measure the AI performance of client systems with GPUs and NPUs on board.

Client AI remains a highly fluid space as hardware and software vendors work to define the types of workloads best suited for local execution and the best compute resources upon which to execute them. To help navigate this rapidly changing environment, the MLCommons consortium and its MLPerf Client working group maintain a client benchmark that's developed in collaboration with major hardware and software vendors.

  • AMD NPU and GPU hybrid support via ONNX Runtime GenAI and the Ryzen AI SDK
  • AMD, Intel, and NVIDIA GPU support via ONNX Runtime GenAI-DirectML
  • Intel NPU and GPU support via OpenVINO
  • Qualcomm Technologies NPU and CPU hybrid support via Qualcomm Genie and the QAIRT SDK
  • Apple Mac GPU support via MLX

Version 1.0 of the benchmark also supports these experimental hardware execution paths:

  • Intel NPU and GPU support via Microsoft Windows ML and the OpenVINO execution provider
  • NVIDIA GPU support via Llama.cpp-CUDA
  • Apple Mac GPU support via Llama.cpp-Metal

Last but certainly not least, MLPerf Client 1.0 now has an available graphical user interface that lets users understand the full range of benchmarks they can run on their hardware and easily choose among them.

The user interface of the MLPerf Client 1.0 benchmark

(Image credit: Future)

The GUI version also provides real-time monitoring of the various hardware resources on a system so that you can see at a glance whether your chosen execution path is exercising the expected GPU or NPU at a glance.

Past versions of MLPerf Client were command-line-only tools, so this new user interface should broaden the appeal of the benchmark both for casual users who just want to kick the AI tires on their GPU or NPU (or both) and for professional hardware testers who need to collect and compare results across a number of different hardware and software configurations.

MLPerf Client 1.0 is available now as a free download for all from GitHub. If you're interested in understanding how your system performs across a broad range of AI workloads, give it a shot. We've already spent some time with version 1.0 and we're excited to continue digging into it as we explore AI performance across a wide range of hardware.

Follow Tom's Hardware on Google News to get our up-to-date news, analysis, and reviews in your feeds. Make sure to click the Follow button.

TOPICS
Jeffrey Kampman
Senior Analyst, Graphics

As the Senior Analyst, Graphics at Tom's Hardware, Jeff Kampman covers everything to do with GPUs, gaming performance, and more. From integrated graphics processors to discrete graphics cards to the hyperscale installations powering our AI future, if it's got a GPU in it, Jeff is on it.